From 7e578390eb15ff40a000f7c18a94ca4628626795 Mon Sep 17 00:00:00 2001 From: Julia Ogris Date: Tue, 15 Feb 2022 13:45:08 +1100 Subject: [PATCH 1/4] Hermitise tools Hermitise tools with hermit upgrade protoc hermit install gosimports protoc-gen-go protoc-gen-go-grpc and by adding a local hermit package for reflect. Signed-off-by: Julia Ogris --- Makefile | 10 ++-------- bin/{.protoc-3.17.3.pkg => .gosimports-0.1.5.pkg} | 0 bin/.protoc-3.19.4.pkg | 1 + bin/.protoc-gen-go-1.27.1.pkg | 1 + bin/.protoc-gen-go-grpc-1.1.0.pkg | 1 + bin/.reflect-0.0.22.pkg | 1 + bin/gosimports | 1 + bin/hermit-packages/reflect.hcl | 7 +++++++ bin/hermit.hcl | 1 + bin/protoc | 2 +- bin/protoc-gen-go | 1 + bin/protoc-gen-go-grpc | 1 + bin/reflect | 1 + httprule/internal/echo.pb.go | 2 +- httprule/internal/test.pb.go | 2 +- 15 files changed, 21 insertions(+), 11 deletions(-) rename bin/{.protoc-3.17.3.pkg => .gosimports-0.1.5.pkg} (100%) create mode 120000 bin/.protoc-3.19.4.pkg create mode 120000 bin/.protoc-gen-go-1.27.1.pkg create mode 120000 bin/.protoc-gen-go-grpc-1.1.0.pkg create mode 120000 bin/.reflect-0.0.22.pkg create mode 120000 bin/gosimports create mode 100644 bin/hermit-packages/reflect.hcl create mode 120000 bin/protoc-gen-go create mode 120000 bin/protoc-gen-go-grpc create mode 120000 bin/reflect diff --git a/Makefile b/Makefile index 4103921..225c6a7 100644 --- a/Makefile +++ b/Makefile @@ -44,7 +44,7 @@ gen-pb = protoc -o $(1:%.proto=%-protoc.pb) $(1) gen-json = reflect fdsf $(1:%.proto=%-protoc.pb) -f json | jq . > $(1:%.proto=%-protoc.json) gen-testdata = $(call gen-pb,$(1))$(nl)$(call gen-json,$(1))$(nl) -gen-testdata: tools +gen-testdata: $(foreach proto,$(wildcard testdata/*.proto),$(call gen-testdata,$(proto))) check-uptodate: gen-testdata protos @@ -70,7 +70,7 @@ protos: --go-grpc_out=. --go-grpc_opt=module=foxygo.at/protog \ -I httprule/internal \ test.proto echo.proto - goimports -w . + gosimports -w . .PHONY: protos @@ -93,12 +93,6 @@ COLOUR_WHITE = $(shell tput setaf 7 2>/dev/null) help: @awk -F ':.*## ' 'NF == 2 && $$1 ~ /^[A-Za-z0-9%_-]+$$/ { printf "$(COLOUR_WHITE)%-25s$(COLOUR_NORMAL)%s\n", $$1, $$2}' $(MAKEFILE_LIST) | sort -tools: - go install google.golang.org/protobuf/cmd/protoc-gen-go@v1.27.1 - go install google.golang.org/grpc/cmd/protoc-gen-go-grpc@v1.1 - go install golang.org/x/tools/cmd/goimports@v0.1.5 - go install github.com/juliaogris/reflect@v0.0.22 - $(O): @mkdir -p $@ diff --git a/bin/.protoc-3.17.3.pkg b/bin/.gosimports-0.1.5.pkg similarity index 100% rename from bin/.protoc-3.17.3.pkg rename to bin/.gosimports-0.1.5.pkg diff --git a/bin/.protoc-3.19.4.pkg b/bin/.protoc-3.19.4.pkg new file mode 120000 index 0000000..383f451 --- /dev/null +++ b/bin/.protoc-3.19.4.pkg @@ -0,0 +1 @@ +hermit \ No newline at end of file diff --git a/bin/.protoc-gen-go-1.27.1.pkg b/bin/.protoc-gen-go-1.27.1.pkg new file mode 120000 index 0000000..383f451 --- /dev/null +++ b/bin/.protoc-gen-go-1.27.1.pkg @@ -0,0 +1 @@ +hermit \ No newline at end of file diff --git a/bin/.protoc-gen-go-grpc-1.1.0.pkg b/bin/.protoc-gen-go-grpc-1.1.0.pkg new file mode 120000 index 0000000..383f451 --- /dev/null +++ b/bin/.protoc-gen-go-grpc-1.1.0.pkg @@ -0,0 +1 @@ +hermit \ No newline at end of file diff --git a/bin/.reflect-0.0.22.pkg b/bin/.reflect-0.0.22.pkg new file mode 120000 index 0000000..383f451 --- /dev/null +++ b/bin/.reflect-0.0.22.pkg @@ -0,0 +1 @@ +hermit \ No newline at end of file diff --git a/bin/gosimports b/bin/gosimports new file mode 120000 index 0000000..f85588e --- /dev/null +++ b/bin/gosimports @@ -0,0 +1 @@ +.gosimports-0.1.5.pkg \ No newline at end of file diff --git a/bin/hermit-packages/reflect.hcl b/bin/hermit-packages/reflect.hcl new file mode 100644 index 0000000..745cdf3 --- /dev/null +++ b/bin/hermit-packages/reflect.hcl @@ -0,0 +1,7 @@ +description = "reflect" +test = "reflect --version" +binaries = ["reflect"] + +version "0.0.22" { + source = "https://github.com/juliaogris/reflect/releases/download/v${version}/reflect_${version}_${os}_${arch}.tar.gz" +} diff --git a/bin/hermit.hcl b/bin/hermit.hcl index a46b470..615243e 100644 --- a/bin/hermit.hcl +++ b/bin/hermit.hcl @@ -1,3 +1,4 @@ +sources = ["env:///bin/hermit-packages", "https://github.com/cashapp/hermit-packages.git"] manage-git = false env = { GOBIN : "${HERMIT_ENV}/out/bin", diff --git a/bin/protoc b/bin/protoc index 2e19f0b..775c89b 120000 --- a/bin/protoc +++ b/bin/protoc @@ -1 +1 @@ -.protoc-3.17.3.pkg \ No newline at end of file +.protoc-3.19.4.pkg \ No newline at end of file diff --git a/bin/protoc-gen-go b/bin/protoc-gen-go new file mode 120000 index 0000000..a81f60a --- /dev/null +++ b/bin/protoc-gen-go @@ -0,0 +1 @@ +.protoc-gen-go-1.27.1.pkg \ No newline at end of file diff --git a/bin/protoc-gen-go-grpc b/bin/protoc-gen-go-grpc new file mode 120000 index 0000000..8c22ed2 --- /dev/null +++ b/bin/protoc-gen-go-grpc @@ -0,0 +1 @@ +.protoc-gen-go-grpc-1.1.0.pkg \ No newline at end of file diff --git a/bin/reflect b/bin/reflect new file mode 120000 index 0000000..2445602 --- /dev/null +++ b/bin/reflect @@ -0,0 +1 @@ +.reflect-0.0.22.pkg \ No newline at end of file diff --git a/httprule/internal/echo.pb.go b/httprule/internal/echo.pb.go index e225672..0dde86f 100644 --- a/httprule/internal/echo.pb.go +++ b/httprule/internal/echo.pb.go @@ -1,7 +1,7 @@ // Code generated by protoc-gen-go. DO NOT EDIT. // versions: // protoc-gen-go v1.27.1 -// protoc v3.17.3 +// protoc v3.19.4 // source: echo.proto package internal diff --git a/httprule/internal/test.pb.go b/httprule/internal/test.pb.go index c295518..6399f58 100644 --- a/httprule/internal/test.pb.go +++ b/httprule/internal/test.pb.go @@ -1,7 +1,7 @@ // Code generated by protoc-gen-go. DO NOT EDIT. // versions: // protoc-gen-go v1.27.1 -// protoc v3.17.3 +// protoc v3.19.4 // source: test.proto package internal From b06b892f1060201711be14742893d89e05462b0e Mon Sep 17 00:00:00 2001 From: Julia Ogris Date: Tue, 15 Feb 2022 13:45:09 +1100 Subject: [PATCH 2/4] Add debug info on up-to-date checks Add debug info on up-to-date checks via `git diff` on Makefile up-to-date check as we have a build failing on CI at `test -z "$ (git status --porcelain)"` that doesn't fail locally. --- Makefile | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/Makefile b/Makefile index 225c6a7..1c17064 100644 --- a/Makefile +++ b/Makefile @@ -48,7 +48,7 @@ gen-testdata: $(foreach proto,$(wildcard testdata/*.proto),$(call gen-testdata,$(proto))) check-uptodate: gen-testdata protos - test -z "$$(git status --porcelain)" + test -z "$$(git status --porcelain)" || { git diff; false; } CHECK_COVERAGE = awk -F '[ \t%]+' '/^total:/ {print; if ($$3 < $(COVERAGE)) exit 1}' FAIL_COVERAGE = { echo '$(COLOUR_RED)FAIL - Coverage below $(COVERAGE)%$(COLOUR_NORMAL)'; exit 1; } From 8ee20e38502d4e3a170e3abbbb8fb88e6f311d80 Mon Sep 17 00:00:00 2001 From: Julia Ogris Date: Tue, 15 Feb 2022 13:45:10 +1100 Subject: [PATCH 3/4] Transplant registry Transplant registry from jig. Use protosync to download google api files. registry.Files is wrapper for the official protoregistry.Files, implementing protoregistry.MessageTypeResolver and protoregistry.ExtensionTypeResolver, so that we can easily use this Files wrapper with prototest and protojson to unmarshal extension. A registry is created from a FileDescriptorSet like so: var fds *descriptorpb.FileDescriptorSet // initialise protoregistryFiles, err := protodesc.NewFiles(&fds) // handle err files := registry.NewFiles(protoregistryFiles) link: https://github.com/foxygoat/jig/tree/v0.0.25/registry Signed-off-by: Julia Ogris --- Makefile | 2 + bin/.protosync-0.2.1.pkg | 1 + bin/protosync | 1 + registry/files.go | 116 +++ registry/files_test.go | 169 ++++ .../testdata/google/api/annotations.proto | 31 + registry/testdata/google/api/http.proto | 375 +++++++ .../testdata/google/protobuf/descriptor.proto | 921 ++++++++++++++++++ registry/testdata/regtest.pb | Bin 0 -> 9425 bytes registry/testdata/regtest.proto | 51 + 10 files changed, 1667 insertions(+) create mode 120000 bin/.protosync-0.2.1.pkg create mode 120000 bin/protosync create mode 100644 registry/files.go create mode 100644 registry/files_test.go create mode 100644 registry/testdata/google/api/annotations.proto create mode 100644 registry/testdata/google/api/http.proto create mode 100644 registry/testdata/google/protobuf/descriptor.proto create mode 100644 registry/testdata/regtest.pb create mode 100644 registry/testdata/regtest.proto diff --git a/Makefile b/Makefile index 1c17064..49cca46 100644 --- a/Makefile +++ b/Makefile @@ -46,6 +46,8 @@ gen-testdata = $(call gen-pb,$(1))$(nl)$(call gen-json,$(1))$(nl) gen-testdata: $(foreach proto,$(wildcard testdata/*.proto),$(call gen-testdata,$(proto))) + protosync --dest registry/testdata google/api/annotations.proto + protoc --include_imports -I registry/testdata -o registry/testdata/regtest.pb registry/testdata/regtest.proto check-uptodate: gen-testdata protos test -z "$$(git status --porcelain)" || { git diff; false; } diff --git a/bin/.protosync-0.2.1.pkg b/bin/.protosync-0.2.1.pkg new file mode 120000 index 0000000..383f451 --- /dev/null +++ b/bin/.protosync-0.2.1.pkg @@ -0,0 +1 @@ +hermit \ No newline at end of file diff --git a/bin/protosync b/bin/protosync new file mode 120000 index 0000000..7f0818f --- /dev/null +++ b/bin/protosync @@ -0,0 +1 @@ +.protosync-0.2.1.pkg \ No newline at end of file diff --git a/registry/files.go b/registry/files.go new file mode 100644 index 0000000..9b7ca6c --- /dev/null +++ b/registry/files.go @@ -0,0 +1,116 @@ +// Package registry provides a type on top of protoregistry.Files that can be +// used as a protoregistry.ExtensionTypeResolver and a +// protoregistry.MessageTypeResolver. This allows a protoregistry.Files to be +// used as Resolver for protobuf encoding marshaling options. +package registry + +import ( + "strings" + + "google.golang.org/protobuf/reflect/protoreflect" + "google.golang.org/protobuf/reflect/protoregistry" + "google.golang.org/protobuf/types/dynamicpb" +) + +type Files struct { + protoregistry.Files +} + +func NewFiles(f *protoregistry.Files) *Files { + return &Files{Files: *f} +} + +type extMatchFn func(protoreflect.ExtensionDescriptor) bool + +// extensionContainer is implemented by FileDescriptor and MessageDescriptor. +// They are both "namespaces" that contain extensions and have "sub-namespaces". +type extensionContainer interface { + Messages() protoreflect.MessageDescriptors + Extensions() protoreflect.ExtensionDescriptors +} + +func (f *Files) FindExtensionByName(field protoreflect.FullName) (protoreflect.ExtensionType, error) { + desc, err := f.FindDescriptorByName(field) + if err != nil { + return nil, err + } + ed, ok := desc.(protoreflect.ExtensionDescriptor) + if !ok { + return nil, protoregistry.NotFound + } + return dynamicpb.NewExtensionType(ed), nil +} + +func (f *Files) FindExtensionByNumber(message protoreflect.FullName, field protoreflect.FieldNumber) (protoreflect.ExtensionType, error) { + return findExtension(&f.Files, func(ed protoreflect.ExtensionDescriptor) bool { + return ed.ContainingMessage().FullName() == message && ed.Number() == field + }) +} + +func (f *Files) GetExtensionsOfMessage(message protoreflect.FullName) []protoreflect.ExtensionType { + return walkExtensions(&f.Files, true, func(ed protoreflect.ExtensionDescriptor) bool { + return ed.ContainingMessage().FullName() == message + }) +} + +func findExtension(files *protoregistry.Files, pred extMatchFn) (protoreflect.ExtensionType, error) { + ets := walkExtensions(files, false, pred) + if len(ets) == 0 { + return nil, protoregistry.NotFound + } + return ets[0], nil +} + +func walkExtensions(files *protoregistry.Files, getAll bool, pred extMatchFn) []protoreflect.ExtensionType { + var result []protoreflect.ExtensionType + + files.RangeFiles(func(fd protoreflect.FileDescriptor) bool { + result = append(result, getExtensions(fd, getAll, pred)...) + // continue if we are getting all extensions or have none so far + return getAll || len(result) == 0 + }) + return result +} + +func getExtensions(ec extensionContainer, getAll bool, pred extMatchFn) []protoreflect.ExtensionType { + var result []protoreflect.ExtensionType + + eds := ec.Extensions() + for i := 0; i < eds.Len() && (getAll || len(result) == 0); i++ { + ed := eds.Get(i) + if pred(ed) { + result = append(result, dynamicpb.NewExtensionType(ed)) + } + } + + mds := ec.Messages() + for i := 0; i < mds.Len() && (getAll || len(result) == 0); i++ { + md := mds.Get(i) + result = append(result, getExtensions(md, getAll, pred)...) + } + + return result +} + +func (f *Files) FindMessageByName(name protoreflect.FullName) (protoreflect.MessageType, error) { + desc, err := f.FindDescriptorByName(name) + if err != nil { + return nil, err + } + md, ok := desc.(protoreflect.MessageDescriptor) + if !ok { + return nil, protoregistry.NotFound + } + return dynamicpb.NewMessageType(md), nil +} + +func (f *Files) FindMessageByURL(url string) (protoreflect.MessageType, error) { + message := protoreflect.FullName(url) + // Strip off before the last slash - we only look locally for the + // message and do not hit the network. The part after the last slash + // must be the full name of the message. + if i := strings.LastIndexByte(url, '/'); i >= 0 { + message = message[i+len("/"):] + } + return f.FindMessageByName(message) +} diff --git a/registry/files_test.go b/registry/files_test.go new file mode 100644 index 0000000..fc0bc24 --- /dev/null +++ b/registry/files_test.go @@ -0,0 +1,169 @@ +package registry + +import ( + "os" + "testing" + + "github.com/stretchr/testify/require" + "google.golang.org/protobuf/proto" + "google.golang.org/protobuf/reflect/protodesc" + "google.golang.org/protobuf/reflect/protoreflect" + "google.golang.org/protobuf/reflect/protoregistry" + "google.golang.org/protobuf/types/descriptorpb" +) + +// ensure Files implments ExtensionTypeResolver +var _ protoregistry.ExtensionTypeResolver = (*Files)(nil) + +// ensure Files implments MessageTypeResolver +var _ protoregistry.MessageTypeResolver = (*Files)(nil) + +func TestFindExtensionByName(t *testing.T) { + tests := map[string]struct { + extName string + err error + }{ + "top-level extension": {"regtest.ef1", nil}, + "nested extension": {"regtest.ExtensionMessage.ef2", nil}, + "deeply nested extension": {"regtest.ExtensionMessage.NestedExtension.ef3", nil}, + "other package extension": {"regtest.base", nil}, + "imported extension": {"google.api.http", nil}, + "unknown extension": {"unknown.extension", protoregistry.NotFound}, + "non-extension descriptor": {"regtest.BaseMessage", protoregistry.NotFound}, + } + + f := newFiles(t) + for name, tc := range tests { + t.Run(name, func(t *testing.T) { + extName := protoreflect.FullName(tc.extName) + et, err := f.FindExtensionByName(extName) + if tc.err != nil { + require.ErrorIs(t, err, tc.err) + } else { + require.NoError(t, err, tc.extName) + require.Equal(t, extName, et.TypeDescriptor().FullName()) + } + }) + } +} + +func TestFindExtensionByNumber(t *testing.T) { + tests := map[string]struct { + message string + fieldNumber int32 + extName string + err error + }{ + "top-level extension": {"regtest.BaseMessage", 1000, "regtest.ef1", nil}, + "nested extension": {"regtest.BaseMessage", 1001, "regtest.ExtensionMessage.ef2", nil}, + "deeply nested extension": {"regtest.BaseMessage", 1002, "regtest.ExtensionMessage.NestedExtension.ef3", nil}, + "other package extension": {"google.protobuf.MethodOptions", 56789, "regtest.base", nil}, + "imported extension": {"google.protobuf.MethodOptions", 72295728, "google.api.http", nil}, + "unknown message": {"regtest.Foo", 999, "unknown.message", protoregistry.NotFound}, + "unknown extension": {"regtest.BaseMessage", 999, "unknown.extension", protoregistry.NotFound}, + } + + f := newFiles(t) + for name, tc := range tests { + t.Run(name, func(t *testing.T) { + messageName := protoreflect.FullName(tc.message) + fieldNumber := protoreflect.FieldNumber(tc.fieldNumber) + et, err := f.FindExtensionByNumber(messageName, fieldNumber) + if tc.err != nil { + require.ErrorIs(t, err, tc.err) + } else { + require.NoError(t, err, tc.extName) + extName := protoreflect.FullName(tc.extName) + require.Equal(t, extName, et.TypeDescriptor().FullName()) + } + }) + } +} + +func TestGetExtensionsOfMessage(t *testing.T) { + tests := map[string]struct { + message string + fields []int32 + }{ + "package message": {"regtest.BaseMessage", []int32{1000, 1001, 1002}}, + "imported message": {"google.protobuf.MethodOptions", []int32{56789, 72295728}}, + "unknown message": {"regtest.Foo", nil}, + } + + f := newFiles(t) + for name, tc := range tests { + t.Run(name, func(t *testing.T) { + messageName := protoreflect.FullName(tc.message) + ets := f.GetExtensionsOfMessage(messageName) + var fields []int32 + for _, et := range ets { + fields = append(fields, int32(et.TypeDescriptor().Number())) + } + require.ElementsMatch(t, tc.fields, fields) + }) + } +} + +func TestFindMessageByName(t *testing.T) { + tests := map[string]struct { + name string + err error + }{ + "top-level message": {"regtest.BaseMessage", nil}, + "nested message": {"regtest.ExtensionMessage.NestedExtension", nil}, + "unknown message": {"regtest.Foo", protoregistry.NotFound}, + "non-message descriptor": {"regtest.ef1", protoregistry.NotFound}, + } + + f := newFiles(t) + for name, tc := range tests { + t.Run(name, func(t *testing.T) { + messageName := protoreflect.FullName(tc.name) + mt, err := f.FindMessageByName(messageName) + if tc.err != nil { + require.ErrorIs(t, err, tc.err) + } else { + require.NoError(t, err, tc.name) + require.Equal(t, messageName, mt.Descriptor().FullName()) + } + }) + } +} + +func TestFindMessageByURL(t *testing.T) { + tests := map[string]struct { + url string + err error + }{ + "simple url": {"regtest.BaseMessage", nil}, + "hostname url": {"example.com/regtest.BaseMessage", nil}, + "multiple slashes": {"example.com/foo/bar/regtest.BaseMessage", nil}, + "unknown message": {"example.com/regtest.Foo", protoregistry.NotFound}, + } + + f := newFiles(t) + for name, tc := range tests { + t.Run(name, func(t *testing.T) { + mt, err := f.FindMessageByURL(tc.url) + if tc.err != nil { + require.ErrorIs(t, err, tc.err) + } else { + require.NoError(t, err, tc.url) + expected := protoreflect.FullName("regtest.BaseMessage") + require.Equal(t, expected, mt.Descriptor().FullName()) + } + }) + } +} + +func newFiles(t *testing.T) *Files { + t.Helper() + b, err := os.ReadFile("testdata/regtest.pb") + require.NoError(t, err) + fds := descriptorpb.FileDescriptorSet{} + err = proto.Unmarshal(b, &fds) + require.NoError(t, err) + files, err := protodesc.NewFiles(&fds) + require.NoError(t, err) + return NewFiles(files) +} diff --git a/registry/testdata/google/api/annotations.proto b/registry/testdata/google/api/annotations.proto new file mode 100644 index 0000000..efdab3d --- /dev/null +++ b/registry/testdata/google/api/annotations.proto @@ -0,0 +1,31 @@ +// Copyright 2015 Google LLC +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// See the License for the specific language governing permissions and +// limitations under the License. + +syntax = "proto3"; + +package google.api; + +import "google/api/http.proto"; +import "google/protobuf/descriptor.proto"; + +option go_package = "google.golang.org/genproto/googleapis/api/annotations;annotations"; +option java_multiple_files = true; +option java_outer_classname = "AnnotationsProto"; +option java_package = "com.google.api"; +option objc_class_prefix = "GAPI"; + +extend google.protobuf.MethodOptions { + // See `HttpRule`. + HttpRule http = 72295728; +} diff --git a/registry/testdata/google/api/http.proto b/registry/testdata/google/api/http.proto new file mode 100644 index 0000000..113fa93 --- /dev/null +++ b/registry/testdata/google/api/http.proto @@ -0,0 +1,375 @@ +// Copyright 2015 Google LLC +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// See the License for the specific language governing permissions and +// limitations under the License. + +syntax = "proto3"; + +package google.api; + +option cc_enable_arenas = true; +option go_package = "google.golang.org/genproto/googleapis/api/annotations;annotations"; +option java_multiple_files = true; +option java_outer_classname = "HttpProto"; +option java_package = "com.google.api"; +option objc_class_prefix = "GAPI"; + +// Defines the HTTP configuration for an API service. It contains a list of +// [HttpRule][google.api.HttpRule], each specifying the mapping of an RPC method +// to one or more HTTP REST API methods. +message Http { + // A list of HTTP configuration rules that apply to individual API methods. + // + // **NOTE:** All service configuration rules follow "last one wins" order. + repeated HttpRule rules = 1; + + // When set to true, URL path parameters will be fully URI-decoded except in + // cases of single segment matches in reserved expansion, where "%2F" will be + // left encoded. + // + // The default behavior is to not decode RFC 6570 reserved characters in multi + // segment matches. + bool fully_decode_reserved_expansion = 2; +} + +// # gRPC Transcoding +// +// gRPC Transcoding is a feature for mapping between a gRPC method and one or +// more HTTP REST endpoints. It allows developers to build a single API service +// that supports both gRPC APIs and REST APIs. Many systems, including [Google +// APIs](https://github.com/googleapis/googleapis), +// [Cloud Endpoints](https://cloud.google.com/endpoints), [gRPC +// Gateway](https://github.com/grpc-ecosystem/grpc-gateway), +// and [Envoy](https://github.com/envoyproxy/envoy) proxy support this feature +// and use it for large scale production services. +// +// `HttpRule` defines the schema of the gRPC/REST mapping. The mapping specifies +// how different portions of the gRPC request message are mapped to the URL +// path, URL query parameters, and HTTP request body. It also controls how the +// gRPC response message is mapped to the HTTP response body. `HttpRule` is +// typically specified as an `google.api.http` annotation on the gRPC method. +// +// Each mapping specifies a URL path template and an HTTP method. The path +// template may refer to one or more fields in the gRPC request message, as long +// as each field is a non-repeated field with a primitive (non-message) type. +// The path template controls how fields of the request message are mapped to +// the URL path. +// +// Example: +// +// service Messaging { +// rpc GetMessage(GetMessageRequest) returns (Message) { +// option (google.api.http) = { +// get: "/v1/{name=messages/*}" +// }; +// } +// } +// message GetMessageRequest { +// string name = 1; // Mapped to URL path. +// } +// message Message { +// string text = 1; // The resource content. +// } +// +// This enables an HTTP REST to gRPC mapping as below: +// +// HTTP | gRPC +// -----|----- +// `GET /v1/messages/123456` | `GetMessage(name: "messages/123456")` +// +// Any fields in the request message which are not bound by the path template +// automatically become HTTP query parameters if there is no HTTP request body. +// For example: +// +// service Messaging { +// rpc GetMessage(GetMessageRequest) returns (Message) { +// option (google.api.http) = { +// get:"/v1/messages/{message_id}" +// }; +// } +// } +// message GetMessageRequest { +// message SubMessage { +// string subfield = 1; +// } +// string message_id = 1; // Mapped to URL path. +// int64 revision = 2; // Mapped to URL query parameter `revision`. +// SubMessage sub = 3; // Mapped to URL query parameter `sub.subfield`. +// } +// +// This enables a HTTP JSON to RPC mapping as below: +// +// HTTP | gRPC +// -----|----- +// `GET /v1/messages/123456?revision=2&sub.subfield=foo` | +// `GetMessage(message_id: "123456" revision: 2 sub: SubMessage(subfield: +// "foo"))` +// +// Note that fields which are mapped to URL query parameters must have a +// primitive type or a repeated primitive type or a non-repeated message type. +// In the case of a repeated type, the parameter can be repeated in the URL +// as `...?param=A¶m=B`. In the case of a message type, each field of the +// message is mapped to a separate parameter, such as +// `...?foo.a=A&foo.b=B&foo.c=C`. +// +// For HTTP methods that allow a request body, the `body` field +// specifies the mapping. Consider a REST update method on the +// message resource collection: +// +// service Messaging { +// rpc UpdateMessage(UpdateMessageRequest) returns (Message) { +// option (google.api.http) = { +// patch: "/v1/messages/{message_id}" +// body: "message" +// }; +// } +// } +// message UpdateMessageRequest { +// string message_id = 1; // mapped to the URL +// Message message = 2; // mapped to the body +// } +// +// The following HTTP JSON to RPC mapping is enabled, where the +// representation of the JSON in the request body is determined by +// protos JSON encoding: +// +// HTTP | gRPC +// -----|----- +// `PATCH /v1/messages/123456 { "text": "Hi!" }` | `UpdateMessage(message_id: +// "123456" message { text: "Hi!" })` +// +// The special name `*` can be used in the body mapping to define that +// every field not bound by the path template should be mapped to the +// request body. This enables the following alternative definition of +// the update method: +// +// service Messaging { +// rpc UpdateMessage(Message) returns (Message) { +// option (google.api.http) = { +// patch: "/v1/messages/{message_id}" +// body: "*" +// }; +// } +// } +// message Message { +// string message_id = 1; +// string text = 2; +// } +// +// +// The following HTTP JSON to RPC mapping is enabled: +// +// HTTP | gRPC +// -----|----- +// `PATCH /v1/messages/123456 { "text": "Hi!" }` | `UpdateMessage(message_id: +// "123456" text: "Hi!")` +// +// Note that when using `*` in the body mapping, it is not possible to +// have HTTP parameters, as all fields not bound by the path end in +// the body. This makes this option more rarely used in practice when +// defining REST APIs. The common usage of `*` is in custom methods +// which don't use the URL at all for transferring data. +// +// It is possible to define multiple HTTP methods for one RPC by using +// the `additional_bindings` option. Example: +// +// service Messaging { +// rpc GetMessage(GetMessageRequest) returns (Message) { +// option (google.api.http) = { +// get: "/v1/messages/{message_id}" +// additional_bindings { +// get: "/v1/users/{user_id}/messages/{message_id}" +// } +// }; +// } +// } +// message GetMessageRequest { +// string message_id = 1; +// string user_id = 2; +// } +// +// This enables the following two alternative HTTP JSON to RPC mappings: +// +// HTTP | gRPC +// -----|----- +// `GET /v1/messages/123456` | `GetMessage(message_id: "123456")` +// `GET /v1/users/me/messages/123456` | `GetMessage(user_id: "me" message_id: +// "123456")` +// +// ## Rules for HTTP mapping +// +// 1. Leaf request fields (recursive expansion nested messages in the request +// message) are classified into three categories: +// - Fields referred by the path template. They are passed via the URL path. +// - Fields referred by the [HttpRule.body][google.api.HttpRule.body]. They are passed via the HTTP +// request body. +// - All other fields are passed via the URL query parameters, and the +// parameter name is the field path in the request message. A repeated +// field can be represented as multiple query parameters under the same +// name. +// 2. If [HttpRule.body][google.api.HttpRule.body] is "*", there is no URL query parameter, all fields +// are passed via URL path and HTTP request body. +// 3. If [HttpRule.body][google.api.HttpRule.body] is omitted, there is no HTTP request body, all +// fields are passed via URL path and URL query parameters. +// +// ### Path template syntax +// +// Template = "/" Segments [ Verb ] ; +// Segments = Segment { "/" Segment } ; +// Segment = "*" | "**" | LITERAL | Variable ; +// Variable = "{" FieldPath [ "=" Segments ] "}" ; +// FieldPath = IDENT { "." IDENT } ; +// Verb = ":" LITERAL ; +// +// The syntax `*` matches a single URL path segment. The syntax `**` matches +// zero or more URL path segments, which must be the last part of the URL path +// except the `Verb`. +// +// The syntax `Variable` matches part of the URL path as specified by its +// template. A variable template must not contain other variables. If a variable +// matches a single path segment, its template may be omitted, e.g. `{var}` +// is equivalent to `{var=*}`. +// +// The syntax `LITERAL` matches literal text in the URL path. If the `LITERAL` +// contains any reserved character, such characters should be percent-encoded +// before the matching. +// +// If a variable contains exactly one path segment, such as `"{var}"` or +// `"{var=*}"`, when such a variable is expanded into a URL path on the client +// side, all characters except `[-_.~0-9a-zA-Z]` are percent-encoded. The +// server side does the reverse decoding. Such variables show up in the +// [Discovery +// Document](https://developers.google.com/discovery/v1/reference/apis) as +// `{var}`. +// +// If a variable contains multiple path segments, such as `"{var=foo/*}"` +// or `"{var=**}"`, when such a variable is expanded into a URL path on the +// client side, all characters except `[-_.~/0-9a-zA-Z]` are percent-encoded. +// The server side does the reverse decoding, except "%2F" and "%2f" are left +// unchanged. Such variables show up in the +// [Discovery +// Document](https://developers.google.com/discovery/v1/reference/apis) as +// `{+var}`. +// +// ## Using gRPC API Service Configuration +// +// gRPC API Service Configuration (service config) is a configuration language +// for configuring a gRPC service to become a user-facing product. The +// service config is simply the YAML representation of the `google.api.Service` +// proto message. +// +// As an alternative to annotating your proto file, you can configure gRPC +// transcoding in your service config YAML files. You do this by specifying a +// `HttpRule` that maps the gRPC method to a REST endpoint, achieving the same +// effect as the proto annotation. This can be particularly useful if you +// have a proto that is reused in multiple services. Note that any transcoding +// specified in the service config will override any matching transcoding +// configuration in the proto. +// +// Example: +// +// http: +// rules: +// # Selects a gRPC method and applies HttpRule to it. +// - selector: example.v1.Messaging.GetMessage +// get: /v1/messages/{message_id}/{sub.subfield} +// +// ## Special notes +// +// When gRPC Transcoding is used to map a gRPC to JSON REST endpoints, the +// proto to JSON conversion must follow the [proto3 +// specification](https://developers.google.com/protocol-buffers/docs/proto3#json). +// +// While the single segment variable follows the semantics of +// [RFC 6570](https://tools.ietf.org/html/rfc6570) Section 3.2.2 Simple String +// Expansion, the multi segment variable **does not** follow RFC 6570 Section +// 3.2.3 Reserved Expansion. The reason is that the Reserved Expansion +// does not expand special characters like `?` and `#`, which would lead +// to invalid URLs. As the result, gRPC Transcoding uses a custom encoding +// for multi segment variables. +// +// The path variables **must not** refer to any repeated or mapped field, +// because client libraries are not capable of handling such variable expansion. +// +// The path variables **must not** capture the leading "/" character. The reason +// is that the most common use case "{var}" does not capture the leading "/" +// character. For consistency, all path variables must share the same behavior. +// +// Repeated message fields must not be mapped to URL query parameters, because +// no client library can support such complicated mapping. +// +// If an API needs to use a JSON array for request or response body, it can map +// the request or response body to a repeated field. However, some gRPC +// Transcoding implementations may not support this feature. +message HttpRule { + // Selects a method to which this rule applies. + // + // Refer to [selector][google.api.DocumentationRule.selector] for syntax details. + string selector = 1; + + // Determines the URL pattern is matched by this rules. This pattern can be + // used with any of the {get|put|post|delete|patch} methods. A custom method + // can be defined using the 'custom' field. + oneof pattern { + // Maps to HTTP GET. Used for listing and getting information about + // resources. + string get = 2; + + // Maps to HTTP PUT. Used for replacing a resource. + string put = 3; + + // Maps to HTTP POST. Used for creating a resource or performing an action. + string post = 4; + + // Maps to HTTP DELETE. Used for deleting a resource. + string delete = 5; + + // Maps to HTTP PATCH. Used for updating a resource. + string patch = 6; + + // The custom pattern is used for specifying an HTTP method that is not + // included in the `pattern` field, such as HEAD, or "*" to leave the + // HTTP method unspecified for this rule. The wild-card rule is useful + // for services that provide content to Web (HTML) clients. + CustomHttpPattern custom = 8; + } + + // The name of the request field whose value is mapped to the HTTP request + // body, or `*` for mapping all request fields not captured by the path + // pattern to the HTTP body, or omitted for not having any HTTP request body. + // + // NOTE: the referred field must be present at the top-level of the request + // message type. + string body = 7; + + // Optional. The name of the response field whose value is mapped to the HTTP + // response body. When omitted, the entire response message will be used + // as the HTTP response body. + // + // NOTE: The referred field must be present at the top-level of the response + // message type. + string response_body = 12; + + // Additional HTTP bindings for the selector. Nested bindings must + // not contain an `additional_bindings` field themselves (that is, + // the nesting may only be one level deep). + repeated HttpRule additional_bindings = 11; +} + +// A custom pattern is used for defining custom HTTP verb. +message CustomHttpPattern { + // The name of this custom HTTP verb. + string kind = 1; + + // The path matched by this custom verb. + string path = 2; +} diff --git a/registry/testdata/google/protobuf/descriptor.proto b/registry/testdata/google/protobuf/descriptor.proto new file mode 100644 index 0000000..298e584 --- /dev/null +++ b/registry/testdata/google/protobuf/descriptor.proto @@ -0,0 +1,921 @@ +// Protocol Buffers - Google's data interchange format +// Copyright 2008 Google Inc. All rights reserved. +// https://developers.google.com/protocol-buffers/ +// +// Redistribution and use in source and binary forms, with or without +// modification, are permitted provided that the following conditions are +// met: +// +// * Redistributions of source code must retain the above copyright +// notice, this list of conditions and the following disclaimer. +// * Redistributions in binary form must reproduce the above +// copyright notice, this list of conditions and the following disclaimer +// in the documentation and/or other materials provided with the +// distribution. +// * Neither the name of Google Inc. nor the names of its +// contributors may be used to endorse or promote products derived from +// this software without specific prior written permission. +// +// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS +// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT +// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR +// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT +// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, +// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT +// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, +// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY +// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT +// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE +// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. + +// Author: kenton@google.com (Kenton Varda) +// Based on original Protocol Buffers design by +// Sanjay Ghemawat, Jeff Dean, and others. +// +// The messages in this file describe the definitions found in .proto files. +// A valid .proto file can be translated directly to a FileDescriptorProto +// without any other information (e.g. without reading its imports). + + +syntax = "proto2"; + +package google.protobuf; + +option go_package = "google.golang.org/protobuf/types/descriptorpb"; +option java_package = "com.google.protobuf"; +option java_outer_classname = "DescriptorProtos"; +option csharp_namespace = "Google.Protobuf.Reflection"; +option objc_class_prefix = "GPB"; +option cc_enable_arenas = true; + +// descriptor.proto must be optimized for speed because reflection-based +// algorithms don't work during bootstrapping. +option optimize_for = SPEED; + +// The protocol compiler can output a FileDescriptorSet containing the .proto +// files it parses. +message FileDescriptorSet { + repeated FileDescriptorProto file = 1; +} + +// Describes a complete .proto file. +message FileDescriptorProto { + optional string name = 1; // file name, relative to root of source tree + optional string package = 2; // e.g. "foo", "foo.bar", etc. + + // Names of files imported by this file. + repeated string dependency = 3; + // Indexes of the public imported files in the dependency list above. + repeated int32 public_dependency = 10; + // Indexes of the weak imported files in the dependency list. + // For Google-internal migration only. Do not use. + repeated int32 weak_dependency = 11; + + // All top-level definitions in this file. + repeated DescriptorProto message_type = 4; + repeated EnumDescriptorProto enum_type = 5; + repeated ServiceDescriptorProto service = 6; + repeated FieldDescriptorProto extension = 7; + + optional FileOptions options = 8; + + // This field contains optional information about the original source code. + // You may safely remove this entire field without harming runtime + // functionality of the descriptors -- the information is needed only by + // development tools. + optional SourceCodeInfo source_code_info = 9; + + // The syntax of the proto file. + // The supported values are "proto2" and "proto3". + optional string syntax = 12; +} + +// Describes a message type. +message DescriptorProto { + optional string name = 1; + + repeated FieldDescriptorProto field = 2; + repeated FieldDescriptorProto extension = 6; + + repeated DescriptorProto nested_type = 3; + repeated EnumDescriptorProto enum_type = 4; + + message ExtensionRange { + optional int32 start = 1; // Inclusive. + optional int32 end = 2; // Exclusive. + + optional ExtensionRangeOptions options = 3; + } + repeated ExtensionRange extension_range = 5; + + repeated OneofDescriptorProto oneof_decl = 8; + + optional MessageOptions options = 7; + + // Range of reserved tag numbers. Reserved tag numbers may not be used by + // fields or extension ranges in the same message. Reserved ranges may + // not overlap. + message ReservedRange { + optional int32 start = 1; // Inclusive. + optional int32 end = 2; // Exclusive. + } + repeated ReservedRange reserved_range = 9; + // Reserved field names, which may not be used by fields in the same message. + // A given name may only be reserved once. + repeated string reserved_name = 10; +} + +message ExtensionRangeOptions { + // The parser stores options it doesn't recognize here. See above. + repeated UninterpretedOption uninterpreted_option = 999; + + + // Clients can define custom options in extensions of this message. See above. + extensions 1000 to max; +} + +// Describes a field within a message. +message FieldDescriptorProto { + enum Type { + // 0 is reserved for errors. + // Order is weird for historical reasons. + TYPE_DOUBLE = 1; + TYPE_FLOAT = 2; + // Not ZigZag encoded. Negative numbers take 10 bytes. Use TYPE_SINT64 if + // negative values are likely. + TYPE_INT64 = 3; + TYPE_UINT64 = 4; + // Not ZigZag encoded. Negative numbers take 10 bytes. Use TYPE_SINT32 if + // negative values are likely. + TYPE_INT32 = 5; + TYPE_FIXED64 = 6; + TYPE_FIXED32 = 7; + TYPE_BOOL = 8; + TYPE_STRING = 9; + // Tag-delimited aggregate. + // Group type is deprecated and not supported in proto3. However, Proto3 + // implementations should still be able to parse the group wire format and + // treat group fields as unknown fields. + TYPE_GROUP = 10; + TYPE_MESSAGE = 11; // Length-delimited aggregate. + + // New in version 2. + TYPE_BYTES = 12; + TYPE_UINT32 = 13; + TYPE_ENUM = 14; + TYPE_SFIXED32 = 15; + TYPE_SFIXED64 = 16; + TYPE_SINT32 = 17; // Uses ZigZag encoding. + TYPE_SINT64 = 18; // Uses ZigZag encoding. + } + + enum Label { + // 0 is reserved for errors + LABEL_OPTIONAL = 1; + LABEL_REQUIRED = 2; + LABEL_REPEATED = 3; + } + + optional string name = 1; + optional int32 number = 3; + optional Label label = 4; + + // If type_name is set, this need not be set. If both this and type_name + // are set, this must be one of TYPE_ENUM, TYPE_MESSAGE or TYPE_GROUP. + optional Type type = 5; + + // For message and enum types, this is the name of the type. If the name + // starts with a '.', it is fully-qualified. Otherwise, C++-like scoping + // rules are used to find the type (i.e. first the nested types within this + // message are searched, then within the parent, on up to the root + // namespace). + optional string type_name = 6; + + // For extensions, this is the name of the type being extended. It is + // resolved in the same manner as type_name. + optional string extendee = 2; + + // For numeric types, contains the original text representation of the value. + // For booleans, "true" or "false". + // For strings, contains the default text contents (not escaped in any way). + // For bytes, contains the C escaped value. All bytes >= 128 are escaped. + optional string default_value = 7; + + // If set, gives the index of a oneof in the containing type's oneof_decl + // list. This field is a member of that oneof. + optional int32 oneof_index = 9; + + // JSON name of this field. The value is set by protocol compiler. If the + // user has set a "json_name" option on this field, that option's value + // will be used. Otherwise, it's deduced from the field's name by converting + // it to camelCase. + optional string json_name = 10; + + optional FieldOptions options = 8; + + // If true, this is a proto3 "optional". When a proto3 field is optional, it + // tracks presence regardless of field type. + // + // When proto3_optional is true, this field must be belong to a oneof to + // signal to old proto3 clients that presence is tracked for this field. This + // oneof is known as a "synthetic" oneof, and this field must be its sole + // member (each proto3 optional field gets its own synthetic oneof). Synthetic + // oneofs exist in the descriptor only, and do not generate any API. Synthetic + // oneofs must be ordered after all "real" oneofs. + // + // For message fields, proto3_optional doesn't create any semantic change, + // since non-repeated message fields always track presence. However it still + // indicates the semantic detail of whether the user wrote "optional" or not. + // This can be useful for round-tripping the .proto file. For consistency we + // give message fields a synthetic oneof also, even though it is not required + // to track presence. This is especially important because the parser can't + // tell if a field is a message or an enum, so it must always create a + // synthetic oneof. + // + // Proto2 optional fields do not set this flag, because they already indicate + // optional with `LABEL_OPTIONAL`. + optional bool proto3_optional = 17; +} + +// Describes a oneof. +message OneofDescriptorProto { + optional string name = 1; + optional OneofOptions options = 2; +} + +// Describes an enum type. +message EnumDescriptorProto { + optional string name = 1; + + repeated EnumValueDescriptorProto value = 2; + + optional EnumOptions options = 3; + + // Range of reserved numeric values. Reserved values may not be used by + // entries in the same enum. Reserved ranges may not overlap. + // + // Note that this is distinct from DescriptorProto.ReservedRange in that it + // is inclusive such that it can appropriately represent the entire int32 + // domain. + message EnumReservedRange { + optional int32 start = 1; // Inclusive. + optional int32 end = 2; // Inclusive. + } + + // Range of reserved numeric values. Reserved numeric values may not be used + // by enum values in the same enum declaration. Reserved ranges may not + // overlap. + repeated EnumReservedRange reserved_range = 4; + + // Reserved enum value names, which may not be reused. A given name may only + // be reserved once. + repeated string reserved_name = 5; +} + +// Describes a value within an enum. +message EnumValueDescriptorProto { + optional string name = 1; + optional int32 number = 2; + + optional EnumValueOptions options = 3; +} + +// Describes a service. +message ServiceDescriptorProto { + optional string name = 1; + repeated MethodDescriptorProto method = 2; + + optional ServiceOptions options = 3; +} + +// Describes a method of a service. +message MethodDescriptorProto { + optional string name = 1; + + // Input and output type names. These are resolved in the same way as + // FieldDescriptorProto.type_name, but must refer to a message type. + optional string input_type = 2; + optional string output_type = 3; + + optional MethodOptions options = 4; + + // Identifies if client streams multiple client messages + optional bool client_streaming = 5 [default = false]; + // Identifies if server streams multiple server messages + optional bool server_streaming = 6 [default = false]; +} + + +// =================================================================== +// Options + +// Each of the definitions above may have "options" attached. These are +// just annotations which may cause code to be generated slightly differently +// or may contain hints for code that manipulates protocol messages. +// +// Clients may define custom options as extensions of the *Options messages. +// These extensions may not yet be known at parsing time, so the parser cannot +// store the values in them. Instead it stores them in a field in the *Options +// message called uninterpreted_option. This field must have the same name +// across all *Options messages. We then use this field to populate the +// extensions when we build a descriptor, at which point all protos have been +// parsed and so all extensions are known. +// +// Extension numbers for custom options may be chosen as follows: +// * For options which will only be used within a single application or +// organization, or for experimental options, use field numbers 50000 +// through 99999. It is up to you to ensure that you do not use the +// same number for multiple options. +// * For options which will be published and used publicly by multiple +// independent entities, e-mail protobuf-global-extension-registry@google.com +// to reserve extension numbers. Simply provide your project name (e.g. +// Objective-C plugin) and your project website (if available) -- there's no +// need to explain how you intend to use them. Usually you only need one +// extension number. You can declare multiple options with only one extension +// number by putting them in a sub-message. See the Custom Options section of +// the docs for examples: +// https://developers.google.com/protocol-buffers/docs/proto#options +// If this turns out to be popular, a web service will be set up +// to automatically assign option numbers. + +message FileOptions { + + // Sets the Java package where classes generated from this .proto will be + // placed. By default, the proto package is used, but this is often + // inappropriate because proto packages do not normally start with backwards + // domain names. + optional string java_package = 1; + + + // Controls the name of the wrapper Java class generated for the .proto file. + // That class will always contain the .proto file's getDescriptor() method as + // well as any top-level extensions defined in the .proto file. + // If java_multiple_files is disabled, then all the other classes from the + // .proto file will be nested inside the single wrapper outer class. + optional string java_outer_classname = 8; + + // If enabled, then the Java code generator will generate a separate .java + // file for each top-level message, enum, and service defined in the .proto + // file. Thus, these types will *not* be nested inside the wrapper class + // named by java_outer_classname. However, the wrapper class will still be + // generated to contain the file's getDescriptor() method as well as any + // top-level extensions defined in the file. + optional bool java_multiple_files = 10 [default = false]; + + // This option does nothing. + optional bool java_generate_equals_and_hash = 20 [deprecated=true]; + + // If set true, then the Java2 code generator will generate code that + // throws an exception whenever an attempt is made to assign a non-UTF-8 + // byte sequence to a string field. + // Message reflection will do the same. + // However, an extension field still accepts non-UTF-8 byte sequences. + // This option has no effect on when used with the lite runtime. + optional bool java_string_check_utf8 = 27 [default = false]; + + + // Generated classes can be optimized for speed or code size. + enum OptimizeMode { + SPEED = 1; // Generate complete code for parsing, serialization, + // etc. + CODE_SIZE = 2; // Use ReflectionOps to implement these methods. + LITE_RUNTIME = 3; // Generate code using MessageLite and the lite runtime. + } + optional OptimizeMode optimize_for = 9 [default = SPEED]; + + // Sets the Go package where structs generated from this .proto will be + // placed. If omitted, the Go package will be derived from the following: + // - The basename of the package import path, if provided. + // - Otherwise, the package statement in the .proto file, if present. + // - Otherwise, the basename of the .proto file, without extension. + optional string go_package = 11; + + + + + // Should generic services be generated in each language? "Generic" services + // are not specific to any particular RPC system. They are generated by the + // main code generators in each language (without additional plugins). + // Generic services were the only kind of service generation supported by + // early versions of google.protobuf. + // + // Generic services are now considered deprecated in favor of using plugins + // that generate code specific to your particular RPC system. Therefore, + // these default to false. Old code which depends on generic services should + // explicitly set them to true. + optional bool cc_generic_services = 16 [default = false]; + optional bool java_generic_services = 17 [default = false]; + optional bool py_generic_services = 18 [default = false]; + optional bool php_generic_services = 42 [default = false]; + + // Is this file deprecated? + // Depending on the target platform, this can emit Deprecated annotations + // for everything in the file, or it will be completely ignored; in the very + // least, this is a formalization for deprecating files. + optional bool deprecated = 23 [default = false]; + + // Enables the use of arenas for the proto messages in this file. This applies + // only to generated classes for C++. + optional bool cc_enable_arenas = 31 [default = true]; + + + // Sets the objective c class prefix which is prepended to all objective c + // generated classes from this .proto. There is no default. + optional string objc_class_prefix = 36; + + // Namespace for generated classes; defaults to the package. + optional string csharp_namespace = 37; + + // By default Swift generators will take the proto package and CamelCase it + // replacing '.' with underscore and use that to prefix the types/symbols + // defined. When this options is provided, they will use this value instead + // to prefix the types/symbols defined. + optional string swift_prefix = 39; + + // Sets the php class prefix which is prepended to all php generated classes + // from this .proto. Default is empty. + optional string php_class_prefix = 40; + + // Use this option to change the namespace of php generated classes. Default + // is empty. When this option is empty, the package name will be used for + // determining the namespace. + optional string php_namespace = 41; + + // Use this option to change the namespace of php generated metadata classes. + // Default is empty. When this option is empty, the proto file name will be + // used for determining the namespace. + optional string php_metadata_namespace = 44; + + // Use this option to change the package of ruby generated classes. Default + // is empty. When this option is not set, the package name will be used for + // determining the ruby package. + optional string ruby_package = 45; + + + // The parser stores options it doesn't recognize here. + // See the documentation for the "Options" section above. + repeated UninterpretedOption uninterpreted_option = 999; + + // Clients can define custom options in extensions of this message. + // See the documentation for the "Options" section above. + extensions 1000 to max; + + reserved 38; +} + +message MessageOptions { + // Set true to use the old proto1 MessageSet wire format for extensions. + // This is provided for backwards-compatibility with the MessageSet wire + // format. You should not use this for any other reason: It's less + // efficient, has fewer features, and is more complicated. + // + // The message must be defined exactly as follows: + // message Foo { + // option message_set_wire_format = true; + // extensions 4 to max; + // } + // Note that the message cannot have any defined fields; MessageSets only + // have extensions. + // + // All extensions of your type must be singular messages; e.g. they cannot + // be int32s, enums, or repeated messages. + // + // Because this is an option, the above two restrictions are not enforced by + // the protocol compiler. + optional bool message_set_wire_format = 1 [default = false]; + + // Disables the generation of the standard "descriptor()" accessor, which can + // conflict with a field of the same name. This is meant to make migration + // from proto1 easier; new code should avoid fields named "descriptor". + optional bool no_standard_descriptor_accessor = 2 [default = false]; + + // Is this message deprecated? + // Depending on the target platform, this can emit Deprecated annotations + // for the message, or it will be completely ignored; in the very least, + // this is a formalization for deprecating messages. + optional bool deprecated = 3 [default = false]; + + reserved 4, 5, 6; + + // Whether the message is an automatically generated map entry type for the + // maps field. + // + // For maps fields: + // map map_field = 1; + // The parsed descriptor looks like: + // message MapFieldEntry { + // option map_entry = true; + // optional KeyType key = 1; + // optional ValueType value = 2; + // } + // repeated MapFieldEntry map_field = 1; + // + // Implementations may choose not to generate the map_entry=true message, but + // use a native map in the target language to hold the keys and values. + // The reflection APIs in such implementations still need to work as + // if the field is a repeated message field. + // + // NOTE: Do not set the option in .proto files. Always use the maps syntax + // instead. The option should only be implicitly set by the proto compiler + // parser. + optional bool map_entry = 7; + + reserved 8; // javalite_serializable + reserved 9; // javanano_as_lite + + + // The parser stores options it doesn't recognize here. See above. + repeated UninterpretedOption uninterpreted_option = 999; + + // Clients can define custom options in extensions of this message. See above. + extensions 1000 to max; +} + +message FieldOptions { + // The ctype option instructs the C++ code generator to use a different + // representation of the field than it normally would. See the specific + // options below. This option is not yet implemented in the open source + // release -- sorry, we'll try to include it in a future version! + optional CType ctype = 1 [default = STRING]; + enum CType { + // Default mode. + STRING = 0; + + CORD = 1; + + STRING_PIECE = 2; + } + // The packed option can be enabled for repeated primitive fields to enable + // a more efficient representation on the wire. Rather than repeatedly + // writing the tag and type for each element, the entire array is encoded as + // a single length-delimited blob. In proto3, only explicit setting it to + // false will avoid using packed encoding. + optional bool packed = 2; + + // The jstype option determines the JavaScript type used for values of the + // field. The option is permitted only for 64 bit integral and fixed types + // (int64, uint64, sint64, fixed64, sfixed64). A field with jstype JS_STRING + // is represented as JavaScript string, which avoids loss of precision that + // can happen when a large value is converted to a floating point JavaScript. + // Specifying JS_NUMBER for the jstype causes the generated JavaScript code to + // use the JavaScript "number" type. The behavior of the default option + // JS_NORMAL is implementation dependent. + // + // This option is an enum to permit additional types to be added, e.g. + // goog.math.Integer. + optional JSType jstype = 6 [default = JS_NORMAL]; + enum JSType { + // Use the default type. + JS_NORMAL = 0; + + // Use JavaScript strings. + JS_STRING = 1; + + // Use JavaScript numbers. + JS_NUMBER = 2; + } + + // Should this field be parsed lazily? Lazy applies only to message-type + // fields. It means that when the outer message is initially parsed, the + // inner message's contents will not be parsed but instead stored in encoded + // form. The inner message will actually be parsed when it is first accessed. + // + // This is only a hint. Implementations are free to choose whether to use + // eager or lazy parsing regardless of the value of this option. However, + // setting this option true suggests that the protocol author believes that + // using lazy parsing on this field is worth the additional bookkeeping + // overhead typically needed to implement it. + // + // This option does not affect the public interface of any generated code; + // all method signatures remain the same. Furthermore, thread-safety of the + // interface is not affected by this option; const methods remain safe to + // call from multiple threads concurrently, while non-const methods continue + // to require exclusive access. + // + // + // Note that implementations may choose not to check required fields within + // a lazy sub-message. That is, calling IsInitialized() on the outer message + // may return true even if the inner message has missing required fields. + // This is necessary because otherwise the inner message would have to be + // parsed in order to perform the check, defeating the purpose of lazy + // parsing. An implementation which chooses not to check required fields + // must be consistent about it. That is, for any particular sub-message, the + // implementation must either *always* check its required fields, or *never* + // check its required fields, regardless of whether or not the message has + // been parsed. + // + // As of 2021, lazy does no correctness checks on the byte stream during + // parsing. This may lead to crashes if and when an invalid byte stream is + // finally parsed upon access. + // + // TODO(b/211906113): Enable validation on lazy fields. + optional bool lazy = 5 [default = false]; + + // unverified_lazy does no correctness checks on the byte stream. This should + // only be used where lazy with verification is prohibitive for performance + // reasons. + optional bool unverified_lazy = 15 [default = false]; + + // Is this field deprecated? + // Depending on the target platform, this can emit Deprecated annotations + // for accessors, or it will be completely ignored; in the very least, this + // is a formalization for deprecating fields. + optional bool deprecated = 3 [default = false]; + + // For Google-internal migration only. Do not use. + optional bool weak = 10 [default = false]; + + + // The parser stores options it doesn't recognize here. See above. + repeated UninterpretedOption uninterpreted_option = 999; + + // Clients can define custom options in extensions of this message. See above. + extensions 1000 to max; + + reserved 4; // removed jtype +} + +message OneofOptions { + // The parser stores options it doesn't recognize here. See above. + repeated UninterpretedOption uninterpreted_option = 999; + + // Clients can define custom options in extensions of this message. See above. + extensions 1000 to max; +} + +message EnumOptions { + + // Set this option to true to allow mapping different tag names to the same + // value. + optional bool allow_alias = 2; + + // Is this enum deprecated? + // Depending on the target platform, this can emit Deprecated annotations + // for the enum, or it will be completely ignored; in the very least, this + // is a formalization for deprecating enums. + optional bool deprecated = 3 [default = false]; + + reserved 5; // javanano_as_lite + + // The parser stores options it doesn't recognize here. See above. + repeated UninterpretedOption uninterpreted_option = 999; + + // Clients can define custom options in extensions of this message. See above. + extensions 1000 to max; +} + +message EnumValueOptions { + // Is this enum value deprecated? + // Depending on the target platform, this can emit Deprecated annotations + // for the enum value, or it will be completely ignored; in the very least, + // this is a formalization for deprecating enum values. + optional bool deprecated = 1 [default = false]; + + // The parser stores options it doesn't recognize here. See above. + repeated UninterpretedOption uninterpreted_option = 999; + + // Clients can define custom options in extensions of this message. See above. + extensions 1000 to max; +} + +message ServiceOptions { + + // Note: Field numbers 1 through 32 are reserved for Google's internal RPC + // framework. We apologize for hoarding these numbers to ourselves, but + // we were already using them long before we decided to release Protocol + // Buffers. + + // Is this service deprecated? + // Depending on the target platform, this can emit Deprecated annotations + // for the service, or it will be completely ignored; in the very least, + // this is a formalization for deprecating services. + optional bool deprecated = 33 [default = false]; + + // The parser stores options it doesn't recognize here. See above. + repeated UninterpretedOption uninterpreted_option = 999; + + // Clients can define custom options in extensions of this message. See above. + extensions 1000 to max; +} + +message MethodOptions { + + // Note: Field numbers 1 through 32 are reserved for Google's internal RPC + // framework. We apologize for hoarding these numbers to ourselves, but + // we were already using them long before we decided to release Protocol + // Buffers. + + // Is this method deprecated? + // Depending on the target platform, this can emit Deprecated annotations + // for the method, or it will be completely ignored; in the very least, + // this is a formalization for deprecating methods. + optional bool deprecated = 33 [default = false]; + + // Is this method side-effect-free (or safe in HTTP parlance), or idempotent, + // or neither? HTTP based RPC implementation may choose GET verb for safe + // methods, and PUT verb for idempotent methods instead of the default POST. + enum IdempotencyLevel { + IDEMPOTENCY_UNKNOWN = 0; + NO_SIDE_EFFECTS = 1; // implies idempotent + IDEMPOTENT = 2; // idempotent, but may have side effects + } + optional IdempotencyLevel idempotency_level = 34 + [default = IDEMPOTENCY_UNKNOWN]; + + // The parser stores options it doesn't recognize here. See above. + repeated UninterpretedOption uninterpreted_option = 999; + + // Clients can define custom options in extensions of this message. See above. + extensions 1000 to max; +} + + +// A message representing a option the parser does not recognize. This only +// appears in options protos created by the compiler::Parser class. +// DescriptorPool resolves these when building Descriptor objects. Therefore, +// options protos in descriptor objects (e.g. returned by Descriptor::options(), +// or produced by Descriptor::CopyTo()) will never have UninterpretedOptions +// in them. +message UninterpretedOption { + // The name of the uninterpreted option. Each string represents a segment in + // a dot-separated name. is_extension is true iff a segment represents an + // extension (denoted with parentheses in options specs in .proto files). + // E.g.,{ ["foo", false], ["bar.baz", true], ["qux", false] } represents + // "foo.(bar.baz).qux". + message NamePart { + required string name_part = 1; + required bool is_extension = 2; + } + repeated NamePart name = 2; + + // The value of the uninterpreted option, in whatever type the tokenizer + // identified it as during parsing. Exactly one of these should be set. + optional string identifier_value = 3; + optional uint64 positive_int_value = 4; + optional int64 negative_int_value = 5; + optional double double_value = 6; + optional bytes string_value = 7; + optional string aggregate_value = 8; +} + +// =================================================================== +// Optional source code info + +// Encapsulates information about the original source file from which a +// FileDescriptorProto was generated. +message SourceCodeInfo { + // A Location identifies a piece of source code in a .proto file which + // corresponds to a particular definition. This information is intended + // to be useful to IDEs, code indexers, documentation generators, and similar + // tools. + // + // For example, say we have a file like: + // message Foo { + // optional string foo = 1; + // } + // Let's look at just the field definition: + // optional string foo = 1; + // ^ ^^ ^^ ^ ^^^ + // a bc de f ghi + // We have the following locations: + // span path represents + // [a,i) [ 4, 0, 2, 0 ] The whole field definition. + // [a,b) [ 4, 0, 2, 0, 4 ] The label (optional). + // [c,d) [ 4, 0, 2, 0, 5 ] The type (string). + // [e,f) [ 4, 0, 2, 0, 1 ] The name (foo). + // [g,h) [ 4, 0, 2, 0, 3 ] The number (1). + // + // Notes: + // - A location may refer to a repeated field itself (i.e. not to any + // particular index within it). This is used whenever a set of elements are + // logically enclosed in a single code segment. For example, an entire + // extend block (possibly containing multiple extension definitions) will + // have an outer location whose path refers to the "extensions" repeated + // field without an index. + // - Multiple locations may have the same path. This happens when a single + // logical declaration is spread out across multiple places. The most + // obvious example is the "extend" block again -- there may be multiple + // extend blocks in the same scope, each of which will have the same path. + // - A location's span is not always a subset of its parent's span. For + // example, the "extendee" of an extension declaration appears at the + // beginning of the "extend" block and is shared by all extensions within + // the block. + // - Just because a location's span is a subset of some other location's span + // does not mean that it is a descendant. For example, a "group" defines + // both a type and a field in a single declaration. Thus, the locations + // corresponding to the type and field and their components will overlap. + // - Code which tries to interpret locations should probably be designed to + // ignore those that it doesn't understand, as more types of locations could + // be recorded in the future. + repeated Location location = 1; + message Location { + // Identifies which part of the FileDescriptorProto was defined at this + // location. + // + // Each element is a field number or an index. They form a path from + // the root FileDescriptorProto to the place where the definition occurs. For + // example, this path: + // [ 4, 3, 2, 7, 1 ] + // refers to: + // file.message_type(3) // 4, 3 + // .field(7) // 2, 7 + // .name() // 1 + // This is because FileDescriptorProto.message_type has field number 4: + // repeated DescriptorProto message_type = 4; + // and DescriptorProto.field has field number 2: + // repeated FieldDescriptorProto field = 2; + // and FieldDescriptorProto.name has field number 1: + // optional string name = 1; + // + // Thus, the above path gives the location of a field name. If we removed + // the last element: + // [ 4, 3, 2, 7 ] + // this path refers to the whole field declaration (from the beginning + // of the label to the terminating semicolon). + repeated int32 path = 1 [packed = true]; + + // Always has exactly three or four elements: start line, start column, + // end line (optional, otherwise assumed same as start line), end column. + // These are packed into a single field for efficiency. Note that line + // and column numbers are zero-based -- typically you will want to add + // 1 to each before displaying to a user. + repeated int32 span = 2 [packed = true]; + + // If this SourceCodeInfo represents a complete declaration, these are any + // comments appearing before and after the declaration which appear to be + // attached to the declaration. + // + // A series of line comments appearing on consecutive lines, with no other + // tokens appearing on those lines, will be treated as a single comment. + // + // leading_detached_comments will keep paragraphs of comments that appear + // before (but not connected to) the current element. Each paragraph, + // separated by empty lines, will be one comment element in the repeated + // field. + // + // Only the comment content is provided; comment markers (e.g. //) are + // stripped out. For block comments, leading whitespace and an asterisk + // will be stripped from the beginning of each line other than the first. + // Newlines are included in the output. + // + // Examples: + // + // optional int32 foo = 1; // Comment attached to foo. + // // Comment attached to bar. + // optional int32 bar = 2; + // + // optional string baz = 3; + // // Comment attached to baz. + // // Another line attached to baz. + // + // // Comment attached to qux. + // // + // // Another line attached to qux. + // optional double qux = 4; + // + // // Detached comment for corge. This is not leading or trailing comments + // // to qux or corge because there are blank lines separating it from + // // both. + // + // // Detached comment for corge paragraph 2. + // + // optional string corge = 5; + // /* Block comment attached + // * to corge. Leading asterisks + // * will be removed. */ + // /* Block comment attached to + // * grault. */ + // optional int32 grault = 6; + // + // // ignored detached comments. + optional string leading_comments = 3; + optional string trailing_comments = 4; + repeated string leading_detached_comments = 6; + } +} + +// Describes the relationship between generated code and its original source +// file. A GeneratedCodeInfo message is associated with only one generated +// source file, but may contain references to different source .proto files. +message GeneratedCodeInfo { + // An Annotation connects some span of text in generated code to an element + // of its generating .proto file. + repeated Annotation annotation = 1; + message Annotation { + // Identifies the element in the original source .proto file. This field + // is formatted the same as SourceCodeInfo.Location.path. + repeated int32 path = 1 [packed = true]; + + // Identifies the filesystem path to the original source .proto. + optional string source_file = 2; + + // Identifies the starting offset in bytes in the generated code + // that relates to the identified object. + optional int32 begin = 3; + + // Identifies the ending offset in bytes in the generated code that + // relates to the identified offset. The end offset should be one past + // the last relevant byte (so the length of the text = end - begin). + optional int32 end = 4; + } +} diff --git a/registry/testdata/regtest.pb b/registry/testdata/regtest.pb new file mode 100644 index 0000000000000000000000000000000000000000..8e5042f69315c93786e595ca387541cfe578f2ac GIT binary patch literal 9425 zcmc&4OKcn0F{DU|e3YrxO15d)i5=!+$4RJ$PU2K9(vs3zrY(u=Qfd+d#bUL5q}E>U zGP_IL4BDb-Pd&E16-5udq$qmmsX)JIuRRnfie7pu&;U()?WM=g%-dZqxm2tIb*h7~ zZ{EzjdGqG`9@$T_>~zm@dKO>T2j=>b?++>i*YO>h32+6#Ta>|%FlStG5(4X493mwViOz&jd$G|+9&bnH_4QtC=hy+CMN zgrx>ZN{CbbOPb|E*76$5dEDZL@3PlXX#2%Rw{+jBUZ}~ZP3iLB@Y>#U+xk!_XA+QbvpwSh*(=mNe zoo;D|rrj~^o>wY-!db{;f_i{o&$9*C5DHbk!Il$}qI~xOH&8HrBMfMcgmzW>uHR!t z!|BI%UC+~o(Rx~`yS2bJdXA;ry^7=Z)_dG0+gcY>U|Aknoo?HXuj9VF8?o##=1Y z1z=#x&xiX6i7Jy6E!>q#6y=XHY-PqQD#_OS$dL{fzrn6R*~Wdn#{=DkL-`rTI()$G z4!4aFnCKiBLkuFn%$5hkL(4S4fg&^pXjxSf6I&5nzRZ>$a{YcB70n{677-jn++>S= z?s*`O=8py(yZ{~fQeyXGU5T9+0;+u=l4~r_?O|VV$P$O7{uIEN=8{9QfW-oa1JmH( zS17{ega~ahKg}bglQ%(%M?M#!WPzBxkYIAtOjy;Dp8!o2^qwcSYrj7dp^W=vMIFXlh_m!L`jnpM)}{G z)VpjcvIotDn1it#C0S3L$dp^|Qn%KZemo9&Wou0{BewVEdZPSMAH*g`_cQoPsrN4^+lC-&lIJOW_U727ZfLHUm zj~Aq`!(Y1myLq-cEAY&8UvPg2iig~#G9{}9BM7otOF!fm6)MH*%L(1hTA{LoB&wj4 z-)0$dNGxfJ)vG7tfbOUgoqV3^0f!L zHRM69vlsxsgfRIcD+sTGVuL@T=$Tc8qcsqM%ir@HTP)9D4^ZwQfR?{LHm<4qE0)0_ zDo)!Jrb@!N79729m6m~bPAv+sAkeMy&(bV|%8-{>VgJ3BqHQ$~>N|>*l8cNH;O0)V zwlAflabs`)&DW*5h~Pi~GI4}!RVfR;vM3ld-c`1MWPTC?I1BPAmM5Hgv$-SXBA)Gi zwXwG?y$QPDM*7dk4ExQC?=JXvOwT z6bmqxra&O66me+_F3S@Dz>v$EEK6nx=3d;X)s-Es+1hV3_i8(!y9lN#Up#22$`t32#D1 zWZBa>K+wfY4G|8Lr3SK?7E=ClBKV8^_~uLrrWUUt%ZV7OJ}wDh(wt5L&?qDroj4eO zGsQ~BDQBjt#~~4qa~u+HjD22c@Y$IvphK8eWBJ1rJ2ShaoGHWGY`)L^BPT4YlDTV_ zrl<6+22}BEFJw(-`#_q6q~A)j(=(aR6!dAvOdA?oQG21n@@k$S6udCX4*h5zWr03J zfC%BOu~M-mE>xdMC{+kEEvNh$CK;B=ZC~?zm+O7#MyY$wZDhN;<#BaMOt-^Xq@c>e zjZt7;8B>Vqh@$-MJS$+Y6)Y}4%NFnH4|FYTszk3=bAXR8P3RWuoa@{L&7o#7T2&b^h%g=BFUB!V>duRi99F^pC&SBYq_iYT;uN# zVc(i=ceEqjJ1VUrgL=A@QqLjcc7Uf4TFvg<0wj5ZogpruLr|b*9C72mHuSr1mCnc4 zCka6>fU}7x2Y?~B*`foTN#FbcWQIEt>_CduWN3Sha5PBLU5(SpfHF<^0(IstH{eV=rL%E#Y|w)VxdP^I zXx!Eh!D4h5KD^RJLeBW^5V{D%P-wD7b8vaK>>S=Rghgt=rfWVbJr4wVbqOcP9$N&K zUjp9Vk?szt`0>EndFcg#D&SZH5g0wfduVq3kmF@SD-bN;c7;i(337!H?Fyk40Ulo$ z3OB?=e-VqmL}-fu9IyB~JA+6N9`%my>oM$C2piQ7A%iBzR3N82JRC)a^{S8)VWDX~ zvEsbE!4~6oFXx%Cd?-ku&f08lDO$U6SAkjtO0%8DzM`oId;5)DMVgBh_O~;+7p2Qs zH54a1HTiSwENG0bRXpx%4^5XGvagFKDT*Si0Tx_9y#qMlZh#{1u#2_>ArS(H?slNd z3@@}a-2mV6z_aP9H1v`Sw$l!f$LXgQpvsfxivd4cK5o8HUHA0?*st%3yU3iH1Av0# zPo+%)ZDyn_{Lf1B@PA%ffd4ru5C8KLD}N`$7OBO-gM*tvY={i(KB(<1R-a3xsW^6J z6O-GT*QlWpJx@7rx+p1w&S9J@8|^W(jTLiA9Xq(&gW#SRq!KW8+-&3bVy~_Wvs0U(@W=L=`-B#>YzcK3g_BygzW@ zqS_c~7JmRYI%Ntn$!p1p6)KG}-42p%tTeWi-Bxp7+1q?iJJ`Fk*L-JBm167(?FqVl zi%HW$$!FOtvB0xe*wS7TvMQuj<>pOgbH6QaPNGe~%}LzLzBk8KW?Z&G*cY!VqV0Y) z5%p)#RIuu3!40KKq*zUXmTVtGuN&M*iAHE?3>TMDuP_;22jPwF0oUM~FhIt}CH)9PY|sf5UvzDohMF^n08|oIkfsOEvf*1w`cfLM z!wRMsU7?mggpz0S83VeXoaI1?47oXYL%WhVn@N(&PJpjQlxu%Tv0Mm-^ekQ(@pU;{ zPfIEB(ui=VM)AEkg5wy^)#5t7g2F3IzYizuiML*LDS&SVV0=yUUEQ=MaOnwHmBNV# zd4rt``F5a=g_f){UH~6S^6J?DVJnyqY0H01v1Mu~uz85ycy<{ZKWql-N}6(lu2PF$ zM%5UDwLW9xnPb%kc}hGrQ9nz03R-^w!j2Fxa1MFTw1b;xF@Wq%7x}0hagksD8nk@l zm)`JBtuBp=&lWz9dbd({UroOHhWP7PV0p3U+QA{c@Xl?w>OW4eZPTw1D#Qt%bt@|G z;!8i67yD6~T?k%%@ka_{O?ttLZ|t`UYqNjOu=dzz8HQ{6$bF* ze*QpA>UQFBTpFL4N1Obgypc~okpCvd&IhWB^bYCAA078LzQD2y&ey*>`;ZyUh60Bk z9(&Y}mJ!E$-*z6_>$K#=`93p!zVC3$x`W7AOMB4^|F1LnQ;41q?}>rP3&C*h6K&w} z2fX%4iYZ@Zg}UzXpszzGIPAVgtwvtO&v>f(1y)mGxC|-a^`G}668c`-gB(Z?qY*Y; z;qft=4d2zv;Bnn+@@ggAcubPgKgI7v5Tja0#Og7K{}8W==OQlAqJfC;D%Y)6@38FF zu-_la_3-0Nc&g!1R_M&vj)9l+-}uu<-%ZbLq^@p|u?+C2pTB#q>pU9uoQm$R-!psbgTr;uoC#TS^dCVqypsR` literal 0 HcmV?d00001 diff --git a/registry/testdata/regtest.proto b/registry/testdata/regtest.proto new file mode 100644 index 0000000..a9ebe60 --- /dev/null +++ b/registry/testdata/regtest.proto @@ -0,0 +1,51 @@ +syntax = "proto2"; + +package regtest; + +option go_package = "foxygo.at/jig/pb/registry"; + +import "google/api/annotations.proto"; +import "google/protobuf/descriptor.proto"; +import "google/protobuf/empty.proto"; + +// A base message to be extended +message BaseMessage { + optional string bf1 = 1; + extensions 1000 to max; +} + +// A simple top-level extension +extend BaseMessage { + optional string ef1 = 1000; +} + +// A message scope for more extensions +message ExtensionMessage { + optional string emf1 = 1; + + // An extension scoped within a message + extend BaseMessage { + optional ExtensionMessage ef2 = 1001; + }; + + message NestedExtension { + optional string nef1 = 1; + + // An extension scoped within a nested message + extend BaseMessage { + optional string ef3 = 1002; + }; + }; +} + +// Extend a message outside this package +extend google.protobuf.MethodOptions { + optional BaseMessage base = 56789; +}; + +// A dummy service so that annotations.proto is not reported as unused +service Dummy { + rpc Dummy(google.protobuf.Empty) returns (google.protobuf.Empty) { + option (google.api.http) = { body:"*" }; + }; +} From a9f5c83e6e2a2e8e97c0e3d0cb6acc2611ff2a4b Mon Sep 17 00:00:00 2001 From: Julia Ogris Date: Tue, 15 Feb 2022 14:30:03 +1100 Subject: [PATCH 4/4] Refactor files.go Refactor files.go in a functionally identical way, without changing any public identifiers or signatures. Inline the walkExtenstion function as it is only used in one place and make walkFiles a receiver for easier readability. Signed-off-by: Julia Ogris --- registry/files.go | 30 ++++++++++++++++-------------- registry/files_test.go | 5 ++--- 2 files changed, 18 insertions(+), 17 deletions(-) diff --git a/registry/files.go b/registry/files.go index 9b7ca6c..32a6596 100644 --- a/registry/files.go +++ b/registry/files.go @@ -7,8 +7,10 @@ package registry import ( "strings" + "google.golang.org/protobuf/reflect/protodesc" "google.golang.org/protobuf/reflect/protoreflect" "google.golang.org/protobuf/reflect/protoregistry" + "google.golang.org/protobuf/types/descriptorpb" "google.golang.org/protobuf/types/dynamicpb" ) @@ -16,8 +18,12 @@ type Files struct { protoregistry.Files } -func NewFiles(f *protoregistry.Files) *Files { - return &Files{Files: *f} +func NewFiles(fds *descriptorpb.FileDescriptorSet) (*Files, error) { + f, err := protodesc.NewFiles(fds) + if err != nil { + return nil, err + } + return &Files{Files: *f}, nil } type extMatchFn func(protoreflect.ExtensionDescriptor) bool @@ -42,29 +48,25 @@ func (f *Files) FindExtensionByName(field protoreflect.FullName) (protoreflect.E } func (f *Files) FindExtensionByNumber(message protoreflect.FullName, field protoreflect.FieldNumber) (protoreflect.ExtensionType, error) { - return findExtension(&f.Files, func(ed protoreflect.ExtensionDescriptor) bool { + ets := f.walkExtensions(false, func(ed protoreflect.ExtensionDescriptor) bool { return ed.ContainingMessage().FullName() == message && ed.Number() == field }) + if len(ets) == 0 { + return nil, protoregistry.NotFound + } + return ets[0], nil } func (f *Files) GetExtensionsOfMessage(message protoreflect.FullName) []protoreflect.ExtensionType { - return walkExtensions(&f.Files, true, func(ed protoreflect.ExtensionDescriptor) bool { + return f.walkExtensions(true, func(ed protoreflect.ExtensionDescriptor) bool { return ed.ContainingMessage().FullName() == message }) } -func findExtension(files *protoregistry.Files, pred extMatchFn) (protoreflect.ExtensionType, error) { - ets := walkExtensions(files, false, pred) - if len(ets) == 0 { - return nil, protoregistry.NotFound - } - return ets[0], nil -} - -func walkExtensions(files *protoregistry.Files, getAll bool, pred extMatchFn) []protoreflect.ExtensionType { +func (f *Files) walkExtensions(getAll bool, pred extMatchFn) []protoreflect.ExtensionType { var result []protoreflect.ExtensionType - files.RangeFiles(func(fd protoreflect.FileDescriptor) bool { + f.RangeFiles(func(fd protoreflect.FileDescriptor) bool { result = append(result, getExtensions(fd, getAll, pred)...) // continue if we are getting all extensions or have none so far return getAll || len(result) == 0 diff --git a/registry/files_test.go b/registry/files_test.go index fc0bc24..9c42e9a 100644 --- a/registry/files_test.go +++ b/registry/files_test.go @@ -6,7 +6,6 @@ import ( "github.com/stretchr/testify/require" "google.golang.org/protobuf/proto" - "google.golang.org/protobuf/reflect/protodesc" "google.golang.org/protobuf/reflect/protoreflect" "google.golang.org/protobuf/reflect/protoregistry" "google.golang.org/protobuf/types/descriptorpb" @@ -163,7 +162,7 @@ func newFiles(t *testing.T) *Files { fds := descriptorpb.FileDescriptorSet{} err = proto.Unmarshal(b, &fds) require.NoError(t, err) - files, err := protodesc.NewFiles(&fds) + files, err := NewFiles(&fds) require.NoError(t, err) - return NewFiles(files) + return files }