Bumping k8s dependencies to 1.13

This commit is contained in:
Cheng Xing
2018-11-16 14:08:25 -08:00
parent 305407125c
commit b4c0b68ec7
8002 changed files with 884099 additions and 276228 deletions

View File

@@ -3,20 +3,28 @@
package datatransfer
import proto "github.com/golang/protobuf/proto"
import fmt "fmt"
import math "math"
import _ "google.golang.org/genproto/googleapis/api/annotations"
import google_protobuf1 "github.com/golang/protobuf/ptypes/struct"
import google_protobuf2 "github.com/golang/protobuf/ptypes/timestamp"
import google_rpc "google.golang.org/genproto/googleapis/rpc/status"
import (
fmt "fmt"
proto "github.com/golang/protobuf/proto"
_struct "github.com/golang/protobuf/ptypes/struct"
timestamp "github.com/golang/protobuf/ptypes/timestamp"
_ "google.golang.org/genproto/googleapis/api/annotations"
status "google.golang.org/genproto/googleapis/rpc/status"
math "math"
)
// Reference imports to suppress errors if they are not otherwise used.
var _ = proto.Marshal
var _ = fmt.Errorf
var _ = math.Inf
// Represents data transfer type.
// This is a compile-time assertion to ensure that this generated file
// is compatible with the proto package it is being compiled against.
// A compilation error at this line likely means your copy of the
// proto package needs to be updated.
const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package
// DEPRECATED. Represents data transfer type.
type TransferType int32
const (
@@ -34,6 +42,7 @@ var TransferType_name = map[int32]string{
1: "BATCH",
2: "STREAMING",
}
var TransferType_value = map[string]int32{
"TRANSFER_TYPE_UNSPECIFIED": 0,
"BATCH": 1,
@@ -43,7 +52,10 @@ var TransferType_value = map[string]int32{
func (x TransferType) String() string {
return proto.EnumName(TransferType_name, int32(x))
}
func (TransferType) EnumDescriptor() ([]byte, []int) { return fileDescriptor1, []int{0} }
func (TransferType) EnumDescriptor() ([]byte, []int) {
return fileDescriptor_90c2574420b818ad, []int{0}
}
// Represents data transfer run state.
type TransferState int32
@@ -72,6 +84,7 @@ var TransferState_name = map[int32]string{
5: "FAILED",
6: "CANCELLED",
}
var TransferState_value = map[string]int32{
"TRANSFER_STATE_UNSPECIFIED": 0,
"PENDING": 2,
@@ -84,7 +97,10 @@ var TransferState_value = map[string]int32{
func (x TransferState) String() string {
return proto.EnumName(TransferState_name, int32(x))
}
func (TransferState) EnumDescriptor() ([]byte, []int) { return fileDescriptor1, []int{1} }
func (TransferState) EnumDescriptor() ([]byte, []int) {
return fileDescriptor_90c2574420b818ad, []int{1}
}
// Represents data transfer user facing message severity.
type TransferMessage_MessageSeverity int32
@@ -106,18 +122,20 @@ var TransferMessage_MessageSeverity_name = map[int32]string{
2: "WARNING",
3: "ERROR",
}
var TransferMessage_MessageSeverity_value = map[string]int32{
"MESSAGE_SEVERITY_UNSPECIFIED": 0,
"INFO": 1,
"WARNING": 2,
"ERROR": 3,
"INFO": 1,
"WARNING": 2,
"ERROR": 3,
}
func (x TransferMessage_MessageSeverity) String() string {
return proto.EnumName(TransferMessage_MessageSeverity_name, int32(x))
}
func (TransferMessage_MessageSeverity) EnumDescriptor() ([]byte, []int) {
return fileDescriptor1, []int{2, 0}
return fileDescriptor_90c2574420b818ad, []int{2, 0}
}
// Represents a data transfer configuration. A transfer configuration
@@ -126,7 +144,6 @@ func (TransferMessage_MessageSeverity) EnumDescriptor() ([]byte, []int) {
// When a new transfer configuration is created, the specified
// `destination_dataset_id` is created when needed and shared with the
// appropriate data source service account.
// Next id: 20
type TransferConfig struct {
// The resource name of the transfer config.
// Transfer config names have the form
@@ -134,15 +151,15 @@ type TransferConfig struct {
// Where `config_id` is usually a uuid, even though it is not
// guaranteed or required. The name is ignored when creating a transfer
// config.
Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"`
Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"`
// The BigQuery target dataset id.
DestinationDatasetId string `protobuf:"bytes,2,opt,name=destination_dataset_id,json=destinationDatasetId" json:"destination_dataset_id,omitempty"`
DestinationDatasetId string `protobuf:"bytes,2,opt,name=destination_dataset_id,json=destinationDatasetId,proto3" json:"destination_dataset_id,omitempty"`
// User specified display name for the data transfer.
DisplayName string `protobuf:"bytes,3,opt,name=display_name,json=displayName" json:"display_name,omitempty"`
DisplayName string `protobuf:"bytes,3,opt,name=display_name,json=displayName,proto3" json:"display_name,omitempty"`
// Data source id. Cannot be changed once data transfer is created.
DataSourceId string `protobuf:"bytes,5,opt,name=data_source_id,json=dataSourceId" json:"data_source_id,omitempty"`
DataSourceId string `protobuf:"bytes,5,opt,name=data_source_id,json=dataSourceId,proto3" json:"data_source_id,omitempty"`
// Data transfer specific parameters.
Params *google_protobuf1.Struct `protobuf:"bytes,9,opt,name=params" json:"params,omitempty"`
Params *_struct.Struct `protobuf:"bytes,9,opt,name=params,proto3" json:"params,omitempty"`
// Data transfer schedule.
// If the data source does not support a custom schedule, this should be
// empty. If it is empty, the default value for the data source will be
@@ -155,37 +172,60 @@ type TransferConfig struct {
// See more explanation about the format here:
// https://cloud.google.com/appengine/docs/flexible/python/scheduling-jobs-with-cron-yaml#the_schedule_format
// NOTE: the granularity should be at least 8 hours, or less frequent.
Schedule string `protobuf:"bytes,7,opt,name=schedule" json:"schedule,omitempty"`
Schedule string `protobuf:"bytes,7,opt,name=schedule,proto3" json:"schedule,omitempty"`
// The number of days to look back to automatically refresh the data.
// For example, if `data_refresh_window_days = 10`, then every day
// BigQuery reingests data for [today-10, today-1], rather than ingesting data
// for just [today-1].
// Only valid if the data source supports the feature. Set the value to 0
// to use the default value.
DataRefreshWindowDays int32 `protobuf:"varint,12,opt,name=data_refresh_window_days,json=dataRefreshWindowDays" json:"data_refresh_window_days,omitempty"`
DataRefreshWindowDays int32 `protobuf:"varint,12,opt,name=data_refresh_window_days,json=dataRefreshWindowDays,proto3" json:"data_refresh_window_days,omitempty"`
// Is this config disabled. When set to true, no runs are scheduled
// for a given transfer.
Disabled bool `protobuf:"varint,13,opt,name=disabled" json:"disabled,omitempty"`
Disabled bool `protobuf:"varint,13,opt,name=disabled,proto3" json:"disabled,omitempty"`
// Output only. Data transfer modification time. Ignored by server on input.
UpdateTime *google_protobuf2.Timestamp `protobuf:"bytes,4,opt,name=update_time,json=updateTime" json:"update_time,omitempty"`
UpdateTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=update_time,json=updateTime,proto3" json:"update_time,omitempty"`
// Output only. Next time when data transfer will run.
NextRunTime *google_protobuf2.Timestamp `protobuf:"bytes,8,opt,name=next_run_time,json=nextRunTime" json:"next_run_time,omitempty"`
NextRunTime *timestamp.Timestamp `protobuf:"bytes,8,opt,name=next_run_time,json=nextRunTime,proto3" json:"next_run_time,omitempty"`
// Output only. State of the most recently updated transfer run.
State TransferState `protobuf:"varint,10,opt,name=state,enum=google.cloud.bigquery.datatransfer.v1.TransferState" json:"state,omitempty"`
State TransferState `protobuf:"varint,10,opt,name=state,proto3,enum=google.cloud.bigquery.datatransfer.v1.TransferState" json:"state,omitempty"`
// Output only. Unique ID of the user on whose behalf transfer is done.
// Applicable only to data sources that do not support service accounts.
// When set to 0, the data source service account credentials are used.
// May be negative. Note, that this identifier is not stable.
// It may change over time even for the same user.
UserId int64 `protobuf:"varint,11,opt,name=user_id,json=userId" json:"user_id,omitempty"`
UserId int64 `protobuf:"varint,11,opt,name=user_id,json=userId,proto3" json:"user_id,omitempty"`
// Output only. Region in which BigQuery dataset is located.
DatasetRegion string `protobuf:"bytes,14,opt,name=dataset_region,json=datasetRegion" json:"dataset_region,omitempty"`
DatasetRegion string `protobuf:"bytes,14,opt,name=dataset_region,json=datasetRegion,proto3" json:"dataset_region,omitempty"`
XXX_NoUnkeyedLiteral struct{} `json:"-"`
XXX_unrecognized []byte `json:"-"`
XXX_sizecache int32 `json:"-"`
}
func (m *TransferConfig) Reset() { *m = TransferConfig{} }
func (m *TransferConfig) String() string { return proto.CompactTextString(m) }
func (*TransferConfig) ProtoMessage() {}
func (*TransferConfig) Descriptor() ([]byte, []int) { return fileDescriptor1, []int{0} }
func (m *TransferConfig) Reset() { *m = TransferConfig{} }
func (m *TransferConfig) String() string { return proto.CompactTextString(m) }
func (*TransferConfig) ProtoMessage() {}
func (*TransferConfig) Descriptor() ([]byte, []int) {
return fileDescriptor_90c2574420b818ad, []int{0}
}
func (m *TransferConfig) XXX_Unmarshal(b []byte) error {
return xxx_messageInfo_TransferConfig.Unmarshal(m, b)
}
func (m *TransferConfig) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) {
return xxx_messageInfo_TransferConfig.Marshal(b, m, deterministic)
}
func (m *TransferConfig) XXX_Merge(src proto.Message) {
xxx_messageInfo_TransferConfig.Merge(m, src)
}
func (m *TransferConfig) XXX_Size() int {
return xxx_messageInfo_TransferConfig.Size(m)
}
func (m *TransferConfig) XXX_DiscardUnknown() {
xxx_messageInfo_TransferConfig.DiscardUnknown(m)
}
var xxx_messageInfo_TransferConfig proto.InternalMessageInfo
func (m *TransferConfig) GetName() string {
if m != nil {
@@ -215,7 +255,7 @@ func (m *TransferConfig) GetDataSourceId() string {
return ""
}
func (m *TransferConfig) GetParams() *google_protobuf1.Struct {
func (m *TransferConfig) GetParams() *_struct.Struct {
if m != nil {
return m.Params
}
@@ -243,14 +283,14 @@ func (m *TransferConfig) GetDisabled() bool {
return false
}
func (m *TransferConfig) GetUpdateTime() *google_protobuf2.Timestamp {
func (m *TransferConfig) GetUpdateTime() *timestamp.Timestamp {
if m != nil {
return m.UpdateTime
}
return nil
}
func (m *TransferConfig) GetNextRunTime() *google_protobuf2.Timestamp {
func (m *TransferConfig) GetNextRunTime() *timestamp.Timestamp {
if m != nil {
return m.NextRunTime
}
@@ -279,54 +319,76 @@ func (m *TransferConfig) GetDatasetRegion() string {
}
// Represents a data transfer run.
// Next id: 27
type TransferRun struct {
// The resource name of the transfer run.
// Transfer run names have the form
// `projects/{project_id}/locations/{location}/transferConfigs/{config_id}/runs/{run_id}`.
// The name is ignored when creating a transfer run.
Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"`
Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"`
// Minimum time after which a transfer run can be started.
ScheduleTime *google_protobuf2.Timestamp `protobuf:"bytes,3,opt,name=schedule_time,json=scheduleTime" json:"schedule_time,omitempty"`
ScheduleTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=schedule_time,json=scheduleTime,proto3" json:"schedule_time,omitempty"`
// For batch transfer runs, specifies the date and time that
// data should be ingested.
RunTime *google_protobuf2.Timestamp `protobuf:"bytes,10,opt,name=run_time,json=runTime" json:"run_time,omitempty"`
RunTime *timestamp.Timestamp `protobuf:"bytes,10,opt,name=run_time,json=runTime,proto3" json:"run_time,omitempty"`
// Status of the transfer run.
ErrorStatus *google_rpc.Status `protobuf:"bytes,21,opt,name=error_status,json=errorStatus" json:"error_status,omitempty"`
ErrorStatus *status.Status `protobuf:"bytes,21,opt,name=error_status,json=errorStatus,proto3" json:"error_status,omitempty"`
// Output only. Time when transfer run was started.
// Parameter ignored by server for input requests.
StartTime *google_protobuf2.Timestamp `protobuf:"bytes,4,opt,name=start_time,json=startTime" json:"start_time,omitempty"`
StartTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=start_time,json=startTime,proto3" json:"start_time,omitempty"`
// Output only. Time when transfer run ended.
// Parameter ignored by server for input requests.
EndTime *google_protobuf2.Timestamp `protobuf:"bytes,5,opt,name=end_time,json=endTime" json:"end_time,omitempty"`
EndTime *timestamp.Timestamp `protobuf:"bytes,5,opt,name=end_time,json=endTime,proto3" json:"end_time,omitempty"`
// Output only. Last time the data transfer run state was updated.
UpdateTime *google_protobuf2.Timestamp `protobuf:"bytes,6,opt,name=update_time,json=updateTime" json:"update_time,omitempty"`
UpdateTime *timestamp.Timestamp `protobuf:"bytes,6,opt,name=update_time,json=updateTime,proto3" json:"update_time,omitempty"`
// Output only. Data transfer specific parameters.
Params *google_protobuf1.Struct `protobuf:"bytes,9,opt,name=params" json:"params,omitempty"`
Params *_struct.Struct `protobuf:"bytes,9,opt,name=params,proto3" json:"params,omitempty"`
// Output only. The BigQuery target dataset id.
DestinationDatasetId string `protobuf:"bytes,2,opt,name=destination_dataset_id,json=destinationDatasetId" json:"destination_dataset_id,omitempty"`
DestinationDatasetId string `protobuf:"bytes,2,opt,name=destination_dataset_id,json=destinationDatasetId,proto3" json:"destination_dataset_id,omitempty"`
// Output only. Data source id.
DataSourceId string `protobuf:"bytes,7,opt,name=data_source_id,json=dataSourceId" json:"data_source_id,omitempty"`
DataSourceId string `protobuf:"bytes,7,opt,name=data_source_id,json=dataSourceId,proto3" json:"data_source_id,omitempty"`
// Data transfer run state. Ignored for input requests.
State TransferState `protobuf:"varint,8,opt,name=state,enum=google.cloud.bigquery.datatransfer.v1.TransferState" json:"state,omitempty"`
State TransferState `protobuf:"varint,8,opt,name=state,proto3,enum=google.cloud.bigquery.datatransfer.v1.TransferState" json:"state,omitempty"`
// Output only. Unique ID of the user on whose behalf transfer is done.
// Applicable only to data sources that do not support service accounts.
// When set to 0, the data source service account credentials are used.
// May be negative. Note, that this identifier is not stable.
// It may change over time even for the same user.
UserId int64 `protobuf:"varint,11,opt,name=user_id,json=userId" json:"user_id,omitempty"`
UserId int64 `protobuf:"varint,11,opt,name=user_id,json=userId,proto3" json:"user_id,omitempty"`
// Output only. Describes the schedule of this transfer run if it was
// created as part of a regular schedule. For batch transfer runs that are
// scheduled manually, this is empty.
// NOTE: the system might choose to delay the schedule depending on the
// current load, so `schedule_time` doesn't always matches this.
Schedule string `protobuf:"bytes,12,opt,name=schedule" json:"schedule,omitempty"`
Schedule string `protobuf:"bytes,12,opt,name=schedule,proto3" json:"schedule,omitempty"`
XXX_NoUnkeyedLiteral struct{} `json:"-"`
XXX_unrecognized []byte `json:"-"`
XXX_sizecache int32 `json:"-"`
}
func (m *TransferRun) Reset() { *m = TransferRun{} }
func (m *TransferRun) String() string { return proto.CompactTextString(m) }
func (*TransferRun) ProtoMessage() {}
func (*TransferRun) Descriptor() ([]byte, []int) { return fileDescriptor1, []int{1} }
func (m *TransferRun) Reset() { *m = TransferRun{} }
func (m *TransferRun) String() string { return proto.CompactTextString(m) }
func (*TransferRun) ProtoMessage() {}
func (*TransferRun) Descriptor() ([]byte, []int) {
return fileDescriptor_90c2574420b818ad, []int{1}
}
func (m *TransferRun) XXX_Unmarshal(b []byte) error {
return xxx_messageInfo_TransferRun.Unmarshal(m, b)
}
func (m *TransferRun) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) {
return xxx_messageInfo_TransferRun.Marshal(b, m, deterministic)
}
func (m *TransferRun) XXX_Merge(src proto.Message) {
xxx_messageInfo_TransferRun.Merge(m, src)
}
func (m *TransferRun) XXX_Size() int {
return xxx_messageInfo_TransferRun.Size(m)
}
func (m *TransferRun) XXX_DiscardUnknown() {
xxx_messageInfo_TransferRun.DiscardUnknown(m)
}
var xxx_messageInfo_TransferRun proto.InternalMessageInfo
func (m *TransferRun) GetName() string {
if m != nil {
@@ -335,49 +397,49 @@ func (m *TransferRun) GetName() string {
return ""
}
func (m *TransferRun) GetScheduleTime() *google_protobuf2.Timestamp {
func (m *TransferRun) GetScheduleTime() *timestamp.Timestamp {
if m != nil {
return m.ScheduleTime
}
return nil
}
func (m *TransferRun) GetRunTime() *google_protobuf2.Timestamp {
func (m *TransferRun) GetRunTime() *timestamp.Timestamp {
if m != nil {
return m.RunTime
}
return nil
}
func (m *TransferRun) GetErrorStatus() *google_rpc.Status {
func (m *TransferRun) GetErrorStatus() *status.Status {
if m != nil {
return m.ErrorStatus
}
return nil
}
func (m *TransferRun) GetStartTime() *google_protobuf2.Timestamp {
func (m *TransferRun) GetStartTime() *timestamp.Timestamp {
if m != nil {
return m.StartTime
}
return nil
}
func (m *TransferRun) GetEndTime() *google_protobuf2.Timestamp {
func (m *TransferRun) GetEndTime() *timestamp.Timestamp {
if m != nil {
return m.EndTime
}
return nil
}
func (m *TransferRun) GetUpdateTime() *google_protobuf2.Timestamp {
func (m *TransferRun) GetUpdateTime() *timestamp.Timestamp {
if m != nil {
return m.UpdateTime
}
return nil
}
func (m *TransferRun) GetParams() *google_protobuf1.Struct {
func (m *TransferRun) GetParams() *_struct.Struct {
if m != nil {
return m.Params
}
@@ -422,19 +484,42 @@ func (m *TransferRun) GetSchedule() string {
// Represents a user facing message for a particular data transfer run.
type TransferMessage struct {
// Time when message was logged.
MessageTime *google_protobuf2.Timestamp `protobuf:"bytes,1,opt,name=message_time,json=messageTime" json:"message_time,omitempty"`
MessageTime *timestamp.Timestamp `protobuf:"bytes,1,opt,name=message_time,json=messageTime,proto3" json:"message_time,omitempty"`
// Message severity.
Severity TransferMessage_MessageSeverity `protobuf:"varint,2,opt,name=severity,enum=google.cloud.bigquery.datatransfer.v1.TransferMessage_MessageSeverity" json:"severity,omitempty"`
Severity TransferMessage_MessageSeverity `protobuf:"varint,2,opt,name=severity,proto3,enum=google.cloud.bigquery.datatransfer.v1.TransferMessage_MessageSeverity" json:"severity,omitempty"`
// Message text.
MessageText string `protobuf:"bytes,3,opt,name=message_text,json=messageText" json:"message_text,omitempty"`
MessageText string `protobuf:"bytes,3,opt,name=message_text,json=messageText,proto3" json:"message_text,omitempty"`
XXX_NoUnkeyedLiteral struct{} `json:"-"`
XXX_unrecognized []byte `json:"-"`
XXX_sizecache int32 `json:"-"`
}
func (m *TransferMessage) Reset() { *m = TransferMessage{} }
func (m *TransferMessage) String() string { return proto.CompactTextString(m) }
func (*TransferMessage) ProtoMessage() {}
func (*TransferMessage) Descriptor() ([]byte, []int) { return fileDescriptor1, []int{2} }
func (m *TransferMessage) Reset() { *m = TransferMessage{} }
func (m *TransferMessage) String() string { return proto.CompactTextString(m) }
func (*TransferMessage) ProtoMessage() {}
func (*TransferMessage) Descriptor() ([]byte, []int) {
return fileDescriptor_90c2574420b818ad, []int{2}
}
func (m *TransferMessage) GetMessageTime() *google_protobuf2.Timestamp {
func (m *TransferMessage) XXX_Unmarshal(b []byte) error {
return xxx_messageInfo_TransferMessage.Unmarshal(m, b)
}
func (m *TransferMessage) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) {
return xxx_messageInfo_TransferMessage.Marshal(b, m, deterministic)
}
func (m *TransferMessage) XXX_Merge(src proto.Message) {
xxx_messageInfo_TransferMessage.Merge(m, src)
}
func (m *TransferMessage) XXX_Size() int {
return xxx_messageInfo_TransferMessage.Size(m)
}
func (m *TransferMessage) XXX_DiscardUnknown() {
xxx_messageInfo_TransferMessage.DiscardUnknown(m)
}
var xxx_messageInfo_TransferMessage proto.InternalMessageInfo
func (m *TransferMessage) GetMessageTime() *timestamp.Timestamp {
if m != nil {
return m.MessageTime
}
@@ -456,19 +541,19 @@ func (m *TransferMessage) GetMessageText() string {
}
func init() {
proto.RegisterType((*TransferConfig)(nil), "google.cloud.bigquery.datatransfer.v1.TransferConfig")
proto.RegisterType((*TransferRun)(nil), "google.cloud.bigquery.datatransfer.v1.TransferRun")
proto.RegisterType((*TransferMessage)(nil), "google.cloud.bigquery.datatransfer.v1.TransferMessage")
proto.RegisterEnum("google.cloud.bigquery.datatransfer.v1.TransferType", TransferType_name, TransferType_value)
proto.RegisterEnum("google.cloud.bigquery.datatransfer.v1.TransferState", TransferState_name, TransferState_value)
proto.RegisterEnum("google.cloud.bigquery.datatransfer.v1.TransferMessage_MessageSeverity", TransferMessage_MessageSeverity_name, TransferMessage_MessageSeverity_value)
proto.RegisterType((*TransferConfig)(nil), "google.cloud.bigquery.datatransfer.v1.TransferConfig")
proto.RegisterType((*TransferRun)(nil), "google.cloud.bigquery.datatransfer.v1.TransferRun")
proto.RegisterType((*TransferMessage)(nil), "google.cloud.bigquery.datatransfer.v1.TransferMessage")
}
func init() {
proto.RegisterFile("google/cloud/bigquery/datatransfer/v1/transfer.proto", fileDescriptor1)
proto.RegisterFile("google/cloud/bigquery/datatransfer/v1/transfer.proto", fileDescriptor_90c2574420b818ad)
}
var fileDescriptor1 = []byte{
var fileDescriptor_90c2574420b818ad = []byte{
// 922 bytes of a gzipped FileDescriptorProto
0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xb4, 0x56, 0xdd, 0x6e, 0xe3, 0x44,
0x14, 0xc6, 0xf9, 0xcf, 0x71, 0x92, 0x8d, 0x46, 0x2c, 0x35, 0xd5, 0x02, 0xa1, 0xa2, 0x52, 0xd8,

File diff suppressed because it is too large Load Diff

View File

@@ -0,0 +1,142 @@
// Code generated by protoc-gen-go. DO NOT EDIT.
// source: google/cloud/bigquery/storage/v1beta1/avro.proto
package storage
import (
fmt "fmt"
proto "github.com/golang/protobuf/proto"
math "math"
)
// Reference imports to suppress errors if they are not otherwise used.
var _ = proto.Marshal
var _ = fmt.Errorf
var _ = math.Inf
// This is a compile-time assertion to ensure that this generated file
// is compatible with the proto package it is being compiled against.
// A compilation error at this line likely means your copy of the
// proto package needs to be updated.
const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package
// Avro schema.
type AvroSchema struct {
// Json serialized schema, as described at
// https://avro.apache.org/docs/1.8.1/spec.html
Schema string `protobuf:"bytes,1,opt,name=schema,proto3" json:"schema,omitempty"`
XXX_NoUnkeyedLiteral struct{} `json:"-"`
XXX_unrecognized []byte `json:"-"`
XXX_sizecache int32 `json:"-"`
}
func (m *AvroSchema) Reset() { *m = AvroSchema{} }
func (m *AvroSchema) String() string { return proto.CompactTextString(m) }
func (*AvroSchema) ProtoMessage() {}
func (*AvroSchema) Descriptor() ([]byte, []int) {
return fileDescriptor_8664acda99313b12, []int{0}
}
func (m *AvroSchema) XXX_Unmarshal(b []byte) error {
return xxx_messageInfo_AvroSchema.Unmarshal(m, b)
}
func (m *AvroSchema) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) {
return xxx_messageInfo_AvroSchema.Marshal(b, m, deterministic)
}
func (m *AvroSchema) XXX_Merge(src proto.Message) {
xxx_messageInfo_AvroSchema.Merge(m, src)
}
func (m *AvroSchema) XXX_Size() int {
return xxx_messageInfo_AvroSchema.Size(m)
}
func (m *AvroSchema) XXX_DiscardUnknown() {
xxx_messageInfo_AvroSchema.DiscardUnknown(m)
}
var xxx_messageInfo_AvroSchema proto.InternalMessageInfo
func (m *AvroSchema) GetSchema() string {
if m != nil {
return m.Schema
}
return ""
}
// Avro rows.
type AvroRows struct {
// Binary serialized rows in a block.
SerializedBinaryRows []byte `protobuf:"bytes,1,opt,name=serialized_binary_rows,json=serializedBinaryRows,proto3" json:"serialized_binary_rows,omitempty"`
// The count of rows in the returning block.
RowCount int64 `protobuf:"varint,2,opt,name=row_count,json=rowCount,proto3" json:"row_count,omitempty"`
XXX_NoUnkeyedLiteral struct{} `json:"-"`
XXX_unrecognized []byte `json:"-"`
XXX_sizecache int32 `json:"-"`
}
func (m *AvroRows) Reset() { *m = AvroRows{} }
func (m *AvroRows) String() string { return proto.CompactTextString(m) }
func (*AvroRows) ProtoMessage() {}
func (*AvroRows) Descriptor() ([]byte, []int) {
return fileDescriptor_8664acda99313b12, []int{1}
}
func (m *AvroRows) XXX_Unmarshal(b []byte) error {
return xxx_messageInfo_AvroRows.Unmarshal(m, b)
}
func (m *AvroRows) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) {
return xxx_messageInfo_AvroRows.Marshal(b, m, deterministic)
}
func (m *AvroRows) XXX_Merge(src proto.Message) {
xxx_messageInfo_AvroRows.Merge(m, src)
}
func (m *AvroRows) XXX_Size() int {
return xxx_messageInfo_AvroRows.Size(m)
}
func (m *AvroRows) XXX_DiscardUnknown() {
xxx_messageInfo_AvroRows.DiscardUnknown(m)
}
var xxx_messageInfo_AvroRows proto.InternalMessageInfo
func (m *AvroRows) GetSerializedBinaryRows() []byte {
if m != nil {
return m.SerializedBinaryRows
}
return nil
}
func (m *AvroRows) GetRowCount() int64 {
if m != nil {
return m.RowCount
}
return 0
}
func init() {
proto.RegisterType((*AvroSchema)(nil), "google.cloud.bigquery.storage.v1beta1.AvroSchema")
proto.RegisterType((*AvroRows)(nil), "google.cloud.bigquery.storage.v1beta1.AvroRows")
}
func init() {
proto.RegisterFile("google/cloud/bigquery/storage/v1beta1/avro.proto", fileDescriptor_8664acda99313b12)
}
var fileDescriptor_8664acda99313b12 = []byte{
// 242 bytes of a gzipped FileDescriptorProto
0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x8c, 0xd0, 0xcf, 0x4b, 0xc3, 0x30,
0x14, 0x07, 0x70, 0xaa, 0x30, 0xd6, 0xe0, 0xa9, 0xc8, 0x18, 0x78, 0x19, 0x43, 0x61, 0x5e, 0x12,
0x87, 0xde, 0x3c, 0x59, 0xaf, 0x1e, 0xa4, 0xde, 0x04, 0x29, 0x69, 0x17, 0x9e, 0x81, 0xae, 0xdf,
0xf9, 0xd2, 0x1f, 0xcc, 0xb3, 0x7f, 0xb8, 0x24, 0x8d, 0x78, 0xd4, 0x5b, 0xbe, 0x7c, 0xf3, 0x81,
0xf7, 0x9e, 0xb8, 0x21, 0x80, 0x1a, 0xa3, 0xea, 0x06, 0xfd, 0x4e, 0x55, 0x96, 0x3e, 0x7a, 0xc3,
0x47, 0xe5, 0x3a, 0xb0, 0x26, 0xa3, 0x86, 0x6d, 0x65, 0x3a, 0xbd, 0x55, 0x7a, 0x60, 0xc8, 0x03,
0xa3, 0x43, 0x76, 0x35, 0x09, 0x19, 0x84, 0xfc, 0x11, 0x32, 0x0a, 0x19, 0xc5, 0xfa, 0x52, 0x88,
0x87, 0x81, 0xf1, 0x52, 0xbf, 0x9b, 0xbd, 0xce, 0x16, 0x62, 0xe6, 0xc2, 0x6b, 0x99, 0xac, 0x92,
0x4d, 0x5a, 0xc4, 0xb4, 0x7e, 0x13, 0x73, 0xff, 0xab, 0xc0, 0xe8, 0xb2, 0x3b, 0xb1, 0x70, 0x86,
0xad, 0x6e, 0xec, 0xa7, 0xd9, 0x95, 0x95, 0x6d, 0x35, 0x1f, 0x4b, 0xc6, 0xe8, 0x82, 0x39, 0x2b,
0xce, 0x7f, 0xdb, 0x3c, 0x94, 0x41, 0x5d, 0x88, 0x94, 0x31, 0x96, 0x35, 0xfa, 0xb6, 0x5b, 0x9e,
0xac, 0x92, 0xcd, 0x69, 0x31, 0x67, 0x8c, 0x8f, 0x3e, 0xe7, 0x5f, 0x89, 0xb8, 0xae, 0xb1, 0x97,
0xff, 0x1a, 0x39, 0x4f, 0xfd, 0x28, 0xcf, 0x7e, 0xc9, 0xd7, 0xa7, 0x28, 0x08, 0x8d, 0x6e, 0x49,
0x82, 0x49, 0x91, 0x69, 0xc3, 0x01, 0xd4, 0x54, 0xe9, 0x83, 0x75, 0x7f, 0x5c, 0xed, 0x3e, 0xe6,
0x6a, 0x16, 0xe0, 0xed, 0x77, 0x00, 0x00, 0x00, 0xff, 0xff, 0x16, 0xaf, 0x83, 0xb1, 0x6d, 0x01,
0x00, 0x00,
}

View File

@@ -0,0 +1,105 @@
// Code generated by protoc-gen-go. DO NOT EDIT.
// source: google/cloud/bigquery/storage/v1beta1/read_options.proto
package storage
import (
fmt "fmt"
proto "github.com/golang/protobuf/proto"
math "math"
)
// Reference imports to suppress errors if they are not otherwise used.
var _ = proto.Marshal
var _ = fmt.Errorf
var _ = math.Inf
// This is a compile-time assertion to ensure that this generated file
// is compatible with the proto package it is being compiled against.
// A compilation error at this line likely means your copy of the
// proto package needs to be updated.
const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package
// Options dictating how we read a table.
type TableReadOptions struct {
// Optional. Names of the fields in the table that should be read. If empty,
// all fields will be read. If the specified field is a nested field, all the
// sub-fields in the field will be selected. The output field order is
// unrelated to the order of fields in selected_fields.
SelectedFields []string `protobuf:"bytes,1,rep,name=selected_fields,json=selectedFields,proto3" json:"selected_fields,omitempty"`
// Optional. SQL text filtering statement, similar to a WHERE clause in
// a query. Currently, we support combinations of predicates that are
// a comparison between a column and a constant value in SQL statement.
// Aggregates are not supported.
//
// Example: "a > DATE '2014-9-27' AND (b > 5 and C LIKE 'date')"
RowRestriction string `protobuf:"bytes,2,opt,name=row_restriction,json=rowRestriction,proto3" json:"row_restriction,omitempty"`
XXX_NoUnkeyedLiteral struct{} `json:"-"`
XXX_unrecognized []byte `json:"-"`
XXX_sizecache int32 `json:"-"`
}
func (m *TableReadOptions) Reset() { *m = TableReadOptions{} }
func (m *TableReadOptions) String() string { return proto.CompactTextString(m) }
func (*TableReadOptions) ProtoMessage() {}
func (*TableReadOptions) Descriptor() ([]byte, []int) {
return fileDescriptor_b5cb6d339ef0ad09, []int{0}
}
func (m *TableReadOptions) XXX_Unmarshal(b []byte) error {
return xxx_messageInfo_TableReadOptions.Unmarshal(m, b)
}
func (m *TableReadOptions) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) {
return xxx_messageInfo_TableReadOptions.Marshal(b, m, deterministic)
}
func (m *TableReadOptions) XXX_Merge(src proto.Message) {
xxx_messageInfo_TableReadOptions.Merge(m, src)
}
func (m *TableReadOptions) XXX_Size() int {
return xxx_messageInfo_TableReadOptions.Size(m)
}
func (m *TableReadOptions) XXX_DiscardUnknown() {
xxx_messageInfo_TableReadOptions.DiscardUnknown(m)
}
var xxx_messageInfo_TableReadOptions proto.InternalMessageInfo
func (m *TableReadOptions) GetSelectedFields() []string {
if m != nil {
return m.SelectedFields
}
return nil
}
func (m *TableReadOptions) GetRowRestriction() string {
if m != nil {
return m.RowRestriction
}
return ""
}
func init() {
proto.RegisterType((*TableReadOptions)(nil), "google.cloud.bigquery.storage.v1beta1.TableReadOptions")
}
func init() {
proto.RegisterFile("google/cloud/bigquery/storage/v1beta1/read_options.proto", fileDescriptor_b5cb6d339ef0ad09)
}
var fileDescriptor_b5cb6d339ef0ad09 = []byte{
// 213 bytes of a gzipped FileDescriptorProto
0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x8c, 0xd0, 0x3f, 0x4b, 0xc6, 0x30,
0x10, 0x06, 0x70, 0x5e, 0x05, 0xc1, 0x0c, 0x2a, 0x9d, 0x3a, 0x16, 0x41, 0xd4, 0x25, 0xa1, 0xb8,
0x08, 0x6e, 0x0e, 0x4e, 0x82, 0x50, 0x9c, 0x5c, 0x4a, 0xfe, 0x9c, 0x47, 0x20, 0xf6, 0xea, 0x25,
0xb5, 0xf4, 0xdb, 0x4b, 0xd3, 0x14, 0x47, 0xdf, 0x31, 0x4f, 0x9e, 0x5f, 0xc8, 0x9d, 0x78, 0x44,
0x22, 0x0c, 0xa0, 0x6c, 0xa0, 0xc9, 0x29, 0xe3, 0xf1, 0x7b, 0x02, 0x5e, 0x54, 0x4c, 0xc4, 0x1a,
0x41, 0xfd, 0xb4, 0x06, 0x92, 0x6e, 0x15, 0x83, 0x76, 0x3d, 0x8d, 0xc9, 0xd3, 0x10, 0xe5, 0xc8,
0x94, 0xa8, 0xba, 0xd9, 0xa4, 0xcc, 0x52, 0xee, 0x52, 0x16, 0x29, 0x8b, 0xbc, 0x76, 0xe2, 0xea,
0x5d, 0x9b, 0x00, 0x1d, 0x68, 0xf7, 0xb6, 0x3d, 0x50, 0xdd, 0x8a, 0xcb, 0x08, 0x01, 0x6c, 0x02,
0xd7, 0x7f, 0x7a, 0x08, 0x2e, 0xd6, 0x87, 0xe6, 0xf4, 0xee, 0xbc, 0xbb, 0xd8, 0xe3, 0x97, 0x9c,
0xae, 0x45, 0xa6, 0xb9, 0x67, 0x88, 0x89, 0xbd, 0x5d, 0x71, 0x7d, 0xd2, 0x1c, 0xd6, 0x22, 0xd3,
0xdc, 0xfd, 0xa5, 0xcf, 0x8b, 0xb8, 0xb7, 0xf4, 0x25, 0x8f, 0xfa, 0xd2, 0xc7, 0x6b, 0xa9, 0x21,
0x05, 0x3d, 0xa0, 0x24, 0x46, 0x85, 0x30, 0xe4, 0xa9, 0xd4, 0x76, 0xa5, 0x47, 0x1f, 0xff, 0x59,
0xc9, 0x53, 0x39, 0x9b, 0xb3, 0x0c, 0x1f, 0x7e, 0x03, 0x00, 0x00, 0xff, 0xff, 0xba, 0xd5, 0x14,
0x41, 0x4a, 0x01, 0x00, 0x00,
}

File diff suppressed because it is too large Load Diff

View File

@@ -0,0 +1,153 @@
// Code generated by protoc-gen-go. DO NOT EDIT.
// source: google/cloud/bigquery/storage/v1beta1/table_reference.proto
package storage
import (
fmt "fmt"
proto "github.com/golang/protobuf/proto"
timestamp "github.com/golang/protobuf/ptypes/timestamp"
math "math"
)
// Reference imports to suppress errors if they are not otherwise used.
var _ = proto.Marshal
var _ = fmt.Errorf
var _ = math.Inf
// This is a compile-time assertion to ensure that this generated file
// is compatible with the proto package it is being compiled against.
// A compilation error at this line likely means your copy of the
// proto package needs to be updated.
const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package
// Table reference that includes just the 3 strings needed to identify a table.
type TableReference struct {
// The assigned project ID of the project.
ProjectId string `protobuf:"bytes,1,opt,name=project_id,json=projectId,proto3" json:"project_id,omitempty"`
// The ID of the dataset in the above project.
DatasetId string `protobuf:"bytes,2,opt,name=dataset_id,json=datasetId,proto3" json:"dataset_id,omitempty"`
// The ID of the table in the above dataset.
TableId string `protobuf:"bytes,3,opt,name=table_id,json=tableId,proto3" json:"table_id,omitempty"`
XXX_NoUnkeyedLiteral struct{} `json:"-"`
XXX_unrecognized []byte `json:"-"`
XXX_sizecache int32 `json:"-"`
}
func (m *TableReference) Reset() { *m = TableReference{} }
func (m *TableReference) String() string { return proto.CompactTextString(m) }
func (*TableReference) ProtoMessage() {}
func (*TableReference) Descriptor() ([]byte, []int) {
return fileDescriptor_00fafb38a9ed74ff, []int{0}
}
func (m *TableReference) XXX_Unmarshal(b []byte) error {
return xxx_messageInfo_TableReference.Unmarshal(m, b)
}
func (m *TableReference) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) {
return xxx_messageInfo_TableReference.Marshal(b, m, deterministic)
}
func (m *TableReference) XXX_Merge(src proto.Message) {
xxx_messageInfo_TableReference.Merge(m, src)
}
func (m *TableReference) XXX_Size() int {
return xxx_messageInfo_TableReference.Size(m)
}
func (m *TableReference) XXX_DiscardUnknown() {
xxx_messageInfo_TableReference.DiscardUnknown(m)
}
var xxx_messageInfo_TableReference proto.InternalMessageInfo
func (m *TableReference) GetProjectId() string {
if m != nil {
return m.ProjectId
}
return ""
}
func (m *TableReference) GetDatasetId() string {
if m != nil {
return m.DatasetId
}
return ""
}
func (m *TableReference) GetTableId() string {
if m != nil {
return m.TableId
}
return ""
}
// All fields in this message optional.
type TableModifiers struct {
// The snapshot time of the table. If not set, interpreted as now.
SnapshotTime *timestamp.Timestamp `protobuf:"bytes,1,opt,name=snapshot_time,json=snapshotTime,proto3" json:"snapshot_time,omitempty"`
XXX_NoUnkeyedLiteral struct{} `json:"-"`
XXX_unrecognized []byte `json:"-"`
XXX_sizecache int32 `json:"-"`
}
func (m *TableModifiers) Reset() { *m = TableModifiers{} }
func (m *TableModifiers) String() string { return proto.CompactTextString(m) }
func (*TableModifiers) ProtoMessage() {}
func (*TableModifiers) Descriptor() ([]byte, []int) {
return fileDescriptor_00fafb38a9ed74ff, []int{1}
}
func (m *TableModifiers) XXX_Unmarshal(b []byte) error {
return xxx_messageInfo_TableModifiers.Unmarshal(m, b)
}
func (m *TableModifiers) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) {
return xxx_messageInfo_TableModifiers.Marshal(b, m, deterministic)
}
func (m *TableModifiers) XXX_Merge(src proto.Message) {
xxx_messageInfo_TableModifiers.Merge(m, src)
}
func (m *TableModifiers) XXX_Size() int {
return xxx_messageInfo_TableModifiers.Size(m)
}
func (m *TableModifiers) XXX_DiscardUnknown() {
xxx_messageInfo_TableModifiers.DiscardUnknown(m)
}
var xxx_messageInfo_TableModifiers proto.InternalMessageInfo
func (m *TableModifiers) GetSnapshotTime() *timestamp.Timestamp {
if m != nil {
return m.SnapshotTime
}
return nil
}
func init() {
proto.RegisterType((*TableReference)(nil), "google.cloud.bigquery.storage.v1beta1.TableReference")
proto.RegisterType((*TableModifiers)(nil), "google.cloud.bigquery.storage.v1beta1.TableModifiers")
}
func init() {
proto.RegisterFile("google/cloud/bigquery/storage/v1beta1/table_reference.proto", fileDescriptor_00fafb38a9ed74ff)
}
var fileDescriptor_00fafb38a9ed74ff = []byte{
// 281 bytes of a gzipped FileDescriptorProto
0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x8c, 0x91, 0x4f, 0x4b, 0xc4, 0x30,
0x10, 0xc5, 0xa9, 0x82, 0xba, 0xf1, 0xcf, 0xa1, 0x5e, 0xd6, 0x05, 0x51, 0x16, 0x04, 0xbd, 0x24,
0xac, 0x1e, 0xf7, 0x20, 0xec, 0xad, 0xa0, 0xa0, 0x65, 0x4f, 0x5e, 0x4a, 0xda, 0x4c, 0x63, 0xa4,
0xed, 0xd4, 0x24, 0x15, 0xfc, 0x12, 0x7e, 0x66, 0xc9, 0xbf, 0x83, 0x27, 0xf7, 0x38, 0xf3, 0xf8,
0xbd, 0xf7, 0x26, 0x21, 0x6b, 0x89, 0x28, 0x3b, 0x60, 0x4d, 0x87, 0x93, 0x60, 0xb5, 0x92, 0x9f,
0x13, 0xe8, 0x6f, 0x66, 0x2c, 0x6a, 0x2e, 0x81, 0x7d, 0xad, 0x6a, 0xb0, 0x7c, 0xc5, 0x2c, 0xaf,
0x3b, 0xa8, 0x34, 0xb4, 0xa0, 0x61, 0x68, 0x80, 0x8e, 0x1a, 0x2d, 0xe6, 0x37, 0x01, 0xa6, 0x1e,
0xa6, 0x09, 0xa6, 0x11, 0xa6, 0x11, 0x5e, 0x5c, 0xc5, 0x0c, 0x0f, 0xd5, 0x53, 0xcb, 0xac, 0xea,
0xc1, 0x58, 0xde, 0x8f, 0xc1, 0x67, 0xa9, 0xc8, 0xd9, 0xd6, 0x05, 0x94, 0xc9, 0x3f, 0xbf, 0x24,
0x64, 0xd4, 0xf8, 0x01, 0x8d, 0xad, 0x94, 0x98, 0x67, 0xd7, 0xd9, 0xed, 0xac, 0x9c, 0xc5, 0x4d,
0x21, 0x9c, 0x2c, 0xb8, 0xe5, 0x06, 0xbc, 0xbc, 0x17, 0xe4, 0xb8, 0x29, 0x44, 0x7e, 0x41, 0x8e,
0x42, 0x61, 0x25, 0xe6, 0xfb, 0x5e, 0x3c, 0xf4, 0x73, 0x21, 0x96, 0xaf, 0x31, 0xea, 0x19, 0x85,
0x6a, 0x15, 0x68, 0x93, 0x3f, 0x92, 0x53, 0x33, 0xf0, 0xd1, 0xbc, 0xa3, 0xad, 0x5c, 0x31, 0x9f,
0x76, 0x7c, 0xbf, 0xa0, 0xf1, 0xb8, 0xd4, 0x9a, 0x6e, 0x53, 0xeb, 0xf2, 0x24, 0x01, 0x6e, 0xb5,
0xf9, 0xc9, 0xc8, 0x5d, 0x83, 0x3d, 0xdd, 0xe9, 0x31, 0x36, 0xe7, 0x7f, 0x2f, 0x7d, 0x71, 0xee,
0x6f, 0x4f, 0x91, 0x95, 0xd8, 0xf1, 0x41, 0x52, 0xd4, 0x92, 0x49, 0x18, 0x7c, 0x32, 0x0b, 0x12,
0x1f, 0x95, 0xf9, 0xe7, 0x93, 0xd6, 0x71, 0xae, 0x0f, 0x3c, 0xf8, 0xf0, 0x1b, 0x00, 0x00, 0xff,
0xff, 0xb3, 0xa1, 0xcb, 0xc3, 0xdc, 0x01, 0x00, 0x00,
}