mirror of
https://github.com/ceph/ceph-csi.git
synced 2025-06-14 18:53:35 +00:00
vendor update for CSI 0.3.0
This commit is contained in:
1194
vendor/google.golang.org/genproto/googleapis/cloud/bigquery/datatransfer/v1/datatransfer.pb.go
generated
vendored
1194
vendor/google.golang.org/genproto/googleapis/cloud/bigquery/datatransfer/v1/datatransfer.pb.go
generated
vendored
File diff suppressed because it is too large
Load Diff
208
vendor/google.golang.org/genproto/googleapis/cloud/bigquery/datatransfer/v1/transfer.pb.go
generated
vendored
208
vendor/google.golang.org/genproto/googleapis/cloud/bigquery/datatransfer/v1/transfer.pb.go
generated
vendored
@ -1,22 +1,28 @@
|
||||
// Code generated by protoc-gen-go. DO NOT EDIT.
|
||||
// source: google/cloud/bigquery/datatransfer/v1/transfer.proto
|
||||
|
||||
package datatransfer
|
||||
package datatransfer // import "google.golang.org/genproto/googleapis/cloud/bigquery/datatransfer/v1"
|
||||
|
||||
import proto "github.com/golang/protobuf/proto"
|
||||
import fmt "fmt"
|
||||
import math "math"
|
||||
import _struct "github.com/golang/protobuf/ptypes/struct"
|
||||
import timestamp "github.com/golang/protobuf/ptypes/timestamp"
|
||||
import _ "google.golang.org/genproto/googleapis/api/annotations"
|
||||
import google_protobuf1 "github.com/golang/protobuf/ptypes/struct"
|
||||
import google_protobuf2 "github.com/golang/protobuf/ptypes/timestamp"
|
||||
import google_rpc "google.golang.org/genproto/googleapis/rpc/status"
|
||||
import status "google.golang.org/genproto/googleapis/rpc/status"
|
||||
|
||||
// Reference imports to suppress errors if they are not otherwise used.
|
||||
var _ = proto.Marshal
|
||||
var _ = fmt.Errorf
|
||||
var _ = math.Inf
|
||||
|
||||
// Represents data transfer type.
|
||||
// This is a compile-time assertion to ensure that this generated file
|
||||
// is compatible with the proto package it is being compiled against.
|
||||
// A compilation error at this line likely means your copy of the
|
||||
// proto package needs to be updated.
|
||||
const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package
|
||||
|
||||
// DEPRECATED. Represents data transfer type.
|
||||
type TransferType int32
|
||||
|
||||
const (
|
||||
@ -43,7 +49,9 @@ var TransferType_value = map[string]int32{
|
||||
func (x TransferType) String() string {
|
||||
return proto.EnumName(TransferType_name, int32(x))
|
||||
}
|
||||
func (TransferType) EnumDescriptor() ([]byte, []int) { return fileDescriptor1, []int{0} }
|
||||
func (TransferType) EnumDescriptor() ([]byte, []int) {
|
||||
return fileDescriptor_transfer_b655bdd8f06d7255, []int{0}
|
||||
}
|
||||
|
||||
// Represents data transfer run state.
|
||||
type TransferState int32
|
||||
@ -84,7 +92,9 @@ var TransferState_value = map[string]int32{
|
||||
func (x TransferState) String() string {
|
||||
return proto.EnumName(TransferState_name, int32(x))
|
||||
}
|
||||
func (TransferState) EnumDescriptor() ([]byte, []int) { return fileDescriptor1, []int{1} }
|
||||
func (TransferState) EnumDescriptor() ([]byte, []int) {
|
||||
return fileDescriptor_transfer_b655bdd8f06d7255, []int{1}
|
||||
}
|
||||
|
||||
// Represents data transfer user facing message severity.
|
||||
type TransferMessage_MessageSeverity int32
|
||||
@ -108,16 +118,16 @@ var TransferMessage_MessageSeverity_name = map[int32]string{
|
||||
}
|
||||
var TransferMessage_MessageSeverity_value = map[string]int32{
|
||||
"MESSAGE_SEVERITY_UNSPECIFIED": 0,
|
||||
"INFO": 1,
|
||||
"WARNING": 2,
|
||||
"ERROR": 3,
|
||||
"INFO": 1,
|
||||
"WARNING": 2,
|
||||
"ERROR": 3,
|
||||
}
|
||||
|
||||
func (x TransferMessage_MessageSeverity) String() string {
|
||||
return proto.EnumName(TransferMessage_MessageSeverity_name, int32(x))
|
||||
}
|
||||
func (TransferMessage_MessageSeverity) EnumDescriptor() ([]byte, []int) {
|
||||
return fileDescriptor1, []int{2, 0}
|
||||
return fileDescriptor_transfer_b655bdd8f06d7255, []int{2, 0}
|
||||
}
|
||||
|
||||
// Represents a data transfer configuration. A transfer configuration
|
||||
@ -126,7 +136,6 @@ func (TransferMessage_MessageSeverity) EnumDescriptor() ([]byte, []int) {
|
||||
// When a new transfer configuration is created, the specified
|
||||
// `destination_dataset_id` is created when needed and shared with the
|
||||
// appropriate data source service account.
|
||||
// Next id: 20
|
||||
type TransferConfig struct {
|
||||
// The resource name of the transfer config.
|
||||
// Transfer config names have the form
|
||||
@ -134,15 +143,15 @@ type TransferConfig struct {
|
||||
// Where `config_id` is usually a uuid, even though it is not
|
||||
// guaranteed or required. The name is ignored when creating a transfer
|
||||
// config.
|
||||
Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"`
|
||||
Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"`
|
||||
// The BigQuery target dataset id.
|
||||
DestinationDatasetId string `protobuf:"bytes,2,opt,name=destination_dataset_id,json=destinationDatasetId" json:"destination_dataset_id,omitempty"`
|
||||
DestinationDatasetId string `protobuf:"bytes,2,opt,name=destination_dataset_id,json=destinationDatasetId,proto3" json:"destination_dataset_id,omitempty"`
|
||||
// User specified display name for the data transfer.
|
||||
DisplayName string `protobuf:"bytes,3,opt,name=display_name,json=displayName" json:"display_name,omitempty"`
|
||||
DisplayName string `protobuf:"bytes,3,opt,name=display_name,json=displayName,proto3" json:"display_name,omitempty"`
|
||||
// Data source id. Cannot be changed once data transfer is created.
|
||||
DataSourceId string `protobuf:"bytes,5,opt,name=data_source_id,json=dataSourceId" json:"data_source_id,omitempty"`
|
||||
DataSourceId string `protobuf:"bytes,5,opt,name=data_source_id,json=dataSourceId,proto3" json:"data_source_id,omitempty"`
|
||||
// Data transfer specific parameters.
|
||||
Params *google_protobuf1.Struct `protobuf:"bytes,9,opt,name=params" json:"params,omitempty"`
|
||||
Params *_struct.Struct `protobuf:"bytes,9,opt,name=params,proto3" json:"params,omitempty"`
|
||||
// Data transfer schedule.
|
||||
// If the data source does not support a custom schedule, this should be
|
||||
// empty. If it is empty, the default value for the data source will be
|
||||
@ -155,37 +164,59 @@ type TransferConfig struct {
|
||||
// See more explanation about the format here:
|
||||
// https://cloud.google.com/appengine/docs/flexible/python/scheduling-jobs-with-cron-yaml#the_schedule_format
|
||||
// NOTE: the granularity should be at least 8 hours, or less frequent.
|
||||
Schedule string `protobuf:"bytes,7,opt,name=schedule" json:"schedule,omitempty"`
|
||||
Schedule string `protobuf:"bytes,7,opt,name=schedule,proto3" json:"schedule,omitempty"`
|
||||
// The number of days to look back to automatically refresh the data.
|
||||
// For example, if `data_refresh_window_days = 10`, then every day
|
||||
// BigQuery reingests data for [today-10, today-1], rather than ingesting data
|
||||
// for just [today-1].
|
||||
// Only valid if the data source supports the feature. Set the value to 0
|
||||
// to use the default value.
|
||||
DataRefreshWindowDays int32 `protobuf:"varint,12,opt,name=data_refresh_window_days,json=dataRefreshWindowDays" json:"data_refresh_window_days,omitempty"`
|
||||
DataRefreshWindowDays int32 `protobuf:"varint,12,opt,name=data_refresh_window_days,json=dataRefreshWindowDays,proto3" json:"data_refresh_window_days,omitempty"`
|
||||
// Is this config disabled. When set to true, no runs are scheduled
|
||||
// for a given transfer.
|
||||
Disabled bool `protobuf:"varint,13,opt,name=disabled" json:"disabled,omitempty"`
|
||||
Disabled bool `protobuf:"varint,13,opt,name=disabled,proto3" json:"disabled,omitempty"`
|
||||
// Output only. Data transfer modification time. Ignored by server on input.
|
||||
UpdateTime *google_protobuf2.Timestamp `protobuf:"bytes,4,opt,name=update_time,json=updateTime" json:"update_time,omitempty"`
|
||||
UpdateTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=update_time,json=updateTime,proto3" json:"update_time,omitempty"`
|
||||
// Output only. Next time when data transfer will run.
|
||||
NextRunTime *google_protobuf2.Timestamp `protobuf:"bytes,8,opt,name=next_run_time,json=nextRunTime" json:"next_run_time,omitempty"`
|
||||
NextRunTime *timestamp.Timestamp `protobuf:"bytes,8,opt,name=next_run_time,json=nextRunTime,proto3" json:"next_run_time,omitempty"`
|
||||
// Output only. State of the most recently updated transfer run.
|
||||
State TransferState `protobuf:"varint,10,opt,name=state,enum=google.cloud.bigquery.datatransfer.v1.TransferState" json:"state,omitempty"`
|
||||
State TransferState `protobuf:"varint,10,opt,name=state,proto3,enum=google.cloud.bigquery.datatransfer.v1.TransferState" json:"state,omitempty"`
|
||||
// Output only. Unique ID of the user on whose behalf transfer is done.
|
||||
// Applicable only to data sources that do not support service accounts.
|
||||
// When set to 0, the data source service account credentials are used.
|
||||
// May be negative. Note, that this identifier is not stable.
|
||||
// It may change over time even for the same user.
|
||||
UserId int64 `protobuf:"varint,11,opt,name=user_id,json=userId" json:"user_id,omitempty"`
|
||||
UserId int64 `protobuf:"varint,11,opt,name=user_id,json=userId,proto3" json:"user_id,omitempty"`
|
||||
// Output only. Region in which BigQuery dataset is located.
|
||||
DatasetRegion string `protobuf:"bytes,14,opt,name=dataset_region,json=datasetRegion" json:"dataset_region,omitempty"`
|
||||
DatasetRegion string `protobuf:"bytes,14,opt,name=dataset_region,json=datasetRegion,proto3" json:"dataset_region,omitempty"`
|
||||
XXX_NoUnkeyedLiteral struct{} `json:"-"`
|
||||
XXX_unrecognized []byte `json:"-"`
|
||||
XXX_sizecache int32 `json:"-"`
|
||||
}
|
||||
|
||||
func (m *TransferConfig) Reset() { *m = TransferConfig{} }
|
||||
func (m *TransferConfig) String() string { return proto.CompactTextString(m) }
|
||||
func (*TransferConfig) ProtoMessage() {}
|
||||
func (*TransferConfig) Descriptor() ([]byte, []int) { return fileDescriptor1, []int{0} }
|
||||
func (m *TransferConfig) Reset() { *m = TransferConfig{} }
|
||||
func (m *TransferConfig) String() string { return proto.CompactTextString(m) }
|
||||
func (*TransferConfig) ProtoMessage() {}
|
||||
func (*TransferConfig) Descriptor() ([]byte, []int) {
|
||||
return fileDescriptor_transfer_b655bdd8f06d7255, []int{0}
|
||||
}
|
||||
func (m *TransferConfig) XXX_Unmarshal(b []byte) error {
|
||||
return xxx_messageInfo_TransferConfig.Unmarshal(m, b)
|
||||
}
|
||||
func (m *TransferConfig) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) {
|
||||
return xxx_messageInfo_TransferConfig.Marshal(b, m, deterministic)
|
||||
}
|
||||
func (dst *TransferConfig) XXX_Merge(src proto.Message) {
|
||||
xxx_messageInfo_TransferConfig.Merge(dst, src)
|
||||
}
|
||||
func (m *TransferConfig) XXX_Size() int {
|
||||
return xxx_messageInfo_TransferConfig.Size(m)
|
||||
}
|
||||
func (m *TransferConfig) XXX_DiscardUnknown() {
|
||||
xxx_messageInfo_TransferConfig.DiscardUnknown(m)
|
||||
}
|
||||
|
||||
var xxx_messageInfo_TransferConfig proto.InternalMessageInfo
|
||||
|
||||
func (m *TransferConfig) GetName() string {
|
||||
if m != nil {
|
||||
@ -215,7 +246,7 @@ func (m *TransferConfig) GetDataSourceId() string {
|
||||
return ""
|
||||
}
|
||||
|
||||
func (m *TransferConfig) GetParams() *google_protobuf1.Struct {
|
||||
func (m *TransferConfig) GetParams() *_struct.Struct {
|
||||
if m != nil {
|
||||
return m.Params
|
||||
}
|
||||
@ -243,14 +274,14 @@ func (m *TransferConfig) GetDisabled() bool {
|
||||
return false
|
||||
}
|
||||
|
||||
func (m *TransferConfig) GetUpdateTime() *google_protobuf2.Timestamp {
|
||||
func (m *TransferConfig) GetUpdateTime() *timestamp.Timestamp {
|
||||
if m != nil {
|
||||
return m.UpdateTime
|
||||
}
|
||||
return nil
|
||||
}
|
||||
|
||||
func (m *TransferConfig) GetNextRunTime() *google_protobuf2.Timestamp {
|
||||
func (m *TransferConfig) GetNextRunTime() *timestamp.Timestamp {
|
||||
if m != nil {
|
||||
return m.NextRunTime
|
||||
}
|
||||
@ -279,54 +310,75 @@ func (m *TransferConfig) GetDatasetRegion() string {
|
||||
}
|
||||
|
||||
// Represents a data transfer run.
|
||||
// Next id: 27
|
||||
type TransferRun struct {
|
||||
// The resource name of the transfer run.
|
||||
// Transfer run names have the form
|
||||
// `projects/{project_id}/locations/{location}/transferConfigs/{config_id}/runs/{run_id}`.
|
||||
// The name is ignored when creating a transfer run.
|
||||
Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"`
|
||||
Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"`
|
||||
// Minimum time after which a transfer run can be started.
|
||||
ScheduleTime *google_protobuf2.Timestamp `protobuf:"bytes,3,opt,name=schedule_time,json=scheduleTime" json:"schedule_time,omitempty"`
|
||||
ScheduleTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=schedule_time,json=scheduleTime,proto3" json:"schedule_time,omitempty"`
|
||||
// For batch transfer runs, specifies the date and time that
|
||||
// data should be ingested.
|
||||
RunTime *google_protobuf2.Timestamp `protobuf:"bytes,10,opt,name=run_time,json=runTime" json:"run_time,omitempty"`
|
||||
RunTime *timestamp.Timestamp `protobuf:"bytes,10,opt,name=run_time,json=runTime,proto3" json:"run_time,omitempty"`
|
||||
// Status of the transfer run.
|
||||
ErrorStatus *google_rpc.Status `protobuf:"bytes,21,opt,name=error_status,json=errorStatus" json:"error_status,omitempty"`
|
||||
ErrorStatus *status.Status `protobuf:"bytes,21,opt,name=error_status,json=errorStatus,proto3" json:"error_status,omitempty"`
|
||||
// Output only. Time when transfer run was started.
|
||||
// Parameter ignored by server for input requests.
|
||||
StartTime *google_protobuf2.Timestamp `protobuf:"bytes,4,opt,name=start_time,json=startTime" json:"start_time,omitempty"`
|
||||
StartTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=start_time,json=startTime,proto3" json:"start_time,omitempty"`
|
||||
// Output only. Time when transfer run ended.
|
||||
// Parameter ignored by server for input requests.
|
||||
EndTime *google_protobuf2.Timestamp `protobuf:"bytes,5,opt,name=end_time,json=endTime" json:"end_time,omitempty"`
|
||||
EndTime *timestamp.Timestamp `protobuf:"bytes,5,opt,name=end_time,json=endTime,proto3" json:"end_time,omitempty"`
|
||||
// Output only. Last time the data transfer run state was updated.
|
||||
UpdateTime *google_protobuf2.Timestamp `protobuf:"bytes,6,opt,name=update_time,json=updateTime" json:"update_time,omitempty"`
|
||||
UpdateTime *timestamp.Timestamp `protobuf:"bytes,6,opt,name=update_time,json=updateTime,proto3" json:"update_time,omitempty"`
|
||||
// Output only. Data transfer specific parameters.
|
||||
Params *google_protobuf1.Struct `protobuf:"bytes,9,opt,name=params" json:"params,omitempty"`
|
||||
Params *_struct.Struct `protobuf:"bytes,9,opt,name=params,proto3" json:"params,omitempty"`
|
||||
// Output only. The BigQuery target dataset id.
|
||||
DestinationDatasetId string `protobuf:"bytes,2,opt,name=destination_dataset_id,json=destinationDatasetId" json:"destination_dataset_id,omitempty"`
|
||||
DestinationDatasetId string `protobuf:"bytes,2,opt,name=destination_dataset_id,json=destinationDatasetId,proto3" json:"destination_dataset_id,omitempty"`
|
||||
// Output only. Data source id.
|
||||
DataSourceId string `protobuf:"bytes,7,opt,name=data_source_id,json=dataSourceId" json:"data_source_id,omitempty"`
|
||||
DataSourceId string `protobuf:"bytes,7,opt,name=data_source_id,json=dataSourceId,proto3" json:"data_source_id,omitempty"`
|
||||
// Data transfer run state. Ignored for input requests.
|
||||
State TransferState `protobuf:"varint,8,opt,name=state,enum=google.cloud.bigquery.datatransfer.v1.TransferState" json:"state,omitempty"`
|
||||
State TransferState `protobuf:"varint,8,opt,name=state,proto3,enum=google.cloud.bigquery.datatransfer.v1.TransferState" json:"state,omitempty"`
|
||||
// Output only. Unique ID of the user on whose behalf transfer is done.
|
||||
// Applicable only to data sources that do not support service accounts.
|
||||
// When set to 0, the data source service account credentials are used.
|
||||
// May be negative. Note, that this identifier is not stable.
|
||||
// It may change over time even for the same user.
|
||||
UserId int64 `protobuf:"varint,11,opt,name=user_id,json=userId" json:"user_id,omitempty"`
|
||||
UserId int64 `protobuf:"varint,11,opt,name=user_id,json=userId,proto3" json:"user_id,omitempty"`
|
||||
// Output only. Describes the schedule of this transfer run if it was
|
||||
// created as part of a regular schedule. For batch transfer runs that are
|
||||
// scheduled manually, this is empty.
|
||||
// NOTE: the system might choose to delay the schedule depending on the
|
||||
// current load, so `schedule_time` doesn't always matches this.
|
||||
Schedule string `protobuf:"bytes,12,opt,name=schedule" json:"schedule,omitempty"`
|
||||
Schedule string `protobuf:"bytes,12,opt,name=schedule,proto3" json:"schedule,omitempty"`
|
||||
XXX_NoUnkeyedLiteral struct{} `json:"-"`
|
||||
XXX_unrecognized []byte `json:"-"`
|
||||
XXX_sizecache int32 `json:"-"`
|
||||
}
|
||||
|
||||
func (m *TransferRun) Reset() { *m = TransferRun{} }
|
||||
func (m *TransferRun) String() string { return proto.CompactTextString(m) }
|
||||
func (*TransferRun) ProtoMessage() {}
|
||||
func (*TransferRun) Descriptor() ([]byte, []int) { return fileDescriptor1, []int{1} }
|
||||
func (m *TransferRun) Reset() { *m = TransferRun{} }
|
||||
func (m *TransferRun) String() string { return proto.CompactTextString(m) }
|
||||
func (*TransferRun) ProtoMessage() {}
|
||||
func (*TransferRun) Descriptor() ([]byte, []int) {
|
||||
return fileDescriptor_transfer_b655bdd8f06d7255, []int{1}
|
||||
}
|
||||
func (m *TransferRun) XXX_Unmarshal(b []byte) error {
|
||||
return xxx_messageInfo_TransferRun.Unmarshal(m, b)
|
||||
}
|
||||
func (m *TransferRun) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) {
|
||||
return xxx_messageInfo_TransferRun.Marshal(b, m, deterministic)
|
||||
}
|
||||
func (dst *TransferRun) XXX_Merge(src proto.Message) {
|
||||
xxx_messageInfo_TransferRun.Merge(dst, src)
|
||||
}
|
||||
func (m *TransferRun) XXX_Size() int {
|
||||
return xxx_messageInfo_TransferRun.Size(m)
|
||||
}
|
||||
func (m *TransferRun) XXX_DiscardUnknown() {
|
||||
xxx_messageInfo_TransferRun.DiscardUnknown(m)
|
||||
}
|
||||
|
||||
var xxx_messageInfo_TransferRun proto.InternalMessageInfo
|
||||
|
||||
func (m *TransferRun) GetName() string {
|
||||
if m != nil {
|
||||
@ -335,49 +387,49 @@ func (m *TransferRun) GetName() string {
|
||||
return ""
|
||||
}
|
||||
|
||||
func (m *TransferRun) GetScheduleTime() *google_protobuf2.Timestamp {
|
||||
func (m *TransferRun) GetScheduleTime() *timestamp.Timestamp {
|
||||
if m != nil {
|
||||
return m.ScheduleTime
|
||||
}
|
||||
return nil
|
||||
}
|
||||
|
||||
func (m *TransferRun) GetRunTime() *google_protobuf2.Timestamp {
|
||||
func (m *TransferRun) GetRunTime() *timestamp.Timestamp {
|
||||
if m != nil {
|
||||
return m.RunTime
|
||||
}
|
||||
return nil
|
||||
}
|
||||
|
||||
func (m *TransferRun) GetErrorStatus() *google_rpc.Status {
|
||||
func (m *TransferRun) GetErrorStatus() *status.Status {
|
||||
if m != nil {
|
||||
return m.ErrorStatus
|
||||
}
|
||||
return nil
|
||||
}
|
||||
|
||||
func (m *TransferRun) GetStartTime() *google_protobuf2.Timestamp {
|
||||
func (m *TransferRun) GetStartTime() *timestamp.Timestamp {
|
||||
if m != nil {
|
||||
return m.StartTime
|
||||
}
|
||||
return nil
|
||||
}
|
||||
|
||||
func (m *TransferRun) GetEndTime() *google_protobuf2.Timestamp {
|
||||
func (m *TransferRun) GetEndTime() *timestamp.Timestamp {
|
||||
if m != nil {
|
||||
return m.EndTime
|
||||
}
|
||||
return nil
|
||||
}
|
||||
|
||||
func (m *TransferRun) GetUpdateTime() *google_protobuf2.Timestamp {
|
||||
func (m *TransferRun) GetUpdateTime() *timestamp.Timestamp {
|
||||
if m != nil {
|
||||
return m.UpdateTime
|
||||
}
|
||||
return nil
|
||||
}
|
||||
|
||||
func (m *TransferRun) GetParams() *google_protobuf1.Struct {
|
||||
func (m *TransferRun) GetParams() *_struct.Struct {
|
||||
if m != nil {
|
||||
return m.Params
|
||||
}
|
||||
@ -422,19 +474,41 @@ func (m *TransferRun) GetSchedule() string {
|
||||
// Represents a user facing message for a particular data transfer run.
|
||||
type TransferMessage struct {
|
||||
// Time when message was logged.
|
||||
MessageTime *google_protobuf2.Timestamp `protobuf:"bytes,1,opt,name=message_time,json=messageTime" json:"message_time,omitempty"`
|
||||
MessageTime *timestamp.Timestamp `protobuf:"bytes,1,opt,name=message_time,json=messageTime,proto3" json:"message_time,omitempty"`
|
||||
// Message severity.
|
||||
Severity TransferMessage_MessageSeverity `protobuf:"varint,2,opt,name=severity,enum=google.cloud.bigquery.datatransfer.v1.TransferMessage_MessageSeverity" json:"severity,omitempty"`
|
||||
Severity TransferMessage_MessageSeverity `protobuf:"varint,2,opt,name=severity,proto3,enum=google.cloud.bigquery.datatransfer.v1.TransferMessage_MessageSeverity" json:"severity,omitempty"`
|
||||
// Message text.
|
||||
MessageText string `protobuf:"bytes,3,opt,name=message_text,json=messageText" json:"message_text,omitempty"`
|
||||
MessageText string `protobuf:"bytes,3,opt,name=message_text,json=messageText,proto3" json:"message_text,omitempty"`
|
||||
XXX_NoUnkeyedLiteral struct{} `json:"-"`
|
||||
XXX_unrecognized []byte `json:"-"`
|
||||
XXX_sizecache int32 `json:"-"`
|
||||
}
|
||||
|
||||
func (m *TransferMessage) Reset() { *m = TransferMessage{} }
|
||||
func (m *TransferMessage) String() string { return proto.CompactTextString(m) }
|
||||
func (*TransferMessage) ProtoMessage() {}
|
||||
func (*TransferMessage) Descriptor() ([]byte, []int) { return fileDescriptor1, []int{2} }
|
||||
func (m *TransferMessage) Reset() { *m = TransferMessage{} }
|
||||
func (m *TransferMessage) String() string { return proto.CompactTextString(m) }
|
||||
func (*TransferMessage) ProtoMessage() {}
|
||||
func (*TransferMessage) Descriptor() ([]byte, []int) {
|
||||
return fileDescriptor_transfer_b655bdd8f06d7255, []int{2}
|
||||
}
|
||||
func (m *TransferMessage) XXX_Unmarshal(b []byte) error {
|
||||
return xxx_messageInfo_TransferMessage.Unmarshal(m, b)
|
||||
}
|
||||
func (m *TransferMessage) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) {
|
||||
return xxx_messageInfo_TransferMessage.Marshal(b, m, deterministic)
|
||||
}
|
||||
func (dst *TransferMessage) XXX_Merge(src proto.Message) {
|
||||
xxx_messageInfo_TransferMessage.Merge(dst, src)
|
||||
}
|
||||
func (m *TransferMessage) XXX_Size() int {
|
||||
return xxx_messageInfo_TransferMessage.Size(m)
|
||||
}
|
||||
func (m *TransferMessage) XXX_DiscardUnknown() {
|
||||
xxx_messageInfo_TransferMessage.DiscardUnknown(m)
|
||||
}
|
||||
|
||||
func (m *TransferMessage) GetMessageTime() *google_protobuf2.Timestamp {
|
||||
var xxx_messageInfo_TransferMessage proto.InternalMessageInfo
|
||||
|
||||
func (m *TransferMessage) GetMessageTime() *timestamp.Timestamp {
|
||||
if m != nil {
|
||||
return m.MessageTime
|
||||
}
|
||||
@ -465,10 +539,10 @@ func init() {
|
||||
}
|
||||
|
||||
func init() {
|
||||
proto.RegisterFile("google/cloud/bigquery/datatransfer/v1/transfer.proto", fileDescriptor1)
|
||||
proto.RegisterFile("google/cloud/bigquery/datatransfer/v1/transfer.proto", fileDescriptor_transfer_b655bdd8f06d7255)
|
||||
}
|
||||
|
||||
var fileDescriptor1 = []byte{
|
||||
var fileDescriptor_transfer_b655bdd8f06d7255 = []byte{
|
||||
// 922 bytes of a gzipped FileDescriptorProto
|
||||
0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xb4, 0x56, 0xdd, 0x6e, 0xe3, 0x44,
|
||||
0x14, 0xc6, 0xf9, 0xcf, 0x71, 0x92, 0x8d, 0x46, 0x2c, 0x35, 0xd5, 0x02, 0xa1, 0xa2, 0x52, 0xd8,
|
||||
|
1657
vendor/google.golang.org/genproto/googleapis/cloud/bigquery/logging/v1/audit_data.pb.go
generated
vendored
1657
vendor/google.golang.org/genproto/googleapis/cloud/bigquery/logging/v1/audit_data.pb.go
generated
vendored
File diff suppressed because it is too large
Load Diff
Reference in New Issue
Block a user