vendor update for CSI 0.3.0

This commit is contained in:
gman
2018-07-18 16:47:22 +02:00
parent 6f484f92fc
commit 8ea659f0d5
6810 changed files with 438061 additions and 193861 deletions

View File

@ -1,22 +1,28 @@
// Code generated by protoc-gen-go. DO NOT EDIT.
// source: google/cloud/bigquery/datatransfer/v1/transfer.proto
package datatransfer
package datatransfer // import "google.golang.org/genproto/googleapis/cloud/bigquery/datatransfer/v1"
import proto "github.com/golang/protobuf/proto"
import fmt "fmt"
import math "math"
import _struct "github.com/golang/protobuf/ptypes/struct"
import timestamp "github.com/golang/protobuf/ptypes/timestamp"
import _ "google.golang.org/genproto/googleapis/api/annotations"
import google_protobuf1 "github.com/golang/protobuf/ptypes/struct"
import google_protobuf2 "github.com/golang/protobuf/ptypes/timestamp"
import google_rpc "google.golang.org/genproto/googleapis/rpc/status"
import status "google.golang.org/genproto/googleapis/rpc/status"
// Reference imports to suppress errors if they are not otherwise used.
var _ = proto.Marshal
var _ = fmt.Errorf
var _ = math.Inf
// Represents data transfer type.
// This is a compile-time assertion to ensure that this generated file
// is compatible with the proto package it is being compiled against.
// A compilation error at this line likely means your copy of the
// proto package needs to be updated.
const _ = proto.ProtoPackageIsVersion2 // please upgrade the proto package
// DEPRECATED. Represents data transfer type.
type TransferType int32
const (
@ -43,7 +49,9 @@ var TransferType_value = map[string]int32{
func (x TransferType) String() string {
return proto.EnumName(TransferType_name, int32(x))
}
func (TransferType) EnumDescriptor() ([]byte, []int) { return fileDescriptor1, []int{0} }
func (TransferType) EnumDescriptor() ([]byte, []int) {
return fileDescriptor_transfer_b655bdd8f06d7255, []int{0}
}
// Represents data transfer run state.
type TransferState int32
@ -84,7 +92,9 @@ var TransferState_value = map[string]int32{
func (x TransferState) String() string {
return proto.EnumName(TransferState_name, int32(x))
}
func (TransferState) EnumDescriptor() ([]byte, []int) { return fileDescriptor1, []int{1} }
func (TransferState) EnumDescriptor() ([]byte, []int) {
return fileDescriptor_transfer_b655bdd8f06d7255, []int{1}
}
// Represents data transfer user facing message severity.
type TransferMessage_MessageSeverity int32
@ -108,16 +118,16 @@ var TransferMessage_MessageSeverity_name = map[int32]string{
}
var TransferMessage_MessageSeverity_value = map[string]int32{
"MESSAGE_SEVERITY_UNSPECIFIED": 0,
"INFO": 1,
"WARNING": 2,
"ERROR": 3,
"INFO": 1,
"WARNING": 2,
"ERROR": 3,
}
func (x TransferMessage_MessageSeverity) String() string {
return proto.EnumName(TransferMessage_MessageSeverity_name, int32(x))
}
func (TransferMessage_MessageSeverity) EnumDescriptor() ([]byte, []int) {
return fileDescriptor1, []int{2, 0}
return fileDescriptor_transfer_b655bdd8f06d7255, []int{2, 0}
}
// Represents a data transfer configuration. A transfer configuration
@ -126,7 +136,6 @@ func (TransferMessage_MessageSeverity) EnumDescriptor() ([]byte, []int) {
// When a new transfer configuration is created, the specified
// `destination_dataset_id` is created when needed and shared with the
// appropriate data source service account.
// Next id: 20
type TransferConfig struct {
// The resource name of the transfer config.
// Transfer config names have the form
@ -134,15 +143,15 @@ type TransferConfig struct {
// Where `config_id` is usually a uuid, even though it is not
// guaranteed or required. The name is ignored when creating a transfer
// config.
Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"`
Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"`
// The BigQuery target dataset id.
DestinationDatasetId string `protobuf:"bytes,2,opt,name=destination_dataset_id,json=destinationDatasetId" json:"destination_dataset_id,omitempty"`
DestinationDatasetId string `protobuf:"bytes,2,opt,name=destination_dataset_id,json=destinationDatasetId,proto3" json:"destination_dataset_id,omitempty"`
// User specified display name for the data transfer.
DisplayName string `protobuf:"bytes,3,opt,name=display_name,json=displayName" json:"display_name,omitempty"`
DisplayName string `protobuf:"bytes,3,opt,name=display_name,json=displayName,proto3" json:"display_name,omitempty"`
// Data source id. Cannot be changed once data transfer is created.
DataSourceId string `protobuf:"bytes,5,opt,name=data_source_id,json=dataSourceId" json:"data_source_id,omitempty"`
DataSourceId string `protobuf:"bytes,5,opt,name=data_source_id,json=dataSourceId,proto3" json:"data_source_id,omitempty"`
// Data transfer specific parameters.
Params *google_protobuf1.Struct `protobuf:"bytes,9,opt,name=params" json:"params,omitempty"`
Params *_struct.Struct `protobuf:"bytes,9,opt,name=params,proto3" json:"params,omitempty"`
// Data transfer schedule.
// If the data source does not support a custom schedule, this should be
// empty. If it is empty, the default value for the data source will be
@ -155,37 +164,59 @@ type TransferConfig struct {
// See more explanation about the format here:
// https://cloud.google.com/appengine/docs/flexible/python/scheduling-jobs-with-cron-yaml#the_schedule_format
// NOTE: the granularity should be at least 8 hours, or less frequent.
Schedule string `protobuf:"bytes,7,opt,name=schedule" json:"schedule,omitempty"`
Schedule string `protobuf:"bytes,7,opt,name=schedule,proto3" json:"schedule,omitempty"`
// The number of days to look back to automatically refresh the data.
// For example, if `data_refresh_window_days = 10`, then every day
// BigQuery reingests data for [today-10, today-1], rather than ingesting data
// for just [today-1].
// Only valid if the data source supports the feature. Set the value to 0
// to use the default value.
DataRefreshWindowDays int32 `protobuf:"varint,12,opt,name=data_refresh_window_days,json=dataRefreshWindowDays" json:"data_refresh_window_days,omitempty"`
DataRefreshWindowDays int32 `protobuf:"varint,12,opt,name=data_refresh_window_days,json=dataRefreshWindowDays,proto3" json:"data_refresh_window_days,omitempty"`
// Is this config disabled. When set to true, no runs are scheduled
// for a given transfer.
Disabled bool `protobuf:"varint,13,opt,name=disabled" json:"disabled,omitempty"`
Disabled bool `protobuf:"varint,13,opt,name=disabled,proto3" json:"disabled,omitempty"`
// Output only. Data transfer modification time. Ignored by server on input.
UpdateTime *google_protobuf2.Timestamp `protobuf:"bytes,4,opt,name=update_time,json=updateTime" json:"update_time,omitempty"`
UpdateTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=update_time,json=updateTime,proto3" json:"update_time,omitempty"`
// Output only. Next time when data transfer will run.
NextRunTime *google_protobuf2.Timestamp `protobuf:"bytes,8,opt,name=next_run_time,json=nextRunTime" json:"next_run_time,omitempty"`
NextRunTime *timestamp.Timestamp `protobuf:"bytes,8,opt,name=next_run_time,json=nextRunTime,proto3" json:"next_run_time,omitempty"`
// Output only. State of the most recently updated transfer run.
State TransferState `protobuf:"varint,10,opt,name=state,enum=google.cloud.bigquery.datatransfer.v1.TransferState" json:"state,omitempty"`
State TransferState `protobuf:"varint,10,opt,name=state,proto3,enum=google.cloud.bigquery.datatransfer.v1.TransferState" json:"state,omitempty"`
// Output only. Unique ID of the user on whose behalf transfer is done.
// Applicable only to data sources that do not support service accounts.
// When set to 0, the data source service account credentials are used.
// May be negative. Note, that this identifier is not stable.
// It may change over time even for the same user.
UserId int64 `protobuf:"varint,11,opt,name=user_id,json=userId" json:"user_id,omitempty"`
UserId int64 `protobuf:"varint,11,opt,name=user_id,json=userId,proto3" json:"user_id,omitempty"`
// Output only. Region in which BigQuery dataset is located.
DatasetRegion string `protobuf:"bytes,14,opt,name=dataset_region,json=datasetRegion" json:"dataset_region,omitempty"`
DatasetRegion string `protobuf:"bytes,14,opt,name=dataset_region,json=datasetRegion,proto3" json:"dataset_region,omitempty"`
XXX_NoUnkeyedLiteral struct{} `json:"-"`
XXX_unrecognized []byte `json:"-"`
XXX_sizecache int32 `json:"-"`
}
func (m *TransferConfig) Reset() { *m = TransferConfig{} }
func (m *TransferConfig) String() string { return proto.CompactTextString(m) }
func (*TransferConfig) ProtoMessage() {}
func (*TransferConfig) Descriptor() ([]byte, []int) { return fileDescriptor1, []int{0} }
func (m *TransferConfig) Reset() { *m = TransferConfig{} }
func (m *TransferConfig) String() string { return proto.CompactTextString(m) }
func (*TransferConfig) ProtoMessage() {}
func (*TransferConfig) Descriptor() ([]byte, []int) {
return fileDescriptor_transfer_b655bdd8f06d7255, []int{0}
}
func (m *TransferConfig) XXX_Unmarshal(b []byte) error {
return xxx_messageInfo_TransferConfig.Unmarshal(m, b)
}
func (m *TransferConfig) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) {
return xxx_messageInfo_TransferConfig.Marshal(b, m, deterministic)
}
func (dst *TransferConfig) XXX_Merge(src proto.Message) {
xxx_messageInfo_TransferConfig.Merge(dst, src)
}
func (m *TransferConfig) XXX_Size() int {
return xxx_messageInfo_TransferConfig.Size(m)
}
func (m *TransferConfig) XXX_DiscardUnknown() {
xxx_messageInfo_TransferConfig.DiscardUnknown(m)
}
var xxx_messageInfo_TransferConfig proto.InternalMessageInfo
func (m *TransferConfig) GetName() string {
if m != nil {
@ -215,7 +246,7 @@ func (m *TransferConfig) GetDataSourceId() string {
return ""
}
func (m *TransferConfig) GetParams() *google_protobuf1.Struct {
func (m *TransferConfig) GetParams() *_struct.Struct {
if m != nil {
return m.Params
}
@ -243,14 +274,14 @@ func (m *TransferConfig) GetDisabled() bool {
return false
}
func (m *TransferConfig) GetUpdateTime() *google_protobuf2.Timestamp {
func (m *TransferConfig) GetUpdateTime() *timestamp.Timestamp {
if m != nil {
return m.UpdateTime
}
return nil
}
func (m *TransferConfig) GetNextRunTime() *google_protobuf2.Timestamp {
func (m *TransferConfig) GetNextRunTime() *timestamp.Timestamp {
if m != nil {
return m.NextRunTime
}
@ -279,54 +310,75 @@ func (m *TransferConfig) GetDatasetRegion() string {
}
// Represents a data transfer run.
// Next id: 27
type TransferRun struct {
// The resource name of the transfer run.
// Transfer run names have the form
// `projects/{project_id}/locations/{location}/transferConfigs/{config_id}/runs/{run_id}`.
// The name is ignored when creating a transfer run.
Name string `protobuf:"bytes,1,opt,name=name" json:"name,omitempty"`
Name string `protobuf:"bytes,1,opt,name=name,proto3" json:"name,omitempty"`
// Minimum time after which a transfer run can be started.
ScheduleTime *google_protobuf2.Timestamp `protobuf:"bytes,3,opt,name=schedule_time,json=scheduleTime" json:"schedule_time,omitempty"`
ScheduleTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=schedule_time,json=scheduleTime,proto3" json:"schedule_time,omitempty"`
// For batch transfer runs, specifies the date and time that
// data should be ingested.
RunTime *google_protobuf2.Timestamp `protobuf:"bytes,10,opt,name=run_time,json=runTime" json:"run_time,omitempty"`
RunTime *timestamp.Timestamp `protobuf:"bytes,10,opt,name=run_time,json=runTime,proto3" json:"run_time,omitempty"`
// Status of the transfer run.
ErrorStatus *google_rpc.Status `protobuf:"bytes,21,opt,name=error_status,json=errorStatus" json:"error_status,omitempty"`
ErrorStatus *status.Status `protobuf:"bytes,21,opt,name=error_status,json=errorStatus,proto3" json:"error_status,omitempty"`
// Output only. Time when transfer run was started.
// Parameter ignored by server for input requests.
StartTime *google_protobuf2.Timestamp `protobuf:"bytes,4,opt,name=start_time,json=startTime" json:"start_time,omitempty"`
StartTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=start_time,json=startTime,proto3" json:"start_time,omitempty"`
// Output only. Time when transfer run ended.
// Parameter ignored by server for input requests.
EndTime *google_protobuf2.Timestamp `protobuf:"bytes,5,opt,name=end_time,json=endTime" json:"end_time,omitempty"`
EndTime *timestamp.Timestamp `protobuf:"bytes,5,opt,name=end_time,json=endTime,proto3" json:"end_time,omitempty"`
// Output only. Last time the data transfer run state was updated.
UpdateTime *google_protobuf2.Timestamp `protobuf:"bytes,6,opt,name=update_time,json=updateTime" json:"update_time,omitempty"`
UpdateTime *timestamp.Timestamp `protobuf:"bytes,6,opt,name=update_time,json=updateTime,proto3" json:"update_time,omitempty"`
// Output only. Data transfer specific parameters.
Params *google_protobuf1.Struct `protobuf:"bytes,9,opt,name=params" json:"params,omitempty"`
Params *_struct.Struct `protobuf:"bytes,9,opt,name=params,proto3" json:"params,omitempty"`
// Output only. The BigQuery target dataset id.
DestinationDatasetId string `protobuf:"bytes,2,opt,name=destination_dataset_id,json=destinationDatasetId" json:"destination_dataset_id,omitempty"`
DestinationDatasetId string `protobuf:"bytes,2,opt,name=destination_dataset_id,json=destinationDatasetId,proto3" json:"destination_dataset_id,omitempty"`
// Output only. Data source id.
DataSourceId string `protobuf:"bytes,7,opt,name=data_source_id,json=dataSourceId" json:"data_source_id,omitempty"`
DataSourceId string `protobuf:"bytes,7,opt,name=data_source_id,json=dataSourceId,proto3" json:"data_source_id,omitempty"`
// Data transfer run state. Ignored for input requests.
State TransferState `protobuf:"varint,8,opt,name=state,enum=google.cloud.bigquery.datatransfer.v1.TransferState" json:"state,omitempty"`
State TransferState `protobuf:"varint,8,opt,name=state,proto3,enum=google.cloud.bigquery.datatransfer.v1.TransferState" json:"state,omitempty"`
// Output only. Unique ID of the user on whose behalf transfer is done.
// Applicable only to data sources that do not support service accounts.
// When set to 0, the data source service account credentials are used.
// May be negative. Note, that this identifier is not stable.
// It may change over time even for the same user.
UserId int64 `protobuf:"varint,11,opt,name=user_id,json=userId" json:"user_id,omitempty"`
UserId int64 `protobuf:"varint,11,opt,name=user_id,json=userId,proto3" json:"user_id,omitempty"`
// Output only. Describes the schedule of this transfer run if it was
// created as part of a regular schedule. For batch transfer runs that are
// scheduled manually, this is empty.
// NOTE: the system might choose to delay the schedule depending on the
// current load, so `schedule_time` doesn't always matches this.
Schedule string `protobuf:"bytes,12,opt,name=schedule" json:"schedule,omitempty"`
Schedule string `protobuf:"bytes,12,opt,name=schedule,proto3" json:"schedule,omitempty"`
XXX_NoUnkeyedLiteral struct{} `json:"-"`
XXX_unrecognized []byte `json:"-"`
XXX_sizecache int32 `json:"-"`
}
func (m *TransferRun) Reset() { *m = TransferRun{} }
func (m *TransferRun) String() string { return proto.CompactTextString(m) }
func (*TransferRun) ProtoMessage() {}
func (*TransferRun) Descriptor() ([]byte, []int) { return fileDescriptor1, []int{1} }
func (m *TransferRun) Reset() { *m = TransferRun{} }
func (m *TransferRun) String() string { return proto.CompactTextString(m) }
func (*TransferRun) ProtoMessage() {}
func (*TransferRun) Descriptor() ([]byte, []int) {
return fileDescriptor_transfer_b655bdd8f06d7255, []int{1}
}
func (m *TransferRun) XXX_Unmarshal(b []byte) error {
return xxx_messageInfo_TransferRun.Unmarshal(m, b)
}
func (m *TransferRun) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) {
return xxx_messageInfo_TransferRun.Marshal(b, m, deterministic)
}
func (dst *TransferRun) XXX_Merge(src proto.Message) {
xxx_messageInfo_TransferRun.Merge(dst, src)
}
func (m *TransferRun) XXX_Size() int {
return xxx_messageInfo_TransferRun.Size(m)
}
func (m *TransferRun) XXX_DiscardUnknown() {
xxx_messageInfo_TransferRun.DiscardUnknown(m)
}
var xxx_messageInfo_TransferRun proto.InternalMessageInfo
func (m *TransferRun) GetName() string {
if m != nil {
@ -335,49 +387,49 @@ func (m *TransferRun) GetName() string {
return ""
}
func (m *TransferRun) GetScheduleTime() *google_protobuf2.Timestamp {
func (m *TransferRun) GetScheduleTime() *timestamp.Timestamp {
if m != nil {
return m.ScheduleTime
}
return nil
}
func (m *TransferRun) GetRunTime() *google_protobuf2.Timestamp {
func (m *TransferRun) GetRunTime() *timestamp.Timestamp {
if m != nil {
return m.RunTime
}
return nil
}
func (m *TransferRun) GetErrorStatus() *google_rpc.Status {
func (m *TransferRun) GetErrorStatus() *status.Status {
if m != nil {
return m.ErrorStatus
}
return nil
}
func (m *TransferRun) GetStartTime() *google_protobuf2.Timestamp {
func (m *TransferRun) GetStartTime() *timestamp.Timestamp {
if m != nil {
return m.StartTime
}
return nil
}
func (m *TransferRun) GetEndTime() *google_protobuf2.Timestamp {
func (m *TransferRun) GetEndTime() *timestamp.Timestamp {
if m != nil {
return m.EndTime
}
return nil
}
func (m *TransferRun) GetUpdateTime() *google_protobuf2.Timestamp {
func (m *TransferRun) GetUpdateTime() *timestamp.Timestamp {
if m != nil {
return m.UpdateTime
}
return nil
}
func (m *TransferRun) GetParams() *google_protobuf1.Struct {
func (m *TransferRun) GetParams() *_struct.Struct {
if m != nil {
return m.Params
}
@ -422,19 +474,41 @@ func (m *TransferRun) GetSchedule() string {
// Represents a user facing message for a particular data transfer run.
type TransferMessage struct {
// Time when message was logged.
MessageTime *google_protobuf2.Timestamp `protobuf:"bytes,1,opt,name=message_time,json=messageTime" json:"message_time,omitempty"`
MessageTime *timestamp.Timestamp `protobuf:"bytes,1,opt,name=message_time,json=messageTime,proto3" json:"message_time,omitempty"`
// Message severity.
Severity TransferMessage_MessageSeverity `protobuf:"varint,2,opt,name=severity,enum=google.cloud.bigquery.datatransfer.v1.TransferMessage_MessageSeverity" json:"severity,omitempty"`
Severity TransferMessage_MessageSeverity `protobuf:"varint,2,opt,name=severity,proto3,enum=google.cloud.bigquery.datatransfer.v1.TransferMessage_MessageSeverity" json:"severity,omitempty"`
// Message text.
MessageText string `protobuf:"bytes,3,opt,name=message_text,json=messageText" json:"message_text,omitempty"`
MessageText string `protobuf:"bytes,3,opt,name=message_text,json=messageText,proto3" json:"message_text,omitempty"`
XXX_NoUnkeyedLiteral struct{} `json:"-"`
XXX_unrecognized []byte `json:"-"`
XXX_sizecache int32 `json:"-"`
}
func (m *TransferMessage) Reset() { *m = TransferMessage{} }
func (m *TransferMessage) String() string { return proto.CompactTextString(m) }
func (*TransferMessage) ProtoMessage() {}
func (*TransferMessage) Descriptor() ([]byte, []int) { return fileDescriptor1, []int{2} }
func (m *TransferMessage) Reset() { *m = TransferMessage{} }
func (m *TransferMessage) String() string { return proto.CompactTextString(m) }
func (*TransferMessage) ProtoMessage() {}
func (*TransferMessage) Descriptor() ([]byte, []int) {
return fileDescriptor_transfer_b655bdd8f06d7255, []int{2}
}
func (m *TransferMessage) XXX_Unmarshal(b []byte) error {
return xxx_messageInfo_TransferMessage.Unmarshal(m, b)
}
func (m *TransferMessage) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) {
return xxx_messageInfo_TransferMessage.Marshal(b, m, deterministic)
}
func (dst *TransferMessage) XXX_Merge(src proto.Message) {
xxx_messageInfo_TransferMessage.Merge(dst, src)
}
func (m *TransferMessage) XXX_Size() int {
return xxx_messageInfo_TransferMessage.Size(m)
}
func (m *TransferMessage) XXX_DiscardUnknown() {
xxx_messageInfo_TransferMessage.DiscardUnknown(m)
}
func (m *TransferMessage) GetMessageTime() *google_protobuf2.Timestamp {
var xxx_messageInfo_TransferMessage proto.InternalMessageInfo
func (m *TransferMessage) GetMessageTime() *timestamp.Timestamp {
if m != nil {
return m.MessageTime
}
@ -465,10 +539,10 @@ func init() {
}
func init() {
proto.RegisterFile("google/cloud/bigquery/datatransfer/v1/transfer.proto", fileDescriptor1)
proto.RegisterFile("google/cloud/bigquery/datatransfer/v1/transfer.proto", fileDescriptor_transfer_b655bdd8f06d7255)
}
var fileDescriptor1 = []byte{
var fileDescriptor_transfer_b655bdd8f06d7255 = []byte{
// 922 bytes of a gzipped FileDescriptorProto
0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xb4, 0x56, 0xdd, 0x6e, 0xe3, 0x44,
0x14, 0xc6, 0xf9, 0xcf, 0x71, 0x92, 0x8d, 0x46, 0x2c, 0x35, 0xd5, 0x02, 0xa1, 0xa2, 0x52, 0xd8,

File diff suppressed because it is too large Load Diff