gcp.storage.TransferJob
Explore with Pulumi AI
Creates a new Transfer Job in Google Cloud Storage Transfer.
To get more information about Google Cloud Storage Transfer, see:
Example Usage
Example creating a nightly Transfer Job from an AWS S3 Bucket to a GCS bucket.
import * as pulumi from "@pulumi/pulumi";
import * as gcp from "@pulumi/gcp";
const _default = gcp.storage.getTransferProjectServiceAccount({
    project: project,
});
const s3_backup_bucket = new gcp.storage.Bucket("s3-backup-bucket", {
    name: `${awsS3Bucket}-backup`,
    storageClass: "NEARLINE",
    project: project,
    location: "US",
});
const s3_backup_bucketBucketIAMMember = new gcp.storage.BucketIAMMember("s3-backup-bucket", {
    bucket: s3_backup_bucket.name,
    role: "roles/storage.admin",
    member: _default.then(_default => `serviceAccount:${_default.email}`),
}, {
    dependsOn: [s3_backup_bucket],
});
const topic = new gcp.pubsub.Topic("topic", {name: pubsubTopicName});
const notificationConfig = new gcp.pubsub.TopicIAMMember("notification_config", {
    topic: topic.id,
    role: "roles/pubsub.publisher",
    member: _default.then(_default => `serviceAccount:${_default.email}`),
});
const s3_bucket_nightly_backup = new gcp.storage.TransferJob("s3-bucket-nightly-backup", {
    description: "Nightly backup of S3 bucket",
    project: project,
    transferSpec: {
        objectConditions: {
            maxTimeElapsedSinceLastModification: "600s",
            excludePrefixes: ["requests.gz"],
        },
        transferOptions: {
            deleteObjectsUniqueInSink: false,
        },
        awsS3DataSource: {
            bucketName: awsS3Bucket,
            awsAccessKey: {
                accessKeyId: awsAccessKey,
                secretAccessKey: awsSecretKey,
            },
        },
        gcsDataSink: {
            bucketName: s3_backup_bucket.name,
            path: "foo/bar/",
        },
    },
    schedule: {
        scheduleStartDate: {
            year: 2018,
            month: 10,
            day: 1,
        },
        scheduleEndDate: {
            year: 2019,
            month: 1,
            day: 15,
        },
        startTimeOfDay: {
            hours: 23,
            minutes: 30,
            seconds: 0,
            nanos: 0,
        },
        repeatInterval: "604800s",
    },
    notificationConfig: {
        pubsubTopic: topic.id,
        eventTypes: [
            "TRANSFER_OPERATION_SUCCESS",
            "TRANSFER_OPERATION_FAILED",
        ],
        payloadFormat: "JSON",
    },
    loggingConfig: {
        logActions: [
            "COPY",
            "DELETE",
        ],
        logActionStates: [
            "SUCCEEDED",
            "FAILED",
        ],
    },
}, {
    dependsOn: [
        s3_backup_bucketBucketIAMMember,
        notificationConfig,
    ],
});
import pulumi
import pulumi_gcp as gcp
default = gcp.storage.get_transfer_project_service_account(project=project)
s3_backup_bucket = gcp.storage.Bucket("s3-backup-bucket",
    name=f"{aws_s3_bucket}-backup",
    storage_class="NEARLINE",
    project=project,
    location="US")
s3_backup_bucket_bucket_iam_member = gcp.storage.BucketIAMMember("s3-backup-bucket",
    bucket=s3_backup_bucket.name,
    role="roles/storage.admin",
    member=f"serviceAccount:{default.email}",
    opts = pulumi.ResourceOptions(depends_on=[s3_backup_bucket]))
topic = gcp.pubsub.Topic("topic", name=pubsub_topic_name)
notification_config = gcp.pubsub.TopicIAMMember("notification_config",
    topic=topic.id,
    role="roles/pubsub.publisher",
    member=f"serviceAccount:{default.email}")
s3_bucket_nightly_backup = gcp.storage.TransferJob("s3-bucket-nightly-backup",
    description="Nightly backup of S3 bucket",
    project=project,
    transfer_spec={
        "object_conditions": {
            "max_time_elapsed_since_last_modification": "600s",
            "exclude_prefixes": ["requests.gz"],
        },
        "transfer_options": {
            "delete_objects_unique_in_sink": False,
        },
        "aws_s3_data_source": {
            "bucket_name": aws_s3_bucket,
            "aws_access_key": {
                "access_key_id": aws_access_key,
                "secret_access_key": aws_secret_key,
            },
        },
        "gcs_data_sink": {
            "bucket_name": s3_backup_bucket.name,
            "path": "foo/bar/",
        },
    },
    schedule={
        "schedule_start_date": {
            "year": 2018,
            "month": 10,
            "day": 1,
        },
        "schedule_end_date": {
            "year": 2019,
            "month": 1,
            "day": 15,
        },
        "start_time_of_day": {
            "hours": 23,
            "minutes": 30,
            "seconds": 0,
            "nanos": 0,
        },
        "repeat_interval": "604800s",
    },
    notification_config={
        "pubsub_topic": topic.id,
        "event_types": [
            "TRANSFER_OPERATION_SUCCESS",
            "TRANSFER_OPERATION_FAILED",
        ],
        "payload_format": "JSON",
    },
    logging_config={
        "log_actions": [
            "COPY",
            "DELETE",
        ],
        "log_action_states": [
            "SUCCEEDED",
            "FAILED",
        ],
    },
    opts = pulumi.ResourceOptions(depends_on=[
            s3_backup_bucket_bucket_iam_member,
            notification_config,
        ]))
package main
import (
	"fmt"
	"github.com/pulumi/pulumi-gcp/sdk/v8/go/gcp/pubsub"
	"github.com/pulumi/pulumi-gcp/sdk/v8/go/gcp/storage"
	"github.com/pulumi/pulumi/sdk/v3/go/pulumi"
)
func main() {
	pulumi.Run(func(ctx *pulumi.Context) error {
		_default, err := storage.GetTransferProjectServiceAccount(ctx, &storage.GetTransferProjectServiceAccountArgs{
			Project: pulumi.StringRef(project),
		}, nil)
		if err != nil {
			return err
		}
		s3_backup_bucket, err := storage.NewBucket(ctx, "s3-backup-bucket", &storage.BucketArgs{
			Name:         pulumi.Sprintf("%v-backup", awsS3Bucket),
			StorageClass: pulumi.String("NEARLINE"),
			Project:      pulumi.Any(project),
			Location:     pulumi.String("US"),
		})
		if err != nil {
			return err
		}
		s3_backup_bucketBucketIAMMember, err := storage.NewBucketIAMMember(ctx, "s3-backup-bucket", &storage.BucketIAMMemberArgs{
			Bucket: s3_backup_bucket.Name,
			Role:   pulumi.String("roles/storage.admin"),
			Member: pulumi.Sprintf("serviceAccount:%v", _default.Email),
		}, pulumi.DependsOn([]pulumi.Resource{
			s3_backup_bucket,
		}))
		if err != nil {
			return err
		}
		topic, err := pubsub.NewTopic(ctx, "topic", &pubsub.TopicArgs{
			Name: pulumi.Any(pubsubTopicName),
		})
		if err != nil {
			return err
		}
		notificationConfig, err := pubsub.NewTopicIAMMember(ctx, "notification_config", &pubsub.TopicIAMMemberArgs{
			Topic:  topic.ID(),
			Role:   pulumi.String("roles/pubsub.publisher"),
			Member: pulumi.Sprintf("serviceAccount:%v", _default.Email),
		})
		if err != nil {
			return err
		}
		_, err = storage.NewTransferJob(ctx, "s3-bucket-nightly-backup", &storage.TransferJobArgs{
			Description: pulumi.String("Nightly backup of S3 bucket"),
			Project:     pulumi.Any(project),
			TransferSpec: &storage.TransferJobTransferSpecArgs{
				ObjectConditions: &storage.TransferJobTransferSpecObjectConditionsArgs{
					MaxTimeElapsedSinceLastModification: pulumi.String("600s"),
					ExcludePrefixes: pulumi.StringArray{
						pulumi.String("requests.gz"),
					},
				},
				TransferOptions: &storage.TransferJobTransferSpecTransferOptionsArgs{
					DeleteObjectsUniqueInSink: pulumi.Bool(false),
				},
				AwsS3DataSource: &storage.TransferJobTransferSpecAwsS3DataSourceArgs{
					BucketName: pulumi.Any(awsS3Bucket),
					AwsAccessKey: &storage.TransferJobTransferSpecAwsS3DataSourceAwsAccessKeyArgs{
						AccessKeyId:     pulumi.Any(awsAccessKey),
						SecretAccessKey: pulumi.Any(awsSecretKey),
					},
				},
				GcsDataSink: &storage.TransferJobTransferSpecGcsDataSinkArgs{
					BucketName: s3_backup_bucket.Name,
					Path:       pulumi.String("foo/bar/"),
				},
			},
			Schedule: &storage.TransferJobScheduleArgs{
				ScheduleStartDate: &storage.TransferJobScheduleScheduleStartDateArgs{
					Year:  pulumi.Int(2018),
					Month: pulumi.Int(10),
					Day:   pulumi.Int(1),
				},
				ScheduleEndDate: &storage.TransferJobScheduleScheduleEndDateArgs{
					Year:  pulumi.Int(2019),
					Month: pulumi.Int(1),
					Day:   pulumi.Int(15),
				},
				StartTimeOfDay: &storage.TransferJobScheduleStartTimeOfDayArgs{
					Hours:   pulumi.Int(23),
					Minutes: pulumi.Int(30),
					Seconds: pulumi.Int(0),
					Nanos:   pulumi.Int(0),
				},
				RepeatInterval: pulumi.String("604800s"),
			},
			NotificationConfig: &storage.TransferJobNotificationConfigArgs{
				PubsubTopic: topic.ID(),
				EventTypes: pulumi.StringArray{
					pulumi.String("TRANSFER_OPERATION_SUCCESS"),
					pulumi.String("TRANSFER_OPERATION_FAILED"),
				},
				PayloadFormat: pulumi.String("JSON"),
			},
			LoggingConfig: &storage.TransferJobLoggingConfigArgs{
				LogActions: pulumi.StringArray{
					pulumi.String("COPY"),
					pulumi.String("DELETE"),
				},
				LogActionStates: pulumi.StringArray{
					pulumi.String("SUCCEEDED"),
					pulumi.String("FAILED"),
				},
			},
		}, pulumi.DependsOn([]pulumi.Resource{
			s3_backup_bucketBucketIAMMember,
			notificationConfig,
		}))
		if err != nil {
			return err
		}
		return nil
	})
}
using System.Collections.Generic;
using System.Linq;
using Pulumi;
using Gcp = Pulumi.Gcp;
return await Deployment.RunAsync(() => 
{
    var @default = Gcp.Storage.GetTransferProjectServiceAccount.Invoke(new()
    {
        Project = project,
    });
    var s3_backup_bucket = new Gcp.Storage.Bucket("s3-backup-bucket", new()
    {
        Name = $"{awsS3Bucket}-backup",
        StorageClass = "NEARLINE",
        Project = project,
        Location = "US",
    });
    var s3_backup_bucketBucketIAMMember = new Gcp.Storage.BucketIAMMember("s3-backup-bucket", new()
    {
        Bucket = s3_backup_bucket.Name,
        Role = "roles/storage.admin",
        Member = @default.Apply(@default => $"serviceAccount:{@default.Apply(getTransferProjectServiceAccountResult => getTransferProjectServiceAccountResult.Email)}"),
    }, new CustomResourceOptions
    {
        DependsOn =
        {
            s3_backup_bucket,
        },
    });
    var topic = new Gcp.PubSub.Topic("topic", new()
    {
        Name = pubsubTopicName,
    });
    var notificationConfig = new Gcp.PubSub.TopicIAMMember("notification_config", new()
    {
        Topic = topic.Id,
        Role = "roles/pubsub.publisher",
        Member = @default.Apply(@default => $"serviceAccount:{@default.Apply(getTransferProjectServiceAccountResult => getTransferProjectServiceAccountResult.Email)}"),
    });
    var s3_bucket_nightly_backup = new Gcp.Storage.TransferJob("s3-bucket-nightly-backup", new()
    {
        Description = "Nightly backup of S3 bucket",
        Project = project,
        TransferSpec = new Gcp.Storage.Inputs.TransferJobTransferSpecArgs
        {
            ObjectConditions = new Gcp.Storage.Inputs.TransferJobTransferSpecObjectConditionsArgs
            {
                MaxTimeElapsedSinceLastModification = "600s",
                ExcludePrefixes = new[]
                {
                    "requests.gz",
                },
            },
            TransferOptions = new Gcp.Storage.Inputs.TransferJobTransferSpecTransferOptionsArgs
            {
                DeleteObjectsUniqueInSink = false,
            },
            AwsS3DataSource = new Gcp.Storage.Inputs.TransferJobTransferSpecAwsS3DataSourceArgs
            {
                BucketName = awsS3Bucket,
                AwsAccessKey = new Gcp.Storage.Inputs.TransferJobTransferSpecAwsS3DataSourceAwsAccessKeyArgs
                {
                    AccessKeyId = awsAccessKey,
                    SecretAccessKey = awsSecretKey,
                },
            },
            GcsDataSink = new Gcp.Storage.Inputs.TransferJobTransferSpecGcsDataSinkArgs
            {
                BucketName = s3_backup_bucket.Name,
                Path = "foo/bar/",
            },
        },
        Schedule = new Gcp.Storage.Inputs.TransferJobScheduleArgs
        {
            ScheduleStartDate = new Gcp.Storage.Inputs.TransferJobScheduleScheduleStartDateArgs
            {
                Year = 2018,
                Month = 10,
                Day = 1,
            },
            ScheduleEndDate = new Gcp.Storage.Inputs.TransferJobScheduleScheduleEndDateArgs
            {
                Year = 2019,
                Month = 1,
                Day = 15,
            },
            StartTimeOfDay = new Gcp.Storage.Inputs.TransferJobScheduleStartTimeOfDayArgs
            {
                Hours = 23,
                Minutes = 30,
                Seconds = 0,
                Nanos = 0,
            },
            RepeatInterval = "604800s",
        },
        NotificationConfig = new Gcp.Storage.Inputs.TransferJobNotificationConfigArgs
        {
            PubsubTopic = topic.Id,
            EventTypes = new[]
            {
                "TRANSFER_OPERATION_SUCCESS",
                "TRANSFER_OPERATION_FAILED",
            },
            PayloadFormat = "JSON",
        },
        LoggingConfig = new Gcp.Storage.Inputs.TransferJobLoggingConfigArgs
        {
            LogActions = new[]
            {
                "COPY",
                "DELETE",
            },
            LogActionStates = new[]
            {
                "SUCCEEDED",
                "FAILED",
            },
        },
    }, new CustomResourceOptions
    {
        DependsOn =
        {
            s3_backup_bucketBucketIAMMember,
            notificationConfig,
        },
    });
});
package generated_program;
import com.pulumi.Context;
import com.pulumi.Pulumi;
import com.pulumi.core.Output;
import com.pulumi.gcp.storage.StorageFunctions;
import com.pulumi.gcp.storage.inputs.GetTransferProjectServiceAccountArgs;
import com.pulumi.gcp.storage.Bucket;
import com.pulumi.gcp.storage.BucketArgs;
import com.pulumi.gcp.storage.BucketIAMMember;
import com.pulumi.gcp.storage.BucketIAMMemberArgs;
import com.pulumi.gcp.pubsub.Topic;
import com.pulumi.gcp.pubsub.TopicArgs;
import com.pulumi.gcp.pubsub.TopicIAMMember;
import com.pulumi.gcp.pubsub.TopicIAMMemberArgs;
import com.pulumi.gcp.storage.TransferJob;
import com.pulumi.gcp.storage.TransferJobArgs;
import com.pulumi.gcp.storage.inputs.TransferJobTransferSpecArgs;
import com.pulumi.gcp.storage.inputs.TransferJobTransferSpecObjectConditionsArgs;
import com.pulumi.gcp.storage.inputs.TransferJobTransferSpecTransferOptionsArgs;
import com.pulumi.gcp.storage.inputs.TransferJobTransferSpecAwsS3DataSourceArgs;
import com.pulumi.gcp.storage.inputs.TransferJobTransferSpecAwsS3DataSourceAwsAccessKeyArgs;
import com.pulumi.gcp.storage.inputs.TransferJobTransferSpecGcsDataSinkArgs;
import com.pulumi.gcp.storage.inputs.TransferJobScheduleArgs;
import com.pulumi.gcp.storage.inputs.TransferJobScheduleScheduleStartDateArgs;
import com.pulumi.gcp.storage.inputs.TransferJobScheduleScheduleEndDateArgs;
import com.pulumi.gcp.storage.inputs.TransferJobScheduleStartTimeOfDayArgs;
import com.pulumi.gcp.storage.inputs.TransferJobNotificationConfigArgs;
import com.pulumi.gcp.storage.inputs.TransferJobLoggingConfigArgs;
import com.pulumi.resources.CustomResourceOptions;
import java.util.List;
import java.util.ArrayList;
import java.util.Map;
import java.io.File;
import java.nio.file.Files;
import java.nio.file.Paths;
public class App {
    public static void main(String[] args) {
        Pulumi.run(App::stack);
    }
    public static void stack(Context ctx) {
        final var default = StorageFunctions.getTransferProjectServiceAccount(GetTransferProjectServiceAccountArgs.builder()
            .project(project)
            .build());
        var s3_backup_bucket = new Bucket("s3-backup-bucket", BucketArgs.builder()
            .name(String.format("%s-backup", awsS3Bucket))
            .storageClass("NEARLINE")
            .project(project)
            .location("US")
            .build());
        var s3_backup_bucketBucketIAMMember = new BucketIAMMember("s3-backup-bucketBucketIAMMember", BucketIAMMemberArgs.builder()
            .bucket(s3_backup_bucket.name())
            .role("roles/storage.admin")
            .member(String.format("serviceAccount:%s", default_.email()))
            .build(), CustomResourceOptions.builder()
                .dependsOn(s3_backup_bucket)
                .build());
        var topic = new Topic("topic", TopicArgs.builder()
            .name(pubsubTopicName)
            .build());
        var notificationConfig = new TopicIAMMember("notificationConfig", TopicIAMMemberArgs.builder()
            .topic(topic.id())
            .role("roles/pubsub.publisher")
            .member(String.format("serviceAccount:%s", default_.email()))
            .build());
        var s3_bucket_nightly_backup = new TransferJob("s3-bucket-nightly-backup", TransferJobArgs.builder()
            .description("Nightly backup of S3 bucket")
            .project(project)
            .transferSpec(TransferJobTransferSpecArgs.builder()
                .objectConditions(TransferJobTransferSpecObjectConditionsArgs.builder()
                    .maxTimeElapsedSinceLastModification("600s")
                    .excludePrefixes("requests.gz")
                    .build())
                .transferOptions(TransferJobTransferSpecTransferOptionsArgs.builder()
                    .deleteObjectsUniqueInSink(false)
                    .build())
                .awsS3DataSource(TransferJobTransferSpecAwsS3DataSourceArgs.builder()
                    .bucketName(awsS3Bucket)
                    .awsAccessKey(TransferJobTransferSpecAwsS3DataSourceAwsAccessKeyArgs.builder()
                        .accessKeyId(awsAccessKey)
                        .secretAccessKey(awsSecretKey)
                        .build())
                    .build())
                .gcsDataSink(TransferJobTransferSpecGcsDataSinkArgs.builder()
                    .bucketName(s3_backup_bucket.name())
                    .path("foo/bar/")
                    .build())
                .build())
            .schedule(TransferJobScheduleArgs.builder()
                .scheduleStartDate(TransferJobScheduleScheduleStartDateArgs.builder()
                    .year(2018)
                    .month(10)
                    .day(1)
                    .build())
                .scheduleEndDate(TransferJobScheduleScheduleEndDateArgs.builder()
                    .year(2019)
                    .month(1)
                    .day(15)
                    .build())
                .startTimeOfDay(TransferJobScheduleStartTimeOfDayArgs.builder()
                    .hours(23)
                    .minutes(30)
                    .seconds(0)
                    .nanos(0)
                    .build())
                .repeatInterval("604800s")
                .build())
            .notificationConfig(TransferJobNotificationConfigArgs.builder()
                .pubsubTopic(topic.id())
                .eventTypes(                
                    "TRANSFER_OPERATION_SUCCESS",
                    "TRANSFER_OPERATION_FAILED")
                .payloadFormat("JSON")
                .build())
            .loggingConfig(TransferJobLoggingConfigArgs.builder()
                .logActions(                
                    "COPY",
                    "DELETE")
                .logActionStates(                
                    "SUCCEEDED",
                    "FAILED")
                .build())
            .build(), CustomResourceOptions.builder()
                .dependsOn(                
                    s3_backup_bucketBucketIAMMember,
                    notificationConfig)
                .build());
    }
}
resources:
  s3-backup-bucket:
    type: gcp:storage:Bucket
    properties:
      name: ${awsS3Bucket}-backup
      storageClass: NEARLINE
      project: ${project}
      location: US
  s3-backup-bucketBucketIAMMember:
    type: gcp:storage:BucketIAMMember
    name: s3-backup-bucket
    properties:
      bucket: ${["s3-backup-bucket"].name}
      role: roles/storage.admin
      member: serviceAccount:${default.email}
    options:
      dependsOn:
        - ${["s3-backup-bucket"]}
  topic:
    type: gcp:pubsub:Topic
    properties:
      name: ${pubsubTopicName}
  notificationConfig:
    type: gcp:pubsub:TopicIAMMember
    name: notification_config
    properties:
      topic: ${topic.id}
      role: roles/pubsub.publisher
      member: serviceAccount:${default.email}
  s3-bucket-nightly-backup:
    type: gcp:storage:TransferJob
    properties:
      description: Nightly backup of S3 bucket
      project: ${project}
      transferSpec:
        objectConditions:
          maxTimeElapsedSinceLastModification: 600s
          excludePrefixes:
            - requests.gz
        transferOptions:
          deleteObjectsUniqueInSink: false
        awsS3DataSource:
          bucketName: ${awsS3Bucket}
          awsAccessKey:
            accessKeyId: ${awsAccessKey}
            secretAccessKey: ${awsSecretKey}
        gcsDataSink:
          bucketName: ${["s3-backup-bucket"].name}
          path: foo/bar/
      schedule:
        scheduleStartDate:
          year: 2018
          month: 10
          day: 1
        scheduleEndDate:
          year: 2019
          month: 1
          day: 15
        startTimeOfDay:
          hours: 23
          minutes: 30
          seconds: 0
          nanos: 0
        repeatInterval: 604800s
      notificationConfig:
        pubsubTopic: ${topic.id}
        eventTypes:
          - TRANSFER_OPERATION_SUCCESS
          - TRANSFER_OPERATION_FAILED
        payloadFormat: JSON
      loggingConfig:
        logActions:
          - COPY
          - DELETE
        logActionStates:
          - SUCCEEDED
          - FAILED
    options:
      dependsOn:
        - ${["s3-backup-bucketBucketIAMMember"]}
        - ${notificationConfig}
variables:
  default:
    fn::invoke:
      function: gcp:storage:getTransferProjectServiceAccount
      arguments:
        project: ${project}
Create TransferJob Resource
Resources are created with functions called constructors. To learn more about declaring and configuring resources, see Resources.
Constructor syntax
new TransferJob(name: string, args: TransferJobArgs, opts?: CustomResourceOptions);@overload
def TransferJob(resource_name: str,
                args: TransferJobArgs,
                opts: Optional[ResourceOptions] = None)
@overload
def TransferJob(resource_name: str,
                opts: Optional[ResourceOptions] = None,
                description: Optional[str] = None,
                event_stream: Optional[TransferJobEventStreamArgs] = None,
                logging_config: Optional[TransferJobLoggingConfigArgs] = None,
                name: Optional[str] = None,
                notification_config: Optional[TransferJobNotificationConfigArgs] = None,
                project: Optional[str] = None,
                replication_spec: Optional[TransferJobReplicationSpecArgs] = None,
                schedule: Optional[TransferJobScheduleArgs] = None,
                status: Optional[str] = None,
                transfer_spec: Optional[TransferJobTransferSpecArgs] = None)func NewTransferJob(ctx *Context, name string, args TransferJobArgs, opts ...ResourceOption) (*TransferJob, error)public TransferJob(string name, TransferJobArgs args, CustomResourceOptions? opts = null)
public TransferJob(String name, TransferJobArgs args)
public TransferJob(String name, TransferJobArgs args, CustomResourceOptions options)
type: gcp:storage:TransferJob
properties: # The arguments to resource properties.
options: # Bag of options to control resource's behavior.
Parameters
- name string
- The unique name of the resource.
- args TransferJobArgs
- The arguments to resource properties.
- opts CustomResourceOptions
- Bag of options to control resource's behavior.
- resource_name str
- The unique name of the resource.
- args TransferJobArgs
- The arguments to resource properties.
- opts ResourceOptions
- Bag of options to control resource's behavior.
- ctx Context
- Context object for the current deployment.
- name string
- The unique name of the resource.
- args TransferJobArgs
- The arguments to resource properties.
- opts ResourceOption
- Bag of options to control resource's behavior.
- name string
- The unique name of the resource.
- args TransferJobArgs
- The arguments to resource properties.
- opts CustomResourceOptions
- Bag of options to control resource's behavior.
- name String
- The unique name of the resource.
- args TransferJobArgs
- The arguments to resource properties.
- options CustomResourceOptions
- Bag of options to control resource's behavior.
Constructor example
The following reference example uses placeholder values for all input properties.
var transferJobResource = new Gcp.Storage.TransferJob("transferJobResource", new()
{
    Description = "string",
    EventStream = new Gcp.Storage.Inputs.TransferJobEventStreamArgs
    {
        Name = "string",
        EventStreamExpirationTime = "string",
        EventStreamStartTime = "string",
    },
    LoggingConfig = new Gcp.Storage.Inputs.TransferJobLoggingConfigArgs
    {
        EnableOnPremGcsTransferLogs = false,
        LogActionStates = new[]
        {
            "string",
        },
        LogActions = new[]
        {
            "string",
        },
    },
    Name = "string",
    NotificationConfig = new Gcp.Storage.Inputs.TransferJobNotificationConfigArgs
    {
        PayloadFormat = "string",
        PubsubTopic = "string",
        EventTypes = new[]
        {
            "string",
        },
    },
    Project = "string",
    ReplicationSpec = new Gcp.Storage.Inputs.TransferJobReplicationSpecArgs
    {
        GcsDataSink = new Gcp.Storage.Inputs.TransferJobReplicationSpecGcsDataSinkArgs
        {
            BucketName = "string",
            Path = "string",
        },
        GcsDataSource = new Gcp.Storage.Inputs.TransferJobReplicationSpecGcsDataSourceArgs
        {
            BucketName = "string",
            Path = "string",
        },
        ObjectConditions = new Gcp.Storage.Inputs.TransferJobReplicationSpecObjectConditionsArgs
        {
            ExcludePrefixes = new[]
            {
                "string",
            },
            IncludePrefixes = new[]
            {
                "string",
            },
            LastModifiedBefore = "string",
            LastModifiedSince = "string",
            MaxTimeElapsedSinceLastModification = "string",
            MinTimeElapsedSinceLastModification = "string",
        },
        TransferOptions = new Gcp.Storage.Inputs.TransferJobReplicationSpecTransferOptionsArgs
        {
            DeleteObjectsFromSourceAfterTransfer = false,
            DeleteObjectsUniqueInSink = false,
            OverwriteObjectsAlreadyExistingInSink = false,
            OverwriteWhen = "string",
        },
    },
    Schedule = new Gcp.Storage.Inputs.TransferJobScheduleArgs
    {
        ScheduleStartDate = new Gcp.Storage.Inputs.TransferJobScheduleScheduleStartDateArgs
        {
            Day = 0,
            Month = 0,
            Year = 0,
        },
        RepeatInterval = "string",
        ScheduleEndDate = new Gcp.Storage.Inputs.TransferJobScheduleScheduleEndDateArgs
        {
            Day = 0,
            Month = 0,
            Year = 0,
        },
        StartTimeOfDay = new Gcp.Storage.Inputs.TransferJobScheduleStartTimeOfDayArgs
        {
            Hours = 0,
            Minutes = 0,
            Nanos = 0,
            Seconds = 0,
        },
    },
    Status = "string",
    TransferSpec = new Gcp.Storage.Inputs.TransferJobTransferSpecArgs
    {
        AwsS3DataSource = new Gcp.Storage.Inputs.TransferJobTransferSpecAwsS3DataSourceArgs
        {
            BucketName = "string",
            AwsAccessKey = new Gcp.Storage.Inputs.TransferJobTransferSpecAwsS3DataSourceAwsAccessKeyArgs
            {
                AccessKeyId = "string",
                SecretAccessKey = "string",
            },
            Path = "string",
            RoleArn = "string",
        },
        AzureBlobStorageDataSource = new Gcp.Storage.Inputs.TransferJobTransferSpecAzureBlobStorageDataSourceArgs
        {
            Container = "string",
            StorageAccount = "string",
            AzureCredentials = new Gcp.Storage.Inputs.TransferJobTransferSpecAzureBlobStorageDataSourceAzureCredentialsArgs
            {
                SasToken = "string",
            },
            CredentialsSecret = "string",
            Path = "string",
        },
        GcsDataSink = new Gcp.Storage.Inputs.TransferJobTransferSpecGcsDataSinkArgs
        {
            BucketName = "string",
            Path = "string",
        },
        GcsDataSource = new Gcp.Storage.Inputs.TransferJobTransferSpecGcsDataSourceArgs
        {
            BucketName = "string",
            Path = "string",
        },
        HdfsDataSource = new Gcp.Storage.Inputs.TransferJobTransferSpecHdfsDataSourceArgs
        {
            Path = "string",
        },
        HttpDataSource = new Gcp.Storage.Inputs.TransferJobTransferSpecHttpDataSourceArgs
        {
            ListUrl = "string",
        },
        ObjectConditions = new Gcp.Storage.Inputs.TransferJobTransferSpecObjectConditionsArgs
        {
            ExcludePrefixes = new[]
            {
                "string",
            },
            IncludePrefixes = new[]
            {
                "string",
            },
            LastModifiedBefore = "string",
            LastModifiedSince = "string",
            MaxTimeElapsedSinceLastModification = "string",
            MinTimeElapsedSinceLastModification = "string",
        },
        PosixDataSink = new Gcp.Storage.Inputs.TransferJobTransferSpecPosixDataSinkArgs
        {
            RootDirectory = "string",
        },
        PosixDataSource = new Gcp.Storage.Inputs.TransferJobTransferSpecPosixDataSourceArgs
        {
            RootDirectory = "string",
        },
        SinkAgentPoolName = "string",
        SourceAgentPoolName = "string",
        TransferOptions = new Gcp.Storage.Inputs.TransferJobTransferSpecTransferOptionsArgs
        {
            DeleteObjectsFromSourceAfterTransfer = false,
            DeleteObjectsUniqueInSink = false,
            OverwriteObjectsAlreadyExistingInSink = false,
            OverwriteWhen = "string",
        },
    },
});
example, err := storage.NewTransferJob(ctx, "transferJobResource", &storage.TransferJobArgs{
	Description: pulumi.String("string"),
	EventStream: &storage.TransferJobEventStreamArgs{
		Name:                      pulumi.String("string"),
		EventStreamExpirationTime: pulumi.String("string"),
		EventStreamStartTime:      pulumi.String("string"),
	},
	LoggingConfig: &storage.TransferJobLoggingConfigArgs{
		EnableOnPremGcsTransferLogs: pulumi.Bool(false),
		LogActionStates: pulumi.StringArray{
			pulumi.String("string"),
		},
		LogActions: pulumi.StringArray{
			pulumi.String("string"),
		},
	},
	Name: pulumi.String("string"),
	NotificationConfig: &storage.TransferJobNotificationConfigArgs{
		PayloadFormat: pulumi.String("string"),
		PubsubTopic:   pulumi.String("string"),
		EventTypes: pulumi.StringArray{
			pulumi.String("string"),
		},
	},
	Project: pulumi.String("string"),
	ReplicationSpec: &storage.TransferJobReplicationSpecArgs{
		GcsDataSink: &storage.TransferJobReplicationSpecGcsDataSinkArgs{
			BucketName: pulumi.String("string"),
			Path:       pulumi.String("string"),
		},
		GcsDataSource: &storage.TransferJobReplicationSpecGcsDataSourceArgs{
			BucketName: pulumi.String("string"),
			Path:       pulumi.String("string"),
		},
		ObjectConditions: &storage.TransferJobReplicationSpecObjectConditionsArgs{
			ExcludePrefixes: pulumi.StringArray{
				pulumi.String("string"),
			},
			IncludePrefixes: pulumi.StringArray{
				pulumi.String("string"),
			},
			LastModifiedBefore:                  pulumi.String("string"),
			LastModifiedSince:                   pulumi.String("string"),
			MaxTimeElapsedSinceLastModification: pulumi.String("string"),
			MinTimeElapsedSinceLastModification: pulumi.String("string"),
		},
		TransferOptions: &storage.TransferJobReplicationSpecTransferOptionsArgs{
			DeleteObjectsFromSourceAfterTransfer:  pulumi.Bool(false),
			DeleteObjectsUniqueInSink:             pulumi.Bool(false),
			OverwriteObjectsAlreadyExistingInSink: pulumi.Bool(false),
			OverwriteWhen:                         pulumi.String("string"),
		},
	},
	Schedule: &storage.TransferJobScheduleArgs{
		ScheduleStartDate: &storage.TransferJobScheduleScheduleStartDateArgs{
			Day:   pulumi.Int(0),
			Month: pulumi.Int(0),
			Year:  pulumi.Int(0),
		},
		RepeatInterval: pulumi.String("string"),
		ScheduleEndDate: &storage.TransferJobScheduleScheduleEndDateArgs{
			Day:   pulumi.Int(0),
			Month: pulumi.Int(0),
			Year:  pulumi.Int(0),
		},
		StartTimeOfDay: &storage.TransferJobScheduleStartTimeOfDayArgs{
			Hours:   pulumi.Int(0),
			Minutes: pulumi.Int(0),
			Nanos:   pulumi.Int(0),
			Seconds: pulumi.Int(0),
		},
	},
	Status: pulumi.String("string"),
	TransferSpec: &storage.TransferJobTransferSpecArgs{
		AwsS3DataSource: &storage.TransferJobTransferSpecAwsS3DataSourceArgs{
			BucketName: pulumi.String("string"),
			AwsAccessKey: &storage.TransferJobTransferSpecAwsS3DataSourceAwsAccessKeyArgs{
				AccessKeyId:     pulumi.String("string"),
				SecretAccessKey: pulumi.String("string"),
			},
			Path:    pulumi.String("string"),
			RoleArn: pulumi.String("string"),
		},
		AzureBlobStorageDataSource: &storage.TransferJobTransferSpecAzureBlobStorageDataSourceArgs{
			Container:      pulumi.String("string"),
			StorageAccount: pulumi.String("string"),
			AzureCredentials: &storage.TransferJobTransferSpecAzureBlobStorageDataSourceAzureCredentialsArgs{
				SasToken: pulumi.String("string"),
			},
			CredentialsSecret: pulumi.String("string"),
			Path:              pulumi.String("string"),
		},
		GcsDataSink: &storage.TransferJobTransferSpecGcsDataSinkArgs{
			BucketName: pulumi.String("string"),
			Path:       pulumi.String("string"),
		},
		GcsDataSource: &storage.TransferJobTransferSpecGcsDataSourceArgs{
			BucketName: pulumi.String("string"),
			Path:       pulumi.String("string"),
		},
		HdfsDataSource: &storage.TransferJobTransferSpecHdfsDataSourceArgs{
			Path: pulumi.String("string"),
		},
		HttpDataSource: &storage.TransferJobTransferSpecHttpDataSourceArgs{
			ListUrl: pulumi.String("string"),
		},
		ObjectConditions: &storage.TransferJobTransferSpecObjectConditionsArgs{
			ExcludePrefixes: pulumi.StringArray{
				pulumi.String("string"),
			},
			IncludePrefixes: pulumi.StringArray{
				pulumi.String("string"),
			},
			LastModifiedBefore:                  pulumi.String("string"),
			LastModifiedSince:                   pulumi.String("string"),
			MaxTimeElapsedSinceLastModification: pulumi.String("string"),
			MinTimeElapsedSinceLastModification: pulumi.String("string"),
		},
		PosixDataSink: &storage.TransferJobTransferSpecPosixDataSinkArgs{
			RootDirectory: pulumi.String("string"),
		},
		PosixDataSource: &storage.TransferJobTransferSpecPosixDataSourceArgs{
			RootDirectory: pulumi.String("string"),
		},
		SinkAgentPoolName:   pulumi.String("string"),
		SourceAgentPoolName: pulumi.String("string"),
		TransferOptions: &storage.TransferJobTransferSpecTransferOptionsArgs{
			DeleteObjectsFromSourceAfterTransfer:  pulumi.Bool(false),
			DeleteObjectsUniqueInSink:             pulumi.Bool(false),
			OverwriteObjectsAlreadyExistingInSink: pulumi.Bool(false),
			OverwriteWhen:                         pulumi.String("string"),
		},
	},
})
var transferJobResource = new TransferJob("transferJobResource", TransferJobArgs.builder()
    .description("string")
    .eventStream(TransferJobEventStreamArgs.builder()
        .name("string")
        .eventStreamExpirationTime("string")
        .eventStreamStartTime("string")
        .build())
    .loggingConfig(TransferJobLoggingConfigArgs.builder()
        .enableOnPremGcsTransferLogs(false)
        .logActionStates("string")
        .logActions("string")
        .build())
    .name("string")
    .notificationConfig(TransferJobNotificationConfigArgs.builder()
        .payloadFormat("string")
        .pubsubTopic("string")
        .eventTypes("string")
        .build())
    .project("string")
    .replicationSpec(TransferJobReplicationSpecArgs.builder()
        .gcsDataSink(TransferJobReplicationSpecGcsDataSinkArgs.builder()
            .bucketName("string")
            .path("string")
            .build())
        .gcsDataSource(TransferJobReplicationSpecGcsDataSourceArgs.builder()
            .bucketName("string")
            .path("string")
            .build())
        .objectConditions(TransferJobReplicationSpecObjectConditionsArgs.builder()
            .excludePrefixes("string")
            .includePrefixes("string")
            .lastModifiedBefore("string")
            .lastModifiedSince("string")
            .maxTimeElapsedSinceLastModification("string")
            .minTimeElapsedSinceLastModification("string")
            .build())
        .transferOptions(TransferJobReplicationSpecTransferOptionsArgs.builder()
            .deleteObjectsFromSourceAfterTransfer(false)
            .deleteObjectsUniqueInSink(false)
            .overwriteObjectsAlreadyExistingInSink(false)
            .overwriteWhen("string")
            .build())
        .build())
    .schedule(TransferJobScheduleArgs.builder()
        .scheduleStartDate(TransferJobScheduleScheduleStartDateArgs.builder()
            .day(0)
            .month(0)
            .year(0)
            .build())
        .repeatInterval("string")
        .scheduleEndDate(TransferJobScheduleScheduleEndDateArgs.builder()
            .day(0)
            .month(0)
            .year(0)
            .build())
        .startTimeOfDay(TransferJobScheduleStartTimeOfDayArgs.builder()
            .hours(0)
            .minutes(0)
            .nanos(0)
            .seconds(0)
            .build())
        .build())
    .status("string")
    .transferSpec(TransferJobTransferSpecArgs.builder()
        .awsS3DataSource(TransferJobTransferSpecAwsS3DataSourceArgs.builder()
            .bucketName("string")
            .awsAccessKey(TransferJobTransferSpecAwsS3DataSourceAwsAccessKeyArgs.builder()
                .accessKeyId("string")
                .secretAccessKey("string")
                .build())
            .path("string")
            .roleArn("string")
            .build())
        .azureBlobStorageDataSource(TransferJobTransferSpecAzureBlobStorageDataSourceArgs.builder()
            .container("string")
            .storageAccount("string")
            .azureCredentials(TransferJobTransferSpecAzureBlobStorageDataSourceAzureCredentialsArgs.builder()
                .sasToken("string")
                .build())
            .credentialsSecret("string")
            .path("string")
            .build())
        .gcsDataSink(TransferJobTransferSpecGcsDataSinkArgs.builder()
            .bucketName("string")
            .path("string")
            .build())
        .gcsDataSource(TransferJobTransferSpecGcsDataSourceArgs.builder()
            .bucketName("string")
            .path("string")
            .build())
        .hdfsDataSource(TransferJobTransferSpecHdfsDataSourceArgs.builder()
            .path("string")
            .build())
        .httpDataSource(TransferJobTransferSpecHttpDataSourceArgs.builder()
            .listUrl("string")
            .build())
        .objectConditions(TransferJobTransferSpecObjectConditionsArgs.builder()
            .excludePrefixes("string")
            .includePrefixes("string")
            .lastModifiedBefore("string")
            .lastModifiedSince("string")
            .maxTimeElapsedSinceLastModification("string")
            .minTimeElapsedSinceLastModification("string")
            .build())
        .posixDataSink(TransferJobTransferSpecPosixDataSinkArgs.builder()
            .rootDirectory("string")
            .build())
        .posixDataSource(TransferJobTransferSpecPosixDataSourceArgs.builder()
            .rootDirectory("string")
            .build())
        .sinkAgentPoolName("string")
        .sourceAgentPoolName("string")
        .transferOptions(TransferJobTransferSpecTransferOptionsArgs.builder()
            .deleteObjectsFromSourceAfterTransfer(false)
            .deleteObjectsUniqueInSink(false)
            .overwriteObjectsAlreadyExistingInSink(false)
            .overwriteWhen("string")
            .build())
        .build())
    .build());
transfer_job_resource = gcp.storage.TransferJob("transferJobResource",
    description="string",
    event_stream={
        "name": "string",
        "event_stream_expiration_time": "string",
        "event_stream_start_time": "string",
    },
    logging_config={
        "enable_on_prem_gcs_transfer_logs": False,
        "log_action_states": ["string"],
        "log_actions": ["string"],
    },
    name="string",
    notification_config={
        "payload_format": "string",
        "pubsub_topic": "string",
        "event_types": ["string"],
    },
    project="string",
    replication_spec={
        "gcs_data_sink": {
            "bucket_name": "string",
            "path": "string",
        },
        "gcs_data_source": {
            "bucket_name": "string",
            "path": "string",
        },
        "object_conditions": {
            "exclude_prefixes": ["string"],
            "include_prefixes": ["string"],
            "last_modified_before": "string",
            "last_modified_since": "string",
            "max_time_elapsed_since_last_modification": "string",
            "min_time_elapsed_since_last_modification": "string",
        },
        "transfer_options": {
            "delete_objects_from_source_after_transfer": False,
            "delete_objects_unique_in_sink": False,
            "overwrite_objects_already_existing_in_sink": False,
            "overwrite_when": "string",
        },
    },
    schedule={
        "schedule_start_date": {
            "day": 0,
            "month": 0,
            "year": 0,
        },
        "repeat_interval": "string",
        "schedule_end_date": {
            "day": 0,
            "month": 0,
            "year": 0,
        },
        "start_time_of_day": {
            "hours": 0,
            "minutes": 0,
            "nanos": 0,
            "seconds": 0,
        },
    },
    status="string",
    transfer_spec={
        "aws_s3_data_source": {
            "bucket_name": "string",
            "aws_access_key": {
                "access_key_id": "string",
                "secret_access_key": "string",
            },
            "path": "string",
            "role_arn": "string",
        },
        "azure_blob_storage_data_source": {
            "container": "string",
            "storage_account": "string",
            "azure_credentials": {
                "sas_token": "string",
            },
            "credentials_secret": "string",
            "path": "string",
        },
        "gcs_data_sink": {
            "bucket_name": "string",
            "path": "string",
        },
        "gcs_data_source": {
            "bucket_name": "string",
            "path": "string",
        },
        "hdfs_data_source": {
            "path": "string",
        },
        "http_data_source": {
            "list_url": "string",
        },
        "object_conditions": {
            "exclude_prefixes": ["string"],
            "include_prefixes": ["string"],
            "last_modified_before": "string",
            "last_modified_since": "string",
            "max_time_elapsed_since_last_modification": "string",
            "min_time_elapsed_since_last_modification": "string",
        },
        "posix_data_sink": {
            "root_directory": "string",
        },
        "posix_data_source": {
            "root_directory": "string",
        },
        "sink_agent_pool_name": "string",
        "source_agent_pool_name": "string",
        "transfer_options": {
            "delete_objects_from_source_after_transfer": False,
            "delete_objects_unique_in_sink": False,
            "overwrite_objects_already_existing_in_sink": False,
            "overwrite_when": "string",
        },
    })
const transferJobResource = new gcp.storage.TransferJob("transferJobResource", {
    description: "string",
    eventStream: {
        name: "string",
        eventStreamExpirationTime: "string",
        eventStreamStartTime: "string",
    },
    loggingConfig: {
        enableOnPremGcsTransferLogs: false,
        logActionStates: ["string"],
        logActions: ["string"],
    },
    name: "string",
    notificationConfig: {
        payloadFormat: "string",
        pubsubTopic: "string",
        eventTypes: ["string"],
    },
    project: "string",
    replicationSpec: {
        gcsDataSink: {
            bucketName: "string",
            path: "string",
        },
        gcsDataSource: {
            bucketName: "string",
            path: "string",
        },
        objectConditions: {
            excludePrefixes: ["string"],
            includePrefixes: ["string"],
            lastModifiedBefore: "string",
            lastModifiedSince: "string",
            maxTimeElapsedSinceLastModification: "string",
            minTimeElapsedSinceLastModification: "string",
        },
        transferOptions: {
            deleteObjectsFromSourceAfterTransfer: false,
            deleteObjectsUniqueInSink: false,
            overwriteObjectsAlreadyExistingInSink: false,
            overwriteWhen: "string",
        },
    },
    schedule: {
        scheduleStartDate: {
            day: 0,
            month: 0,
            year: 0,
        },
        repeatInterval: "string",
        scheduleEndDate: {
            day: 0,
            month: 0,
            year: 0,
        },
        startTimeOfDay: {
            hours: 0,
            minutes: 0,
            nanos: 0,
            seconds: 0,
        },
    },
    status: "string",
    transferSpec: {
        awsS3DataSource: {
            bucketName: "string",
            awsAccessKey: {
                accessKeyId: "string",
                secretAccessKey: "string",
            },
            path: "string",
            roleArn: "string",
        },
        azureBlobStorageDataSource: {
            container: "string",
            storageAccount: "string",
            azureCredentials: {
                sasToken: "string",
            },
            credentialsSecret: "string",
            path: "string",
        },
        gcsDataSink: {
            bucketName: "string",
            path: "string",
        },
        gcsDataSource: {
            bucketName: "string",
            path: "string",
        },
        hdfsDataSource: {
            path: "string",
        },
        httpDataSource: {
            listUrl: "string",
        },
        objectConditions: {
            excludePrefixes: ["string"],
            includePrefixes: ["string"],
            lastModifiedBefore: "string",
            lastModifiedSince: "string",
            maxTimeElapsedSinceLastModification: "string",
            minTimeElapsedSinceLastModification: "string",
        },
        posixDataSink: {
            rootDirectory: "string",
        },
        posixDataSource: {
            rootDirectory: "string",
        },
        sinkAgentPoolName: "string",
        sourceAgentPoolName: "string",
        transferOptions: {
            deleteObjectsFromSourceAfterTransfer: false,
            deleteObjectsUniqueInSink: false,
            overwriteObjectsAlreadyExistingInSink: false,
            overwriteWhen: "string",
        },
    },
});
type: gcp:storage:TransferJob
properties:
    description: string
    eventStream:
        eventStreamExpirationTime: string
        eventStreamStartTime: string
        name: string
    loggingConfig:
        enableOnPremGcsTransferLogs: false
        logActionStates:
            - string
        logActions:
            - string
    name: string
    notificationConfig:
        eventTypes:
            - string
        payloadFormat: string
        pubsubTopic: string
    project: string
    replicationSpec:
        gcsDataSink:
            bucketName: string
            path: string
        gcsDataSource:
            bucketName: string
            path: string
        objectConditions:
            excludePrefixes:
                - string
            includePrefixes:
                - string
            lastModifiedBefore: string
            lastModifiedSince: string
            maxTimeElapsedSinceLastModification: string
            minTimeElapsedSinceLastModification: string
        transferOptions:
            deleteObjectsFromSourceAfterTransfer: false
            deleteObjectsUniqueInSink: false
            overwriteObjectsAlreadyExistingInSink: false
            overwriteWhen: string
    schedule:
        repeatInterval: string
        scheduleEndDate:
            day: 0
            month: 0
            year: 0
        scheduleStartDate:
            day: 0
            month: 0
            year: 0
        startTimeOfDay:
            hours: 0
            minutes: 0
            nanos: 0
            seconds: 0
    status: string
    transferSpec:
        awsS3DataSource:
            awsAccessKey:
                accessKeyId: string
                secretAccessKey: string
            bucketName: string
            path: string
            roleArn: string
        azureBlobStorageDataSource:
            azureCredentials:
                sasToken: string
            container: string
            credentialsSecret: string
            path: string
            storageAccount: string
        gcsDataSink:
            bucketName: string
            path: string
        gcsDataSource:
            bucketName: string
            path: string
        hdfsDataSource:
            path: string
        httpDataSource:
            listUrl: string
        objectConditions:
            excludePrefixes:
                - string
            includePrefixes:
                - string
            lastModifiedBefore: string
            lastModifiedSince: string
            maxTimeElapsedSinceLastModification: string
            minTimeElapsedSinceLastModification: string
        posixDataSink:
            rootDirectory: string
        posixDataSource:
            rootDirectory: string
        sinkAgentPoolName: string
        sourceAgentPoolName: string
        transferOptions:
            deleteObjectsFromSourceAfterTransfer: false
            deleteObjectsUniqueInSink: false
            overwriteObjectsAlreadyExistingInSink: false
            overwriteWhen: string
TransferJob Resource Properties
To learn more about resource properties and how to use them, see Inputs and Outputs in the Architecture and Concepts docs.
Inputs
In Python, inputs that are objects can be passed either as argument classes or as dictionary literals.
The TransferJob resource accepts the following input properties:
- Description string
- Unique description to identify the Transfer Job.
- EventStream TransferJob Event Stream 
- Specifies the Event-driven transfer options. Event-driven transfers listen to an event stream to transfer updated files. Structure documented below Either event_streamorschedulemust be set.
- LoggingConfig TransferJob Logging Config 
- Logging configuration. Structure documented below.
- Name string
- The name of the Transfer Job. This name must start with "transferJobs/" prefix and end with a letter or a number, and should be no more than 128 characters ( transferJobs/^(?!OPI)[A-Za-z0-9-._~]*[A-Za-z0-9]$). For transfers involving PosixFilesystem, this name must start with transferJobs/OPI specifically (transferJobs/OPI^[A-Za-z0-9-._~]*[A-Za-z0-9]$). For all other transfer types, this name must not start with transferJobs/OPI. Default the provider will assign a random unique name withtransferJobs/{{name}}format, wherenameis a numeric value.
- NotificationConfig TransferJob Notification Config 
- Notification configuration. This is not supported for transfers involving PosixFilesystem. Structure documented below.
- Project string
- The project in which the resource belongs. If it is not provided, the provider project is used.
- ReplicationSpec TransferJob Replication Spec 
- Replication specification. Structure documented below. User should not configure schedule,event_streamwith this argument. One oftransfer_spec, orreplication_specmust be specified.
- Schedule
TransferJob Schedule 
- Schedule specification defining when the Transfer Job should be scheduled to start, end and what time to run. Structure documented below. Either scheduleorevent_streammust be set.
- Status string
- Status of the job. Default: ENABLED. NOTE: The effect of the new job status takes place during a subsequent job run. For example, if you change the job status from ENABLED to DISABLED, and an operation spawned by the transfer is running, the status change would not affect the current operation.
- TransferSpec TransferJob Transfer Spec 
- Transfer specification. Structure documented below. One of transfer_spec, orreplication_speccan be specified.
- Description string
- Unique description to identify the Transfer Job.
- EventStream TransferJob Event Stream Args 
- Specifies the Event-driven transfer options. Event-driven transfers listen to an event stream to transfer updated files. Structure documented below Either event_streamorschedulemust be set.
- LoggingConfig TransferJob Logging Config Args 
- Logging configuration. Structure documented below.
- Name string
- The name of the Transfer Job. This name must start with "transferJobs/" prefix and end with a letter or a number, and should be no more than 128 characters ( transferJobs/^(?!OPI)[A-Za-z0-9-._~]*[A-Za-z0-9]$). For transfers involving PosixFilesystem, this name must start with transferJobs/OPI specifically (transferJobs/OPI^[A-Za-z0-9-._~]*[A-Za-z0-9]$). For all other transfer types, this name must not start with transferJobs/OPI. Default the provider will assign a random unique name withtransferJobs/{{name}}format, wherenameis a numeric value.
- NotificationConfig TransferJob Notification Config Args 
- Notification configuration. This is not supported for transfers involving PosixFilesystem. Structure documented below.
- Project string
- The project in which the resource belongs. If it is not provided, the provider project is used.
- ReplicationSpec TransferJob Replication Spec Args 
- Replication specification. Structure documented below. User should not configure schedule,event_streamwith this argument. One oftransfer_spec, orreplication_specmust be specified.
- Schedule
TransferJob Schedule Args 
- Schedule specification defining when the Transfer Job should be scheduled to start, end and what time to run. Structure documented below. Either scheduleorevent_streammust be set.
- Status string
- Status of the job. Default: ENABLED. NOTE: The effect of the new job status takes place during a subsequent job run. For example, if you change the job status from ENABLED to DISABLED, and an operation spawned by the transfer is running, the status change would not affect the current operation.
- TransferSpec TransferJob Transfer Spec Args 
- Transfer specification. Structure documented below. One of transfer_spec, orreplication_speccan be specified.
- description String
- Unique description to identify the Transfer Job.
- eventStream TransferJob Event Stream 
- Specifies the Event-driven transfer options. Event-driven transfers listen to an event stream to transfer updated files. Structure documented below Either event_streamorschedulemust be set.
- loggingConfig TransferJob Logging Config 
- Logging configuration. Structure documented below.
- name String
- The name of the Transfer Job. This name must start with "transferJobs/" prefix and end with a letter or a number, and should be no more than 128 characters ( transferJobs/^(?!OPI)[A-Za-z0-9-._~]*[A-Za-z0-9]$). For transfers involving PosixFilesystem, this name must start with transferJobs/OPI specifically (transferJobs/OPI^[A-Za-z0-9-._~]*[A-Za-z0-9]$). For all other transfer types, this name must not start with transferJobs/OPI. Default the provider will assign a random unique name withtransferJobs/{{name}}format, wherenameis a numeric value.
- notificationConfig TransferJob Notification Config 
- Notification configuration. This is not supported for transfers involving PosixFilesystem. Structure documented below.
- project String
- The project in which the resource belongs. If it is not provided, the provider project is used.
- replicationSpec TransferJob Replication Spec 
- Replication specification. Structure documented below. User should not configure schedule,event_streamwith this argument. One oftransfer_spec, orreplication_specmust be specified.
- schedule
TransferJob Schedule 
- Schedule specification defining when the Transfer Job should be scheduled to start, end and what time to run. Structure documented below. Either scheduleorevent_streammust be set.
- status String
- Status of the job. Default: ENABLED. NOTE: The effect of the new job status takes place during a subsequent job run. For example, if you change the job status from ENABLED to DISABLED, and an operation spawned by the transfer is running, the status change would not affect the current operation.
- transferSpec TransferJob Transfer Spec 
- Transfer specification. Structure documented below. One of transfer_spec, orreplication_speccan be specified.
- description string
- Unique description to identify the Transfer Job.
- eventStream TransferJob Event Stream 
- Specifies the Event-driven transfer options. Event-driven transfers listen to an event stream to transfer updated files. Structure documented below Either event_streamorschedulemust be set.
- loggingConfig TransferJob Logging Config 
- Logging configuration. Structure documented below.
- name string
- The name of the Transfer Job. This name must start with "transferJobs/" prefix and end with a letter or a number, and should be no more than 128 characters ( transferJobs/^(?!OPI)[A-Za-z0-9-._~]*[A-Za-z0-9]$). For transfers involving PosixFilesystem, this name must start with transferJobs/OPI specifically (transferJobs/OPI^[A-Za-z0-9-._~]*[A-Za-z0-9]$). For all other transfer types, this name must not start with transferJobs/OPI. Default the provider will assign a random unique name withtransferJobs/{{name}}format, wherenameis a numeric value.
- notificationConfig TransferJob Notification Config 
- Notification configuration. This is not supported for transfers involving PosixFilesystem. Structure documented below.
- project string
- The project in which the resource belongs. If it is not provided, the provider project is used.
- replicationSpec TransferJob Replication Spec 
- Replication specification. Structure documented below. User should not configure schedule,event_streamwith this argument. One oftransfer_spec, orreplication_specmust be specified.
- schedule
TransferJob Schedule 
- Schedule specification defining when the Transfer Job should be scheduled to start, end and what time to run. Structure documented below. Either scheduleorevent_streammust be set.
- status string
- Status of the job. Default: ENABLED. NOTE: The effect of the new job status takes place during a subsequent job run. For example, if you change the job status from ENABLED to DISABLED, and an operation spawned by the transfer is running, the status change would not affect the current operation.
- transferSpec TransferJob Transfer Spec 
- Transfer specification. Structure documented below. One of transfer_spec, orreplication_speccan be specified.
- description str
- Unique description to identify the Transfer Job.
- event_stream TransferJob Event Stream Args 
- Specifies the Event-driven transfer options. Event-driven transfers listen to an event stream to transfer updated files. Structure documented below Either event_streamorschedulemust be set.
- logging_config TransferJob Logging Config Args 
- Logging configuration. Structure documented below.
- name str
- The name of the Transfer Job. This name must start with "transferJobs/" prefix and end with a letter or a number, and should be no more than 128 characters ( transferJobs/^(?!OPI)[A-Za-z0-9-._~]*[A-Za-z0-9]$). For transfers involving PosixFilesystem, this name must start with transferJobs/OPI specifically (transferJobs/OPI^[A-Za-z0-9-._~]*[A-Za-z0-9]$). For all other transfer types, this name must not start with transferJobs/OPI. Default the provider will assign a random unique name withtransferJobs/{{name}}format, wherenameis a numeric value.
- notification_config TransferJob Notification Config Args 
- Notification configuration. This is not supported for transfers involving PosixFilesystem. Structure documented below.
- project str
- The project in which the resource belongs. If it is not provided, the provider project is used.
- replication_spec TransferJob Replication Spec Args 
- Replication specification. Structure documented below. User should not configure schedule,event_streamwith this argument. One oftransfer_spec, orreplication_specmust be specified.
- schedule
TransferJob Schedule Args 
- Schedule specification defining when the Transfer Job should be scheduled to start, end and what time to run. Structure documented below. Either scheduleorevent_streammust be set.
- status str
- Status of the job. Default: ENABLED. NOTE: The effect of the new job status takes place during a subsequent job run. For example, if you change the job status from ENABLED to DISABLED, and an operation spawned by the transfer is running, the status change would not affect the current operation.
- transfer_spec TransferJob Transfer Spec Args 
- Transfer specification. Structure documented below. One of transfer_spec, orreplication_speccan be specified.
- description String
- Unique description to identify the Transfer Job.
- eventStream Property Map
- Specifies the Event-driven transfer options. Event-driven transfers listen to an event stream to transfer updated files. Structure documented below Either event_streamorschedulemust be set.
- loggingConfig Property Map
- Logging configuration. Structure documented below.
- name String
- The name of the Transfer Job. This name must start with "transferJobs/" prefix and end with a letter or a number, and should be no more than 128 characters ( transferJobs/^(?!OPI)[A-Za-z0-9-._~]*[A-Za-z0-9]$). For transfers involving PosixFilesystem, this name must start with transferJobs/OPI specifically (transferJobs/OPI^[A-Za-z0-9-._~]*[A-Za-z0-9]$). For all other transfer types, this name must not start with transferJobs/OPI. Default the provider will assign a random unique name withtransferJobs/{{name}}format, wherenameis a numeric value.
- notificationConfig Property Map
- Notification configuration. This is not supported for transfers involving PosixFilesystem. Structure documented below.
- project String
- The project in which the resource belongs. If it is not provided, the provider project is used.
- replicationSpec Property Map
- Replication specification. Structure documented below. User should not configure schedule,event_streamwith this argument. One oftransfer_spec, orreplication_specmust be specified.
- schedule Property Map
- Schedule specification defining when the Transfer Job should be scheduled to start, end and what time to run. Structure documented below. Either scheduleorevent_streammust be set.
- status String
- Status of the job. Default: ENABLED. NOTE: The effect of the new job status takes place during a subsequent job run. For example, if you change the job status from ENABLED to DISABLED, and an operation spawned by the transfer is running, the status change would not affect the current operation.
- transferSpec Property Map
- Transfer specification. Structure documented below. One of transfer_spec, orreplication_speccan be specified.
Outputs
All input properties are implicitly available as output properties. Additionally, the TransferJob resource produces the following output properties:
- CreationTime string
- When the Transfer Job was created.
- DeletionTime string
- When the Transfer Job was deleted.
- Id string
- The provider-assigned unique ID for this managed resource.
- LastModification stringTime 
- When the Transfer Job was last modified.
- CreationTime string
- When the Transfer Job was created.
- DeletionTime string
- When the Transfer Job was deleted.
- Id string
- The provider-assigned unique ID for this managed resource.
- LastModification stringTime 
- When the Transfer Job was last modified.
- creationTime String
- When the Transfer Job was created.
- deletionTime String
- When the Transfer Job was deleted.
- id String
- The provider-assigned unique ID for this managed resource.
- lastModification StringTime 
- When the Transfer Job was last modified.
- creationTime string
- When the Transfer Job was created.
- deletionTime string
- When the Transfer Job was deleted.
- id string
- The provider-assigned unique ID for this managed resource.
- lastModification stringTime 
- When the Transfer Job was last modified.
- creation_time str
- When the Transfer Job was created.
- deletion_time str
- When the Transfer Job was deleted.
- id str
- The provider-assigned unique ID for this managed resource.
- last_modification_ strtime 
- When the Transfer Job was last modified.
- creationTime String
- When the Transfer Job was created.
- deletionTime String
- When the Transfer Job was deleted.
- id String
- The provider-assigned unique ID for this managed resource.
- lastModification StringTime 
- When the Transfer Job was last modified.
Look up Existing TransferJob Resource
Get an existing TransferJob resource’s state with the given name, ID, and optional extra properties used to qualify the lookup.
public static get(name: string, id: Input<ID>, state?: TransferJobState, opts?: CustomResourceOptions): TransferJob@staticmethod
def get(resource_name: str,
        id: str,
        opts: Optional[ResourceOptions] = None,
        creation_time: Optional[str] = None,
        deletion_time: Optional[str] = None,
        description: Optional[str] = None,
        event_stream: Optional[TransferJobEventStreamArgs] = None,
        last_modification_time: Optional[str] = None,
        logging_config: Optional[TransferJobLoggingConfigArgs] = None,
        name: Optional[str] = None,
        notification_config: Optional[TransferJobNotificationConfigArgs] = None,
        project: Optional[str] = None,
        replication_spec: Optional[TransferJobReplicationSpecArgs] = None,
        schedule: Optional[TransferJobScheduleArgs] = None,
        status: Optional[str] = None,
        transfer_spec: Optional[TransferJobTransferSpecArgs] = None) -> TransferJobfunc GetTransferJob(ctx *Context, name string, id IDInput, state *TransferJobState, opts ...ResourceOption) (*TransferJob, error)public static TransferJob Get(string name, Input<string> id, TransferJobState? state, CustomResourceOptions? opts = null)public static TransferJob get(String name, Output<String> id, TransferJobState state, CustomResourceOptions options)resources:  _:    type: gcp:storage:TransferJob    get:      id: ${id}- name
- The unique name of the resulting resource.
- id
- The unique provider ID of the resource to lookup.
- state
- Any extra arguments used during the lookup.
- opts
- A bag of options that control this resource's behavior.
- resource_name
- The unique name of the resulting resource.
- id
- The unique provider ID of the resource to lookup.
- name
- The unique name of the resulting resource.
- id
- The unique provider ID of the resource to lookup.
- state
- Any extra arguments used during the lookup.
- opts
- A bag of options that control this resource's behavior.
- name
- The unique name of the resulting resource.
- id
- The unique provider ID of the resource to lookup.
- state
- Any extra arguments used during the lookup.
- opts
- A bag of options that control this resource's behavior.
- name
- The unique name of the resulting resource.
- id
- The unique provider ID of the resource to lookup.
- state
- Any extra arguments used during the lookup.
- opts
- A bag of options that control this resource's behavior.
- CreationTime string
- When the Transfer Job was created.
- DeletionTime string
- When the Transfer Job was deleted.
- Description string
- Unique description to identify the Transfer Job.
- EventStream TransferJob Event Stream 
- Specifies the Event-driven transfer options. Event-driven transfers listen to an event stream to transfer updated files. Structure documented below Either event_streamorschedulemust be set.
- LastModification stringTime 
- When the Transfer Job was last modified.
- LoggingConfig TransferJob Logging Config 
- Logging configuration. Structure documented below.
- Name string
- The name of the Transfer Job. This name must start with "transferJobs/" prefix and end with a letter or a number, and should be no more than 128 characters ( transferJobs/^(?!OPI)[A-Za-z0-9-._~]*[A-Za-z0-9]$). For transfers involving PosixFilesystem, this name must start with transferJobs/OPI specifically (transferJobs/OPI^[A-Za-z0-9-._~]*[A-Za-z0-9]$). For all other transfer types, this name must not start with transferJobs/OPI. Default the provider will assign a random unique name withtransferJobs/{{name}}format, wherenameis a numeric value.
- NotificationConfig TransferJob Notification Config 
- Notification configuration. This is not supported for transfers involving PosixFilesystem. Structure documented below.
- Project string
- The project in which the resource belongs. If it is not provided, the provider project is used.
- ReplicationSpec TransferJob Replication Spec 
- Replication specification. Structure documented below. User should not configure schedule,event_streamwith this argument. One oftransfer_spec, orreplication_specmust be specified.
- Schedule
TransferJob Schedule 
- Schedule specification defining when the Transfer Job should be scheduled to start, end and what time to run. Structure documented below. Either scheduleorevent_streammust be set.
- Status string
- Status of the job. Default: ENABLED. NOTE: The effect of the new job status takes place during a subsequent job run. For example, if you change the job status from ENABLED to DISABLED, and an operation spawned by the transfer is running, the status change would not affect the current operation.
- TransferSpec TransferJob Transfer Spec 
- Transfer specification. Structure documented below. One of transfer_spec, orreplication_speccan be specified.
- CreationTime string
- When the Transfer Job was created.
- DeletionTime string
- When the Transfer Job was deleted.
- Description string
- Unique description to identify the Transfer Job.
- EventStream TransferJob Event Stream Args 
- Specifies the Event-driven transfer options. Event-driven transfers listen to an event stream to transfer updated files. Structure documented below Either event_streamorschedulemust be set.
- LastModification stringTime 
- When the Transfer Job was last modified.
- LoggingConfig TransferJob Logging Config Args 
- Logging configuration. Structure documented below.
- Name string
- The name of the Transfer Job. This name must start with "transferJobs/" prefix and end with a letter or a number, and should be no more than 128 characters ( transferJobs/^(?!OPI)[A-Za-z0-9-._~]*[A-Za-z0-9]$). For transfers involving PosixFilesystem, this name must start with transferJobs/OPI specifically (transferJobs/OPI^[A-Za-z0-9-._~]*[A-Za-z0-9]$). For all other transfer types, this name must not start with transferJobs/OPI. Default the provider will assign a random unique name withtransferJobs/{{name}}format, wherenameis a numeric value.
- NotificationConfig TransferJob Notification Config Args 
- Notification configuration. This is not supported for transfers involving PosixFilesystem. Structure documented below.
- Project string
- The project in which the resource belongs. If it is not provided, the provider project is used.
- ReplicationSpec TransferJob Replication Spec Args 
- Replication specification. Structure documented below. User should not configure schedule,event_streamwith this argument. One oftransfer_spec, orreplication_specmust be specified.
- Schedule
TransferJob Schedule Args 
- Schedule specification defining when the Transfer Job should be scheduled to start, end and what time to run. Structure documented below. Either scheduleorevent_streammust be set.
- Status string
- Status of the job. Default: ENABLED. NOTE: The effect of the new job status takes place during a subsequent job run. For example, if you change the job status from ENABLED to DISABLED, and an operation spawned by the transfer is running, the status change would not affect the current operation.
- TransferSpec TransferJob Transfer Spec Args 
- Transfer specification. Structure documented below. One of transfer_spec, orreplication_speccan be specified.
- creationTime String
- When the Transfer Job was created.
- deletionTime String
- When the Transfer Job was deleted.
- description String
- Unique description to identify the Transfer Job.
- eventStream TransferJob Event Stream 
- Specifies the Event-driven transfer options. Event-driven transfers listen to an event stream to transfer updated files. Structure documented below Either event_streamorschedulemust be set.
- lastModification StringTime 
- When the Transfer Job was last modified.
- loggingConfig TransferJob Logging Config 
- Logging configuration. Structure documented below.
- name String
- The name of the Transfer Job. This name must start with "transferJobs/" prefix and end with a letter or a number, and should be no more than 128 characters ( transferJobs/^(?!OPI)[A-Za-z0-9-._~]*[A-Za-z0-9]$). For transfers involving PosixFilesystem, this name must start with transferJobs/OPI specifically (transferJobs/OPI^[A-Za-z0-9-._~]*[A-Za-z0-9]$). For all other transfer types, this name must not start with transferJobs/OPI. Default the provider will assign a random unique name withtransferJobs/{{name}}format, wherenameis a numeric value.
- notificationConfig TransferJob Notification Config 
- Notification configuration. This is not supported for transfers involving PosixFilesystem. Structure documented below.
- project String
- The project in which the resource belongs. If it is not provided, the provider project is used.
- replicationSpec TransferJob Replication Spec 
- Replication specification. Structure documented below. User should not configure schedule,event_streamwith this argument. One oftransfer_spec, orreplication_specmust be specified.
- schedule
TransferJob Schedule 
- Schedule specification defining when the Transfer Job should be scheduled to start, end and what time to run. Structure documented below. Either scheduleorevent_streammust be set.
- status String
- Status of the job. Default: ENABLED. NOTE: The effect of the new job status takes place during a subsequent job run. For example, if you change the job status from ENABLED to DISABLED, and an operation spawned by the transfer is running, the status change would not affect the current operation.
- transferSpec TransferJob Transfer Spec 
- Transfer specification. Structure documented below. One of transfer_spec, orreplication_speccan be specified.
- creationTime string
- When the Transfer Job was created.
- deletionTime string
- When the Transfer Job was deleted.
- description string
- Unique description to identify the Transfer Job.
- eventStream TransferJob Event Stream 
- Specifies the Event-driven transfer options. Event-driven transfers listen to an event stream to transfer updated files. Structure documented below Either event_streamorschedulemust be set.
- lastModification stringTime 
- When the Transfer Job was last modified.
- loggingConfig TransferJob Logging Config 
- Logging configuration. Structure documented below.
- name string
- The name of the Transfer Job. This name must start with "transferJobs/" prefix and end with a letter or a number, and should be no more than 128 characters ( transferJobs/^(?!OPI)[A-Za-z0-9-._~]*[A-Za-z0-9]$). For transfers involving PosixFilesystem, this name must start with transferJobs/OPI specifically (transferJobs/OPI^[A-Za-z0-9-._~]*[A-Za-z0-9]$). For all other transfer types, this name must not start with transferJobs/OPI. Default the provider will assign a random unique name withtransferJobs/{{name}}format, wherenameis a numeric value.
- notificationConfig TransferJob Notification Config 
- Notification configuration. This is not supported for transfers involving PosixFilesystem. Structure documented below.
- project string
- The project in which the resource belongs. If it is not provided, the provider project is used.
- replicationSpec TransferJob Replication Spec 
- Replication specification. Structure documented below. User should not configure schedule,event_streamwith this argument. One oftransfer_spec, orreplication_specmust be specified.
- schedule
TransferJob Schedule 
- Schedule specification defining when the Transfer Job should be scheduled to start, end and what time to run. Structure documented below. Either scheduleorevent_streammust be set.
- status string
- Status of the job. Default: ENABLED. NOTE: The effect of the new job status takes place during a subsequent job run. For example, if you change the job status from ENABLED to DISABLED, and an operation spawned by the transfer is running, the status change would not affect the current operation.
- transferSpec TransferJob Transfer Spec 
- Transfer specification. Structure documented below. One of transfer_spec, orreplication_speccan be specified.
- creation_time str
- When the Transfer Job was created.
- deletion_time str
- When the Transfer Job was deleted.
- description str
- Unique description to identify the Transfer Job.
- event_stream TransferJob Event Stream Args 
- Specifies the Event-driven transfer options. Event-driven transfers listen to an event stream to transfer updated files. Structure documented below Either event_streamorschedulemust be set.
- last_modification_ strtime 
- When the Transfer Job was last modified.
- logging_config TransferJob Logging Config Args 
- Logging configuration. Structure documented below.
- name str
- The name of the Transfer Job. This name must start with "transferJobs/" prefix and end with a letter or a number, and should be no more than 128 characters ( transferJobs/^(?!OPI)[A-Za-z0-9-._~]*[A-Za-z0-9]$). For transfers involving PosixFilesystem, this name must start with transferJobs/OPI specifically (transferJobs/OPI^[A-Za-z0-9-._~]*[A-Za-z0-9]$). For all other transfer types, this name must not start with transferJobs/OPI. Default the provider will assign a random unique name withtransferJobs/{{name}}format, wherenameis a numeric value.
- notification_config TransferJob Notification Config Args 
- Notification configuration. This is not supported for transfers involving PosixFilesystem. Structure documented below.
- project str
- The project in which the resource belongs. If it is not provided, the provider project is used.
- replication_spec TransferJob Replication Spec Args 
- Replication specification. Structure documented below. User should not configure schedule,event_streamwith this argument. One oftransfer_spec, orreplication_specmust be specified.
- schedule
TransferJob Schedule Args 
- Schedule specification defining when the Transfer Job should be scheduled to start, end and what time to run. Structure documented below. Either scheduleorevent_streammust be set.
- status str
- Status of the job. Default: ENABLED. NOTE: The effect of the new job status takes place during a subsequent job run. For example, if you change the job status from ENABLED to DISABLED, and an operation spawned by the transfer is running, the status change would not affect the current operation.
- transfer_spec TransferJob Transfer Spec Args 
- Transfer specification. Structure documented below. One of transfer_spec, orreplication_speccan be specified.
- creationTime String
- When the Transfer Job was created.
- deletionTime String
- When the Transfer Job was deleted.
- description String
- Unique description to identify the Transfer Job.
- eventStream Property Map
- Specifies the Event-driven transfer options. Event-driven transfers listen to an event stream to transfer updated files. Structure documented below Either event_streamorschedulemust be set.
- lastModification StringTime 
- When the Transfer Job was last modified.
- loggingConfig Property Map
- Logging configuration. Structure documented below.
- name String
- The name of the Transfer Job. This name must start with "transferJobs/" prefix and end with a letter or a number, and should be no more than 128 characters ( transferJobs/^(?!OPI)[A-Za-z0-9-._~]*[A-Za-z0-9]$). For transfers involving PosixFilesystem, this name must start with transferJobs/OPI specifically (transferJobs/OPI^[A-Za-z0-9-._~]*[A-Za-z0-9]$). For all other transfer types, this name must not start with transferJobs/OPI. Default the provider will assign a random unique name withtransferJobs/{{name}}format, wherenameis a numeric value.
- notificationConfig Property Map
- Notification configuration. This is not supported for transfers involving PosixFilesystem. Structure documented below.
- project String
- The project in which the resource belongs. If it is not provided, the provider project is used.
- replicationSpec Property Map
- Replication specification. Structure documented below. User should not configure schedule,event_streamwith this argument. One oftransfer_spec, orreplication_specmust be specified.
- schedule Property Map
- Schedule specification defining when the Transfer Job should be scheduled to start, end and what time to run. Structure documented below. Either scheduleorevent_streammust be set.
- status String
- Status of the job. Default: ENABLED. NOTE: The effect of the new job status takes place during a subsequent job run. For example, if you change the job status from ENABLED to DISABLED, and an operation spawned by the transfer is running, the status change would not affect the current operation.
- transferSpec Property Map
- Transfer specification. Structure documented below. One of transfer_spec, orreplication_speccan be specified.
Supporting Types
TransferJobEventStream, TransferJobEventStreamArgs        
- Name string
- Specifies a unique name of the resource such as AWS SQS ARN in the form 'arn:aws:sqs:region:account_id:queue_name', or Pub/Sub subscription resource name in the form 'projects/{project}/subscriptions/{sub}'.
- EventStream stringExpiration Time 
- Specifies the data and time at which Storage Transfer Service stops listening for events from this stream. After this time, any transfers in progress will complete, but no new transfers are initiated.A timestamp in RFC3339 UTC "Zulu" format, with nanosecond resolution and up to nine fractional digits. Examples: "2014-10-02T15:01:23Z" and "2014-10-02T15:01:23.045123456Z".
- EventStream stringStart Time 
- Specifies the date and time that Storage Transfer Service starts listening for events from this stream. If no start time is specified or start time is in the past, Storage Transfer Service starts listening immediately. A timestamp in RFC3339 UTC "Zulu" format, with nanosecond resolution and up to nine fractional digits. Examples: "2014-10-02T15:01:23Z" and "2014-10-02T15:01:23.045123456Z".
- Name string
- Specifies a unique name of the resource such as AWS SQS ARN in the form 'arn:aws:sqs:region:account_id:queue_name', or Pub/Sub subscription resource name in the form 'projects/{project}/subscriptions/{sub}'.
- EventStream stringExpiration Time 
- Specifies the data and time at which Storage Transfer Service stops listening for events from this stream. After this time, any transfers in progress will complete, but no new transfers are initiated.A timestamp in RFC3339 UTC "Zulu" format, with nanosecond resolution and up to nine fractional digits. Examples: "2014-10-02T15:01:23Z" and "2014-10-02T15:01:23.045123456Z".
- EventStream stringStart Time 
- Specifies the date and time that Storage Transfer Service starts listening for events from this stream. If no start time is specified or start time is in the past, Storage Transfer Service starts listening immediately. A timestamp in RFC3339 UTC "Zulu" format, with nanosecond resolution and up to nine fractional digits. Examples: "2014-10-02T15:01:23Z" and "2014-10-02T15:01:23.045123456Z".
- name String
- Specifies a unique name of the resource such as AWS SQS ARN in the form 'arn:aws:sqs:region:account_id:queue_name', or Pub/Sub subscription resource name in the form 'projects/{project}/subscriptions/{sub}'.
- eventStream StringExpiration Time 
- Specifies the data and time at which Storage Transfer Service stops listening for events from this stream. After this time, any transfers in progress will complete, but no new transfers are initiated.A timestamp in RFC3339 UTC "Zulu" format, with nanosecond resolution and up to nine fractional digits. Examples: "2014-10-02T15:01:23Z" and "2014-10-02T15:01:23.045123456Z".
- eventStream StringStart Time 
- Specifies the date and time that Storage Transfer Service starts listening for events from this stream. If no start time is specified or start time is in the past, Storage Transfer Service starts listening immediately. A timestamp in RFC3339 UTC "Zulu" format, with nanosecond resolution and up to nine fractional digits. Examples: "2014-10-02T15:01:23Z" and "2014-10-02T15:01:23.045123456Z".
- name string
- Specifies a unique name of the resource such as AWS SQS ARN in the form 'arn:aws:sqs:region:account_id:queue_name', or Pub/Sub subscription resource name in the form 'projects/{project}/subscriptions/{sub}'.
- eventStream stringExpiration Time 
- Specifies the data and time at which Storage Transfer Service stops listening for events from this stream. After this time, any transfers in progress will complete, but no new transfers are initiated.A timestamp in RFC3339 UTC "Zulu" format, with nanosecond resolution and up to nine fractional digits. Examples: "2014-10-02T15:01:23Z" and "2014-10-02T15:01:23.045123456Z".
- eventStream stringStart Time 
- Specifies the date and time that Storage Transfer Service starts listening for events from this stream. If no start time is specified or start time is in the past, Storage Transfer Service starts listening immediately. A timestamp in RFC3339 UTC "Zulu" format, with nanosecond resolution and up to nine fractional digits. Examples: "2014-10-02T15:01:23Z" and "2014-10-02T15:01:23.045123456Z".
- name str
- Specifies a unique name of the resource such as AWS SQS ARN in the form 'arn:aws:sqs:region:account_id:queue_name', or Pub/Sub subscription resource name in the form 'projects/{project}/subscriptions/{sub}'.
- event_stream_ strexpiration_ time 
- Specifies the data and time at which Storage Transfer Service stops listening for events from this stream. After this time, any transfers in progress will complete, but no new transfers are initiated.A timestamp in RFC3339 UTC "Zulu" format, with nanosecond resolution and up to nine fractional digits. Examples: "2014-10-02T15:01:23Z" and "2014-10-02T15:01:23.045123456Z".
- event_stream_ strstart_ time 
- Specifies the date and time that Storage Transfer Service starts listening for events from this stream. If no start time is specified or start time is in the past, Storage Transfer Service starts listening immediately. A timestamp in RFC3339 UTC "Zulu" format, with nanosecond resolution and up to nine fractional digits. Examples: "2014-10-02T15:01:23Z" and "2014-10-02T15:01:23.045123456Z".
- name String
- Specifies a unique name of the resource such as AWS SQS ARN in the form 'arn:aws:sqs:region:account_id:queue_name', or Pub/Sub subscription resource name in the form 'projects/{project}/subscriptions/{sub}'.
- eventStream StringExpiration Time 
- Specifies the data and time at which Storage Transfer Service stops listening for events from this stream. After this time, any transfers in progress will complete, but no new transfers are initiated.A timestamp in RFC3339 UTC "Zulu" format, with nanosecond resolution and up to nine fractional digits. Examples: "2014-10-02T15:01:23Z" and "2014-10-02T15:01:23.045123456Z".
- eventStream StringStart Time 
- Specifies the date and time that Storage Transfer Service starts listening for events from this stream. If no start time is specified or start time is in the past, Storage Transfer Service starts listening immediately. A timestamp in RFC3339 UTC "Zulu" format, with nanosecond resolution and up to nine fractional digits. Examples: "2014-10-02T15:01:23Z" and "2014-10-02T15:01:23.045123456Z".
TransferJobLoggingConfig, TransferJobLoggingConfigArgs        
- EnableOn boolPrem Gcs Transfer Logs 
- For transfers with a PosixFilesystem source, this option enables the Cloud Storage transfer logs for this transfer.
- LogAction List<string>States 
- States in which logActions are logged. Not supported for transfers with PosifxFilesystem data sources; use enable_on_prem_gcs_transfer_logs instead.
- LogActions List<string>
- Specifies the actions to be logged. Not supported for transfers with PosifxFilesystem data sources; use enable_on_prem_gcs_transfer_logs instead.
- EnableOn boolPrem Gcs Transfer Logs 
- For transfers with a PosixFilesystem source, this option enables the Cloud Storage transfer logs for this transfer.
- LogAction []stringStates 
- States in which logActions are logged. Not supported for transfers with PosifxFilesystem data sources; use enable_on_prem_gcs_transfer_logs instead.
- LogActions []string
- Specifies the actions to be logged. Not supported for transfers with PosifxFilesystem data sources; use enable_on_prem_gcs_transfer_logs instead.
- enableOn BooleanPrem Gcs Transfer Logs 
- For transfers with a PosixFilesystem source, this option enables the Cloud Storage transfer logs for this transfer.
- logAction List<String>States 
- States in which logActions are logged. Not supported for transfers with PosifxFilesystem data sources; use enable_on_prem_gcs_transfer_logs instead.
- logActions List<String>
- Specifies the actions to be logged. Not supported for transfers with PosifxFilesystem data sources; use enable_on_prem_gcs_transfer_logs instead.
- enableOn booleanPrem Gcs Transfer Logs 
- For transfers with a PosixFilesystem source, this option enables the Cloud Storage transfer logs for this transfer.
- logAction string[]States 
- States in which logActions are logged. Not supported for transfers with PosifxFilesystem data sources; use enable_on_prem_gcs_transfer_logs instead.
- logActions string[]
- Specifies the actions to be logged. Not supported for transfers with PosifxFilesystem data sources; use enable_on_prem_gcs_transfer_logs instead.
- enable_on_ boolprem_ gcs_ transfer_ logs 
- For transfers with a PosixFilesystem source, this option enables the Cloud Storage transfer logs for this transfer.
- log_action_ Sequence[str]states 
- States in which logActions are logged. Not supported for transfers with PosifxFilesystem data sources; use enable_on_prem_gcs_transfer_logs instead.
- log_actions Sequence[str]
- Specifies the actions to be logged. Not supported for transfers with PosifxFilesystem data sources; use enable_on_prem_gcs_transfer_logs instead.
- enableOn BooleanPrem Gcs Transfer Logs 
- For transfers with a PosixFilesystem source, this option enables the Cloud Storage transfer logs for this transfer.
- logAction List<String>States 
- States in which logActions are logged. Not supported for transfers with PosifxFilesystem data sources; use enable_on_prem_gcs_transfer_logs instead.
- logActions List<String>
- Specifies the actions to be logged. Not supported for transfers with PosifxFilesystem data sources; use enable_on_prem_gcs_transfer_logs instead.
TransferJobNotificationConfig, TransferJobNotificationConfigArgs        
- PayloadFormat string
- The desired format of the notification message payloads. One of "NONE" or "JSON".
- PubsubTopic string
- The Topic.name of the Pub/Sub topic to which to publish notifications. Must be of the format: projects/{project}/topics/{topic}. Not matching this format results in an INVALID_ARGUMENT error.
- EventTypes List<string>
- Event types for which a notification is desired. If empty, send notifications for all event types. The valid types are "TRANSFER_OPERATION_SUCCESS", "TRANSFER_OPERATION_FAILED", "TRANSFER_OPERATION_ABORTED".
- PayloadFormat string
- The desired format of the notification message payloads. One of "NONE" or "JSON".
- PubsubTopic string
- The Topic.name of the Pub/Sub topic to which to publish notifications. Must be of the format: projects/{project}/topics/{topic}. Not matching this format results in an INVALID_ARGUMENT error.
- EventTypes []string
- Event types for which a notification is desired. If empty, send notifications for all event types. The valid types are "TRANSFER_OPERATION_SUCCESS", "TRANSFER_OPERATION_FAILED", "TRANSFER_OPERATION_ABORTED".
- payloadFormat String
- The desired format of the notification message payloads. One of "NONE" or "JSON".
- pubsubTopic String
- The Topic.name of the Pub/Sub topic to which to publish notifications. Must be of the format: projects/{project}/topics/{topic}. Not matching this format results in an INVALID_ARGUMENT error.
- eventTypes List<String>
- Event types for which a notification is desired. If empty, send notifications for all event types. The valid types are "TRANSFER_OPERATION_SUCCESS", "TRANSFER_OPERATION_FAILED", "TRANSFER_OPERATION_ABORTED".
- payloadFormat string
- The desired format of the notification message payloads. One of "NONE" or "JSON".
- pubsubTopic string
- The Topic.name of the Pub/Sub topic to which to publish notifications. Must be of the format: projects/{project}/topics/{topic}. Not matching this format results in an INVALID_ARGUMENT error.
- eventTypes string[]
- Event types for which a notification is desired. If empty, send notifications for all event types. The valid types are "TRANSFER_OPERATION_SUCCESS", "TRANSFER_OPERATION_FAILED", "TRANSFER_OPERATION_ABORTED".
- payload_format str
- The desired format of the notification message payloads. One of "NONE" or "JSON".
- pubsub_topic str
- The Topic.name of the Pub/Sub topic to which to publish notifications. Must be of the format: projects/{project}/topics/{topic}. Not matching this format results in an INVALID_ARGUMENT error.
- event_types Sequence[str]
- Event types for which a notification is desired. If empty, send notifications for all event types. The valid types are "TRANSFER_OPERATION_SUCCESS", "TRANSFER_OPERATION_FAILED", "TRANSFER_OPERATION_ABORTED".
- payloadFormat String
- The desired format of the notification message payloads. One of "NONE" or "JSON".
- pubsubTopic String
- The Topic.name of the Pub/Sub topic to which to publish notifications. Must be of the format: projects/{project}/topics/{topic}. Not matching this format results in an INVALID_ARGUMENT error.
- eventTypes List<String>
- Event types for which a notification is desired. If empty, send notifications for all event types. The valid types are "TRANSFER_OPERATION_SUCCESS", "TRANSFER_OPERATION_FAILED", "TRANSFER_OPERATION_ABORTED".
TransferJobReplicationSpec, TransferJobReplicationSpecArgs        
- GcsData TransferSink Job Replication Spec Gcs Data Sink 
- A Google Cloud Storage data sink. Structure documented below.
- GcsData TransferSource Job Replication Spec Gcs Data Source 
- A Google Cloud Storage data source. Structure documented below.
- ObjectConditions TransferJob Replication Spec Object Conditions 
- Only objects that satisfy these object conditions are included in the set of data source and data sink objects. Object conditions based on objects' last_modification_timedo not exclude objects in a data sink. Structure documented below.
- TransferOptions TransferJob Replication Spec Transfer Options 
- Characteristics of how to treat files from datasource and sink during job. If the option delete_objects_unique_in_sinkis true, object conditions based on objects'last_modification_timeare ignored and do not exclude objects in a data source or a data sink. Structure documented below.
- GcsData TransferSink Job Replication Spec Gcs Data Sink 
- A Google Cloud Storage data sink. Structure documented below.
- GcsData TransferSource Job Replication Spec Gcs Data Source 
- A Google Cloud Storage data source. Structure documented below.
- ObjectConditions TransferJob Replication Spec Object Conditions 
- Only objects that satisfy these object conditions are included in the set of data source and data sink objects. Object conditions based on objects' last_modification_timedo not exclude objects in a data sink. Structure documented below.
- TransferOptions TransferJob Replication Spec Transfer Options 
- Characteristics of how to treat files from datasource and sink during job. If the option delete_objects_unique_in_sinkis true, object conditions based on objects'last_modification_timeare ignored and do not exclude objects in a data source or a data sink. Structure documented below.
- gcsData TransferSink Job Replication Spec Gcs Data Sink 
- A Google Cloud Storage data sink. Structure documented below.
- gcsData TransferSource Job Replication Spec Gcs Data Source 
- A Google Cloud Storage data source. Structure documented below.
- objectConditions TransferJob Replication Spec Object Conditions 
- Only objects that satisfy these object conditions are included in the set of data source and data sink objects. Object conditions based on objects' last_modification_timedo not exclude objects in a data sink. Structure documented below.
- transferOptions TransferJob Replication Spec Transfer Options 
- Characteristics of how to treat files from datasource and sink during job. If the option delete_objects_unique_in_sinkis true, object conditions based on objects'last_modification_timeare ignored and do not exclude objects in a data source or a data sink. Structure documented below.
- gcsData TransferSink Job Replication Spec Gcs Data Sink 
- A Google Cloud Storage data sink. Structure documented below.
- gcsData TransferSource Job Replication Spec Gcs Data Source 
- A Google Cloud Storage data source. Structure documented below.
- objectConditions TransferJob Replication Spec Object Conditions 
- Only objects that satisfy these object conditions are included in the set of data source and data sink objects. Object conditions based on objects' last_modification_timedo not exclude objects in a data sink. Structure documented below.
- transferOptions TransferJob Replication Spec Transfer Options 
- Characteristics of how to treat files from datasource and sink during job. If the option delete_objects_unique_in_sinkis true, object conditions based on objects'last_modification_timeare ignored and do not exclude objects in a data source or a data sink. Structure documented below.
- gcs_data_ Transfersink Job Replication Spec Gcs Data Sink 
- A Google Cloud Storage data sink. Structure documented below.
- gcs_data_ Transfersource Job Replication Spec Gcs Data Source 
- A Google Cloud Storage data source. Structure documented below.
- object_conditions TransferJob Replication Spec Object Conditions 
- Only objects that satisfy these object conditions are included in the set of data source and data sink objects. Object conditions based on objects' last_modification_timedo not exclude objects in a data sink. Structure documented below.
- transfer_options TransferJob Replication Spec Transfer Options 
- Characteristics of how to treat files from datasource and sink during job. If the option delete_objects_unique_in_sinkis true, object conditions based on objects'last_modification_timeare ignored and do not exclude objects in a data source or a data sink. Structure documented below.
- gcsData Property MapSink 
- A Google Cloud Storage data sink. Structure documented below.
- gcsData Property MapSource 
- A Google Cloud Storage data source. Structure documented below.
- objectConditions Property Map
- Only objects that satisfy these object conditions are included in the set of data source and data sink objects. Object conditions based on objects' last_modification_timedo not exclude objects in a data sink. Structure documented below.
- transferOptions Property Map
- Characteristics of how to treat files from datasource and sink during job. If the option delete_objects_unique_in_sinkis true, object conditions based on objects'last_modification_timeare ignored and do not exclude objects in a data source or a data sink. Structure documented below.
TransferJobReplicationSpecGcsDataSink, TransferJobReplicationSpecGcsDataSinkArgs              
- BucketName string
- Google Cloud Storage bucket name.
- Path string
- Root path to transfer objects. Must be an empty string or full path name that ends with a '/'. This field is treated as an object prefix. As such, it should generally not begin with a '/'.
- BucketName string
- Google Cloud Storage bucket name.
- Path string
- Root path to transfer objects. Must be an empty string or full path name that ends with a '/'. This field is treated as an object prefix. As such, it should generally not begin with a '/'.
- bucketName String
- Google Cloud Storage bucket name.
- path String
- Root path to transfer objects. Must be an empty string or full path name that ends with a '/'. This field is treated as an object prefix. As such, it should generally not begin with a '/'.
- bucketName string
- Google Cloud Storage bucket name.
- path string
- Root path to transfer objects. Must be an empty string or full path name that ends with a '/'. This field is treated as an object prefix. As such, it should generally not begin with a '/'.
- bucket_name str
- Google Cloud Storage bucket name.
- path str
- Root path to transfer objects. Must be an empty string or full path name that ends with a '/'. This field is treated as an object prefix. As such, it should generally not begin with a '/'.
- bucketName String
- Google Cloud Storage bucket name.
- path String
- Root path to transfer objects. Must be an empty string or full path name that ends with a '/'. This field is treated as an object prefix. As such, it should generally not begin with a '/'.
TransferJobReplicationSpecGcsDataSource, TransferJobReplicationSpecGcsDataSourceArgs              
- BucketName string
- Google Cloud Storage bucket name.
- Path string
- Root path to transfer objects. Must be an empty string or full path name that ends with a '/'. This field is treated as an object prefix. As such, it should generally not begin with a '/'.
- BucketName string
- Google Cloud Storage bucket name.
- Path string
- Root path to transfer objects. Must be an empty string or full path name that ends with a '/'. This field is treated as an object prefix. As such, it should generally not begin with a '/'.
- bucketName String
- Google Cloud Storage bucket name.
- path String
- Root path to transfer objects. Must be an empty string or full path name that ends with a '/'. This field is treated as an object prefix. As such, it should generally not begin with a '/'.
- bucketName string
- Google Cloud Storage bucket name.
- path string
- Root path to transfer objects. Must be an empty string or full path name that ends with a '/'. This field is treated as an object prefix. As such, it should generally not begin with a '/'.
- bucket_name str
- Google Cloud Storage bucket name.
- path str
- Root path to transfer objects. Must be an empty string or full path name that ends with a '/'. This field is treated as an object prefix. As such, it should generally not begin with a '/'.
- bucketName String
- Google Cloud Storage bucket name.
- path String
- Root path to transfer objects. Must be an empty string or full path name that ends with a '/'. This field is treated as an object prefix. As such, it should generally not begin with a '/'.
TransferJobReplicationSpecObjectConditions, TransferJobReplicationSpecObjectConditionsArgs            
- ExcludePrefixes List<string>
- exclude_prefixesmust follow the requirements described for- include_prefixes. See Requirements.
- IncludePrefixes List<string>
- If include_prefixesis specified, objects that satisfy the object conditions must have names that start with one of theinclude_prefixesand that do not start with any of theexclude_prefixes. Ifinclude_prefixesis not specified, all objects except those that have names starting with one of theexclude_prefixesmust satisfy the object conditions. See Requirements.
- LastModified stringBefore 
- If specified, only objects with a "last modification time" before this timestamp and objects that don't have a "last modification time" are transferred. A timestamp in RFC3339 UTC "Zulu" format, with nanosecond resolution and up to nine fractional digits. Examples: "2014-10-02T15:01:23Z" and "2014-10-02T15:01:23.045123456Z".
- LastModified stringSince 
- If specified, only objects with a "last modification time" on or after this timestamp and objects that don't have a "last modification time" are transferred. A timestamp in RFC3339 UTC "Zulu" format, with nanosecond resolution and up to nine fractional digits. Examples: "2014-10-02T15:01:23Z" and "2014-10-02T15:01:23.045123456Z".
- MaxTime stringElapsed Since Last Modification 
- A duration in seconds with up to nine fractional digits, terminated by 's'. Example: "3.5s".
- MinTime stringElapsed Since Last Modification 
- A duration in seconds with up to nine fractional digits, terminated by 's'. Example: "3.5s".
- ExcludePrefixes []string
- exclude_prefixesmust follow the requirements described for- include_prefixes. See Requirements.
- IncludePrefixes []string
- If include_prefixesis specified, objects that satisfy the object conditions must have names that start with one of theinclude_prefixesand that do not start with any of theexclude_prefixes. Ifinclude_prefixesis not specified, all objects except those that have names starting with one of theexclude_prefixesmust satisfy the object conditions. See Requirements.
- LastModified stringBefore 
- If specified, only objects with a "last modification time" before this timestamp and objects that don't have a "last modification time" are transferred. A timestamp in RFC3339 UTC "Zulu" format, with nanosecond resolution and up to nine fractional digits. Examples: "2014-10-02T15:01:23Z" and "2014-10-02T15:01:23.045123456Z".
- LastModified stringSince 
- If specified, only objects with a "last modification time" on or after this timestamp and objects that don't have a "last modification time" are transferred. A timestamp in RFC3339 UTC "Zulu" format, with nanosecond resolution and up to nine fractional digits. Examples: "2014-10-02T15:01:23Z" and "2014-10-02T15:01:23.045123456Z".
- MaxTime stringElapsed Since Last Modification 
- A duration in seconds with up to nine fractional digits, terminated by 's'. Example: "3.5s".
- MinTime stringElapsed Since Last Modification 
- A duration in seconds with up to nine fractional digits, terminated by 's'. Example: "3.5s".
- excludePrefixes List<String>
- exclude_prefixesmust follow the requirements described for- include_prefixes. See Requirements.
- includePrefixes List<String>
- If include_prefixesis specified, objects that satisfy the object conditions must have names that start with one of theinclude_prefixesand that do not start with any of theexclude_prefixes. Ifinclude_prefixesis not specified, all objects except those that have names starting with one of theexclude_prefixesmust satisfy the object conditions. See Requirements.
- lastModified StringBefore 
- If specified, only objects with a "last modification time" before this timestamp and objects that don't have a "last modification time" are transferred. A timestamp in RFC3339 UTC "Zulu" format, with nanosecond resolution and up to nine fractional digits. Examples: "2014-10-02T15:01:23Z" and "2014-10-02T15:01:23.045123456Z".
- lastModified StringSince 
- If specified, only objects with a "last modification time" on or after this timestamp and objects that don't have a "last modification time" are transferred. A timestamp in RFC3339 UTC "Zulu" format, with nanosecond resolution and up to nine fractional digits. Examples: "2014-10-02T15:01:23Z" and "2014-10-02T15:01:23.045123456Z".
- maxTime StringElapsed Since Last Modification 
- A duration in seconds with up to nine fractional digits, terminated by 's'. Example: "3.5s".
- minTime StringElapsed Since Last Modification 
- A duration in seconds with up to nine fractional digits, terminated by 's'. Example: "3.5s".
- excludePrefixes string[]
- exclude_prefixesmust follow the requirements described for- include_prefixes. See Requirements.
- includePrefixes string[]
- If include_prefixesis specified, objects that satisfy the object conditions must have names that start with one of theinclude_prefixesand that do not start with any of theexclude_prefixes. Ifinclude_prefixesis not specified, all objects except those that have names starting with one of theexclude_prefixesmust satisfy the object conditions. See Requirements.
- lastModified stringBefore 
- If specified, only objects with a "last modification time" before this timestamp and objects that don't have a "last modification time" are transferred. A timestamp in RFC3339 UTC "Zulu" format, with nanosecond resolution and up to nine fractional digits. Examples: "2014-10-02T15:01:23Z" and "2014-10-02T15:01:23.045123456Z".
- lastModified stringSince 
- If specified, only objects with a "last modification time" on or after this timestamp and objects that don't have a "last modification time" are transferred. A timestamp in RFC3339 UTC "Zulu" format, with nanosecond resolution and up to nine fractional digits. Examples: "2014-10-02T15:01:23Z" and "2014-10-02T15:01:23.045123456Z".
- maxTime stringElapsed Since Last Modification 
- A duration in seconds with up to nine fractional digits, terminated by 's'. Example: "3.5s".
- minTime stringElapsed Since Last Modification 
- A duration in seconds with up to nine fractional digits, terminated by 's'. Example: "3.5s".
- exclude_prefixes Sequence[str]
- exclude_prefixesmust follow the requirements described for- include_prefixes. See Requirements.
- include_prefixes Sequence[str]
- If include_prefixesis specified, objects that satisfy the object conditions must have names that start with one of theinclude_prefixesand that do not start with any of theexclude_prefixes. Ifinclude_prefixesis not specified, all objects except those that have names starting with one of theexclude_prefixesmust satisfy the object conditions. See Requirements.
- last_modified_ strbefore 
- If specified, only objects with a "last modification time" before this timestamp and objects that don't have a "last modification time" are transferred. A timestamp in RFC3339 UTC "Zulu" format, with nanosecond resolution and up to nine fractional digits. Examples: "2014-10-02T15:01:23Z" and "2014-10-02T15:01:23.045123456Z".
- last_modified_ strsince 
- If specified, only objects with a "last modification time" on or after this timestamp and objects that don't have a "last modification time" are transferred. A timestamp in RFC3339 UTC "Zulu" format, with nanosecond resolution and up to nine fractional digits. Examples: "2014-10-02T15:01:23Z" and "2014-10-02T15:01:23.045123456Z".
- max_time_ strelapsed_ since_ last_ modification 
- A duration in seconds with up to nine fractional digits, terminated by 's'. Example: "3.5s".
- min_time_ strelapsed_ since_ last_ modification 
- A duration in seconds with up to nine fractional digits, terminated by 's'. Example: "3.5s".
- excludePrefixes List<String>
- exclude_prefixesmust follow the requirements described for- include_prefixes. See Requirements.
- includePrefixes List<String>
- If include_prefixesis specified, objects that satisfy the object conditions must have names that start with one of theinclude_prefixesand that do not start with any of theexclude_prefixes. Ifinclude_prefixesis not specified, all objects except those that have names starting with one of theexclude_prefixesmust satisfy the object conditions. See Requirements.
- lastModified StringBefore 
- If specified, only objects with a "last modification time" before this timestamp and objects that don't have a "last modification time" are transferred. A timestamp in RFC3339 UTC "Zulu" format, with nanosecond resolution and up to nine fractional digits. Examples: "2014-10-02T15:01:23Z" and "2014-10-02T15:01:23.045123456Z".
- lastModified StringSince 
- If specified, only objects with a "last modification time" on or after this timestamp and objects that don't have a "last modification time" are transferred. A timestamp in RFC3339 UTC "Zulu" format, with nanosecond resolution and up to nine fractional digits. Examples: "2014-10-02T15:01:23Z" and "2014-10-02T15:01:23.045123456Z".
- maxTime StringElapsed Since Last Modification 
- A duration in seconds with up to nine fractional digits, terminated by 's'. Example: "3.5s".
- minTime StringElapsed Since Last Modification 
- A duration in seconds with up to nine fractional digits, terminated by 's'. Example: "3.5s".
TransferJobReplicationSpecTransferOptions, TransferJobReplicationSpecTransferOptionsArgs            
- DeleteObjects boolFrom Source After Transfer 
- Whether objects should be deleted from the source after they are transferred to the sink. Note that this option and delete_objects_unique_in_sinkare mutually exclusive.
- DeleteObjects boolUnique In Sink 
- Whether objects that exist only in the sink should be deleted. Note that this option and
delete_objects_from_source_after_transferare mutually exclusive.
- OverwriteObjects boolAlready Existing In Sink 
- Whether overwriting objects that already exist in the sink is allowed.
- OverwriteWhen string
- When to overwrite objects that already exist in the sink. If not set, overwrite behavior is determined by overwrite_objects_already_existing_in_sink. Possible values: ALWAYS, DIFFERENT, NEVER.
- DeleteObjects boolFrom Source After Transfer 
- Whether objects should be deleted from the source after they are transferred to the sink. Note that this option and delete_objects_unique_in_sinkare mutually exclusive.
- DeleteObjects boolUnique In Sink 
- Whether objects that exist only in the sink should be deleted. Note that this option and
delete_objects_from_source_after_transferare mutually exclusive.
- OverwriteObjects boolAlready Existing In Sink 
- Whether overwriting objects that already exist in the sink is allowed.
- OverwriteWhen string
- When to overwrite objects that already exist in the sink. If not set, overwrite behavior is determined by overwrite_objects_already_existing_in_sink. Possible values: ALWAYS, DIFFERENT, NEVER.
- deleteObjects BooleanFrom Source After Transfer 
- Whether objects should be deleted from the source after they are transferred to the sink. Note that this option and delete_objects_unique_in_sinkare mutually exclusive.
- deleteObjects BooleanUnique In Sink 
- Whether objects that exist only in the sink should be deleted. Note that this option and
delete_objects_from_source_after_transferare mutually exclusive.
- overwriteObjects BooleanAlready Existing In Sink 
- Whether overwriting objects that already exist in the sink is allowed.
- overwriteWhen String
- When to overwrite objects that already exist in the sink. If not set, overwrite behavior is determined by overwrite_objects_already_existing_in_sink. Possible values: ALWAYS, DIFFERENT, NEVER.
- deleteObjects booleanFrom Source After Transfer 
- Whether objects should be deleted from the source after they are transferred to the sink. Note that this option and delete_objects_unique_in_sinkare mutually exclusive.
- deleteObjects booleanUnique In Sink 
- Whether objects that exist only in the sink should be deleted. Note that this option and
delete_objects_from_source_after_transferare mutually exclusive.
- overwriteObjects booleanAlready Existing In Sink 
- Whether overwriting objects that already exist in the sink is allowed.
- overwriteWhen string
- When to overwrite objects that already exist in the sink. If not set, overwrite behavior is determined by overwrite_objects_already_existing_in_sink. Possible values: ALWAYS, DIFFERENT, NEVER.
- delete_objects_ boolfrom_ source_ after_ transfer 
- Whether objects should be deleted from the source after they are transferred to the sink. Note that this option and delete_objects_unique_in_sinkare mutually exclusive.
- delete_objects_ boolunique_ in_ sink 
- Whether objects that exist only in the sink should be deleted. Note that this option and
delete_objects_from_source_after_transferare mutually exclusive.
- overwrite_objects_ boolalready_ existing_ in_ sink 
- Whether overwriting objects that already exist in the sink is allowed.
- overwrite_when str
- When to overwrite objects that already exist in the sink. If not set, overwrite behavior is determined by overwrite_objects_already_existing_in_sink. Possible values: ALWAYS, DIFFERENT, NEVER.
- deleteObjects BooleanFrom Source After Transfer 
- Whether objects should be deleted from the source after they are transferred to the sink. Note that this option and delete_objects_unique_in_sinkare mutually exclusive.
- deleteObjects BooleanUnique In Sink 
- Whether objects that exist only in the sink should be deleted. Note that this option and
delete_objects_from_source_after_transferare mutually exclusive.
- overwriteObjects BooleanAlready Existing In Sink 
- Whether overwriting objects that already exist in the sink is allowed.
- overwriteWhen String
- When to overwrite objects that already exist in the sink. If not set, overwrite behavior is determined by overwrite_objects_already_existing_in_sink. Possible values: ALWAYS, DIFFERENT, NEVER.
TransferJobSchedule, TransferJobScheduleArgs      
- ScheduleStart TransferDate Job Schedule Schedule Start Date 
- The first day the recurring transfer is scheduled to run. If schedule_start_dateis in the past, the transfer will run for the first time on the following day. Structure documented below.
- RepeatInterval string
- Interval between the start of each scheduled transfer. If unspecified, the default value is 24 hours. This value may not be less than 1 hour. A duration in seconds with up to nine fractional digits, terminated by 's'. Example: "3.5s".
- ScheduleEnd TransferDate Job Schedule Schedule End Date 
- The last day the recurring transfer will be run. If schedule_end_dateis the same asschedule_start_date, the transfer will be executed only once. Structure documented below.
- StartTime TransferOf Day Job Schedule Start Time Of Day 
- The time in UTC at which the transfer will be scheduled to start in a day. Transfers may start later than this time. If not specified, recurring and one-time transfers that are scheduled to run today will run immediately; recurring transfers that are scheduled to run on a future date will start at approximately midnight UTC on that date. Note that when configuring a transfer with the Cloud Platform Console, the transfer's start time in a day is specified in your local timezone. Structure documented below.
- ScheduleStart TransferDate Job Schedule Schedule Start Date 
- The first day the recurring transfer is scheduled to run. If schedule_start_dateis in the past, the transfer will run for the first time on the following day. Structure documented below.
- RepeatInterval string
- Interval between the start of each scheduled transfer. If unspecified, the default value is 24 hours. This value may not be less than 1 hour. A duration in seconds with up to nine fractional digits, terminated by 's'. Example: "3.5s".
- ScheduleEnd TransferDate Job Schedule Schedule End Date 
- The last day the recurring transfer will be run. If schedule_end_dateis the same asschedule_start_date, the transfer will be executed only once. Structure documented below.
- StartTime TransferOf Day Job Schedule Start Time Of Day 
- The time in UTC at which the transfer will be scheduled to start in a day. Transfers may start later than this time. If not specified, recurring and one-time transfers that are scheduled to run today will run immediately; recurring transfers that are scheduled to run on a future date will start at approximately midnight UTC on that date. Note that when configuring a transfer with the Cloud Platform Console, the transfer's start time in a day is specified in your local timezone. Structure documented below.
- scheduleStart TransferDate Job Schedule Schedule Start Date 
- The first day the recurring transfer is scheduled to run. If schedule_start_dateis in the past, the transfer will run for the first time on the following day. Structure documented below.
- repeatInterval String
- Interval between the start of each scheduled transfer. If unspecified, the default value is 24 hours. This value may not be less than 1 hour. A duration in seconds with up to nine fractional digits, terminated by 's'. Example: "3.5s".
- scheduleEnd TransferDate Job Schedule Schedule End Date 
- The last day the recurring transfer will be run. If schedule_end_dateis the same asschedule_start_date, the transfer will be executed only once. Structure documented below.
- startTime TransferOf Day Job Schedule Start Time Of Day 
- The time in UTC at which the transfer will be scheduled to start in a day. Transfers may start later than this time. If not specified, recurring and one-time transfers that are scheduled to run today will run immediately; recurring transfers that are scheduled to run on a future date will start at approximately midnight UTC on that date. Note that when configuring a transfer with the Cloud Platform Console, the transfer's start time in a day is specified in your local timezone. Structure documented below.
- scheduleStart TransferDate Job Schedule Schedule Start Date 
- The first day the recurring transfer is scheduled to run. If schedule_start_dateis in the past, the transfer will run for the first time on the following day. Structure documented below.
- repeatInterval string
- Interval between the start of each scheduled transfer. If unspecified, the default value is 24 hours. This value may not be less than 1 hour. A duration in seconds with up to nine fractional digits, terminated by 's'. Example: "3.5s".
- scheduleEnd TransferDate Job Schedule Schedule End Date 
- The last day the recurring transfer will be run. If schedule_end_dateis the same asschedule_start_date, the transfer will be executed only once. Structure documented below.
- startTime TransferOf Day Job Schedule Start Time Of Day 
- The time in UTC at which the transfer will be scheduled to start in a day. Transfers may start later than this time. If not specified, recurring and one-time transfers that are scheduled to run today will run immediately; recurring transfers that are scheduled to run on a future date will start at approximately midnight UTC on that date. Note that when configuring a transfer with the Cloud Platform Console, the transfer's start time in a day is specified in your local timezone. Structure documented below.
- schedule_start_ Transferdate Job Schedule Schedule Start Date 
- The first day the recurring transfer is scheduled to run. If schedule_start_dateis in the past, the transfer will run for the first time on the following day. Structure documented below.
- repeat_interval str
- Interval between the start of each scheduled transfer. If unspecified, the default value is 24 hours. This value may not be less than 1 hour. A duration in seconds with up to nine fractional digits, terminated by 's'. Example: "3.5s".
- schedule_end_ Transferdate Job Schedule Schedule End Date 
- The last day the recurring transfer will be run. If schedule_end_dateis the same asschedule_start_date, the transfer will be executed only once. Structure documented below.
- start_time_ Transferof_ day Job Schedule Start Time Of Day 
- The time in UTC at which the transfer will be scheduled to start in a day. Transfers may start later than this time. If not specified, recurring and one-time transfers that are scheduled to run today will run immediately; recurring transfers that are scheduled to run on a future date will start at approximately midnight UTC on that date. Note that when configuring a transfer with the Cloud Platform Console, the transfer's start time in a day is specified in your local timezone. Structure documented below.
- scheduleStart Property MapDate 
- The first day the recurring transfer is scheduled to run. If schedule_start_dateis in the past, the transfer will run for the first time on the following day. Structure documented below.
- repeatInterval String
- Interval between the start of each scheduled transfer. If unspecified, the default value is 24 hours. This value may not be less than 1 hour. A duration in seconds with up to nine fractional digits, terminated by 's'. Example: "3.5s".
- scheduleEnd Property MapDate 
- The last day the recurring transfer will be run. If schedule_end_dateis the same asschedule_start_date, the transfer will be executed only once. Structure documented below.
- startTime Property MapOf Day 
- The time in UTC at which the transfer will be scheduled to start in a day. Transfers may start later than this time. If not specified, recurring and one-time transfers that are scheduled to run today will run immediately; recurring transfers that are scheduled to run on a future date will start at approximately midnight UTC on that date. Note that when configuring a transfer with the Cloud Platform Console, the transfer's start time in a day is specified in your local timezone. Structure documented below.
TransferJobScheduleScheduleEndDate, TransferJobScheduleScheduleEndDateArgs            
TransferJobScheduleScheduleStartDate, TransferJobScheduleScheduleStartDateArgs            
TransferJobScheduleStartTimeOfDay, TransferJobScheduleStartTimeOfDayArgs              
TransferJobTransferSpec, TransferJobTransferSpecArgs        
- AwsS3Data TransferSource Job Transfer Spec Aws S3Data Source 
- An AWS S3 data source. Structure documented below.
- AzureBlob TransferStorage Data Source Job Transfer Spec Azure Blob Storage Data Source 
- An Azure Blob Storage data source. Structure documented below.
- GcsData TransferSink Job Transfer Spec Gcs Data Sink 
- A Google Cloud Storage data sink. Structure documented below.
- GcsData TransferSource Job Transfer Spec Gcs Data Source 
- A Google Cloud Storage data source. Structure documented below.
- HdfsData TransferSource Job Transfer Spec Hdfs Data Source 
- An HDFS data source. Structure documented below.
- HttpData TransferSource Job Transfer Spec Http Data Source 
- A HTTP URL data source. Structure documented below.
- ObjectConditions TransferJob Transfer Spec Object Conditions 
- Only objects that satisfy these object conditions are included in the set of data source and data sink objects. Object conditions based on objects' last_modification_timedo not exclude objects in a data sink. Structure documented below.
- PosixData TransferSink Job Transfer Spec Posix Data Sink 
- A POSIX data sink. Structure documented below.
- PosixData TransferSource Job Transfer Spec Posix Data Source 
- A POSIX filesystem data source. Structure documented below.
- SinkAgent stringPool Name 
- Specifies the agent pool name associated with the posix data sink. When unspecified, the default name is used.
- SourceAgent stringPool Name 
- Specifies the agent pool name associated with the posix data source. When unspecified, the default name is used.
- TransferOptions TransferJob Transfer Spec Transfer Options 
- Characteristics of how to treat files from datasource and sink during job. If the option delete_objects_unique_in_sinkis true, object conditions based on objects'last_modification_timeare ignored and do not exclude objects in a data source or a data sink. Structure documented below.
- AwsS3Data TransferSource Job Transfer Spec Aws S3Data Source 
- An AWS S3 data source. Structure documented below.
- AzureBlob TransferStorage Data Source Job Transfer Spec Azure Blob Storage Data Source 
- An Azure Blob Storage data source. Structure documented below.
- GcsData TransferSink Job Transfer Spec Gcs Data Sink 
- A Google Cloud Storage data sink. Structure documented below.
- GcsData TransferSource Job Transfer Spec Gcs Data Source 
- A Google Cloud Storage data source. Structure documented below.
- HdfsData TransferSource Job Transfer Spec Hdfs Data Source 
- An HDFS data source. Structure documented below.
- HttpData TransferSource Job Transfer Spec Http Data Source 
- A HTTP URL data source. Structure documented below.
- ObjectConditions TransferJob Transfer Spec Object Conditions 
- Only objects that satisfy these object conditions are included in the set of data source and data sink objects. Object conditions based on objects' last_modification_timedo not exclude objects in a data sink. Structure documented below.
- PosixData TransferSink Job Transfer Spec Posix Data Sink 
- A POSIX data sink. Structure documented below.
- PosixData TransferSource Job Transfer Spec Posix Data Source 
- A POSIX filesystem data source. Structure documented below.
- SinkAgent stringPool Name 
- Specifies the agent pool name associated with the posix data sink. When unspecified, the default name is used.
- SourceAgent stringPool Name 
- Specifies the agent pool name associated with the posix data source. When unspecified, the default name is used.
- TransferOptions TransferJob Transfer Spec Transfer Options 
- Characteristics of how to treat files from datasource and sink during job. If the option delete_objects_unique_in_sinkis true, object conditions based on objects'last_modification_timeare ignored and do not exclude objects in a data source or a data sink. Structure documented below.
- awsS3Data TransferSource Job Transfer Spec Aws S3Data Source 
- An AWS S3 data source. Structure documented below.
- azureBlob TransferStorage Data Source Job Transfer Spec Azure Blob Storage Data Source 
- An Azure Blob Storage data source. Structure documented below.
- gcsData TransferSink Job Transfer Spec Gcs Data Sink 
- A Google Cloud Storage data sink. Structure documented below.
- gcsData TransferSource Job Transfer Spec Gcs Data Source 
- A Google Cloud Storage data source. Structure documented below.
- hdfsData TransferSource Job Transfer Spec Hdfs Data Source 
- An HDFS data source. Structure documented below.
- httpData TransferSource Job Transfer Spec Http Data Source 
- A HTTP URL data source. Structure documented below.
- objectConditions TransferJob Transfer Spec Object Conditions 
- Only objects that satisfy these object conditions are included in the set of data source and data sink objects. Object conditions based on objects' last_modification_timedo not exclude objects in a data sink. Structure documented below.
- posixData TransferSink Job Transfer Spec Posix Data Sink 
- A POSIX data sink. Structure documented below.
- posixData TransferSource Job Transfer Spec Posix Data Source 
- A POSIX filesystem data source. Structure documented below.
- sinkAgent StringPool Name 
- Specifies the agent pool name associated with the posix data sink. When unspecified, the default name is used.
- sourceAgent StringPool Name 
- Specifies the agent pool name associated with the posix data source. When unspecified, the default name is used.
- transferOptions TransferJob Transfer Spec Transfer Options 
- Characteristics of how to treat files from datasource and sink during job. If the option delete_objects_unique_in_sinkis true, object conditions based on objects'last_modification_timeare ignored and do not exclude objects in a data source or a data sink. Structure documented below.
- awsS3Data TransferSource Job Transfer Spec Aws S3Data Source 
- An AWS S3 data source. Structure documented below.
- azureBlob TransferStorage Data Source Job Transfer Spec Azure Blob Storage Data Source 
- An Azure Blob Storage data source. Structure documented below.
- gcsData TransferSink Job Transfer Spec Gcs Data Sink 
- A Google Cloud Storage data sink. Structure documented below.
- gcsData TransferSource Job Transfer Spec Gcs Data Source 
- A Google Cloud Storage data source. Structure documented below.
- hdfsData TransferSource Job Transfer Spec Hdfs Data Source 
- An HDFS data source. Structure documented below.
- httpData TransferSource Job Transfer Spec Http Data Source 
- A HTTP URL data source. Structure documented below.
- objectConditions TransferJob Transfer Spec Object Conditions 
- Only objects that satisfy these object conditions are included in the set of data source and data sink objects. Object conditions based on objects' last_modification_timedo not exclude objects in a data sink. Structure documented below.
- posixData TransferSink Job Transfer Spec Posix Data Sink 
- A POSIX data sink. Structure documented below.
- posixData TransferSource Job Transfer Spec Posix Data Source 
- A POSIX filesystem data source. Structure documented below.
- sinkAgent stringPool Name 
- Specifies the agent pool name associated with the posix data sink. When unspecified, the default name is used.
- sourceAgent stringPool Name 
- Specifies the agent pool name associated with the posix data source. When unspecified, the default name is used.
- transferOptions TransferJob Transfer Spec Transfer Options 
- Characteristics of how to treat files from datasource and sink during job. If the option delete_objects_unique_in_sinkis true, object conditions based on objects'last_modification_timeare ignored and do not exclude objects in a data source or a data sink. Structure documented below.
- aws_s3_ Transferdata_ source Job Transfer Spec Aws S3Data Source 
- An AWS S3 data source. Structure documented below.
- azure_blob_ Transferstorage_ data_ source Job Transfer Spec Azure Blob Storage Data Source 
- An Azure Blob Storage data source. Structure documented below.
- gcs_data_ Transfersink Job Transfer Spec Gcs Data Sink 
- A Google Cloud Storage data sink. Structure documented below.
- gcs_data_ Transfersource Job Transfer Spec Gcs Data Source 
- A Google Cloud Storage data source. Structure documented below.
- hdfs_data_ Transfersource Job Transfer Spec Hdfs Data Source 
- An HDFS data source. Structure documented below.
- http_data_ Transfersource Job Transfer Spec Http Data Source 
- A HTTP URL data source. Structure documented below.
- object_conditions TransferJob Transfer Spec Object Conditions 
- Only objects that satisfy these object conditions are included in the set of data source and data sink objects. Object conditions based on objects' last_modification_timedo not exclude objects in a data sink. Structure documented below.
- posix_data_ Transfersink Job Transfer Spec Posix Data Sink 
- A POSIX data sink. Structure documented below.
- posix_data_ Transfersource Job Transfer Spec Posix Data Source 
- A POSIX filesystem data source. Structure documented below.
- sink_agent_ strpool_ name 
- Specifies the agent pool name associated with the posix data sink. When unspecified, the default name is used.
- source_agent_ strpool_ name 
- Specifies the agent pool name associated with the posix data source. When unspecified, the default name is used.
- transfer_options TransferJob Transfer Spec Transfer Options 
- Characteristics of how to treat files from datasource and sink during job. If the option delete_objects_unique_in_sinkis true, object conditions based on objects'last_modification_timeare ignored and do not exclude objects in a data source or a data sink. Structure documented below.
- awsS3Data Property MapSource 
- An AWS S3 data source. Structure documented below.
- azureBlob Property MapStorage Data Source 
- An Azure Blob Storage data source. Structure documented below.
- gcsData Property MapSink 
- A Google Cloud Storage data sink. Structure documented below.
- gcsData Property MapSource 
- A Google Cloud Storage data source. Structure documented below.
- hdfsData Property MapSource 
- An HDFS data source. Structure documented below.
- httpData Property MapSource 
- A HTTP URL data source. Structure documented below.
- objectConditions Property Map
- Only objects that satisfy these object conditions are included in the set of data source and data sink objects. Object conditions based on objects' last_modification_timedo not exclude objects in a data sink. Structure documented below.
- posixData Property MapSink 
- A POSIX data sink. Structure documented below.
- posixData Property MapSource 
- A POSIX filesystem data source. Structure documented below.
- sinkAgent StringPool Name 
- Specifies the agent pool name associated with the posix data sink. When unspecified, the default name is used.
- sourceAgent StringPool Name 
- Specifies the agent pool name associated with the posix data source. When unspecified, the default name is used.
- transferOptions Property Map
- Characteristics of how to treat files from datasource and sink during job. If the option delete_objects_unique_in_sinkis true, object conditions based on objects'last_modification_timeare ignored and do not exclude objects in a data source or a data sink. Structure documented below.
TransferJobTransferSpecAwsS3DataSource, TransferJobTransferSpecAwsS3DataSourceArgs              
- BucketName string
- S3 Bucket name.
- AwsAccess TransferKey Job Transfer Spec Aws S3Data Source Aws Access Key 
- AWS credentials block.
- Path string
- S3 Bucket path in bucket to transfer.
- RoleArn string
- The Amazon Resource Name (ARN) of the role to support temporary credentials via 'AssumeRoleWithWebIdentity'. For more information about ARNs, see IAM ARNs. When a role ARN is provided, Transfer Service fetches temporary credentials for the session using a 'AssumeRoleWithWebIdentity' call for the provided role using the [GoogleServiceAccount][] for this project.
- BucketName string
- S3 Bucket name.
- AwsAccess TransferKey Job Transfer Spec Aws S3Data Source Aws Access Key 
- AWS credentials block.
- Path string
- S3 Bucket path in bucket to transfer.
- RoleArn string
- The Amazon Resource Name (ARN) of the role to support temporary credentials via 'AssumeRoleWithWebIdentity'. For more information about ARNs, see IAM ARNs. When a role ARN is provided, Transfer Service fetches temporary credentials for the session using a 'AssumeRoleWithWebIdentity' call for the provided role using the [GoogleServiceAccount][] for this project.
- bucketName String
- S3 Bucket name.
- awsAccess TransferKey Job Transfer Spec Aws S3Data Source Aws Access Key 
- AWS credentials block.
- path String
- S3 Bucket path in bucket to transfer.
- roleArn String
- The Amazon Resource Name (ARN) of the role to support temporary credentials via 'AssumeRoleWithWebIdentity'. For more information about ARNs, see IAM ARNs. When a role ARN is provided, Transfer Service fetches temporary credentials for the session using a 'AssumeRoleWithWebIdentity' call for the provided role using the [GoogleServiceAccount][] for this project.
- bucketName string
- S3 Bucket name.
- awsAccess TransferKey Job Transfer Spec Aws S3Data Source Aws Access Key 
- AWS credentials block.
- path string
- S3 Bucket path in bucket to transfer.
- roleArn string
- The Amazon Resource Name (ARN) of the role to support temporary credentials via 'AssumeRoleWithWebIdentity'. For more information about ARNs, see IAM ARNs. When a role ARN is provided, Transfer Service fetches temporary credentials for the session using a 'AssumeRoleWithWebIdentity' call for the provided role using the [GoogleServiceAccount][] for this project.
- bucket_name str
- S3 Bucket name.
- aws_access_ Transferkey Job Transfer Spec Aws S3Data Source Aws Access Key 
- AWS credentials block.
- path str
- S3 Bucket path in bucket to transfer.
- role_arn str
- The Amazon Resource Name (ARN) of the role to support temporary credentials via 'AssumeRoleWithWebIdentity'. For more information about ARNs, see IAM ARNs. When a role ARN is provided, Transfer Service fetches temporary credentials for the session using a 'AssumeRoleWithWebIdentity' call for the provided role using the [GoogleServiceAccount][] for this project.
- bucketName String
- S3 Bucket name.
- awsAccess Property MapKey 
- AWS credentials block.
- path String
- S3 Bucket path in bucket to transfer.
- roleArn String
- The Amazon Resource Name (ARN) of the role to support temporary credentials via 'AssumeRoleWithWebIdentity'. For more information about ARNs, see IAM ARNs. When a role ARN is provided, Transfer Service fetches temporary credentials for the session using a 'AssumeRoleWithWebIdentity' call for the provided role using the [GoogleServiceAccount][] for this project.
TransferJobTransferSpecAwsS3DataSourceAwsAccessKey, TransferJobTransferSpecAwsS3DataSourceAwsAccessKeyArgs                    
- AccessKey stringId 
- AWS Key ID.
- SecretAccess stringKey 
- AWS Secret Access Key.
- AccessKey stringId 
- AWS Key ID.
- SecretAccess stringKey 
- AWS Secret Access Key.
- accessKey StringId 
- AWS Key ID.
- secretAccess StringKey 
- AWS Secret Access Key.
- accessKey stringId 
- AWS Key ID.
- secretAccess stringKey 
- AWS Secret Access Key.
- access_key_ strid 
- AWS Key ID.
- secret_access_ strkey 
- AWS Secret Access Key.
- accessKey StringId 
- AWS Key ID.
- secretAccess StringKey 
- AWS Secret Access Key.
TransferJobTransferSpecAzureBlobStorageDataSource, TransferJobTransferSpecAzureBlobStorageDataSourceArgs                  
- Container string
- The container to transfer from the Azure Storage account.`
- StorageAccount string
- The name of the Azure Storage account.
- AzureCredentials TransferJob Transfer Spec Azure Blob Storage Data Source Azure Credentials 
- Credentials used to authenticate API requests to Azure block.
- CredentialsSecret string
- Full Resource name of a secret in Secret Manager containing SAS Credentials in JSON form. Service Agent for Storage Transfer must have permissions to access secret. If credentials_secret is specified, do not specify azure_credentials.`,
- Path string
- Root path to transfer objects. Must be an empty string or full path name that ends with a '/'. This field is treated as an object prefix. As such, it should generally not begin with a '/'.
- Container string
- The container to transfer from the Azure Storage account.`
- StorageAccount string
- The name of the Azure Storage account.
- AzureCredentials TransferJob Transfer Spec Azure Blob Storage Data Source Azure Credentials 
- Credentials used to authenticate API requests to Azure block.
- CredentialsSecret string
- Full Resource name of a secret in Secret Manager containing SAS Credentials in JSON form. Service Agent for Storage Transfer must have permissions to access secret. If credentials_secret is specified, do not specify azure_credentials.`,
- Path string
- Root path to transfer objects. Must be an empty string or full path name that ends with a '/'. This field is treated as an object prefix. As such, it should generally not begin with a '/'.
- container String
- The container to transfer from the Azure Storage account.`
- storageAccount String
- The name of the Azure Storage account.
- azureCredentials TransferJob Transfer Spec Azure Blob Storage Data Source Azure Credentials 
- Credentials used to authenticate API requests to Azure block.
- credentialsSecret String
- Full Resource name of a secret in Secret Manager containing SAS Credentials in JSON form. Service Agent for Storage Transfer must have permissions to access secret. If credentials_secret is specified, do not specify azure_credentials.`,
- path String
- Root path to transfer objects. Must be an empty string or full path name that ends with a '/'. This field is treated as an object prefix. As such, it should generally not begin with a '/'.
- container string
- The container to transfer from the Azure Storage account.`
- storageAccount string
- The name of the Azure Storage account.
- azureCredentials TransferJob Transfer Spec Azure Blob Storage Data Source Azure Credentials 
- Credentials used to authenticate API requests to Azure block.
- credentialsSecret string
- Full Resource name of a secret in Secret Manager containing SAS Credentials in JSON form. Service Agent for Storage Transfer must have permissions to access secret. If credentials_secret is specified, do not specify azure_credentials.`,
- path string
- Root path to transfer objects. Must be an empty string or full path name that ends with a '/'. This field is treated as an object prefix. As such, it should generally not begin with a '/'.
- container str
- The container to transfer from the Azure Storage account.`
- storage_account str
- The name of the Azure Storage account.
- azure_credentials TransferJob Transfer Spec Azure Blob Storage Data Source Azure Credentials 
- Credentials used to authenticate API requests to Azure block.
- credentials_secret str
- Full Resource name of a secret in Secret Manager containing SAS Credentials in JSON form. Service Agent for Storage Transfer must have permissions to access secret. If credentials_secret is specified, do not specify azure_credentials.`,
- path str
- Root path to transfer objects. Must be an empty string or full path name that ends with a '/'. This field is treated as an object prefix. As such, it should generally not begin with a '/'.
- container String
- The container to transfer from the Azure Storage account.`
- storageAccount String
- The name of the Azure Storage account.
- azureCredentials Property Map
- Credentials used to authenticate API requests to Azure block.
- credentialsSecret String
- Full Resource name of a secret in Secret Manager containing SAS Credentials in JSON form. Service Agent for Storage Transfer must have permissions to access secret. If credentials_secret is specified, do not specify azure_credentials.`,
- path String
- Root path to transfer objects. Must be an empty string or full path name that ends with a '/'. This field is treated as an object prefix. As such, it should generally not begin with a '/'.
TransferJobTransferSpecAzureBlobStorageDataSourceAzureCredentials, TransferJobTransferSpecAzureBlobStorageDataSourceAzureCredentialsArgs                      
- SasToken string
- Azure shared access signature. See Grant limited access to Azure Storage resources using shared access signatures (SAS). - The - schedule_start_dateand- schedule_end_dateblocks support:
- SasToken string
- Azure shared access signature. See Grant limited access to Azure Storage resources using shared access signatures (SAS). - The - schedule_start_dateand- schedule_end_dateblocks support:
- sasToken String
- Azure shared access signature. See Grant limited access to Azure Storage resources using shared access signatures (SAS). - The - schedule_start_dateand- schedule_end_dateblocks support:
- sasToken string
- Azure shared access signature. See Grant limited access to Azure Storage resources using shared access signatures (SAS). - The - schedule_start_dateand- schedule_end_dateblocks support:
- sas_token str
- Azure shared access signature. See Grant limited access to Azure Storage resources using shared access signatures (SAS). - The - schedule_start_dateand- schedule_end_dateblocks support:
- sasToken String
- Azure shared access signature. See Grant limited access to Azure Storage resources using shared access signatures (SAS). - The - schedule_start_dateand- schedule_end_dateblocks support:
TransferJobTransferSpecGcsDataSink, TransferJobTransferSpecGcsDataSinkArgs              
- BucketName string
- Google Cloud Storage bucket name.
- Path string
- Root path to transfer objects. Must be an empty string or full path name that ends with a '/'. This field is treated as an object prefix. As such, it should generally not begin with a '/'.
- BucketName string
- Google Cloud Storage bucket name.
- Path string
- Root path to transfer objects. Must be an empty string or full path name that ends with a '/'. This field is treated as an object prefix. As such, it should generally not begin with a '/'.
- bucketName String
- Google Cloud Storage bucket name.
- path String
- Root path to transfer objects. Must be an empty string or full path name that ends with a '/'. This field is treated as an object prefix. As such, it should generally not begin with a '/'.
- bucketName string
- Google Cloud Storage bucket name.
- path string
- Root path to transfer objects. Must be an empty string or full path name that ends with a '/'. This field is treated as an object prefix. As such, it should generally not begin with a '/'.
- bucket_name str
- Google Cloud Storage bucket name.
- path str
- Root path to transfer objects. Must be an empty string or full path name that ends with a '/'. This field is treated as an object prefix. As such, it should generally not begin with a '/'.
- bucketName String
- Google Cloud Storage bucket name.
- path String
- Root path to transfer objects. Must be an empty string or full path name that ends with a '/'. This field is treated as an object prefix. As such, it should generally not begin with a '/'.
TransferJobTransferSpecGcsDataSource, TransferJobTransferSpecGcsDataSourceArgs              
- BucketName string
- Google Cloud Storage bucket name.
- Path string
- Root path to transfer objects. Must be an empty string or full path name that ends with a '/'. This field is treated as an object prefix. As such, it should generally not begin with a '/'.
- BucketName string
- Google Cloud Storage bucket name.
- Path string
- Root path to transfer objects. Must be an empty string or full path name that ends with a '/'. This field is treated as an object prefix. As such, it should generally not begin with a '/'.
- bucketName String
- Google Cloud Storage bucket name.
- path String
- Root path to transfer objects. Must be an empty string or full path name that ends with a '/'. This field is treated as an object prefix. As such, it should generally not begin with a '/'.
- bucketName string
- Google Cloud Storage bucket name.
- path string
- Root path to transfer objects. Must be an empty string or full path name that ends with a '/'. This field is treated as an object prefix. As such, it should generally not begin with a '/'.
- bucket_name str
- Google Cloud Storage bucket name.
- path str
- Root path to transfer objects. Must be an empty string or full path name that ends with a '/'. This field is treated as an object prefix. As such, it should generally not begin with a '/'.
- bucketName String
- Google Cloud Storage bucket name.
- path String
- Root path to transfer objects. Must be an empty string or full path name that ends with a '/'. This field is treated as an object prefix. As such, it should generally not begin with a '/'.
TransferJobTransferSpecHdfsDataSource, TransferJobTransferSpecHdfsDataSourceArgs              
- Path string
- Root path to transfer objects. Must be an empty string or full path name that ends with a '/'. This field is treated as an object prefix. As such, it should generally not begin with a '/'.
- Path string
- Root path to transfer objects. Must be an empty string or full path name that ends with a '/'. This field is treated as an object prefix. As such, it should generally not begin with a '/'.
- path String
- Root path to transfer objects. Must be an empty string or full path name that ends with a '/'. This field is treated as an object prefix. As such, it should generally not begin with a '/'.
- path string
- Root path to transfer objects. Must be an empty string or full path name that ends with a '/'. This field is treated as an object prefix. As such, it should generally not begin with a '/'.
- path str
- Root path to transfer objects. Must be an empty string or full path name that ends with a '/'. This field is treated as an object prefix. As such, it should generally not begin with a '/'.
- path String
- Root path to transfer objects. Must be an empty string or full path name that ends with a '/'. This field is treated as an object prefix. As such, it should generally not begin with a '/'.
TransferJobTransferSpecHttpDataSource, TransferJobTransferSpecHttpDataSourceArgs              
- ListUrl string
- The URL that points to the file that stores the object list entries. This file must allow public access. Currently, only URLs with HTTP and HTTPS schemes are supported.
- ListUrl string
- The URL that points to the file that stores the object list entries. This file must allow public access. Currently, only URLs with HTTP and HTTPS schemes are supported.
- listUrl String
- The URL that points to the file that stores the object list entries. This file must allow public access. Currently, only URLs with HTTP and HTTPS schemes are supported.
- listUrl string
- The URL that points to the file that stores the object list entries. This file must allow public access. Currently, only URLs with HTTP and HTTPS schemes are supported.
- list_url str
- The URL that points to the file that stores the object list entries. This file must allow public access. Currently, only URLs with HTTP and HTTPS schemes are supported.
- listUrl String
- The URL that points to the file that stores the object list entries. This file must allow public access. Currently, only URLs with HTTP and HTTPS schemes are supported.
TransferJobTransferSpecObjectConditions, TransferJobTransferSpecObjectConditionsArgs            
- ExcludePrefixes List<string>
- exclude_prefixesmust follow the requirements described for- include_prefixes. See Requirements.
- IncludePrefixes List<string>
- If include_prefixesis specified, objects that satisfy the object conditions must have names that start with one of theinclude_prefixesand that do not start with any of theexclude_prefixes. Ifinclude_prefixesis not specified, all objects except those that have names starting with one of theexclude_prefixesmust satisfy the object conditions. See Requirements.
- LastModified stringBefore 
- If specified, only objects with a "last modification time" before this timestamp and objects that don't have a "last modification time" are transferred. A timestamp in RFC3339 UTC "Zulu" format, with nanosecond resolution and up to nine fractional digits. Examples: "2014-10-02T15:01:23Z" and "2014-10-02T15:01:23.045123456Z".
- LastModified stringSince 
- If specified, only objects with a "last modification time" on or after this timestamp and objects that don't have a "last modification time" are transferred. A timestamp in RFC3339 UTC "Zulu" format, with nanosecond resolution and up to nine fractional digits. Examples: "2014-10-02T15:01:23Z" and "2014-10-02T15:01:23.045123456Z".
- MaxTime stringElapsed Since Last Modification 
- A duration in seconds with up to nine fractional digits, terminated by 's'. Example: "3.5s".
- MinTime stringElapsed Since Last Modification 
- A duration in seconds with up to nine fractional digits, terminated by 's'. Example: "3.5s".
- ExcludePrefixes []string
- exclude_prefixesmust follow the requirements described for- include_prefixes. See Requirements.
- IncludePrefixes []string
- If include_prefixesis specified, objects that satisfy the object conditions must have names that start with one of theinclude_prefixesand that do not start with any of theexclude_prefixes. Ifinclude_prefixesis not specified, all objects except those that have names starting with one of theexclude_prefixesmust satisfy the object conditions. See Requirements.
- LastModified stringBefore 
- If specified, only objects with a "last modification time" before this timestamp and objects that don't have a "last modification time" are transferred. A timestamp in RFC3339 UTC "Zulu" format, with nanosecond resolution and up to nine fractional digits. Examples: "2014-10-02T15:01:23Z" and "2014-10-02T15:01:23.045123456Z".
- LastModified stringSince 
- If specified, only objects with a "last modification time" on or after this timestamp and objects that don't have a "last modification time" are transferred. A timestamp in RFC3339 UTC "Zulu" format, with nanosecond resolution and up to nine fractional digits. Examples: "2014-10-02T15:01:23Z" and "2014-10-02T15:01:23.045123456Z".
- MaxTime stringElapsed Since Last Modification 
- A duration in seconds with up to nine fractional digits, terminated by 's'. Example: "3.5s".
- MinTime stringElapsed Since Last Modification 
- A duration in seconds with up to nine fractional digits, terminated by 's'. Example: "3.5s".
- excludePrefixes List<String>
- exclude_prefixesmust follow the requirements described for- include_prefixes. See Requirements.
- includePrefixes List<String>
- If include_prefixesis specified, objects that satisfy the object conditions must have names that start with one of theinclude_prefixesand that do not start with any of theexclude_prefixes. Ifinclude_prefixesis not specified, all objects except those that have names starting with one of theexclude_prefixesmust satisfy the object conditions. See Requirements.
- lastModified StringBefore 
- If specified, only objects with a "last modification time" before this timestamp and objects that don't have a "last modification time" are transferred. A timestamp in RFC3339 UTC "Zulu" format, with nanosecond resolution and up to nine fractional digits. Examples: "2014-10-02T15:01:23Z" and "2014-10-02T15:01:23.045123456Z".
- lastModified StringSince 
- If specified, only objects with a "last modification time" on or after this timestamp and objects that don't have a "last modification time" are transferred. A timestamp in RFC3339 UTC "Zulu" format, with nanosecond resolution and up to nine fractional digits. Examples: "2014-10-02T15:01:23Z" and "2014-10-02T15:01:23.045123456Z".
- maxTime StringElapsed Since Last Modification 
- A duration in seconds with up to nine fractional digits, terminated by 's'. Example: "3.5s".
- minTime StringElapsed Since Last Modification 
- A duration in seconds with up to nine fractional digits, terminated by 's'. Example: "3.5s".
- excludePrefixes string[]
- exclude_prefixesmust follow the requirements described for- include_prefixes. See Requirements.
- includePrefixes string[]
- If include_prefixesis specified, objects that satisfy the object conditions must have names that start with one of theinclude_prefixesand that do not start with any of theexclude_prefixes. Ifinclude_prefixesis not specified, all objects except those that have names starting with one of theexclude_prefixesmust satisfy the object conditions. See Requirements.
- lastModified stringBefore 
- If specified, only objects with a "last modification time" before this timestamp and objects that don't have a "last modification time" are transferred. A timestamp in RFC3339 UTC "Zulu" format, with nanosecond resolution and up to nine fractional digits. Examples: "2014-10-02T15:01:23Z" and "2014-10-02T15:01:23.045123456Z".
- lastModified stringSince 
- If specified, only objects with a "last modification time" on or after this timestamp and objects that don't have a "last modification time" are transferred. A timestamp in RFC3339 UTC "Zulu" format, with nanosecond resolution and up to nine fractional digits. Examples: "2014-10-02T15:01:23Z" and "2014-10-02T15:01:23.045123456Z".
- maxTime stringElapsed Since Last Modification 
- A duration in seconds with up to nine fractional digits, terminated by 's'. Example: "3.5s".
- minTime stringElapsed Since Last Modification 
- A duration in seconds with up to nine fractional digits, terminated by 's'. Example: "3.5s".
- exclude_prefixes Sequence[str]
- exclude_prefixesmust follow the requirements described for- include_prefixes. See Requirements.
- include_prefixes Sequence[str]
- If include_prefixesis specified, objects that satisfy the object conditions must have names that start with one of theinclude_prefixesand that do not start with any of theexclude_prefixes. Ifinclude_prefixesis not specified, all objects except those that have names starting with one of theexclude_prefixesmust satisfy the object conditions. See Requirements.
- last_modified_ strbefore 
- If specified, only objects with a "last modification time" before this timestamp and objects that don't have a "last modification time" are transferred. A timestamp in RFC3339 UTC "Zulu" format, with nanosecond resolution and up to nine fractional digits. Examples: "2014-10-02T15:01:23Z" and "2014-10-02T15:01:23.045123456Z".
- last_modified_ strsince 
- If specified, only objects with a "last modification time" on or after this timestamp and objects that don't have a "last modification time" are transferred. A timestamp in RFC3339 UTC "Zulu" format, with nanosecond resolution and up to nine fractional digits. Examples: "2014-10-02T15:01:23Z" and "2014-10-02T15:01:23.045123456Z".
- max_time_ strelapsed_ since_ last_ modification 
- A duration in seconds with up to nine fractional digits, terminated by 's'. Example: "3.5s".
- min_time_ strelapsed_ since_ last_ modification 
- A duration in seconds with up to nine fractional digits, terminated by 's'. Example: "3.5s".
- excludePrefixes List<String>
- exclude_prefixesmust follow the requirements described for- include_prefixes. See Requirements.
- includePrefixes List<String>
- If include_prefixesis specified, objects that satisfy the object conditions must have names that start with one of theinclude_prefixesand that do not start with any of theexclude_prefixes. Ifinclude_prefixesis not specified, all objects except those that have names starting with one of theexclude_prefixesmust satisfy the object conditions. See Requirements.
- lastModified StringBefore 
- If specified, only objects with a "last modification time" before this timestamp and objects that don't have a "last modification time" are transferred. A timestamp in RFC3339 UTC "Zulu" format, with nanosecond resolution and up to nine fractional digits. Examples: "2014-10-02T15:01:23Z" and "2014-10-02T15:01:23.045123456Z".
- lastModified StringSince 
- If specified, only objects with a "last modification time" on or after this timestamp and objects that don't have a "last modification time" are transferred. A timestamp in RFC3339 UTC "Zulu" format, with nanosecond resolution and up to nine fractional digits. Examples: "2014-10-02T15:01:23Z" and "2014-10-02T15:01:23.045123456Z".
- maxTime StringElapsed Since Last Modification 
- A duration in seconds with up to nine fractional digits, terminated by 's'. Example: "3.5s".
- minTime StringElapsed Since Last Modification 
- A duration in seconds with up to nine fractional digits, terminated by 's'. Example: "3.5s".
TransferJobTransferSpecPosixDataSink, TransferJobTransferSpecPosixDataSinkArgs              
- RootDirectory string
- Root directory path to the filesystem.
- RootDirectory string
- Root directory path to the filesystem.
- rootDirectory String
- Root directory path to the filesystem.
- rootDirectory string
- Root directory path to the filesystem.
- root_directory str
- Root directory path to the filesystem.
- rootDirectory String
- Root directory path to the filesystem.
TransferJobTransferSpecPosixDataSource, TransferJobTransferSpecPosixDataSourceArgs              
- RootDirectory string
- Root directory path to the filesystem.
- RootDirectory string
- Root directory path to the filesystem.
- rootDirectory String
- Root directory path to the filesystem.
- rootDirectory string
- Root directory path to the filesystem.
- root_directory str
- Root directory path to the filesystem.
- rootDirectory String
- Root directory path to the filesystem.
TransferJobTransferSpecTransferOptions, TransferJobTransferSpecTransferOptionsArgs            
- DeleteObjects boolFrom Source After Transfer 
- Whether objects should be deleted from the source after they are transferred to the sink. Note that this option and delete_objects_unique_in_sinkare mutually exclusive.
- DeleteObjects boolUnique In Sink 
- Whether objects that exist only in the sink should be deleted. Note that this option and
delete_objects_from_source_after_transferare mutually exclusive.
- OverwriteObjects boolAlready Existing In Sink 
- Whether overwriting objects that already exist in the sink is allowed.
- OverwriteWhen string
- When to overwrite objects that already exist in the sink. If not set, overwrite behavior is determined by overwrite_objects_already_existing_in_sink. Possible values: ALWAYS, DIFFERENT, NEVER.
- DeleteObjects boolFrom Source After Transfer 
- Whether objects should be deleted from the source after they are transferred to the sink. Note that this option and delete_objects_unique_in_sinkare mutually exclusive.
- DeleteObjects boolUnique In Sink 
- Whether objects that exist only in the sink should be deleted. Note that this option and
delete_objects_from_source_after_transferare mutually exclusive.
- OverwriteObjects boolAlready Existing In Sink 
- Whether overwriting objects that already exist in the sink is allowed.
- OverwriteWhen string
- When to overwrite objects that already exist in the sink. If not set, overwrite behavior is determined by overwrite_objects_already_existing_in_sink. Possible values: ALWAYS, DIFFERENT, NEVER.
- deleteObjects BooleanFrom Source After Transfer 
- Whether objects should be deleted from the source after they are transferred to the sink. Note that this option and delete_objects_unique_in_sinkare mutually exclusive.
- deleteObjects BooleanUnique In Sink 
- Whether objects that exist only in the sink should be deleted. Note that this option and
delete_objects_from_source_after_transferare mutually exclusive.
- overwriteObjects BooleanAlready Existing In Sink 
- Whether overwriting objects that already exist in the sink is allowed.
- overwriteWhen String
- When to overwrite objects that already exist in the sink. If not set, overwrite behavior is determined by overwrite_objects_already_existing_in_sink. Possible values: ALWAYS, DIFFERENT, NEVER.
- deleteObjects booleanFrom Source After Transfer 
- Whether objects should be deleted from the source after they are transferred to the sink. Note that this option and delete_objects_unique_in_sinkare mutually exclusive.
- deleteObjects booleanUnique In Sink 
- Whether objects that exist only in the sink should be deleted. Note that this option and
delete_objects_from_source_after_transferare mutually exclusive.
- overwriteObjects booleanAlready Existing In Sink 
- Whether overwriting objects that already exist in the sink is allowed.
- overwriteWhen string
- When to overwrite objects that already exist in the sink. If not set, overwrite behavior is determined by overwrite_objects_already_existing_in_sink. Possible values: ALWAYS, DIFFERENT, NEVER.
- delete_objects_ boolfrom_ source_ after_ transfer 
- Whether objects should be deleted from the source after they are transferred to the sink. Note that this option and delete_objects_unique_in_sinkare mutually exclusive.
- delete_objects_ boolunique_ in_ sink 
- Whether objects that exist only in the sink should be deleted. Note that this option and
delete_objects_from_source_after_transferare mutually exclusive.
- overwrite_objects_ boolalready_ existing_ in_ sink 
- Whether overwriting objects that already exist in the sink is allowed.
- overwrite_when str
- When to overwrite objects that already exist in the sink. If not set, overwrite behavior is determined by overwrite_objects_already_existing_in_sink. Possible values: ALWAYS, DIFFERENT, NEVER.
- deleteObjects BooleanFrom Source After Transfer 
- Whether objects should be deleted from the source after they are transferred to the sink. Note that this option and delete_objects_unique_in_sinkare mutually exclusive.
- deleteObjects BooleanUnique In Sink 
- Whether objects that exist only in the sink should be deleted. Note that this option and
delete_objects_from_source_after_transferare mutually exclusive.
- overwriteObjects BooleanAlready Existing In Sink 
- Whether overwriting objects that already exist in the sink is allowed.
- overwriteWhen String
- When to overwrite objects that already exist in the sink. If not set, overwrite behavior is determined by overwrite_objects_already_existing_in_sink. Possible values: ALWAYS, DIFFERENT, NEVER.
Import
Storage Transfer Jobs can be imported using the Transfer Job’s project and name (without the transferJob/ prefix), e.g.
- {{project_id}}/{{name}}, where- nameis a numeric value.
When using the pulumi import command, Storage Transfer Jobs can be imported using one of the formats above. For example:
$ pulumi import gcp:storage/transferJob:TransferJob default {{project_id}}/123456789
To learn more about importing existing cloud resources, see Importing resources.
Package Details
- Repository
- Google Cloud (GCP) Classic pulumi/pulumi-gcp
- License
- Apache-2.0
- Notes
- This Pulumi package is based on the google-betaTerraform Provider.