We recommend using Azure Native.
azure.streamanalytics.StreamInputBlob
Explore with Pulumi AI
Manages a Stream Analytics Stream Input Blob.
Example Usage
import * as pulumi from "@pulumi/pulumi";
import * as azure from "@pulumi/azure";
const exampleResourceGroup = new azure.core.ResourceGroup("example", {
    name: "example-resources",
    location: "West Europe",
});
const example = azure.streamanalytics.getJobOutput({
    name: "example-job",
    resourceGroupName: exampleResourceGroup.name,
});
const exampleAccount = new azure.storage.Account("example", {
    name: "examplestoracc",
    resourceGroupName: exampleResourceGroup.name,
    location: exampleResourceGroup.location,
    accountTier: "Standard",
    accountReplicationType: "LRS",
});
const exampleContainer = new azure.storage.Container("example", {
    name: "example",
    storageAccountName: exampleAccount.name,
    containerAccessType: "private",
});
const exampleStreamInputBlob = new azure.streamanalytics.StreamInputBlob("example", {
    name: "blob-stream-input",
    streamAnalyticsJobName: example.apply(example => example.name),
    resourceGroupName: example.apply(example => example.resourceGroupName),
    storageAccountName: exampleAccount.name,
    storageAccountKey: exampleAccount.primaryAccessKey,
    storageContainerName: exampleContainer.name,
    pathPattern: "some-random-pattern",
    dateFormat: "yyyy/MM/dd",
    timeFormat: "HH",
    serialization: {
        type: "Json",
        encoding: "UTF8",
    },
});
import pulumi
import pulumi_azure as azure
example_resource_group = azure.core.ResourceGroup("example",
    name="example-resources",
    location="West Europe")
example = azure.streamanalytics.get_job_output(name="example-job",
    resource_group_name=example_resource_group.name)
example_account = azure.storage.Account("example",
    name="examplestoracc",
    resource_group_name=example_resource_group.name,
    location=example_resource_group.location,
    account_tier="Standard",
    account_replication_type="LRS")
example_container = azure.storage.Container("example",
    name="example",
    storage_account_name=example_account.name,
    container_access_type="private")
example_stream_input_blob = azure.streamanalytics.StreamInputBlob("example",
    name="blob-stream-input",
    stream_analytics_job_name=example.name,
    resource_group_name=example.resource_group_name,
    storage_account_name=example_account.name,
    storage_account_key=example_account.primary_access_key,
    storage_container_name=example_container.name,
    path_pattern="some-random-pattern",
    date_format="yyyy/MM/dd",
    time_format="HH",
    serialization={
        "type": "Json",
        "encoding": "UTF8",
    })
package main
import (
	"github.com/pulumi/pulumi-azure/sdk/v6/go/azure/core"
	"github.com/pulumi/pulumi-azure/sdk/v6/go/azure/storage"
	"github.com/pulumi/pulumi-azure/sdk/v6/go/azure/streamanalytics"
	"github.com/pulumi/pulumi/sdk/v3/go/pulumi"
)
func main() {
	pulumi.Run(func(ctx *pulumi.Context) error {
		exampleResourceGroup, err := core.NewResourceGroup(ctx, "example", &core.ResourceGroupArgs{
			Name:     pulumi.String("example-resources"),
			Location: pulumi.String("West Europe"),
		})
		if err != nil {
			return err
		}
		example := streamanalytics.LookupJobOutput(ctx, streamanalytics.GetJobOutputArgs{
			Name:              pulumi.String("example-job"),
			ResourceGroupName: exampleResourceGroup.Name,
		}, nil)
		exampleAccount, err := storage.NewAccount(ctx, "example", &storage.AccountArgs{
			Name:                   pulumi.String("examplestoracc"),
			ResourceGroupName:      exampleResourceGroup.Name,
			Location:               exampleResourceGroup.Location,
			AccountTier:            pulumi.String("Standard"),
			AccountReplicationType: pulumi.String("LRS"),
		})
		if err != nil {
			return err
		}
		exampleContainer, err := storage.NewContainer(ctx, "example", &storage.ContainerArgs{
			Name:                pulumi.String("example"),
			StorageAccountName:  exampleAccount.Name,
			ContainerAccessType: pulumi.String("private"),
		})
		if err != nil {
			return err
		}
		_, err = streamanalytics.NewStreamInputBlob(ctx, "example", &streamanalytics.StreamInputBlobArgs{
			Name: pulumi.String("blob-stream-input"),
			StreamAnalyticsJobName: pulumi.String(example.ApplyT(func(example streamanalytics.GetJobResult) (*string, error) {
				return &example.Name, nil
			}).(pulumi.StringPtrOutput)),
			ResourceGroupName: pulumi.String(example.ApplyT(func(example streamanalytics.GetJobResult) (*string, error) {
				return &example.ResourceGroupName, nil
			}).(pulumi.StringPtrOutput)),
			StorageAccountName:   exampleAccount.Name,
			StorageAccountKey:    exampleAccount.PrimaryAccessKey,
			StorageContainerName: exampleContainer.Name,
			PathPattern:          pulumi.String("some-random-pattern"),
			DateFormat:           pulumi.String("yyyy/MM/dd"),
			TimeFormat:           pulumi.String("HH"),
			Serialization: &streamanalytics.StreamInputBlobSerializationArgs{
				Type:     pulumi.String("Json"),
				Encoding: pulumi.String("UTF8"),
			},
		})
		if err != nil {
			return err
		}
		return nil
	})
}
using System.Collections.Generic;
using System.Linq;
using Pulumi;
using Azure = Pulumi.Azure;
return await Deployment.RunAsync(() => 
{
    var exampleResourceGroup = new Azure.Core.ResourceGroup("example", new()
    {
        Name = "example-resources",
        Location = "West Europe",
    });
    var example = Azure.StreamAnalytics.GetJob.Invoke(new()
    {
        Name = "example-job",
        ResourceGroupName = exampleResourceGroup.Name,
    });
    var exampleAccount = new Azure.Storage.Account("example", new()
    {
        Name = "examplestoracc",
        ResourceGroupName = exampleResourceGroup.Name,
        Location = exampleResourceGroup.Location,
        AccountTier = "Standard",
        AccountReplicationType = "LRS",
    });
    var exampleContainer = new Azure.Storage.Container("example", new()
    {
        Name = "example",
        StorageAccountName = exampleAccount.Name,
        ContainerAccessType = "private",
    });
    var exampleStreamInputBlob = new Azure.StreamAnalytics.StreamInputBlob("example", new()
    {
        Name = "blob-stream-input",
        StreamAnalyticsJobName = example.Apply(getJobResult => getJobResult.Name),
        ResourceGroupName = example.Apply(getJobResult => getJobResult.ResourceGroupName),
        StorageAccountName = exampleAccount.Name,
        StorageAccountKey = exampleAccount.PrimaryAccessKey,
        StorageContainerName = exampleContainer.Name,
        PathPattern = "some-random-pattern",
        DateFormat = "yyyy/MM/dd",
        TimeFormat = "HH",
        Serialization = new Azure.StreamAnalytics.Inputs.StreamInputBlobSerializationArgs
        {
            Type = "Json",
            Encoding = "UTF8",
        },
    });
});
package generated_program;
import com.pulumi.Context;
import com.pulumi.Pulumi;
import com.pulumi.core.Output;
import com.pulumi.azure.core.ResourceGroup;
import com.pulumi.azure.core.ResourceGroupArgs;
import com.pulumi.azure.streamanalytics.StreamanalyticsFunctions;
import com.pulumi.azure.streamanalytics.inputs.GetJobArgs;
import com.pulumi.azure.storage.Account;
import com.pulumi.azure.storage.AccountArgs;
import com.pulumi.azure.storage.Container;
import com.pulumi.azure.storage.ContainerArgs;
import com.pulumi.azure.streamanalytics.StreamInputBlob;
import com.pulumi.azure.streamanalytics.StreamInputBlobArgs;
import com.pulumi.azure.streamanalytics.inputs.StreamInputBlobSerializationArgs;
import java.util.List;
import java.util.ArrayList;
import java.util.Map;
import java.io.File;
import java.nio.file.Files;
import java.nio.file.Paths;
public class App {
    public static void main(String[] args) {
        Pulumi.run(App::stack);
    }
    public static void stack(Context ctx) {
        var exampleResourceGroup = new ResourceGroup("exampleResourceGroup", ResourceGroupArgs.builder()
            .name("example-resources")
            .location("West Europe")
            .build());
        final var example = StreamanalyticsFunctions.getJob(GetJobArgs.builder()
            .name("example-job")
            .resourceGroupName(exampleResourceGroup.name())
            .build());
        var exampleAccount = new Account("exampleAccount", AccountArgs.builder()
            .name("examplestoracc")
            .resourceGroupName(exampleResourceGroup.name())
            .location(exampleResourceGroup.location())
            .accountTier("Standard")
            .accountReplicationType("LRS")
            .build());
        var exampleContainer = new Container("exampleContainer", ContainerArgs.builder()
            .name("example")
            .storageAccountName(exampleAccount.name())
            .containerAccessType("private")
            .build());
        var exampleStreamInputBlob = new StreamInputBlob("exampleStreamInputBlob", StreamInputBlobArgs.builder()
            .name("blob-stream-input")
            .streamAnalyticsJobName(example.applyValue(getJobResult -> getJobResult).applyValue(example -> example.applyValue(getJobResult -> getJobResult.name())))
            .resourceGroupName(example.applyValue(getJobResult -> getJobResult).applyValue(example -> example.applyValue(getJobResult -> getJobResult.resourceGroupName())))
            .storageAccountName(exampleAccount.name())
            .storageAccountKey(exampleAccount.primaryAccessKey())
            .storageContainerName(exampleContainer.name())
            .pathPattern("some-random-pattern")
            .dateFormat("yyyy/MM/dd")
            .timeFormat("HH")
            .serialization(StreamInputBlobSerializationArgs.builder()
                .type("Json")
                .encoding("UTF8")
                .build())
            .build());
    }
}
resources:
  exampleResourceGroup:
    type: azure:core:ResourceGroup
    name: example
    properties:
      name: example-resources
      location: West Europe
  exampleAccount:
    type: azure:storage:Account
    name: example
    properties:
      name: examplestoracc
      resourceGroupName: ${exampleResourceGroup.name}
      location: ${exampleResourceGroup.location}
      accountTier: Standard
      accountReplicationType: LRS
  exampleContainer:
    type: azure:storage:Container
    name: example
    properties:
      name: example
      storageAccountName: ${exampleAccount.name}
      containerAccessType: private
  exampleStreamInputBlob:
    type: azure:streamanalytics:StreamInputBlob
    name: example
    properties:
      name: blob-stream-input
      streamAnalyticsJobName: ${example.name}
      resourceGroupName: ${example.resourceGroupName}
      storageAccountName: ${exampleAccount.name}
      storageAccountKey: ${exampleAccount.primaryAccessKey}
      storageContainerName: ${exampleContainer.name}
      pathPattern: some-random-pattern
      dateFormat: yyyy/MM/dd
      timeFormat: HH
      serialization:
        type: Json
        encoding: UTF8
variables:
  example:
    fn::invoke:
      function: azure:streamanalytics:getJob
      arguments:
        name: example-job
        resourceGroupName: ${exampleResourceGroup.name}
Create StreamInputBlob Resource
Resources are created with functions called constructors. To learn more about declaring and configuring resources, see Resources.
Constructor syntax
new StreamInputBlob(name: string, args: StreamInputBlobArgs, opts?: CustomResourceOptions);@overload
def StreamInputBlob(resource_name: str,
                    args: StreamInputBlobArgs,
                    opts: Optional[ResourceOptions] = None)
@overload
def StreamInputBlob(resource_name: str,
                    opts: Optional[ResourceOptions] = None,
                    date_format: Optional[str] = None,
                    path_pattern: Optional[str] = None,
                    resource_group_name: Optional[str] = None,
                    serialization: Optional[StreamInputBlobSerializationArgs] = None,
                    storage_account_key: Optional[str] = None,
                    storage_account_name: Optional[str] = None,
                    storage_container_name: Optional[str] = None,
                    stream_analytics_job_name: Optional[str] = None,
                    time_format: Optional[str] = None,
                    authentication_mode: Optional[str] = None,
                    name: Optional[str] = None)func NewStreamInputBlob(ctx *Context, name string, args StreamInputBlobArgs, opts ...ResourceOption) (*StreamInputBlob, error)public StreamInputBlob(string name, StreamInputBlobArgs args, CustomResourceOptions? opts = null)
public StreamInputBlob(String name, StreamInputBlobArgs args)
public StreamInputBlob(String name, StreamInputBlobArgs args, CustomResourceOptions options)
type: azure:streamanalytics:StreamInputBlob
properties: # The arguments to resource properties.
options: # Bag of options to control resource's behavior.
Parameters
- name string
- The unique name of the resource.
- args StreamInputBlobArgs
- The arguments to resource properties.
- opts CustomResourceOptions
- Bag of options to control resource's behavior.
- resource_name str
- The unique name of the resource.
- args StreamInputBlobArgs
- The arguments to resource properties.
- opts ResourceOptions
- Bag of options to control resource's behavior.
- ctx Context
- Context object for the current deployment.
- name string
- The unique name of the resource.
- args StreamInputBlobArgs
- The arguments to resource properties.
- opts ResourceOption
- Bag of options to control resource's behavior.
- name string
- The unique name of the resource.
- args StreamInputBlobArgs
- The arguments to resource properties.
- opts CustomResourceOptions
- Bag of options to control resource's behavior.
- name String
- The unique name of the resource.
- args StreamInputBlobArgs
- The arguments to resource properties.
- options CustomResourceOptions
- Bag of options to control resource's behavior.
Constructor example
The following reference example uses placeholder values for all input properties.
var streamInputBlobResource = new Azure.StreamAnalytics.StreamInputBlob("streamInputBlobResource", new()
{
    DateFormat = "string",
    PathPattern = "string",
    ResourceGroupName = "string",
    Serialization = new Azure.StreamAnalytics.Inputs.StreamInputBlobSerializationArgs
    {
        Type = "string",
        Encoding = "string",
        FieldDelimiter = "string",
    },
    StorageAccountKey = "string",
    StorageAccountName = "string",
    StorageContainerName = "string",
    StreamAnalyticsJobName = "string",
    TimeFormat = "string",
    AuthenticationMode = "string",
    Name = "string",
});
example, err := streamanalytics.NewStreamInputBlob(ctx, "streamInputBlobResource", &streamanalytics.StreamInputBlobArgs{
	DateFormat:        pulumi.String("string"),
	PathPattern:       pulumi.String("string"),
	ResourceGroupName: pulumi.String("string"),
	Serialization: &streamanalytics.StreamInputBlobSerializationArgs{
		Type:           pulumi.String("string"),
		Encoding:       pulumi.String("string"),
		FieldDelimiter: pulumi.String("string"),
	},
	StorageAccountKey:      pulumi.String("string"),
	StorageAccountName:     pulumi.String("string"),
	StorageContainerName:   pulumi.String("string"),
	StreamAnalyticsJobName: pulumi.String("string"),
	TimeFormat:             pulumi.String("string"),
	AuthenticationMode:     pulumi.String("string"),
	Name:                   pulumi.String("string"),
})
var streamInputBlobResource = new StreamInputBlob("streamInputBlobResource", StreamInputBlobArgs.builder()
    .dateFormat("string")
    .pathPattern("string")
    .resourceGroupName("string")
    .serialization(StreamInputBlobSerializationArgs.builder()
        .type("string")
        .encoding("string")
        .fieldDelimiter("string")
        .build())
    .storageAccountKey("string")
    .storageAccountName("string")
    .storageContainerName("string")
    .streamAnalyticsJobName("string")
    .timeFormat("string")
    .authenticationMode("string")
    .name("string")
    .build());
stream_input_blob_resource = azure.streamanalytics.StreamInputBlob("streamInputBlobResource",
    date_format="string",
    path_pattern="string",
    resource_group_name="string",
    serialization={
        "type": "string",
        "encoding": "string",
        "field_delimiter": "string",
    },
    storage_account_key="string",
    storage_account_name="string",
    storage_container_name="string",
    stream_analytics_job_name="string",
    time_format="string",
    authentication_mode="string",
    name="string")
const streamInputBlobResource = new azure.streamanalytics.StreamInputBlob("streamInputBlobResource", {
    dateFormat: "string",
    pathPattern: "string",
    resourceGroupName: "string",
    serialization: {
        type: "string",
        encoding: "string",
        fieldDelimiter: "string",
    },
    storageAccountKey: "string",
    storageAccountName: "string",
    storageContainerName: "string",
    streamAnalyticsJobName: "string",
    timeFormat: "string",
    authenticationMode: "string",
    name: "string",
});
type: azure:streamanalytics:StreamInputBlob
properties:
    authenticationMode: string
    dateFormat: string
    name: string
    pathPattern: string
    resourceGroupName: string
    serialization:
        encoding: string
        fieldDelimiter: string
        type: string
    storageAccountKey: string
    storageAccountName: string
    storageContainerName: string
    streamAnalyticsJobName: string
    timeFormat: string
StreamInputBlob Resource Properties
To learn more about resource properties and how to use them, see Inputs and Outputs in the Architecture and Concepts docs.
Inputs
In Python, inputs that are objects can be passed either as argument classes or as dictionary literals.
The StreamInputBlob resource accepts the following input properties:
- DateFormat string
- The date format. Wherever {date}appears inpath_pattern, the value of this property is used as the date format instead.
- PathPattern string
- The blob path pattern. Not a regular expression. It represents a pattern against which blob names will be matched to determine whether or not they should be included as input or output to the job.
- ResourceGroup stringName 
- The name of the Resource Group where the Stream Analytics Job exists. Changing this forces a new resource to be created.
- Serialization
StreamInput Blob Serialization 
- A serializationblock as defined below.
- StorageAccount stringKey 
- The Access Key which should be used to connect to this Storage Account.
- StorageAccount stringName 
- The name of the Storage Account.
- StorageContainer stringName 
- The name of the Container within the Storage Account.
- StreamAnalytics stringJob Name 
- The name of the Stream Analytics Job. Changing this forces a new resource to be created.
- TimeFormat string
- The time format. Wherever {time}appears inpath_pattern, the value of this property is used as the time format instead.
- AuthenticationMode string
- The authentication mode for the Stream Analytics Input. Possible values are MsiandConnectionString. Defaults toConnectionString.
- Name string
- The name of the Stream Input Blob. Changing this forces a new resource to be created.
- DateFormat string
- The date format. Wherever {date}appears inpath_pattern, the value of this property is used as the date format instead.
- PathPattern string
- The blob path pattern. Not a regular expression. It represents a pattern against which blob names will be matched to determine whether or not they should be included as input or output to the job.
- ResourceGroup stringName 
- The name of the Resource Group where the Stream Analytics Job exists. Changing this forces a new resource to be created.
- Serialization
StreamInput Blob Serialization Args 
- A serializationblock as defined below.
- StorageAccount stringKey 
- The Access Key which should be used to connect to this Storage Account.
- StorageAccount stringName 
- The name of the Storage Account.
- StorageContainer stringName 
- The name of the Container within the Storage Account.
- StreamAnalytics stringJob Name 
- The name of the Stream Analytics Job. Changing this forces a new resource to be created.
- TimeFormat string
- The time format. Wherever {time}appears inpath_pattern, the value of this property is used as the time format instead.
- AuthenticationMode string
- The authentication mode for the Stream Analytics Input. Possible values are MsiandConnectionString. Defaults toConnectionString.
- Name string
- The name of the Stream Input Blob. Changing this forces a new resource to be created.
- dateFormat String
- The date format. Wherever {date}appears inpath_pattern, the value of this property is used as the date format instead.
- pathPattern String
- The blob path pattern. Not a regular expression. It represents a pattern against which blob names will be matched to determine whether or not they should be included as input or output to the job.
- resourceGroup StringName 
- The name of the Resource Group where the Stream Analytics Job exists. Changing this forces a new resource to be created.
- serialization
StreamInput Blob Serialization 
- A serializationblock as defined below.
- storageAccount StringKey 
- The Access Key which should be used to connect to this Storage Account.
- storageAccount StringName 
- The name of the Storage Account.
- storageContainer StringName 
- The name of the Container within the Storage Account.
- streamAnalytics StringJob Name 
- The name of the Stream Analytics Job. Changing this forces a new resource to be created.
- timeFormat String
- The time format. Wherever {time}appears inpath_pattern, the value of this property is used as the time format instead.
- authenticationMode String
- The authentication mode for the Stream Analytics Input. Possible values are MsiandConnectionString. Defaults toConnectionString.
- name String
- The name of the Stream Input Blob. Changing this forces a new resource to be created.
- dateFormat string
- The date format. Wherever {date}appears inpath_pattern, the value of this property is used as the date format instead.
- pathPattern string
- The blob path pattern. Not a regular expression. It represents a pattern against which blob names will be matched to determine whether or not they should be included as input or output to the job.
- resourceGroup stringName 
- The name of the Resource Group where the Stream Analytics Job exists. Changing this forces a new resource to be created.
- serialization
StreamInput Blob Serialization 
- A serializationblock as defined below.
- storageAccount stringKey 
- The Access Key which should be used to connect to this Storage Account.
- storageAccount stringName 
- The name of the Storage Account.
- storageContainer stringName 
- The name of the Container within the Storage Account.
- streamAnalytics stringJob Name 
- The name of the Stream Analytics Job. Changing this forces a new resource to be created.
- timeFormat string
- The time format. Wherever {time}appears inpath_pattern, the value of this property is used as the time format instead.
- authenticationMode string
- The authentication mode for the Stream Analytics Input. Possible values are MsiandConnectionString. Defaults toConnectionString.
- name string
- The name of the Stream Input Blob. Changing this forces a new resource to be created.
- date_format str
- The date format. Wherever {date}appears inpath_pattern, the value of this property is used as the date format instead.
- path_pattern str
- The blob path pattern. Not a regular expression. It represents a pattern against which blob names will be matched to determine whether or not they should be included as input or output to the job.
- resource_group_ strname 
- The name of the Resource Group where the Stream Analytics Job exists. Changing this forces a new resource to be created.
- serialization
StreamInput Blob Serialization Args 
- A serializationblock as defined below.
- storage_account_ strkey 
- The Access Key which should be used to connect to this Storage Account.
- storage_account_ strname 
- The name of the Storage Account.
- storage_container_ strname 
- The name of the Container within the Storage Account.
- stream_analytics_ strjob_ name 
- The name of the Stream Analytics Job. Changing this forces a new resource to be created.
- time_format str
- The time format. Wherever {time}appears inpath_pattern, the value of this property is used as the time format instead.
- authentication_mode str
- The authentication mode for the Stream Analytics Input. Possible values are MsiandConnectionString. Defaults toConnectionString.
- name str
- The name of the Stream Input Blob. Changing this forces a new resource to be created.
- dateFormat String
- The date format. Wherever {date}appears inpath_pattern, the value of this property is used as the date format instead.
- pathPattern String
- The blob path pattern. Not a regular expression. It represents a pattern against which blob names will be matched to determine whether or not they should be included as input or output to the job.
- resourceGroup StringName 
- The name of the Resource Group where the Stream Analytics Job exists. Changing this forces a new resource to be created.
- serialization Property Map
- A serializationblock as defined below.
- storageAccount StringKey 
- The Access Key which should be used to connect to this Storage Account.
- storageAccount StringName 
- The name of the Storage Account.
- storageContainer StringName 
- The name of the Container within the Storage Account.
- streamAnalytics StringJob Name 
- The name of the Stream Analytics Job. Changing this forces a new resource to be created.
- timeFormat String
- The time format. Wherever {time}appears inpath_pattern, the value of this property is used as the time format instead.
- authenticationMode String
- The authentication mode for the Stream Analytics Input. Possible values are MsiandConnectionString. Defaults toConnectionString.
- name String
- The name of the Stream Input Blob. Changing this forces a new resource to be created.
Outputs
All input properties are implicitly available as output properties. Additionally, the StreamInputBlob resource produces the following output properties:
- Id string
- The provider-assigned unique ID for this managed resource.
- Id string
- The provider-assigned unique ID for this managed resource.
- id String
- The provider-assigned unique ID for this managed resource.
- id string
- The provider-assigned unique ID for this managed resource.
- id str
- The provider-assigned unique ID for this managed resource.
- id String
- The provider-assigned unique ID for this managed resource.
Look up Existing StreamInputBlob Resource
Get an existing StreamInputBlob resource’s state with the given name, ID, and optional extra properties used to qualify the lookup.
public static get(name: string, id: Input<ID>, state?: StreamInputBlobState, opts?: CustomResourceOptions): StreamInputBlob@staticmethod
def get(resource_name: str,
        id: str,
        opts: Optional[ResourceOptions] = None,
        authentication_mode: Optional[str] = None,
        date_format: Optional[str] = None,
        name: Optional[str] = None,
        path_pattern: Optional[str] = None,
        resource_group_name: Optional[str] = None,
        serialization: Optional[StreamInputBlobSerializationArgs] = None,
        storage_account_key: Optional[str] = None,
        storage_account_name: Optional[str] = None,
        storage_container_name: Optional[str] = None,
        stream_analytics_job_name: Optional[str] = None,
        time_format: Optional[str] = None) -> StreamInputBlobfunc GetStreamInputBlob(ctx *Context, name string, id IDInput, state *StreamInputBlobState, opts ...ResourceOption) (*StreamInputBlob, error)public static StreamInputBlob Get(string name, Input<string> id, StreamInputBlobState? state, CustomResourceOptions? opts = null)public static StreamInputBlob get(String name, Output<String> id, StreamInputBlobState state, CustomResourceOptions options)resources:  _:    type: azure:streamanalytics:StreamInputBlob    get:      id: ${id}- name
- The unique name of the resulting resource.
- id
- The unique provider ID of the resource to lookup.
- state
- Any extra arguments used during the lookup.
- opts
- A bag of options that control this resource's behavior.
- resource_name
- The unique name of the resulting resource.
- id
- The unique provider ID of the resource to lookup.
- name
- The unique name of the resulting resource.
- id
- The unique provider ID of the resource to lookup.
- state
- Any extra arguments used during the lookup.
- opts
- A bag of options that control this resource's behavior.
- name
- The unique name of the resulting resource.
- id
- The unique provider ID of the resource to lookup.
- state
- Any extra arguments used during the lookup.
- opts
- A bag of options that control this resource's behavior.
- name
- The unique name of the resulting resource.
- id
- The unique provider ID of the resource to lookup.
- state
- Any extra arguments used during the lookup.
- opts
- A bag of options that control this resource's behavior.
- AuthenticationMode string
- The authentication mode for the Stream Analytics Input. Possible values are MsiandConnectionString. Defaults toConnectionString.
- DateFormat string
- The date format. Wherever {date}appears inpath_pattern, the value of this property is used as the date format instead.
- Name string
- The name of the Stream Input Blob. Changing this forces a new resource to be created.
- PathPattern string
- The blob path pattern. Not a regular expression. It represents a pattern against which blob names will be matched to determine whether or not they should be included as input or output to the job.
- ResourceGroup stringName 
- The name of the Resource Group where the Stream Analytics Job exists. Changing this forces a new resource to be created.
- Serialization
StreamInput Blob Serialization 
- A serializationblock as defined below.
- StorageAccount stringKey 
- The Access Key which should be used to connect to this Storage Account.
- StorageAccount stringName 
- The name of the Storage Account.
- StorageContainer stringName 
- The name of the Container within the Storage Account.
- StreamAnalytics stringJob Name 
- The name of the Stream Analytics Job. Changing this forces a new resource to be created.
- TimeFormat string
- The time format. Wherever {time}appears inpath_pattern, the value of this property is used as the time format instead.
- AuthenticationMode string
- The authentication mode for the Stream Analytics Input. Possible values are MsiandConnectionString. Defaults toConnectionString.
- DateFormat string
- The date format. Wherever {date}appears inpath_pattern, the value of this property is used as the date format instead.
- Name string
- The name of the Stream Input Blob. Changing this forces a new resource to be created.
- PathPattern string
- The blob path pattern. Not a regular expression. It represents a pattern against which blob names will be matched to determine whether or not they should be included as input or output to the job.
- ResourceGroup stringName 
- The name of the Resource Group where the Stream Analytics Job exists. Changing this forces a new resource to be created.
- Serialization
StreamInput Blob Serialization Args 
- A serializationblock as defined below.
- StorageAccount stringKey 
- The Access Key which should be used to connect to this Storage Account.
- StorageAccount stringName 
- The name of the Storage Account.
- StorageContainer stringName 
- The name of the Container within the Storage Account.
- StreamAnalytics stringJob Name 
- The name of the Stream Analytics Job. Changing this forces a new resource to be created.
- TimeFormat string
- The time format. Wherever {time}appears inpath_pattern, the value of this property is used as the time format instead.
- authenticationMode String
- The authentication mode for the Stream Analytics Input. Possible values are MsiandConnectionString. Defaults toConnectionString.
- dateFormat String
- The date format. Wherever {date}appears inpath_pattern, the value of this property is used as the date format instead.
- name String
- The name of the Stream Input Blob. Changing this forces a new resource to be created.
- pathPattern String
- The blob path pattern. Not a regular expression. It represents a pattern against which blob names will be matched to determine whether or not they should be included as input or output to the job.
- resourceGroup StringName 
- The name of the Resource Group where the Stream Analytics Job exists. Changing this forces a new resource to be created.
- serialization
StreamInput Blob Serialization 
- A serializationblock as defined below.
- storageAccount StringKey 
- The Access Key which should be used to connect to this Storage Account.
- storageAccount StringName 
- The name of the Storage Account.
- storageContainer StringName 
- The name of the Container within the Storage Account.
- streamAnalytics StringJob Name 
- The name of the Stream Analytics Job. Changing this forces a new resource to be created.
- timeFormat String
- The time format. Wherever {time}appears inpath_pattern, the value of this property is used as the time format instead.
- authenticationMode string
- The authentication mode for the Stream Analytics Input. Possible values are MsiandConnectionString. Defaults toConnectionString.
- dateFormat string
- The date format. Wherever {date}appears inpath_pattern, the value of this property is used as the date format instead.
- name string
- The name of the Stream Input Blob. Changing this forces a new resource to be created.
- pathPattern string
- The blob path pattern. Not a regular expression. It represents a pattern against which blob names will be matched to determine whether or not they should be included as input or output to the job.
- resourceGroup stringName 
- The name of the Resource Group where the Stream Analytics Job exists. Changing this forces a new resource to be created.
- serialization
StreamInput Blob Serialization 
- A serializationblock as defined below.
- storageAccount stringKey 
- The Access Key which should be used to connect to this Storage Account.
- storageAccount stringName 
- The name of the Storage Account.
- storageContainer stringName 
- The name of the Container within the Storage Account.
- streamAnalytics stringJob Name 
- The name of the Stream Analytics Job. Changing this forces a new resource to be created.
- timeFormat string
- The time format. Wherever {time}appears inpath_pattern, the value of this property is used as the time format instead.
- authentication_mode str
- The authentication mode for the Stream Analytics Input. Possible values are MsiandConnectionString. Defaults toConnectionString.
- date_format str
- The date format. Wherever {date}appears inpath_pattern, the value of this property is used as the date format instead.
- name str
- The name of the Stream Input Blob. Changing this forces a new resource to be created.
- path_pattern str
- The blob path pattern. Not a regular expression. It represents a pattern against which blob names will be matched to determine whether or not they should be included as input or output to the job.
- resource_group_ strname 
- The name of the Resource Group where the Stream Analytics Job exists. Changing this forces a new resource to be created.
- serialization
StreamInput Blob Serialization Args 
- A serializationblock as defined below.
- storage_account_ strkey 
- The Access Key which should be used to connect to this Storage Account.
- storage_account_ strname 
- The name of the Storage Account.
- storage_container_ strname 
- The name of the Container within the Storage Account.
- stream_analytics_ strjob_ name 
- The name of the Stream Analytics Job. Changing this forces a new resource to be created.
- time_format str
- The time format. Wherever {time}appears inpath_pattern, the value of this property is used as the time format instead.
- authenticationMode String
- The authentication mode for the Stream Analytics Input. Possible values are MsiandConnectionString. Defaults toConnectionString.
- dateFormat String
- The date format. Wherever {date}appears inpath_pattern, the value of this property is used as the date format instead.
- name String
- The name of the Stream Input Blob. Changing this forces a new resource to be created.
- pathPattern String
- The blob path pattern. Not a regular expression. It represents a pattern against which blob names will be matched to determine whether or not they should be included as input or output to the job.
- resourceGroup StringName 
- The name of the Resource Group where the Stream Analytics Job exists. Changing this forces a new resource to be created.
- serialization Property Map
- A serializationblock as defined below.
- storageAccount StringKey 
- The Access Key which should be used to connect to this Storage Account.
- storageAccount StringName 
- The name of the Storage Account.
- storageContainer StringName 
- The name of the Container within the Storage Account.
- streamAnalytics StringJob Name 
- The name of the Stream Analytics Job. Changing this forces a new resource to be created.
- timeFormat String
- The time format. Wherever {time}appears inpath_pattern, the value of this property is used as the time format instead.
Supporting Types
StreamInputBlobSerialization, StreamInputBlobSerializationArgs        
- Type string
- The serialization format used for incoming data streams. Possible values are Avro,CsvandJson.
- Encoding string
- The encoding of the incoming data in the case of input and the encoding of outgoing data in the case of output. It currently can only be set to - UTF8.- NOTE: This is required when - typeis set to- Csvor- Json.
- FieldDelimiter string
- The delimiter that will be used to separate comma-separated value (CSV) records. Possible values are - (space),- ,(comma),- (tab),- |(pipe) and- ;.- NOTE: This is required when - typeis set to- Csv.
- Type string
- The serialization format used for incoming data streams. Possible values are Avro,CsvandJson.
- Encoding string
- The encoding of the incoming data in the case of input and the encoding of outgoing data in the case of output. It currently can only be set to - UTF8.- NOTE: This is required when - typeis set to- Csvor- Json.
- FieldDelimiter string
- The delimiter that will be used to separate comma-separated value (CSV) records. Possible values are - (space),- ,(comma),- (tab),- |(pipe) and- ;.- NOTE: This is required when - typeis set to- Csv.
- type String
- The serialization format used for incoming data streams. Possible values are Avro,CsvandJson.
- encoding String
- The encoding of the incoming data in the case of input and the encoding of outgoing data in the case of output. It currently can only be set to - UTF8.- NOTE: This is required when - typeis set to- Csvor- Json.
- fieldDelimiter String
- The delimiter that will be used to separate comma-separated value (CSV) records. Possible values are - (space),- ,(comma),- (tab),- |(pipe) and- ;.- NOTE: This is required when - typeis set to- Csv.
- type string
- The serialization format used for incoming data streams. Possible values are Avro,CsvandJson.
- encoding string
- The encoding of the incoming data in the case of input and the encoding of outgoing data in the case of output. It currently can only be set to - UTF8.- NOTE: This is required when - typeis set to- Csvor- Json.
- fieldDelimiter string
- The delimiter that will be used to separate comma-separated value (CSV) records. Possible values are - (space),- ,(comma),- (tab),- |(pipe) and- ;.- NOTE: This is required when - typeis set to- Csv.
- type str
- The serialization format used for incoming data streams. Possible values are Avro,CsvandJson.
- encoding str
- The encoding of the incoming data in the case of input and the encoding of outgoing data in the case of output. It currently can only be set to - UTF8.- NOTE: This is required when - typeis set to- Csvor- Json.
- field_delimiter str
- The delimiter that will be used to separate comma-separated value (CSV) records. Possible values are - (space),- ,(comma),- (tab),- |(pipe) and- ;.- NOTE: This is required when - typeis set to- Csv.
- type String
- The serialization format used for incoming data streams. Possible values are Avro,CsvandJson.
- encoding String
- The encoding of the incoming data in the case of input and the encoding of outgoing data in the case of output. It currently can only be set to - UTF8.- NOTE: This is required when - typeis set to- Csvor- Json.
- fieldDelimiter String
- The delimiter that will be used to separate comma-separated value (CSV) records. Possible values are - (space),- ,(comma),- (tab),- |(pipe) and- ;.- NOTE: This is required when - typeis set to- Csv.
Import
Stream Analytics Stream Input Blob’s can be imported using the resource id, e.g.
$ pulumi import azure:streamanalytics/streamInputBlob:StreamInputBlob example /subscriptions/00000000-0000-0000-0000-000000000000/resourceGroups/group1/providers/Microsoft.StreamAnalytics/streamingJobs/job1/inputs/input1
To learn more about importing existing cloud resources, see Importing resources.
Package Details
- Repository
- Azure Classic pulumi/pulumi-azure
- License
- Apache-2.0
- Notes
- This Pulumi package is based on the azurermTerraform Provider.