1. Packages
  2. Azure Classic
  3. API Docs
  4. streamanalytics
  5. JobSchedule

We recommend using Azure Native.

Azure v6.22.0 published on Tuesday, Apr 1, 2025 by Pulumi

azure.streamanalytics.JobSchedule

Explore with Pulumi AI

Manages a Stream Analytics Job Schedule.

Example Usage

import * as pulumi from "@pulumi/pulumi";
import * as azure from "@pulumi/azure";

const example = new azure.core.ResourceGroup("example", {
    name: "example-resources",
    location: "West Europe",
});
const exampleAccount = new azure.storage.Account("example", {
    name: "example",
    resourceGroupName: example.name,
    location: example.location,
    accountTier: "Standard",
    accountReplicationType: "LRS",
});
const exampleContainer = new azure.storage.Container("example", {
    name: "example",
    storageAccountName: exampleAccount.name,
    containerAccessType: "private",
});
const exampleBlob = new azure.storage.Blob("example", {
    name: "example",
    storageAccountName: exampleAccount.name,
    storageContainerName: exampleContainer.name,
    type: "Block",
    source: new pulumi.asset.FileAsset("example.csv"),
});
const exampleJob = new azure.streamanalytics.Job("example", {
    name: "example-job",
    resourceGroupName: example.name,
    location: example.location,
    compatibilityLevel: "1.2",
    dataLocale: "en-GB",
    eventsLateArrivalMaxDelayInSeconds: 60,
    eventsOutOfOrderMaxDelayInSeconds: 50,
    eventsOutOfOrderPolicy: "Adjust",
    outputErrorPolicy: "Drop",
    streamingUnits: 3,
    tags: {
        environment: "Example",
    },
    transformationQuery: `    SELECT *
    INTO [exampleoutput]
    FROM [exampleinput]
`,
});
const exampleStreamInputBlob = new azure.streamanalytics.StreamInputBlob("example", {
    name: "exampleinput",
    streamAnalyticsJobName: exampleJob.name,
    resourceGroupName: exampleJob.resourceGroupName,
    storageAccountName: exampleAccount.name,
    storageAccountKey: exampleAccount.primaryAccessKey,
    storageContainerName: exampleContainer.name,
    pathPattern: "",
    dateFormat: "yyyy/MM/dd",
    timeFormat: "HH",
    serialization: {
        type: "Csv",
        encoding: "UTF8",
        fieldDelimiter: ",",
    },
});
const exampleOutputBlob = new azure.streamanalytics.OutputBlob("example", {
    name: "exampleoutput",
    streamAnalyticsJobName: exampleJob.name,
    resourceGroupName: exampleJob.resourceGroupName,
    storageAccountName: exampleAccount.name,
    storageAccountKey: exampleAccount.primaryAccessKey,
    storageContainerName: exampleContainer.name,
    pathPattern: "example-{date}-{time}",
    dateFormat: "yyyy-MM-dd",
    timeFormat: "HH",
    serialization: {
        type: "Avro",
    },
});
const exampleJobSchedule = new azure.streamanalytics.JobSchedule("example", {
    streamAnalyticsJobId: exampleJob.id,
    startMode: "CustomTime",
    startTime: "2022-09-21T00:00:00Z",
}, {
    dependsOn: [
        exampleJob,
        exampleStreamInputBlob,
        exampleOutputBlob,
    ],
});
Copy
import pulumi
import pulumi_azure as azure

example = azure.core.ResourceGroup("example",
    name="example-resources",
    location="West Europe")
example_account = azure.storage.Account("example",
    name="example",
    resource_group_name=example.name,
    location=example.location,
    account_tier="Standard",
    account_replication_type="LRS")
example_container = azure.storage.Container("example",
    name="example",
    storage_account_name=example_account.name,
    container_access_type="private")
example_blob = azure.storage.Blob("example",
    name="example",
    storage_account_name=example_account.name,
    storage_container_name=example_container.name,
    type="Block",
    source=pulumi.FileAsset("example.csv"))
example_job = azure.streamanalytics.Job("example",
    name="example-job",
    resource_group_name=example.name,
    location=example.location,
    compatibility_level="1.2",
    data_locale="en-GB",
    events_late_arrival_max_delay_in_seconds=60,
    events_out_of_order_max_delay_in_seconds=50,
    events_out_of_order_policy="Adjust",
    output_error_policy="Drop",
    streaming_units=3,
    tags={
        "environment": "Example",
    },
    transformation_query="""    SELECT *
    INTO [exampleoutput]
    FROM [exampleinput]
""")
example_stream_input_blob = azure.streamanalytics.StreamInputBlob("example",
    name="exampleinput",
    stream_analytics_job_name=example_job.name,
    resource_group_name=example_job.resource_group_name,
    storage_account_name=example_account.name,
    storage_account_key=example_account.primary_access_key,
    storage_container_name=example_container.name,
    path_pattern="",
    date_format="yyyy/MM/dd",
    time_format="HH",
    serialization={
        "type": "Csv",
        "encoding": "UTF8",
        "field_delimiter": ",",
    })
example_output_blob = azure.streamanalytics.OutputBlob("example",
    name="exampleoutput",
    stream_analytics_job_name=example_job.name,
    resource_group_name=example_job.resource_group_name,
    storage_account_name=example_account.name,
    storage_account_key=example_account.primary_access_key,
    storage_container_name=example_container.name,
    path_pattern="example-{date}-{time}",
    date_format="yyyy-MM-dd",
    time_format="HH",
    serialization={
        "type": "Avro",
    })
example_job_schedule = azure.streamanalytics.JobSchedule("example",
    stream_analytics_job_id=example_job.id,
    start_mode="CustomTime",
    start_time="2022-09-21T00:00:00Z",
    opts = pulumi.ResourceOptions(depends_on=[
            example_job,
            example_stream_input_blob,
            example_output_blob,
        ]))
Copy
package main

import (
	"github.com/pulumi/pulumi-azure/sdk/v6/go/azure/core"
	"github.com/pulumi/pulumi-azure/sdk/v6/go/azure/storage"
	"github.com/pulumi/pulumi-azure/sdk/v6/go/azure/streamanalytics"
	"github.com/pulumi/pulumi/sdk/v3/go/pulumi"
)

func main() {
	pulumi.Run(func(ctx *pulumi.Context) error {
		example, err := core.NewResourceGroup(ctx, "example", &core.ResourceGroupArgs{
			Name:     pulumi.String("example-resources"),
			Location: pulumi.String("West Europe"),
		})
		if err != nil {
			return err
		}
		exampleAccount, err := storage.NewAccount(ctx, "example", &storage.AccountArgs{
			Name:                   pulumi.String("example"),
			ResourceGroupName:      example.Name,
			Location:               example.Location,
			AccountTier:            pulumi.String("Standard"),
			AccountReplicationType: pulumi.String("LRS"),
		})
		if err != nil {
			return err
		}
		exampleContainer, err := storage.NewContainer(ctx, "example", &storage.ContainerArgs{
			Name:                pulumi.String("example"),
			StorageAccountName:  exampleAccount.Name,
			ContainerAccessType: pulumi.String("private"),
		})
		if err != nil {
			return err
		}
		_, err = storage.NewBlob(ctx, "example", &storage.BlobArgs{
			Name:                 pulumi.String("example"),
			StorageAccountName:   exampleAccount.Name,
			StorageContainerName: exampleContainer.Name,
			Type:                 pulumi.String("Block"),
			Source:               pulumi.NewFileAsset("example.csv"),
		})
		if err != nil {
			return err
		}
		exampleJob, err := streamanalytics.NewJob(ctx, "example", &streamanalytics.JobArgs{
			Name:                               pulumi.String("example-job"),
			ResourceGroupName:                  example.Name,
			Location:                           example.Location,
			CompatibilityLevel:                 pulumi.String("1.2"),
			DataLocale:                         pulumi.String("en-GB"),
			EventsLateArrivalMaxDelayInSeconds: pulumi.Int(60),
			EventsOutOfOrderMaxDelayInSeconds:  pulumi.Int(50),
			EventsOutOfOrderPolicy:             pulumi.String("Adjust"),
			OutputErrorPolicy:                  pulumi.String("Drop"),
			StreamingUnits:                     pulumi.Int(3),
			Tags: pulumi.StringMap{
				"environment": pulumi.String("Example"),
			},
			TransformationQuery: pulumi.String("    SELECT *\n    INTO [exampleoutput]\n    FROM [exampleinput]\n"),
		})
		if err != nil {
			return err
		}
		exampleStreamInputBlob, err := streamanalytics.NewStreamInputBlob(ctx, "example", &streamanalytics.StreamInputBlobArgs{
			Name:                   pulumi.String("exampleinput"),
			StreamAnalyticsJobName: exampleJob.Name,
			ResourceGroupName:      exampleJob.ResourceGroupName,
			StorageAccountName:     exampleAccount.Name,
			StorageAccountKey:      exampleAccount.PrimaryAccessKey,
			StorageContainerName:   exampleContainer.Name,
			PathPattern:            pulumi.String(""),
			DateFormat:             pulumi.String("yyyy/MM/dd"),
			TimeFormat:             pulumi.String("HH"),
			Serialization: &streamanalytics.StreamInputBlobSerializationArgs{
				Type:           pulumi.String("Csv"),
				Encoding:       pulumi.String("UTF8"),
				FieldDelimiter: pulumi.String(","),
			},
		})
		if err != nil {
			return err
		}
		exampleOutputBlob, err := streamanalytics.NewOutputBlob(ctx, "example", &streamanalytics.OutputBlobArgs{
			Name:                   pulumi.String("exampleoutput"),
			StreamAnalyticsJobName: exampleJob.Name,
			ResourceGroupName:      exampleJob.ResourceGroupName,
			StorageAccountName:     exampleAccount.Name,
			StorageAccountKey:      exampleAccount.PrimaryAccessKey,
			StorageContainerName:   exampleContainer.Name,
			PathPattern:            pulumi.String("example-{date}-{time}"),
			DateFormat:             pulumi.String("yyyy-MM-dd"),
			TimeFormat:             pulumi.String("HH"),
			Serialization: &streamanalytics.OutputBlobSerializationArgs{
				Type: pulumi.String("Avro"),
			},
		})
		if err != nil {
			return err
		}
		_, err = streamanalytics.NewJobSchedule(ctx, "example", &streamanalytics.JobScheduleArgs{
			StreamAnalyticsJobId: exampleJob.ID(),
			StartMode:            pulumi.String("CustomTime"),
			StartTime:            pulumi.String("2022-09-21T00:00:00Z"),
		}, pulumi.DependsOn([]pulumi.Resource{
			exampleJob,
			exampleStreamInputBlob,
			exampleOutputBlob,
		}))
		if err != nil {
			return err
		}
		return nil
	})
}
Copy
using System.Collections.Generic;
using System.Linq;
using Pulumi;
using Azure = Pulumi.Azure;

return await Deployment.RunAsync(() => 
{
    var example = new Azure.Core.ResourceGroup("example", new()
    {
        Name = "example-resources",
        Location = "West Europe",
    });

    var exampleAccount = new Azure.Storage.Account("example", new()
    {
        Name = "example",
        ResourceGroupName = example.Name,
        Location = example.Location,
        AccountTier = "Standard",
        AccountReplicationType = "LRS",
    });

    var exampleContainer = new Azure.Storage.Container("example", new()
    {
        Name = "example",
        StorageAccountName = exampleAccount.Name,
        ContainerAccessType = "private",
    });

    var exampleBlob = new Azure.Storage.Blob("example", new()
    {
        Name = "example",
        StorageAccountName = exampleAccount.Name,
        StorageContainerName = exampleContainer.Name,
        Type = "Block",
        Source = new FileAsset("example.csv"),
    });

    var exampleJob = new Azure.StreamAnalytics.Job("example", new()
    {
        Name = "example-job",
        ResourceGroupName = example.Name,
        Location = example.Location,
        CompatibilityLevel = "1.2",
        DataLocale = "en-GB",
        EventsLateArrivalMaxDelayInSeconds = 60,
        EventsOutOfOrderMaxDelayInSeconds = 50,
        EventsOutOfOrderPolicy = "Adjust",
        OutputErrorPolicy = "Drop",
        StreamingUnits = 3,
        Tags = 
        {
            { "environment", "Example" },
        },
        TransformationQuery = @"    SELECT *
    INTO [exampleoutput]
    FROM [exampleinput]
",
    });

    var exampleStreamInputBlob = new Azure.StreamAnalytics.StreamInputBlob("example", new()
    {
        Name = "exampleinput",
        StreamAnalyticsJobName = exampleJob.Name,
        ResourceGroupName = exampleJob.ResourceGroupName,
        StorageAccountName = exampleAccount.Name,
        StorageAccountKey = exampleAccount.PrimaryAccessKey,
        StorageContainerName = exampleContainer.Name,
        PathPattern = "",
        DateFormat = "yyyy/MM/dd",
        TimeFormat = "HH",
        Serialization = new Azure.StreamAnalytics.Inputs.StreamInputBlobSerializationArgs
        {
            Type = "Csv",
            Encoding = "UTF8",
            FieldDelimiter = ",",
        },
    });

    var exampleOutputBlob = new Azure.StreamAnalytics.OutputBlob("example", new()
    {
        Name = "exampleoutput",
        StreamAnalyticsJobName = exampleJob.Name,
        ResourceGroupName = exampleJob.ResourceGroupName,
        StorageAccountName = exampleAccount.Name,
        StorageAccountKey = exampleAccount.PrimaryAccessKey,
        StorageContainerName = exampleContainer.Name,
        PathPattern = "example-{date}-{time}",
        DateFormat = "yyyy-MM-dd",
        TimeFormat = "HH",
        Serialization = new Azure.StreamAnalytics.Inputs.OutputBlobSerializationArgs
        {
            Type = "Avro",
        },
    });

    var exampleJobSchedule = new Azure.StreamAnalytics.JobSchedule("example", new()
    {
        StreamAnalyticsJobId = exampleJob.Id,
        StartMode = "CustomTime",
        StartTime = "2022-09-21T00:00:00Z",
    }, new CustomResourceOptions
    {
        DependsOn =
        {
            exampleJob,
            exampleStreamInputBlob,
            exampleOutputBlob,
        },
    });

});
Copy
package generated_program;

import com.pulumi.Context;
import com.pulumi.Pulumi;
import com.pulumi.core.Output;
import com.pulumi.azure.core.ResourceGroup;
import com.pulumi.azure.core.ResourceGroupArgs;
import com.pulumi.azure.storage.Account;
import com.pulumi.azure.storage.AccountArgs;
import com.pulumi.azure.storage.Container;
import com.pulumi.azure.storage.ContainerArgs;
import com.pulumi.azure.storage.Blob;
import com.pulumi.azure.storage.BlobArgs;
import com.pulumi.azure.streamanalytics.Job;
import com.pulumi.azure.streamanalytics.JobArgs;
import com.pulumi.azure.streamanalytics.StreamInputBlob;
import com.pulumi.azure.streamanalytics.StreamInputBlobArgs;
import com.pulumi.azure.streamanalytics.inputs.StreamInputBlobSerializationArgs;
import com.pulumi.azure.streamanalytics.OutputBlob;
import com.pulumi.azure.streamanalytics.OutputBlobArgs;
import com.pulumi.azure.streamanalytics.inputs.OutputBlobSerializationArgs;
import com.pulumi.azure.streamanalytics.JobSchedule;
import com.pulumi.azure.streamanalytics.JobScheduleArgs;
import com.pulumi.resources.CustomResourceOptions;
import com.pulumi.asset.FileAsset;
import java.util.List;
import java.util.ArrayList;
import java.util.Map;
import java.io.File;
import java.nio.file.Files;
import java.nio.file.Paths;

public class App {
    public static void main(String[] args) {
        Pulumi.run(App::stack);
    }

    public static void stack(Context ctx) {
        var example = new ResourceGroup("example", ResourceGroupArgs.builder()
            .name("example-resources")
            .location("West Europe")
            .build());

        var exampleAccount = new Account("exampleAccount", AccountArgs.builder()
            .name("example")
            .resourceGroupName(example.name())
            .location(example.location())
            .accountTier("Standard")
            .accountReplicationType("LRS")
            .build());

        var exampleContainer = new Container("exampleContainer", ContainerArgs.builder()
            .name("example")
            .storageAccountName(exampleAccount.name())
            .containerAccessType("private")
            .build());

        var exampleBlob = new Blob("exampleBlob", BlobArgs.builder()
            .name("example")
            .storageAccountName(exampleAccount.name())
            .storageContainerName(exampleContainer.name())
            .type("Block")
            .source(new FileAsset("example.csv"))
            .build());

        var exampleJob = new Job("exampleJob", JobArgs.builder()
            .name("example-job")
            .resourceGroupName(example.name())
            .location(example.location())
            .compatibilityLevel("1.2")
            .dataLocale("en-GB")
            .eventsLateArrivalMaxDelayInSeconds(60)
            .eventsOutOfOrderMaxDelayInSeconds(50)
            .eventsOutOfOrderPolicy("Adjust")
            .outputErrorPolicy("Drop")
            .streamingUnits(3)
            .tags(Map.of("environment", "Example"))
            .transformationQuery("""
    SELECT *
    INTO [exampleoutput]
    FROM [exampleinput]
            """)
            .build());

        var exampleStreamInputBlob = new StreamInputBlob("exampleStreamInputBlob", StreamInputBlobArgs.builder()
            .name("exampleinput")
            .streamAnalyticsJobName(exampleJob.name())
            .resourceGroupName(exampleJob.resourceGroupName())
            .storageAccountName(exampleAccount.name())
            .storageAccountKey(exampleAccount.primaryAccessKey())
            .storageContainerName(exampleContainer.name())
            .pathPattern("")
            .dateFormat("yyyy/MM/dd")
            .timeFormat("HH")
            .serialization(StreamInputBlobSerializationArgs.builder()
                .type("Csv")
                .encoding("UTF8")
                .fieldDelimiter(",")
                .build())
            .build());

        var exampleOutputBlob = new OutputBlob("exampleOutputBlob", OutputBlobArgs.builder()
            .name("exampleoutput")
            .streamAnalyticsJobName(exampleJob.name())
            .resourceGroupName(exampleJob.resourceGroupName())
            .storageAccountName(exampleAccount.name())
            .storageAccountKey(exampleAccount.primaryAccessKey())
            .storageContainerName(exampleContainer.name())
            .pathPattern("example-{date}-{time}")
            .dateFormat("yyyy-MM-dd")
            .timeFormat("HH")
            .serialization(OutputBlobSerializationArgs.builder()
                .type("Avro")
                .build())
            .build());

        var exampleJobSchedule = new JobSchedule("exampleJobSchedule", JobScheduleArgs.builder()
            .streamAnalyticsJobId(exampleJob.id())
            .startMode("CustomTime")
            .startTime("2022-09-21T00:00:00Z")
            .build(), CustomResourceOptions.builder()
                .dependsOn(                
                    exampleJob,
                    exampleStreamInputBlob,
                    exampleOutputBlob)
                .build());

    }
}
Copy
resources:
  example:
    type: azure:core:ResourceGroup
    properties:
      name: example-resources
      location: West Europe
  exampleAccount:
    type: azure:storage:Account
    name: example
    properties:
      name: example
      resourceGroupName: ${example.name}
      location: ${example.location}
      accountTier: Standard
      accountReplicationType: LRS
  exampleContainer:
    type: azure:storage:Container
    name: example
    properties:
      name: example
      storageAccountName: ${exampleAccount.name}
      containerAccessType: private
  exampleBlob:
    type: azure:storage:Blob
    name: example
    properties:
      name: example
      storageAccountName: ${exampleAccount.name}
      storageContainerName: ${exampleContainer.name}
      type: Block
      source:
        fn::FileAsset: example.csv
  exampleJob:
    type: azure:streamanalytics:Job
    name: example
    properties:
      name: example-job
      resourceGroupName: ${example.name}
      location: ${example.location}
      compatibilityLevel: '1.2'
      dataLocale: en-GB
      eventsLateArrivalMaxDelayInSeconds: 60
      eventsOutOfOrderMaxDelayInSeconds: 50
      eventsOutOfOrderPolicy: Adjust
      outputErrorPolicy: Drop
      streamingUnits: 3
      tags:
        environment: Example
      transformationQuery: |2
            SELECT *
            INTO [exampleoutput]
            FROM [exampleinput]
  exampleStreamInputBlob:
    type: azure:streamanalytics:StreamInputBlob
    name: example
    properties:
      name: exampleinput
      streamAnalyticsJobName: ${exampleJob.name}
      resourceGroupName: ${exampleJob.resourceGroupName}
      storageAccountName: ${exampleAccount.name}
      storageAccountKey: ${exampleAccount.primaryAccessKey}
      storageContainerName: ${exampleContainer.name}
      pathPattern: ""
      dateFormat: yyyy/MM/dd
      timeFormat: HH
      serialization:
        type: Csv
        encoding: UTF8
        fieldDelimiter: ','
  exampleOutputBlob:
    type: azure:streamanalytics:OutputBlob
    name: example
    properties:
      name: exampleoutput
      streamAnalyticsJobName: ${exampleJob.name}
      resourceGroupName: ${exampleJob.resourceGroupName}
      storageAccountName: ${exampleAccount.name}
      storageAccountKey: ${exampleAccount.primaryAccessKey}
      storageContainerName: ${exampleContainer.name}
      pathPattern: example-{date}-{time}
      dateFormat: yyyy-MM-dd
      timeFormat: HH
      serialization:
        type: Avro
  exampleJobSchedule:
    type: azure:streamanalytics:JobSchedule
    name: example
    properties:
      streamAnalyticsJobId: ${exampleJob.id}
      startMode: CustomTime
      startTime: 2022-09-21T00:00:00Z
    options:
      dependsOn:
        - ${exampleJob}
        - ${exampleStreamInputBlob}
        - ${exampleOutputBlob}
Copy

Create JobSchedule Resource

Resources are created with functions called constructors. To learn more about declaring and configuring resources, see Resources.

Constructor syntax

new JobSchedule(name: string, args: JobScheduleArgs, opts?: CustomResourceOptions);
@overload
def JobSchedule(resource_name: str,
                args: JobScheduleArgs,
                opts: Optional[ResourceOptions] = None)

@overload
def JobSchedule(resource_name: str,
                opts: Optional[ResourceOptions] = None,
                start_mode: Optional[str] = None,
                stream_analytics_job_id: Optional[str] = None,
                start_time: Optional[str] = None)
func NewJobSchedule(ctx *Context, name string, args JobScheduleArgs, opts ...ResourceOption) (*JobSchedule, error)
public JobSchedule(string name, JobScheduleArgs args, CustomResourceOptions? opts = null)
public JobSchedule(String name, JobScheduleArgs args)
public JobSchedule(String name, JobScheduleArgs args, CustomResourceOptions options)
type: azure:streamanalytics:JobSchedule
properties: # The arguments to resource properties.
options: # Bag of options to control resource's behavior.

Parameters

name This property is required. string
The unique name of the resource.
args This property is required. JobScheduleArgs
The arguments to resource properties.
opts CustomResourceOptions
Bag of options to control resource's behavior.
resource_name This property is required. str
The unique name of the resource.
args This property is required. JobScheduleArgs
The arguments to resource properties.
opts ResourceOptions
Bag of options to control resource's behavior.
ctx Context
Context object for the current deployment.
name This property is required. string
The unique name of the resource.
args This property is required. JobScheduleArgs
The arguments to resource properties.
opts ResourceOption
Bag of options to control resource's behavior.
name This property is required. string
The unique name of the resource.
args This property is required. JobScheduleArgs
The arguments to resource properties.
opts CustomResourceOptions
Bag of options to control resource's behavior.
name This property is required. String
The unique name of the resource.
args This property is required. JobScheduleArgs
The arguments to resource properties.
options CustomResourceOptions
Bag of options to control resource's behavior.

Constructor example

The following reference example uses placeholder values for all input properties.

var examplejobScheduleResourceResourceFromStreamanalyticsjobSchedule = new Azure.StreamAnalytics.JobSchedule("examplejobScheduleResourceResourceFromStreamanalyticsjobSchedule", new()
{
    StartMode = "string",
    StreamAnalyticsJobId = "string",
    StartTime = "string",
});
Copy
example, err := streamanalytics.NewJobSchedule(ctx, "examplejobScheduleResourceResourceFromStreamanalyticsjobSchedule", &streamanalytics.JobScheduleArgs{
	StartMode:            pulumi.String("string"),
	StreamAnalyticsJobId: pulumi.String("string"),
	StartTime:            pulumi.String("string"),
})
Copy
var examplejobScheduleResourceResourceFromStreamanalyticsjobSchedule = new JobSchedule("examplejobScheduleResourceResourceFromStreamanalyticsjobSchedule", JobScheduleArgs.builder()
    .startMode("string")
    .streamAnalyticsJobId("string")
    .startTime("string")
    .build());
Copy
examplejob_schedule_resource_resource_from_streamanalyticsjob_schedule = azure.streamanalytics.JobSchedule("examplejobScheduleResourceResourceFromStreamanalyticsjobSchedule",
    start_mode="string",
    stream_analytics_job_id="string",
    start_time="string")
Copy
const examplejobScheduleResourceResourceFromStreamanalyticsjobSchedule = new azure.streamanalytics.JobSchedule("examplejobScheduleResourceResourceFromStreamanalyticsjobSchedule", {
    startMode: "string",
    streamAnalyticsJobId: "string",
    startTime: "string",
});
Copy
type: azure:streamanalytics:JobSchedule
properties:
    startMode: string
    startTime: string
    streamAnalyticsJobId: string
Copy

JobSchedule Resource Properties

To learn more about resource properties and how to use them, see Inputs and Outputs in the Architecture and Concepts docs.

Inputs

In Python, inputs that are objects can be passed either as argument classes or as dictionary literals.

The JobSchedule resource accepts the following input properties:

StartMode This property is required. string

The starting mode of the Stream Analytics Job. Possible values are JobStartTime, CustomTime and LastOutputEventTime.

Note: Setting start_mode to LastOutputEventTime is only possible if the job had been previously started and produced output.

StreamAnalyticsJobId
This property is required.
Changes to this property will trigger replacement.
string
The ID of the Stream Analytics Job that should be scheduled or started. Changing this forces a new resource to be created.
StartTime string
The time in ISO8601 format at which the Stream Analytics Job should be started e.g. 2022-04-01T00:00:00Z. This property can only be specified if start_mode is set to CustomTime
StartMode This property is required. string

The starting mode of the Stream Analytics Job. Possible values are JobStartTime, CustomTime and LastOutputEventTime.

Note: Setting start_mode to LastOutputEventTime is only possible if the job had been previously started and produced output.

StreamAnalyticsJobId
This property is required.
Changes to this property will trigger replacement.
string
The ID of the Stream Analytics Job that should be scheduled or started. Changing this forces a new resource to be created.
StartTime string
The time in ISO8601 format at which the Stream Analytics Job should be started e.g. 2022-04-01T00:00:00Z. This property can only be specified if start_mode is set to CustomTime
startMode This property is required. String

The starting mode of the Stream Analytics Job. Possible values are JobStartTime, CustomTime and LastOutputEventTime.

Note: Setting start_mode to LastOutputEventTime is only possible if the job had been previously started and produced output.

streamAnalyticsJobId
This property is required.
Changes to this property will trigger replacement.
String
The ID of the Stream Analytics Job that should be scheduled or started. Changing this forces a new resource to be created.
startTime String
The time in ISO8601 format at which the Stream Analytics Job should be started e.g. 2022-04-01T00:00:00Z. This property can only be specified if start_mode is set to CustomTime
startMode This property is required. string

The starting mode of the Stream Analytics Job. Possible values are JobStartTime, CustomTime and LastOutputEventTime.

Note: Setting start_mode to LastOutputEventTime is only possible if the job had been previously started and produced output.

streamAnalyticsJobId
This property is required.
Changes to this property will trigger replacement.
string
The ID of the Stream Analytics Job that should be scheduled or started. Changing this forces a new resource to be created.
startTime string
The time in ISO8601 format at which the Stream Analytics Job should be started e.g. 2022-04-01T00:00:00Z. This property can only be specified if start_mode is set to CustomTime
start_mode This property is required. str

The starting mode of the Stream Analytics Job. Possible values are JobStartTime, CustomTime and LastOutputEventTime.

Note: Setting start_mode to LastOutputEventTime is only possible if the job had been previously started and produced output.

stream_analytics_job_id
This property is required.
Changes to this property will trigger replacement.
str
The ID of the Stream Analytics Job that should be scheduled or started. Changing this forces a new resource to be created.
start_time str
The time in ISO8601 format at which the Stream Analytics Job should be started e.g. 2022-04-01T00:00:00Z. This property can only be specified if start_mode is set to CustomTime
startMode This property is required. String

The starting mode of the Stream Analytics Job. Possible values are JobStartTime, CustomTime and LastOutputEventTime.

Note: Setting start_mode to LastOutputEventTime is only possible if the job had been previously started and produced output.

streamAnalyticsJobId
This property is required.
Changes to this property will trigger replacement.
String
The ID of the Stream Analytics Job that should be scheduled or started. Changing this forces a new resource to be created.
startTime String
The time in ISO8601 format at which the Stream Analytics Job should be started e.g. 2022-04-01T00:00:00Z. This property can only be specified if start_mode is set to CustomTime

Outputs

All input properties are implicitly available as output properties. Additionally, the JobSchedule resource produces the following output properties:

Id string
The provider-assigned unique ID for this managed resource.
LastOutputTime string
The time at which the Stream Analytics job last produced an output.
Id string
The provider-assigned unique ID for this managed resource.
LastOutputTime string
The time at which the Stream Analytics job last produced an output.
id String
The provider-assigned unique ID for this managed resource.
lastOutputTime String
The time at which the Stream Analytics job last produced an output.
id string
The provider-assigned unique ID for this managed resource.
lastOutputTime string
The time at which the Stream Analytics job last produced an output.
id str
The provider-assigned unique ID for this managed resource.
last_output_time str
The time at which the Stream Analytics job last produced an output.
id String
The provider-assigned unique ID for this managed resource.
lastOutputTime String
The time at which the Stream Analytics job last produced an output.

Look up Existing JobSchedule Resource

Get an existing JobSchedule resource’s state with the given name, ID, and optional extra properties used to qualify the lookup.

public static get(name: string, id: Input<ID>, state?: JobScheduleState, opts?: CustomResourceOptions): JobSchedule
@staticmethod
def get(resource_name: str,
        id: str,
        opts: Optional[ResourceOptions] = None,
        last_output_time: Optional[str] = None,
        start_mode: Optional[str] = None,
        start_time: Optional[str] = None,
        stream_analytics_job_id: Optional[str] = None) -> JobSchedule
func GetJobSchedule(ctx *Context, name string, id IDInput, state *JobScheduleState, opts ...ResourceOption) (*JobSchedule, error)
public static JobSchedule Get(string name, Input<string> id, JobScheduleState? state, CustomResourceOptions? opts = null)
public static JobSchedule get(String name, Output<String> id, JobScheduleState state, CustomResourceOptions options)
resources:  _:    type: azure:streamanalytics:JobSchedule    get:      id: ${id}
name This property is required.
The unique name of the resulting resource.
id This property is required.
The unique provider ID of the resource to lookup.
state
Any extra arguments used during the lookup.
opts
A bag of options that control this resource's behavior.
resource_name This property is required.
The unique name of the resulting resource.
id This property is required.
The unique provider ID of the resource to lookup.
name This property is required.
The unique name of the resulting resource.
id This property is required.
The unique provider ID of the resource to lookup.
state
Any extra arguments used during the lookup.
opts
A bag of options that control this resource's behavior.
name This property is required.
The unique name of the resulting resource.
id This property is required.
The unique provider ID of the resource to lookup.
state
Any extra arguments used during the lookup.
opts
A bag of options that control this resource's behavior.
name This property is required.
The unique name of the resulting resource.
id This property is required.
The unique provider ID of the resource to lookup.
state
Any extra arguments used during the lookup.
opts
A bag of options that control this resource's behavior.
The following state arguments are supported:
LastOutputTime string
The time at which the Stream Analytics job last produced an output.
StartMode string

The starting mode of the Stream Analytics Job. Possible values are JobStartTime, CustomTime and LastOutputEventTime.

Note: Setting start_mode to LastOutputEventTime is only possible if the job had been previously started and produced output.

StartTime string
The time in ISO8601 format at which the Stream Analytics Job should be started e.g. 2022-04-01T00:00:00Z. This property can only be specified if start_mode is set to CustomTime
StreamAnalyticsJobId Changes to this property will trigger replacement. string
The ID of the Stream Analytics Job that should be scheduled or started. Changing this forces a new resource to be created.
LastOutputTime string
The time at which the Stream Analytics job last produced an output.
StartMode string

The starting mode of the Stream Analytics Job. Possible values are JobStartTime, CustomTime and LastOutputEventTime.

Note: Setting start_mode to LastOutputEventTime is only possible if the job had been previously started and produced output.

StartTime string
The time in ISO8601 format at which the Stream Analytics Job should be started e.g. 2022-04-01T00:00:00Z. This property can only be specified if start_mode is set to CustomTime
StreamAnalyticsJobId Changes to this property will trigger replacement. string
The ID of the Stream Analytics Job that should be scheduled or started. Changing this forces a new resource to be created.
lastOutputTime String
The time at which the Stream Analytics job last produced an output.
startMode String

The starting mode of the Stream Analytics Job. Possible values are JobStartTime, CustomTime and LastOutputEventTime.

Note: Setting start_mode to LastOutputEventTime is only possible if the job had been previously started and produced output.

startTime String
The time in ISO8601 format at which the Stream Analytics Job should be started e.g. 2022-04-01T00:00:00Z. This property can only be specified if start_mode is set to CustomTime
streamAnalyticsJobId Changes to this property will trigger replacement. String
The ID of the Stream Analytics Job that should be scheduled or started. Changing this forces a new resource to be created.
lastOutputTime string
The time at which the Stream Analytics job last produced an output.
startMode string

The starting mode of the Stream Analytics Job. Possible values are JobStartTime, CustomTime and LastOutputEventTime.

Note: Setting start_mode to LastOutputEventTime is only possible if the job had been previously started and produced output.

startTime string
The time in ISO8601 format at which the Stream Analytics Job should be started e.g. 2022-04-01T00:00:00Z. This property can only be specified if start_mode is set to CustomTime
streamAnalyticsJobId Changes to this property will trigger replacement. string
The ID of the Stream Analytics Job that should be scheduled or started. Changing this forces a new resource to be created.
last_output_time str
The time at which the Stream Analytics job last produced an output.
start_mode str

The starting mode of the Stream Analytics Job. Possible values are JobStartTime, CustomTime and LastOutputEventTime.

Note: Setting start_mode to LastOutputEventTime is only possible if the job had been previously started and produced output.

start_time str
The time in ISO8601 format at which the Stream Analytics Job should be started e.g. 2022-04-01T00:00:00Z. This property can only be specified if start_mode is set to CustomTime
stream_analytics_job_id Changes to this property will trigger replacement. str
The ID of the Stream Analytics Job that should be scheduled or started. Changing this forces a new resource to be created.
lastOutputTime String
The time at which the Stream Analytics job last produced an output.
startMode String

The starting mode of the Stream Analytics Job. Possible values are JobStartTime, CustomTime and LastOutputEventTime.

Note: Setting start_mode to LastOutputEventTime is only possible if the job had been previously started and produced output.

startTime String
The time in ISO8601 format at which the Stream Analytics Job should be started e.g. 2022-04-01T00:00:00Z. This property can only be specified if start_mode is set to CustomTime
streamAnalyticsJobId Changes to this property will trigger replacement. String
The ID of the Stream Analytics Job that should be scheduled or started. Changing this forces a new resource to be created.

Import

Stream Analytics Job’s can be imported using the resource id, e.g.

$ pulumi import azure:streamanalytics/jobSchedule:JobSchedule example /subscriptions/00000000-0000-0000-0000-000000000000/resourceGroups/group1/providers/Microsoft.StreamAnalytics/streamingJobs/job1/schedule/default
Copy

To learn more about importing existing cloud resources, see Importing resources.

Package Details

Repository
Azure Classic pulumi/pulumi-azure
License
Apache-2.0
Notes
This Pulumi package is based on the azurerm Terraform Provider.