We recommend using Azure Native.
azure.streamanalytics.JobSchedule
Explore with Pulumi AI
Manages a Stream Analytics Job Schedule.
Example Usage
import * as pulumi from "@pulumi/pulumi";
import * as azure from "@pulumi/azure";
const example = new azure.core.ResourceGroup("example", {
name: "example-resources",
location: "West Europe",
});
const exampleAccount = new azure.storage.Account("example", {
name: "example",
resourceGroupName: example.name,
location: example.location,
accountTier: "Standard",
accountReplicationType: "LRS",
});
const exampleContainer = new azure.storage.Container("example", {
name: "example",
storageAccountName: exampleAccount.name,
containerAccessType: "private",
});
const exampleBlob = new azure.storage.Blob("example", {
name: "example",
storageAccountName: exampleAccount.name,
storageContainerName: exampleContainer.name,
type: "Block",
source: new pulumi.asset.FileAsset("example.csv"),
});
const exampleJob = new azure.streamanalytics.Job("example", {
name: "example-job",
resourceGroupName: example.name,
location: example.location,
compatibilityLevel: "1.2",
dataLocale: "en-GB",
eventsLateArrivalMaxDelayInSeconds: 60,
eventsOutOfOrderMaxDelayInSeconds: 50,
eventsOutOfOrderPolicy: "Adjust",
outputErrorPolicy: "Drop",
streamingUnits: 3,
tags: {
environment: "Example",
},
transformationQuery: ` SELECT *
INTO [exampleoutput]
FROM [exampleinput]
`,
});
const exampleStreamInputBlob = new azure.streamanalytics.StreamInputBlob("example", {
name: "exampleinput",
streamAnalyticsJobName: exampleJob.name,
resourceGroupName: exampleJob.resourceGroupName,
storageAccountName: exampleAccount.name,
storageAccountKey: exampleAccount.primaryAccessKey,
storageContainerName: exampleContainer.name,
pathPattern: "",
dateFormat: "yyyy/MM/dd",
timeFormat: "HH",
serialization: {
type: "Csv",
encoding: "UTF8",
fieldDelimiter: ",",
},
});
const exampleOutputBlob = new azure.streamanalytics.OutputBlob("example", {
name: "exampleoutput",
streamAnalyticsJobName: exampleJob.name,
resourceGroupName: exampleJob.resourceGroupName,
storageAccountName: exampleAccount.name,
storageAccountKey: exampleAccount.primaryAccessKey,
storageContainerName: exampleContainer.name,
pathPattern: "example-{date}-{time}",
dateFormat: "yyyy-MM-dd",
timeFormat: "HH",
serialization: {
type: "Avro",
},
});
const exampleJobSchedule = new azure.streamanalytics.JobSchedule("example", {
streamAnalyticsJobId: exampleJob.id,
startMode: "CustomTime",
startTime: "2022-09-21T00:00:00Z",
}, {
dependsOn: [
exampleJob,
exampleStreamInputBlob,
exampleOutputBlob,
],
});
import pulumi
import pulumi_azure as azure
example = azure.core.ResourceGroup("example",
name="example-resources",
location="West Europe")
example_account = azure.storage.Account("example",
name="example",
resource_group_name=example.name,
location=example.location,
account_tier="Standard",
account_replication_type="LRS")
example_container = azure.storage.Container("example",
name="example",
storage_account_name=example_account.name,
container_access_type="private")
example_blob = azure.storage.Blob("example",
name="example",
storage_account_name=example_account.name,
storage_container_name=example_container.name,
type="Block",
source=pulumi.FileAsset("example.csv"))
example_job = azure.streamanalytics.Job("example",
name="example-job",
resource_group_name=example.name,
location=example.location,
compatibility_level="1.2",
data_locale="en-GB",
events_late_arrival_max_delay_in_seconds=60,
events_out_of_order_max_delay_in_seconds=50,
events_out_of_order_policy="Adjust",
output_error_policy="Drop",
streaming_units=3,
tags={
"environment": "Example",
},
transformation_query=""" SELECT *
INTO [exampleoutput]
FROM [exampleinput]
""")
example_stream_input_blob = azure.streamanalytics.StreamInputBlob("example",
name="exampleinput",
stream_analytics_job_name=example_job.name,
resource_group_name=example_job.resource_group_name,
storage_account_name=example_account.name,
storage_account_key=example_account.primary_access_key,
storage_container_name=example_container.name,
path_pattern="",
date_format="yyyy/MM/dd",
time_format="HH",
serialization={
"type": "Csv",
"encoding": "UTF8",
"field_delimiter": ",",
})
example_output_blob = azure.streamanalytics.OutputBlob("example",
name="exampleoutput",
stream_analytics_job_name=example_job.name,
resource_group_name=example_job.resource_group_name,
storage_account_name=example_account.name,
storage_account_key=example_account.primary_access_key,
storage_container_name=example_container.name,
path_pattern="example-{date}-{time}",
date_format="yyyy-MM-dd",
time_format="HH",
serialization={
"type": "Avro",
})
example_job_schedule = azure.streamanalytics.JobSchedule("example",
stream_analytics_job_id=example_job.id,
start_mode="CustomTime",
start_time="2022-09-21T00:00:00Z",
opts = pulumi.ResourceOptions(depends_on=[
example_job,
example_stream_input_blob,
example_output_blob,
]))
package main
import (
"github.com/pulumi/pulumi-azure/sdk/v6/go/azure/core"
"github.com/pulumi/pulumi-azure/sdk/v6/go/azure/storage"
"github.com/pulumi/pulumi-azure/sdk/v6/go/azure/streamanalytics"
"github.com/pulumi/pulumi/sdk/v3/go/pulumi"
)
func main() {
pulumi.Run(func(ctx *pulumi.Context) error {
example, err := core.NewResourceGroup(ctx, "example", &core.ResourceGroupArgs{
Name: pulumi.String("example-resources"),
Location: pulumi.String("West Europe"),
})
if err != nil {
return err
}
exampleAccount, err := storage.NewAccount(ctx, "example", &storage.AccountArgs{
Name: pulumi.String("example"),
ResourceGroupName: example.Name,
Location: example.Location,
AccountTier: pulumi.String("Standard"),
AccountReplicationType: pulumi.String("LRS"),
})
if err != nil {
return err
}
exampleContainer, err := storage.NewContainer(ctx, "example", &storage.ContainerArgs{
Name: pulumi.String("example"),
StorageAccountName: exampleAccount.Name,
ContainerAccessType: pulumi.String("private"),
})
if err != nil {
return err
}
_, err = storage.NewBlob(ctx, "example", &storage.BlobArgs{
Name: pulumi.String("example"),
StorageAccountName: exampleAccount.Name,
StorageContainerName: exampleContainer.Name,
Type: pulumi.String("Block"),
Source: pulumi.NewFileAsset("example.csv"),
})
if err != nil {
return err
}
exampleJob, err := streamanalytics.NewJob(ctx, "example", &streamanalytics.JobArgs{
Name: pulumi.String("example-job"),
ResourceGroupName: example.Name,
Location: example.Location,
CompatibilityLevel: pulumi.String("1.2"),
DataLocale: pulumi.String("en-GB"),
EventsLateArrivalMaxDelayInSeconds: pulumi.Int(60),
EventsOutOfOrderMaxDelayInSeconds: pulumi.Int(50),
EventsOutOfOrderPolicy: pulumi.String("Adjust"),
OutputErrorPolicy: pulumi.String("Drop"),
StreamingUnits: pulumi.Int(3),
Tags: pulumi.StringMap{
"environment": pulumi.String("Example"),
},
TransformationQuery: pulumi.String(" SELECT *\n INTO [exampleoutput]\n FROM [exampleinput]\n"),
})
if err != nil {
return err
}
exampleStreamInputBlob, err := streamanalytics.NewStreamInputBlob(ctx, "example", &streamanalytics.StreamInputBlobArgs{
Name: pulumi.String("exampleinput"),
StreamAnalyticsJobName: exampleJob.Name,
ResourceGroupName: exampleJob.ResourceGroupName,
StorageAccountName: exampleAccount.Name,
StorageAccountKey: exampleAccount.PrimaryAccessKey,
StorageContainerName: exampleContainer.Name,
PathPattern: pulumi.String(""),
DateFormat: pulumi.String("yyyy/MM/dd"),
TimeFormat: pulumi.String("HH"),
Serialization: &streamanalytics.StreamInputBlobSerializationArgs{
Type: pulumi.String("Csv"),
Encoding: pulumi.String("UTF8"),
FieldDelimiter: pulumi.String(","),
},
})
if err != nil {
return err
}
exampleOutputBlob, err := streamanalytics.NewOutputBlob(ctx, "example", &streamanalytics.OutputBlobArgs{
Name: pulumi.String("exampleoutput"),
StreamAnalyticsJobName: exampleJob.Name,
ResourceGroupName: exampleJob.ResourceGroupName,
StorageAccountName: exampleAccount.Name,
StorageAccountKey: exampleAccount.PrimaryAccessKey,
StorageContainerName: exampleContainer.Name,
PathPattern: pulumi.String("example-{date}-{time}"),
DateFormat: pulumi.String("yyyy-MM-dd"),
TimeFormat: pulumi.String("HH"),
Serialization: &streamanalytics.OutputBlobSerializationArgs{
Type: pulumi.String("Avro"),
},
})
if err != nil {
return err
}
_, err = streamanalytics.NewJobSchedule(ctx, "example", &streamanalytics.JobScheduleArgs{
StreamAnalyticsJobId: exampleJob.ID(),
StartMode: pulumi.String("CustomTime"),
StartTime: pulumi.String("2022-09-21T00:00:00Z"),
}, pulumi.DependsOn([]pulumi.Resource{
exampleJob,
exampleStreamInputBlob,
exampleOutputBlob,
}))
if err != nil {
return err
}
return nil
})
}
using System.Collections.Generic;
using System.Linq;
using Pulumi;
using Azure = Pulumi.Azure;
return await Deployment.RunAsync(() =>
{
var example = new Azure.Core.ResourceGroup("example", new()
{
Name = "example-resources",
Location = "West Europe",
});
var exampleAccount = new Azure.Storage.Account("example", new()
{
Name = "example",
ResourceGroupName = example.Name,
Location = example.Location,
AccountTier = "Standard",
AccountReplicationType = "LRS",
});
var exampleContainer = new Azure.Storage.Container("example", new()
{
Name = "example",
StorageAccountName = exampleAccount.Name,
ContainerAccessType = "private",
});
var exampleBlob = new Azure.Storage.Blob("example", new()
{
Name = "example",
StorageAccountName = exampleAccount.Name,
StorageContainerName = exampleContainer.Name,
Type = "Block",
Source = new FileAsset("example.csv"),
});
var exampleJob = new Azure.StreamAnalytics.Job("example", new()
{
Name = "example-job",
ResourceGroupName = example.Name,
Location = example.Location,
CompatibilityLevel = "1.2",
DataLocale = "en-GB",
EventsLateArrivalMaxDelayInSeconds = 60,
EventsOutOfOrderMaxDelayInSeconds = 50,
EventsOutOfOrderPolicy = "Adjust",
OutputErrorPolicy = "Drop",
StreamingUnits = 3,
Tags =
{
{ "environment", "Example" },
},
TransformationQuery = @" SELECT *
INTO [exampleoutput]
FROM [exampleinput]
",
});
var exampleStreamInputBlob = new Azure.StreamAnalytics.StreamInputBlob("example", new()
{
Name = "exampleinput",
StreamAnalyticsJobName = exampleJob.Name,
ResourceGroupName = exampleJob.ResourceGroupName,
StorageAccountName = exampleAccount.Name,
StorageAccountKey = exampleAccount.PrimaryAccessKey,
StorageContainerName = exampleContainer.Name,
PathPattern = "",
DateFormat = "yyyy/MM/dd",
TimeFormat = "HH",
Serialization = new Azure.StreamAnalytics.Inputs.StreamInputBlobSerializationArgs
{
Type = "Csv",
Encoding = "UTF8",
FieldDelimiter = ",",
},
});
var exampleOutputBlob = new Azure.StreamAnalytics.OutputBlob("example", new()
{
Name = "exampleoutput",
StreamAnalyticsJobName = exampleJob.Name,
ResourceGroupName = exampleJob.ResourceGroupName,
StorageAccountName = exampleAccount.Name,
StorageAccountKey = exampleAccount.PrimaryAccessKey,
StorageContainerName = exampleContainer.Name,
PathPattern = "example-{date}-{time}",
DateFormat = "yyyy-MM-dd",
TimeFormat = "HH",
Serialization = new Azure.StreamAnalytics.Inputs.OutputBlobSerializationArgs
{
Type = "Avro",
},
});
var exampleJobSchedule = new Azure.StreamAnalytics.JobSchedule("example", new()
{
StreamAnalyticsJobId = exampleJob.Id,
StartMode = "CustomTime",
StartTime = "2022-09-21T00:00:00Z",
}, new CustomResourceOptions
{
DependsOn =
{
exampleJob,
exampleStreamInputBlob,
exampleOutputBlob,
},
});
});
package generated_program;
import com.pulumi.Context;
import com.pulumi.Pulumi;
import com.pulumi.core.Output;
import com.pulumi.azure.core.ResourceGroup;
import com.pulumi.azure.core.ResourceGroupArgs;
import com.pulumi.azure.storage.Account;
import com.pulumi.azure.storage.AccountArgs;
import com.pulumi.azure.storage.Container;
import com.pulumi.azure.storage.ContainerArgs;
import com.pulumi.azure.storage.Blob;
import com.pulumi.azure.storage.BlobArgs;
import com.pulumi.azure.streamanalytics.Job;
import com.pulumi.azure.streamanalytics.JobArgs;
import com.pulumi.azure.streamanalytics.StreamInputBlob;
import com.pulumi.azure.streamanalytics.StreamInputBlobArgs;
import com.pulumi.azure.streamanalytics.inputs.StreamInputBlobSerializationArgs;
import com.pulumi.azure.streamanalytics.OutputBlob;
import com.pulumi.azure.streamanalytics.OutputBlobArgs;
import com.pulumi.azure.streamanalytics.inputs.OutputBlobSerializationArgs;
import com.pulumi.azure.streamanalytics.JobSchedule;
import com.pulumi.azure.streamanalytics.JobScheduleArgs;
import com.pulumi.resources.CustomResourceOptions;
import com.pulumi.asset.FileAsset;
import java.util.List;
import java.util.ArrayList;
import java.util.Map;
import java.io.File;
import java.nio.file.Files;
import java.nio.file.Paths;
public class App {
public static void main(String[] args) {
Pulumi.run(App::stack);
}
public static void stack(Context ctx) {
var example = new ResourceGroup("example", ResourceGroupArgs.builder()
.name("example-resources")
.location("West Europe")
.build());
var exampleAccount = new Account("exampleAccount", AccountArgs.builder()
.name("example")
.resourceGroupName(example.name())
.location(example.location())
.accountTier("Standard")
.accountReplicationType("LRS")
.build());
var exampleContainer = new Container("exampleContainer", ContainerArgs.builder()
.name("example")
.storageAccountName(exampleAccount.name())
.containerAccessType("private")
.build());
var exampleBlob = new Blob("exampleBlob", BlobArgs.builder()
.name("example")
.storageAccountName(exampleAccount.name())
.storageContainerName(exampleContainer.name())
.type("Block")
.source(new FileAsset("example.csv"))
.build());
var exampleJob = new Job("exampleJob", JobArgs.builder()
.name("example-job")
.resourceGroupName(example.name())
.location(example.location())
.compatibilityLevel("1.2")
.dataLocale("en-GB")
.eventsLateArrivalMaxDelayInSeconds(60)
.eventsOutOfOrderMaxDelayInSeconds(50)
.eventsOutOfOrderPolicy("Adjust")
.outputErrorPolicy("Drop")
.streamingUnits(3)
.tags(Map.of("environment", "Example"))
.transformationQuery("""
SELECT *
INTO [exampleoutput]
FROM [exampleinput]
""")
.build());
var exampleStreamInputBlob = new StreamInputBlob("exampleStreamInputBlob", StreamInputBlobArgs.builder()
.name("exampleinput")
.streamAnalyticsJobName(exampleJob.name())
.resourceGroupName(exampleJob.resourceGroupName())
.storageAccountName(exampleAccount.name())
.storageAccountKey(exampleAccount.primaryAccessKey())
.storageContainerName(exampleContainer.name())
.pathPattern("")
.dateFormat("yyyy/MM/dd")
.timeFormat("HH")
.serialization(StreamInputBlobSerializationArgs.builder()
.type("Csv")
.encoding("UTF8")
.fieldDelimiter(",")
.build())
.build());
var exampleOutputBlob = new OutputBlob("exampleOutputBlob", OutputBlobArgs.builder()
.name("exampleoutput")
.streamAnalyticsJobName(exampleJob.name())
.resourceGroupName(exampleJob.resourceGroupName())
.storageAccountName(exampleAccount.name())
.storageAccountKey(exampleAccount.primaryAccessKey())
.storageContainerName(exampleContainer.name())
.pathPattern("example-{date}-{time}")
.dateFormat("yyyy-MM-dd")
.timeFormat("HH")
.serialization(OutputBlobSerializationArgs.builder()
.type("Avro")
.build())
.build());
var exampleJobSchedule = new JobSchedule("exampleJobSchedule", JobScheduleArgs.builder()
.streamAnalyticsJobId(exampleJob.id())
.startMode("CustomTime")
.startTime("2022-09-21T00:00:00Z")
.build(), CustomResourceOptions.builder()
.dependsOn(
exampleJob,
exampleStreamInputBlob,
exampleOutputBlob)
.build());
}
}
resources:
example:
type: azure:core:ResourceGroup
properties:
name: example-resources
location: West Europe
exampleAccount:
type: azure:storage:Account
name: example
properties:
name: example
resourceGroupName: ${example.name}
location: ${example.location}
accountTier: Standard
accountReplicationType: LRS
exampleContainer:
type: azure:storage:Container
name: example
properties:
name: example
storageAccountName: ${exampleAccount.name}
containerAccessType: private
exampleBlob:
type: azure:storage:Blob
name: example
properties:
name: example
storageAccountName: ${exampleAccount.name}
storageContainerName: ${exampleContainer.name}
type: Block
source:
fn::FileAsset: example.csv
exampleJob:
type: azure:streamanalytics:Job
name: example
properties:
name: example-job
resourceGroupName: ${example.name}
location: ${example.location}
compatibilityLevel: '1.2'
dataLocale: en-GB
eventsLateArrivalMaxDelayInSeconds: 60
eventsOutOfOrderMaxDelayInSeconds: 50
eventsOutOfOrderPolicy: Adjust
outputErrorPolicy: Drop
streamingUnits: 3
tags:
environment: Example
transformationQuery: |2
SELECT *
INTO [exampleoutput]
FROM [exampleinput]
exampleStreamInputBlob:
type: azure:streamanalytics:StreamInputBlob
name: example
properties:
name: exampleinput
streamAnalyticsJobName: ${exampleJob.name}
resourceGroupName: ${exampleJob.resourceGroupName}
storageAccountName: ${exampleAccount.name}
storageAccountKey: ${exampleAccount.primaryAccessKey}
storageContainerName: ${exampleContainer.name}
pathPattern:
dateFormat: yyyy/MM/dd
timeFormat: HH
serialization:
type: Csv
encoding: UTF8
fieldDelimiter: ','
exampleOutputBlob:
type: azure:streamanalytics:OutputBlob
name: example
properties:
name: exampleoutput
streamAnalyticsJobName: ${exampleJob.name}
resourceGroupName: ${exampleJob.resourceGroupName}
storageAccountName: ${exampleAccount.name}
storageAccountKey: ${exampleAccount.primaryAccessKey}
storageContainerName: ${exampleContainer.name}
pathPattern: example-{date}-{time}
dateFormat: yyyy-MM-dd
timeFormat: HH
serialization:
type: Avro
exampleJobSchedule:
type: azure:streamanalytics:JobSchedule
name: example
properties:
streamAnalyticsJobId: ${exampleJob.id}
startMode: CustomTime
startTime: 2022-09-21T00:00:00Z
options:
dependson:
- ${exampleJob}
- ${exampleStreamInputBlob}
- ${exampleOutputBlob}
Create JobSchedule Resource
Resources are created with functions called constructors. To learn more about declaring and configuring resources, see Resources.
Constructor syntax
new JobSchedule(name: string, args: JobScheduleArgs, opts?: CustomResourceOptions);
@overload
def JobSchedule(resource_name: str,
args: JobScheduleArgs,
opts: Optional[ResourceOptions] = None)
@overload
def JobSchedule(resource_name: str,
opts: Optional[ResourceOptions] = None,
start_mode: Optional[str] = None,
stream_analytics_job_id: Optional[str] = None,
start_time: Optional[str] = None)
func NewJobSchedule(ctx *Context, name string, args JobScheduleArgs, opts ...ResourceOption) (*JobSchedule, error)
public JobSchedule(string name, JobScheduleArgs args, CustomResourceOptions? opts = null)
public JobSchedule(String name, JobScheduleArgs args)
public JobSchedule(String name, JobScheduleArgs args, CustomResourceOptions options)
type: azure:streamanalytics:JobSchedule
properties: # The arguments to resource properties.
options: # Bag of options to control resource's behavior.
Parameters
- name string
- The unique name of the resource.
- args JobScheduleArgs
- The arguments to resource properties.
- opts CustomResourceOptions
- Bag of options to control resource's behavior.
- resource_name str
- The unique name of the resource.
- args JobScheduleArgs
- The arguments to resource properties.
- opts ResourceOptions
- Bag of options to control resource's behavior.
- ctx Context
- Context object for the current deployment.
- name string
- The unique name of the resource.
- args JobScheduleArgs
- The arguments to resource properties.
- opts ResourceOption
- Bag of options to control resource's behavior.
- name string
- The unique name of the resource.
- args JobScheduleArgs
- The arguments to resource properties.
- opts CustomResourceOptions
- Bag of options to control resource's behavior.
- name String
- The unique name of the resource.
- args JobScheduleArgs
- The arguments to resource properties.
- options CustomResourceOptions
- Bag of options to control resource's behavior.
Constructor example
The following reference example uses placeholder values for all input properties.
var azureJobScheduleResource = new Azure.StreamAnalytics.JobSchedule("azureJobScheduleResource", new()
{
StartMode = "string",
StreamAnalyticsJobId = "string",
StartTime = "string",
});
example, err := streamanalytics.NewJobSchedule(ctx, "azureJobScheduleResource", &streamanalytics.JobScheduleArgs{
StartMode: pulumi.String("string"),
StreamAnalyticsJobId: pulumi.String("string"),
StartTime: pulumi.String("string"),
})
var azureJobScheduleResource = new JobSchedule("azureJobScheduleResource", JobScheduleArgs.builder()
.startMode("string")
.streamAnalyticsJobId("string")
.startTime("string")
.build());
azure_job_schedule_resource = azure.streamanalytics.JobSchedule("azureJobScheduleResource",
start_mode="string",
stream_analytics_job_id="string",
start_time="string")
const azureJobScheduleResource = new azure.streamanalytics.JobSchedule("azureJobScheduleResource", {
startMode: "string",
streamAnalyticsJobId: "string",
startTime: "string",
});
type: azure:streamanalytics:JobSchedule
properties:
startMode: string
startTime: string
streamAnalyticsJobId: string
JobSchedule Resource Properties
To learn more about resource properties and how to use them, see Inputs and Outputs in the Architecture and Concepts docs.
Inputs
In Python, inputs that are objects can be passed either as argument classes or as dictionary literals.
The JobSchedule resource accepts the following input properties:
- Start
Mode string The starting mode of the Stream Analytics Job. Possible values are
JobStartTime
,CustomTime
andLastOutputEventTime
.Note: Setting
start_mode
toLastOutputEventTime
is only possible if the job had been previously started and produced output.- Stream
Analytics stringJob Id - The ID of the Stream Analytics Job that should be scheduled or started. Changing this forces a new resource to be created.
- Start
Time string - The time in ISO8601 format at which the Stream Analytics Job should be started e.g.
2022-04-01T00:00:00Z
. This property can only be specified ifstart_mode
is set toCustomTime
- Start
Mode string The starting mode of the Stream Analytics Job. Possible values are
JobStartTime
,CustomTime
andLastOutputEventTime
.Note: Setting
start_mode
toLastOutputEventTime
is only possible if the job had been previously started and produced output.- Stream
Analytics stringJob Id - The ID of the Stream Analytics Job that should be scheduled or started. Changing this forces a new resource to be created.
- Start
Time string - The time in ISO8601 format at which the Stream Analytics Job should be started e.g.
2022-04-01T00:00:00Z
. This property can only be specified ifstart_mode
is set toCustomTime
- start
Mode String The starting mode of the Stream Analytics Job. Possible values are
JobStartTime
,CustomTime
andLastOutputEventTime
.Note: Setting
start_mode
toLastOutputEventTime
is only possible if the job had been previously started and produced output.- stream
Analytics StringJob Id - The ID of the Stream Analytics Job that should be scheduled or started. Changing this forces a new resource to be created.
- start
Time String - The time in ISO8601 format at which the Stream Analytics Job should be started e.g.
2022-04-01T00:00:00Z
. This property can only be specified ifstart_mode
is set toCustomTime
- start
Mode string The starting mode of the Stream Analytics Job. Possible values are
JobStartTime
,CustomTime
andLastOutputEventTime
.Note: Setting
start_mode
toLastOutputEventTime
is only possible if the job had been previously started and produced output.- stream
Analytics stringJob Id - The ID of the Stream Analytics Job that should be scheduled or started. Changing this forces a new resource to be created.
- start
Time string - The time in ISO8601 format at which the Stream Analytics Job should be started e.g.
2022-04-01T00:00:00Z
. This property can only be specified ifstart_mode
is set toCustomTime
- start_
mode str The starting mode of the Stream Analytics Job. Possible values are
JobStartTime
,CustomTime
andLastOutputEventTime
.Note: Setting
start_mode
toLastOutputEventTime
is only possible if the job had been previously started and produced output.- stream_
analytics_ strjob_ id - The ID of the Stream Analytics Job that should be scheduled or started. Changing this forces a new resource to be created.
- start_
time str - The time in ISO8601 format at which the Stream Analytics Job should be started e.g.
2022-04-01T00:00:00Z
. This property can only be specified ifstart_mode
is set toCustomTime
- start
Mode String The starting mode of the Stream Analytics Job. Possible values are
JobStartTime
,CustomTime
andLastOutputEventTime
.Note: Setting
start_mode
toLastOutputEventTime
is only possible if the job had been previously started and produced output.- stream
Analytics StringJob Id - The ID of the Stream Analytics Job that should be scheduled or started. Changing this forces a new resource to be created.
- start
Time String - The time in ISO8601 format at which the Stream Analytics Job should be started e.g.
2022-04-01T00:00:00Z
. This property can only be specified ifstart_mode
is set toCustomTime
Outputs
All input properties are implicitly available as output properties. Additionally, the JobSchedule resource produces the following output properties:
- Id string
- The provider-assigned unique ID for this managed resource.
- Last
Output stringTime - The time at which the Stream Analytics job last produced an output.
- Id string
- The provider-assigned unique ID for this managed resource.
- Last
Output stringTime - The time at which the Stream Analytics job last produced an output.
- id String
- The provider-assigned unique ID for this managed resource.
- last
Output StringTime - The time at which the Stream Analytics job last produced an output.
- id string
- The provider-assigned unique ID for this managed resource.
- last
Output stringTime - The time at which the Stream Analytics job last produced an output.
- id str
- The provider-assigned unique ID for this managed resource.
- last_
output_ strtime - The time at which the Stream Analytics job last produced an output.
- id String
- The provider-assigned unique ID for this managed resource.
- last
Output StringTime - The time at which the Stream Analytics job last produced an output.
Look up Existing JobSchedule Resource
Get an existing JobSchedule resource’s state with the given name, ID, and optional extra properties used to qualify the lookup.
public static get(name: string, id: Input<ID>, state?: JobScheduleState, opts?: CustomResourceOptions): JobSchedule
@staticmethod
def get(resource_name: str,
id: str,
opts: Optional[ResourceOptions] = None,
last_output_time: Optional[str] = None,
start_mode: Optional[str] = None,
start_time: Optional[str] = None,
stream_analytics_job_id: Optional[str] = None) -> JobSchedule
func GetJobSchedule(ctx *Context, name string, id IDInput, state *JobScheduleState, opts ...ResourceOption) (*JobSchedule, error)
public static JobSchedule Get(string name, Input<string> id, JobScheduleState? state, CustomResourceOptions? opts = null)
public static JobSchedule get(String name, Output<String> id, JobScheduleState state, CustomResourceOptions options)
Resource lookup is not supported in YAML
- name
- The unique name of the resulting resource.
- id
- The unique provider ID of the resource to lookup.
- state
- Any extra arguments used during the lookup.
- opts
- A bag of options that control this resource's behavior.
- resource_name
- The unique name of the resulting resource.
- id
- The unique provider ID of the resource to lookup.
- name
- The unique name of the resulting resource.
- id
- The unique provider ID of the resource to lookup.
- state
- Any extra arguments used during the lookup.
- opts
- A bag of options that control this resource's behavior.
- name
- The unique name of the resulting resource.
- id
- The unique provider ID of the resource to lookup.
- state
- Any extra arguments used during the lookup.
- opts
- A bag of options that control this resource's behavior.
- name
- The unique name of the resulting resource.
- id
- The unique provider ID of the resource to lookup.
- state
- Any extra arguments used during the lookup.
- opts
- A bag of options that control this resource's behavior.
- Last
Output stringTime - The time at which the Stream Analytics job last produced an output.
- Start
Mode string The starting mode of the Stream Analytics Job. Possible values are
JobStartTime
,CustomTime
andLastOutputEventTime
.Note: Setting
start_mode
toLastOutputEventTime
is only possible if the job had been previously started and produced output.- Start
Time string - The time in ISO8601 format at which the Stream Analytics Job should be started e.g.
2022-04-01T00:00:00Z
. This property can only be specified ifstart_mode
is set toCustomTime
- Stream
Analytics stringJob Id - The ID of the Stream Analytics Job that should be scheduled or started. Changing this forces a new resource to be created.
- Last
Output stringTime - The time at which the Stream Analytics job last produced an output.
- Start
Mode string The starting mode of the Stream Analytics Job. Possible values are
JobStartTime
,CustomTime
andLastOutputEventTime
.Note: Setting
start_mode
toLastOutputEventTime
is only possible if the job had been previously started and produced output.- Start
Time string - The time in ISO8601 format at which the Stream Analytics Job should be started e.g.
2022-04-01T00:00:00Z
. This property can only be specified ifstart_mode
is set toCustomTime
- Stream
Analytics stringJob Id - The ID of the Stream Analytics Job that should be scheduled or started. Changing this forces a new resource to be created.
- last
Output StringTime - The time at which the Stream Analytics job last produced an output.
- start
Mode String The starting mode of the Stream Analytics Job. Possible values are
JobStartTime
,CustomTime
andLastOutputEventTime
.Note: Setting
start_mode
toLastOutputEventTime
is only possible if the job had been previously started and produced output.- start
Time String - The time in ISO8601 format at which the Stream Analytics Job should be started e.g.
2022-04-01T00:00:00Z
. This property can only be specified ifstart_mode
is set toCustomTime
- stream
Analytics StringJob Id - The ID of the Stream Analytics Job that should be scheduled or started. Changing this forces a new resource to be created.
- last
Output stringTime - The time at which the Stream Analytics job last produced an output.
- start
Mode string The starting mode of the Stream Analytics Job. Possible values are
JobStartTime
,CustomTime
andLastOutputEventTime
.Note: Setting
start_mode
toLastOutputEventTime
is only possible if the job had been previously started and produced output.- start
Time string - The time in ISO8601 format at which the Stream Analytics Job should be started e.g.
2022-04-01T00:00:00Z
. This property can only be specified ifstart_mode
is set toCustomTime
- stream
Analytics stringJob Id - The ID of the Stream Analytics Job that should be scheduled or started. Changing this forces a new resource to be created.
- last_
output_ strtime - The time at which the Stream Analytics job last produced an output.
- start_
mode str The starting mode of the Stream Analytics Job. Possible values are
JobStartTime
,CustomTime
andLastOutputEventTime
.Note: Setting
start_mode
toLastOutputEventTime
is only possible if the job had been previously started and produced output.- start_
time str - The time in ISO8601 format at which the Stream Analytics Job should be started e.g.
2022-04-01T00:00:00Z
. This property can only be specified ifstart_mode
is set toCustomTime
- stream_
analytics_ strjob_ id - The ID of the Stream Analytics Job that should be scheduled or started. Changing this forces a new resource to be created.
- last
Output StringTime - The time at which the Stream Analytics job last produced an output.
- start
Mode String The starting mode of the Stream Analytics Job. Possible values are
JobStartTime
,CustomTime
andLastOutputEventTime
.Note: Setting
start_mode
toLastOutputEventTime
is only possible if the job had been previously started and produced output.- start
Time String - The time in ISO8601 format at which the Stream Analytics Job should be started e.g.
2022-04-01T00:00:00Z
. This property can only be specified ifstart_mode
is set toCustomTime
- stream
Analytics StringJob Id - The ID of the Stream Analytics Job that should be scheduled or started. Changing this forces a new resource to be created.
Import
Stream Analytics Job’s can be imported using the resource id
, e.g.
$ pulumi import azure:streamanalytics/jobSchedule:JobSchedule example /subscriptions/00000000-0000-0000-0000-000000000000/resourceGroups/group1/providers/Microsoft.StreamAnalytics/streamingJobs/job1/schedule/default
To learn more about importing existing cloud resources, see Importing resources.
Package Details
- Repository
- Azure Classic pulumi/pulumi-azure
- License
- Apache-2.0
- Notes
- This Pulumi package is based on the
azurerm
Terraform Provider.