Job Schedule Args
data class JobScheduleArgs(val startMode: Output<String>? = null, val startTime: Output<String>? = null, val streamAnalyticsJobId: Output<String>? = null) : ConvertibleToJava<JobScheduleArgs>
Manages a Stream Analytics Job Schedule.
Example Usage
package generated_program;
import com.pulumi.Context;
import com.pulumi.Pulumi;
import com.pulumi.core.Output;
import com.pulumi.azure.core.ResourceGroup;
import com.pulumi.azure.core.ResourceGroupArgs;
import com.pulumi.azure.storage.Account;
import com.pulumi.azure.storage.AccountArgs;
import com.pulumi.azure.storage.Container;
import com.pulumi.azure.storage.ContainerArgs;
import com.pulumi.azure.storage.Blob;
import com.pulumi.azure.storage.BlobArgs;
import com.pulumi.azure.streamanalytics.Job;
import com.pulumi.azure.streamanalytics.JobArgs;
import com.pulumi.azure.streamanalytics.StreamInputBlob;
import com.pulumi.azure.streamanalytics.StreamInputBlobArgs;
import com.pulumi.azure.streamanalytics.inputs.StreamInputBlobSerializationArgs;
import com.pulumi.azure.streamanalytics.OutputBlob;
import com.pulumi.azure.streamanalytics.OutputBlobArgs;
import com.pulumi.azure.streamanalytics.inputs.OutputBlobSerializationArgs;
import com.pulumi.azure.streamanalytics.JobSchedule;
import com.pulumi.azure.streamanalytics.JobScheduleArgs;
import com.pulumi.resources.CustomResourceOptions;
import com.pulumi.asset.FileAsset;
import java.util.List;
import java.util.ArrayList;
import java.util.Map;
import java.io.File;
import java.nio.file.Files;
import java.nio.file.Paths;
public class App {
public static void main(String[] args) {
Pulumi.run(App::stack);
}
public static void stack(Context ctx) {
var exampleResourceGroup = new ResourceGroup("exampleResourceGroup", ResourceGroupArgs.builder()
.location("West Europe")
.build());
var exampleAccount = new Account("exampleAccount", AccountArgs.builder()
.resourceGroupName(exampleResourceGroup.name())
.location(exampleResourceGroup.location())
.accountTier("Standard")
.accountReplicationType("LRS")
.build());
var exampleContainer = new Container("exampleContainer", ContainerArgs.builder()
.storageAccountName(exampleAccount.name())
.containerAccessType("private")
.build());
var exampleBlob = new Blob("exampleBlob", BlobArgs.builder()
.storageAccountName(exampleAccount.name())
.storageContainerName(exampleContainer.name())
.type("Block")
.source(new FileAsset("example.csv"))
.build());
var exampleJob = new Job("exampleJob", JobArgs.builder()
.resourceGroupName(exampleResourceGroup.name())
.location(exampleResourceGroup.location())
.compatibilityLevel("1.2")
.dataLocale("en-GB")
.eventsLateArrivalMaxDelayInSeconds(60)
.eventsOutOfOrderMaxDelayInSeconds(50)
.eventsOutOfOrderPolicy("Adjust")
.outputErrorPolicy("Drop")
.streamingUnits(3)
.tags(Map.of("environment", "Example"))
.transformationQuery("""
SELECT *
INTO [exampleoutput]
FROM [exampleinput]
""")
.build());
var exampleStreamInputBlob = new StreamInputBlob("exampleStreamInputBlob", StreamInputBlobArgs.builder()
.streamAnalyticsJobName(exampleJob.name())
.resourceGroupName(exampleJob.resourceGroupName())
.storageAccountName(exampleAccount.name())
.storageAccountKey(exampleAccount.primaryAccessKey())
.storageContainerName(exampleContainer.name())
.pathPattern("")
.dateFormat("yyyy/MM/dd")
.timeFormat("HH")
.serialization(StreamInputBlobSerializationArgs.builder()
.type("Csv")
.encoding("UTF8")
.fieldDelimiter(",")
.build())
.build());
var exampleOutputBlob = new OutputBlob("exampleOutputBlob", OutputBlobArgs.builder()
.streamAnalyticsJobName(exampleJob.name())
.resourceGroupName(exampleJob.resourceGroupName())
.storageAccountName(exampleAccount.name())
.storageAccountKey(exampleAccount.primaryAccessKey())
.storageContainerName(exampleContainer.name())
.pathPattern("example-{date}-{time}")
.dateFormat("yyyy-MM-dd")
.timeFormat("HH")
.serialization(OutputBlobSerializationArgs.builder()
.type("Avro")
.build())
.build());
var exampleJobSchedule = new JobSchedule("exampleJobSchedule", JobScheduleArgs.builder()
.streamAnalyticsJobId(exampleJob.id())
.startMode("CustomTime")
.startTime("2022-09-21T00:00:00Z")
.build(), CustomResourceOptions.builder()
.dependsOn(
exampleJob,
exampleStreamInputBlob,
exampleOutputBlob)
.build());
}
}
Content copied to clipboard
Import
Stream Analytics Job's can be imported using the resource id
, e.g.
$ pulumi import azure:streamanalytics/jobSchedule:JobSchedule example /subscriptions/00000000-0000-0000-0000-000000000000/resourceGroups/group1/providers/Microsoft.StreamAnalytics/streamingJobs/job1/schedule/default
Content copied to clipboard
Constructors
Link copied to clipboard
fun JobScheduleArgs(startMode: Output<String>? = null, startTime: Output<String>? = null, streamAnalyticsJobId: Output<String>? = null)