Big Data Pool Args
A Big Data pool Uses Azure REST API version 2021-06-01. In version 2.x of the Azure Native provider, it used API version 2021-06-01. Other available API versions: 2021-04-01-preview, 2021-05-01, 2021-06-01-preview. These can be accessed by generating a local SDK package using the CLI command pulumi package add azure-native synapse [ApiVersion]
. See the ../../../version-guide/#accessing-any-api-version-via-local-packages for details.
Example Usage
Create or update a Big Data pool
using System.Collections.Generic;
using System.Linq;
using Pulumi;
using AzureNative = Pulumi.AzureNative;
return await Deployment.RunAsync(() =>
{
var bigDataPool = new AzureNative.Synapse.BigDataPool("bigDataPool", new()
{
AutoPause = new AzureNative.Synapse.Inputs.AutoPausePropertiesArgs
{
DelayInMinutes = 15,
Enabled = true,
},
AutoScale = new AzureNative.Synapse.Inputs.AutoScalePropertiesArgs
{
Enabled = true,
MaxNodeCount = 50,
MinNodeCount = 3,
},
BigDataPoolName = "ExamplePool",
DefaultSparkLogFolder = "/logs",
IsAutotuneEnabled = false,
LibraryRequirements = new AzureNative.Synapse.Inputs.LibraryRequirementsArgs
{
Content = "",
Filename = "requirements.txt",
},
Location = "West US 2",
NodeCount = 4,
NodeSize = AzureNative.Synapse.NodeSize.Medium,
NodeSizeFamily = AzureNative.Synapse.NodeSizeFamily.MemoryOptimized,
ResourceGroupName = "ExampleResourceGroup",
SparkEventsFolder = "/events",
SparkVersion = "3.3",
Tags =
{
{ "key", "value" },
},
WorkspaceName = "ExampleWorkspace",
});
});
package main
import (
synapse "github.com/pulumi/pulumi-azure-native-sdk/synapse/v2"
"github.com/pulumi/pulumi/sdk/v3/go/pulumi"
)
func main() {
pulumi.Run(func(ctx *pulumi.Context) error {
_, err := synapse.NewBigDataPool(ctx, "bigDataPool", &synapse.BigDataPoolArgs{
AutoPause: &synapse.AutoPausePropertiesArgs{
DelayInMinutes: pulumi.Int(15),
Enabled: pulumi.Bool(true),
},
AutoScale: &synapse.AutoScalePropertiesArgs{
Enabled: pulumi.Bool(true),
MaxNodeCount: pulumi.Int(50),
MinNodeCount: pulumi.Int(3),
},
BigDataPoolName: pulumi.String("ExamplePool"),
DefaultSparkLogFolder: pulumi.String("/logs"),
IsAutotuneEnabled: pulumi.Bool(false),
LibraryRequirements: &synapse.LibraryRequirementsArgs{
Content: pulumi.String(""),
Filename: pulumi.String("requirements.txt"),
},
Location: pulumi.String("West US 2"),
NodeCount: pulumi.Int(4),
NodeSize: pulumi.String(synapse.NodeSizeMedium),
NodeSizeFamily: pulumi.String(synapse.NodeSizeFamilyMemoryOptimized),
ResourceGroupName: pulumi.String("ExampleResourceGroup"),
SparkEventsFolder: pulumi.String("/events"),
SparkVersion: pulumi.String("3.3"),
Tags: pulumi.StringMap{
"key": pulumi.String("value"),
},
WorkspaceName: pulumi.String("ExampleWorkspace"),
})
if err != nil {
return err
}
return nil
})
}
package generated_program;
import com.pulumi.Context;
import com.pulumi.Pulumi;
import com.pulumi.core.Output;
import com.pulumi.azurenative.synapse.BigDataPool;
import com.pulumi.azurenative.synapse.BigDataPoolArgs;
import com.pulumi.azurenative.synapse.inputs.AutoPausePropertiesArgs;
import com.pulumi.azurenative.synapse.inputs.AutoScalePropertiesArgs;
import com.pulumi.azurenative.synapse.inputs.LibraryRequirementsArgs;
import java.util.List;
import java.util.ArrayList;
import java.util.Map;
import java.io.File;
import java.nio.file.Files;
import java.nio.file.Paths;
public class App {
public static void main(String[] args) {
Pulumi.run(App::stack);
}
public static void stack(Context ctx) {
var bigDataPool = new BigDataPool("bigDataPool", BigDataPoolArgs.builder()
.autoPause(AutoPausePropertiesArgs.builder()
.delayInMinutes(15)
.enabled(true)
.build())
.autoScale(AutoScalePropertiesArgs.builder()
.enabled(true)
.maxNodeCount(50)
.minNodeCount(3)
.build())
.bigDataPoolName("ExamplePool")
.defaultSparkLogFolder("/logs")
.isAutotuneEnabled(false)
.libraryRequirements(LibraryRequirementsArgs.builder()
.content("")
.filename("requirements.txt")
.build())
.location("West US 2")
.nodeCount(4)
.nodeSize("Medium")
.nodeSizeFamily("MemoryOptimized")
.resourceGroupName("ExampleResourceGroup")
.sparkEventsFolder("/events")
.sparkVersion("3.3")
.tags(Map.of("key", "value"))
.workspaceName("ExampleWorkspace")
.build());
}
}
Import
An existing resource can be imported using its type token, name, and identifier, e.g.
$ pulumi import azure-native:synapse:BigDataPool ExamplePool /subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Synapse/workspaces/{workspaceName}/bigDataPools/{bigDataPoolName}
Constructors
Properties
Auto-pausing properties
Auto-scaling properties
Big Data pool name
List of custom libraries/packages associated with the spark pool.
The default folder where Spark logs will be written.
Dynamic Executor Allocation
Whether autotune is required or not.
Whether compute isolation is required or not.
Library version requirements
The kind of nodes that the Big Data pool provides.
The state of the Big Data pool.
The name of the resource group. The name is case insensitive.
Whether session level packages enabled.
Spark configuration file to specify additional properties
The Spark events folder
The Apache Spark version.
The name of the workspace.