File Cache Args
Resource for managing an Amazon File Cache cache. See the Create File Cache for more information.
Example Usage
package generated_program;
import com.pulumi.Context;
import com.pulumi.Pulumi;
import com.pulumi.core.Output;
import com.pulumi.aws.fsx.FileCache;
import com.pulumi.aws.fsx.FileCacheArgs;
import com.pulumi.aws.fsx.inputs.FileCacheDataRepositoryAssociationArgs;
import com.pulumi.aws.fsx.inputs.FileCacheLustreConfigurationArgs;
import java.util.List;
import java.util.ArrayList;
import java.util.Map;
import java.io.File;
import java.nio.file.Files;
import java.nio.file.Paths;
public class App {
public static void main(String[] args) {
Pulumi.run(App::stack);
}
public static void stack(Context ctx) {
var example = new FileCache("example", FileCacheArgs.builder()
.dataRepositoryAssociations(FileCacheDataRepositoryAssociationArgs.builder()
.dataRepositoryPath("nfs://filer.domain.com")
.dataRepositorySubdirectories(
"test",
"test2")
.fileCachePath("/ns1")
.nfs(FileCacheDataRepositoryAssociationNfArgs.builder()
.dnsIps(
"192.168.0.1",
"192.168.0.2")
.version("NFS3")
.build())
.build())
.fileCacheType("LUSTRE")
.fileCacheTypeVersion("2.12")
.lustreConfigurations(FileCacheLustreConfigurationArgs.builder()
.deploymentType("CACHE_1")
.metadataConfigurations(FileCacheLustreConfigurationMetadataConfigurationArgs.builder()
.storageCapacity(2400)
.build())
.perUnitStorageThroughput(1000)
.weeklyMaintenanceStartTime("2:05:00")
.build())
.subnetIds(aws_subnet.test1().id())
.storageCapacity(1200)
.build());
}
}
Import
Using pulumi import
, import Amazon File Cache cache using the resource id
. For example:
$ pulumi import aws:fsx/fileCache:FileCache example fc-8012925589
Constructors
Functions
Properties
See the data_repository_association
configuration block. Max of 8. A list of up to 8 configurations for data repository associations (DRAs) to be created during the cache creation. The DRAs link the cache to either an Amazon S3 data repository or a Network File System (NFS) data repository that supports the NFSv3 protocol. The DRA configurations must meet the following requirements: 1) All configurations on the list must be of the same data repository type, either all S3 or all NFS. A cache can't link to different data repository types at the same time. 2) An NFS DRA must link to an NFS file system that supports the NFSv3 protocol. DRA automatic import and automatic export is not supported.