# frozen_string_literal: true # WARNING ABOUT GENERATED CODE # # This file is generated. See the contributing guide for more information: # https://github.com/aws/aws-sdk-ruby/blob/version-3/CONTRIBUTING.md # # WARNING ABOUT GENERATED CODE require 'seahorse/client/plugins/content_length.rb' require 'aws-sdk-core/plugins/credentials_configuration.rb' require 'aws-sdk-core/plugins/logging.rb' require 'aws-sdk-core/plugins/param_converter.rb' require 'aws-sdk-core/plugins/param_validator.rb' require 'aws-sdk-core/plugins/user_agent.rb' require 'aws-sdk-core/plugins/helpful_socket_errors.rb' require 'aws-sdk-core/plugins/retry_errors.rb' require 'aws-sdk-core/plugins/global_configuration.rb' require 'aws-sdk-core/plugins/regional_endpoint.rb' require 'aws-sdk-core/plugins/endpoint_discovery.rb' require 'aws-sdk-core/plugins/endpoint_pattern.rb' require 'aws-sdk-core/plugins/response_paging.rb' require 'aws-sdk-core/plugins/stub_responses.rb' require 'aws-sdk-core/plugins/idempotency_token.rb' require 'aws-sdk-core/plugins/jsonvalue_converter.rb' require 'aws-sdk-core/plugins/client_metrics_plugin.rb' require 'aws-sdk-core/plugins/client_metrics_send_plugin.rb' require 'aws-sdk-core/plugins/transfer_encoding.rb' require 'aws-sdk-core/plugins/http_checksum.rb' require 'aws-sdk-core/plugins/signature_v4.rb' require 'aws-sdk-core/plugins/protocols/json_rpc.rb' Aws::Plugins::GlobalConfiguration.add_identifier(:fsx) module Aws::FSx # An API client for FSx. To construct a client, you need to configure a `:region` and `:credentials`. # # client = Aws::FSx::Client.new( # region: region_name, # credentials: credentials, # # ... # ) # # For details on configuring region and credentials see # the [developer guide](/sdk-for-ruby/v3/developer-guide/setup-config.html). # # See {#initialize} for a full list of supported configuration options. class Client < Seahorse::Client::Base include Aws::ClientStubs @identifier = :fsx set_api(ClientApi::API) add_plugin(Seahorse::Client::Plugins::ContentLength) add_plugin(Aws::Plugins::CredentialsConfiguration) add_plugin(Aws::Plugins::Logging) add_plugin(Aws::Plugins::ParamConverter) add_plugin(Aws::Plugins::ParamValidator) add_plugin(Aws::Plugins::UserAgent) add_plugin(Aws::Plugins::HelpfulSocketErrors) add_plugin(Aws::Plugins::RetryErrors) add_plugin(Aws::Plugins::GlobalConfiguration) add_plugin(Aws::Plugins::RegionalEndpoint) add_plugin(Aws::Plugins::EndpointDiscovery) add_plugin(Aws::Plugins::EndpointPattern) add_plugin(Aws::Plugins::ResponsePaging) add_plugin(Aws::Plugins::StubResponses) add_plugin(Aws::Plugins::IdempotencyToken) add_plugin(Aws::Plugins::JsonvalueConverter) add_plugin(Aws::Plugins::ClientMetricsPlugin) add_plugin(Aws::Plugins::ClientMetricsSendPlugin) add_plugin(Aws::Plugins::TransferEncoding) add_plugin(Aws::Plugins::HttpChecksum) add_plugin(Aws::Plugins::SignatureV4) add_plugin(Aws::Plugins::Protocols::JsonRpc) # @overload initialize(options) # @param [Hash] options # @option options [required, Aws::CredentialProvider] :credentials # Your AWS credentials. This can be an instance of any one of the # following classes: # # * `Aws::Credentials` - Used for configuring static, non-refreshing # credentials. # # * `Aws::SharedCredentials` - Used for loading static credentials from a # shared file, such as `~/.aws/config`. # # * `Aws::AssumeRoleCredentials` - Used when you need to assume a role. # # * `Aws::AssumeRoleWebIdentityCredentials` - Used when you need to # assume a role after providing credentials via the web. # # * `Aws::SSOCredentials` - Used for loading credentials from AWS SSO using an # access token generated from `aws login`. # # * `Aws::ProcessCredentials` - Used for loading credentials from a # process that outputs to stdout. # # * `Aws::InstanceProfileCredentials` - Used for loading credentials # from an EC2 IMDS on an EC2 instance. # # * `Aws::ECSCredentials` - Used for loading credentials from # instances running in ECS. # # * `Aws::CognitoIdentityCredentials` - Used for loading credentials # from the Cognito Identity service. # # When `:credentials` are not configured directly, the following # locations will be searched for credentials: # # * `Aws.config[:credentials]` # * The `:access_key_id`, `:secret_access_key`, and `:session_token` options. # * ENV['AWS_ACCESS_KEY_ID'], ENV['AWS_SECRET_ACCESS_KEY'] # * `~/.aws/credentials` # * `~/.aws/config` # * EC2/ECS IMDS instance profile - When used by default, the timeouts # are very aggressive. Construct and pass an instance of # `Aws::InstanceProfileCredentails` or `Aws::ECSCredentials` to # enable retries and extended timeouts. # # @option options [required, String] :region # The AWS region to connect to. The configured `:region` is # used to determine the service `:endpoint`. When not passed, # a default `:region` is searched for in the following locations: # # * `Aws.config[:region]` # * `ENV['AWS_REGION']` # * `ENV['AMAZON_REGION']` # * `ENV['AWS_DEFAULT_REGION']` # * `~/.aws/credentials` # * `~/.aws/config` # # @option options [String] :access_key_id # # @option options [Boolean] :active_endpoint_cache (false) # When set to `true`, a thread polling for endpoints will be running in # the background every 60 secs (default). Defaults to `false`. # # @option options [Boolean] :adaptive_retry_wait_to_fill (true) # Used only in `adaptive` retry mode. When true, the request will sleep # until there is sufficent client side capacity to retry the request. # When false, the request will raise a `RetryCapacityNotAvailableError` and will # not retry instead of sleeping. # # @option options [Boolean] :client_side_monitoring (false) # When `true`, client-side metrics will be collected for all API requests from # this client. # # @option options [String] :client_side_monitoring_client_id ("") # Allows you to provide an identifier for this client which will be attached to # all generated client side metrics. Defaults to an empty string. # # @option options [String] :client_side_monitoring_host ("127.0.0.1") # Allows you to specify the DNS hostname or IPv4 or IPv6 address that the client # side monitoring agent is running on, where client metrics will be published via UDP. # # @option options [Integer] :client_side_monitoring_port (31000) # Required for publishing client metrics. The port that the client side monitoring # agent is running on, where client metrics will be published via UDP. # # @option options [Aws::ClientSideMonitoring::Publisher] :client_side_monitoring_publisher (Aws::ClientSideMonitoring::Publisher) # Allows you to provide a custom client-side monitoring publisher class. By default, # will use the Client Side Monitoring Agent Publisher. # # @option options [Boolean] :convert_params (true) # When `true`, an attempt is made to coerce request parameters into # the required types. # # @option options [Boolean] :correct_clock_skew (true) # Used only in `standard` and adaptive retry modes. Specifies whether to apply # a clock skew correction and retry requests with skewed client clocks. # # @option options [Boolean] :disable_host_prefix_injection (false) # Set to true to disable SDK automatically adding host prefix # to default service endpoint when available. # # @option options [String] :endpoint # The client endpoint is normally constructed from the `:region` # option. You should only configure an `:endpoint` when connecting # to test or custom endpoints. This should be a valid HTTP(S) URI. # # @option options [Integer] :endpoint_cache_max_entries (1000) # Used for the maximum size limit of the LRU cache storing endpoints data # for endpoint discovery enabled operations. Defaults to 1000. # # @option options [Integer] :endpoint_cache_max_threads (10) # Used for the maximum threads in use for polling endpoints to be cached, defaults to 10. # # @option options [Integer] :endpoint_cache_poll_interval (60) # When :endpoint_discovery and :active_endpoint_cache is enabled, # Use this option to config the time interval in seconds for making # requests fetching endpoints information. Defaults to 60 sec. # # @option options [Boolean] :endpoint_discovery (false) # When set to `true`, endpoint discovery will be enabled for operations when available. # # @option options [Aws::Log::Formatter] :log_formatter (Aws::Log::Formatter.default) # The log formatter. # # @option options [Symbol] :log_level (:info) # The log level to send messages to the `:logger` at. # # @option options [Logger] :logger # The Logger instance to send log messages to. If this option # is not set, logging will be disabled. # # @option options [Integer] :max_attempts (3) # An integer representing the maximum number attempts that will be made for # a single request, including the initial attempt. For example, # setting this value to 5 will result in a request being retried up to # 4 times. Used in `standard` and `adaptive` retry modes. # # @option options [String] :profile ("default") # Used when loading credentials from the shared credentials file # at HOME/.aws/credentials. When not specified, 'default' is used. # # @option options [Proc] :retry_backoff # A proc or lambda used for backoff. Defaults to 2**retries * retry_base_delay. # This option is only used in the `legacy` retry mode. # # @option options [Float] :retry_base_delay (0.3) # The base delay in seconds used by the default backoff function. This option # is only used in the `legacy` retry mode. # # @option options [Symbol] :retry_jitter (:none) # A delay randomiser function used by the default backoff function. # Some predefined functions can be referenced by name - :none, :equal, :full, # otherwise a Proc that takes and returns a number. This option is only used # in the `legacy` retry mode. # # @see https://www.awsarchitectureblog.com/2015/03/backoff.html # # @option options [Integer] :retry_limit (3) # The maximum number of times to retry failed requests. Only # ~ 500 level server errors and certain ~ 400 level client errors # are retried. Generally, these are throttling errors, data # checksum errors, networking errors, timeout errors, auth errors, # endpoint discovery, and errors from expired credentials. # This option is only used in the `legacy` retry mode. # # @option options [Integer] :retry_max_delay (0) # The maximum number of seconds to delay between retries (0 for no limit) # used by the default backoff function. This option is only used in the # `legacy` retry mode. # # @option options [String] :retry_mode ("legacy") # Specifies which retry algorithm to use. Values are: # # * `legacy` - The pre-existing retry behavior. This is default value if # no retry mode is provided. # # * `standard` - A standardized set of retry rules across the AWS SDKs. # This includes support for retry quotas, which limit the number of # unsuccessful retries a client can make. # # * `adaptive` - An experimental retry mode that includes all the # functionality of `standard` mode along with automatic client side # throttling. This is a provisional mode that may change behavior # in the future. # # # @option options [String] :secret_access_key # # @option options [String] :session_token # # @option options [Boolean] :simple_json (false) # Disables request parameter conversion, validation, and formatting. # Also disable response data type conversions. This option is useful # when you want to ensure the highest level of performance by # avoiding overhead of walking request parameters and response data # structures. # # When `:simple_json` is enabled, the request parameters hash must # be formatted exactly as the DynamoDB API expects. # # @option options [Boolean] :stub_responses (false) # Causes the client to return stubbed responses. By default # fake responses are generated and returned. You can specify # the response data to return or errors to raise by calling # {ClientStubs#stub_responses}. See {ClientStubs} for more information. # # ** Please note ** When response stubbing is enabled, no HTTP # requests are made, and retries are disabled. # # @option options [Boolean] :validate_params (true) # When `true`, request parameters are validated before # sending the request. # # @option options [URI::HTTP,String] :http_proxy A proxy to send # requests through. Formatted like 'http://proxy.com:123'. # # @option options [Float] :http_open_timeout (15) The number of # seconds to wait when opening a HTTP session before raising a # `Timeout::Error`. # # @option options [Integer] :http_read_timeout (60) The default # number of seconds to wait for response data. This value can # safely be set per-request on the session. # # @option options [Float] :http_idle_timeout (5) The number of # seconds a connection is allowed to sit idle before it is # considered stale. Stale connections are closed and removed # from the pool before making a request. # # @option options [Float] :http_continue_timeout (1) The number of # seconds to wait for a 100-continue response before sending the # request body. This option has no effect unless the request has # "Expect" header set to "100-continue". Defaults to `nil` which # disables this behaviour. This value can safely be set per # request on the session. # # @option options [Boolean] :http_wire_trace (false) When `true`, # HTTP debug output will be sent to the `:logger`. # # @option options [Boolean] :ssl_verify_peer (true) When `true`, # SSL peer certificates are verified when establishing a # connection. # # @option options [String] :ssl_ca_bundle Full path to the SSL # certificate authority bundle file that should be used when # verifying peer certificates. If you do not pass # `:ssl_ca_bundle` or `:ssl_ca_directory` the the system default # will be used if available. # # @option options [String] :ssl_ca_directory Full path of the # directory that contains the unbundled SSL certificate # authority files for verifying peer certificates. If you do # not pass `:ssl_ca_bundle` or `:ssl_ca_directory` the the # system default will be used if available. # def initialize(*args) super end # @!group API Operations # Use this action to associate one or more Domain Name Server (DNS) # aliases with an existing Amazon FSx for Windows File Server file # system. A file systen can have a maximum of 50 DNS aliases associated # with it at any one time. If you try to associate a DNS alias that is # already associated with the file system, FSx takes no action on that # alias in the request. For more information, see [Working with DNS # Aliases][1] and [Walkthrough 5: Using DNS aliases to access your file # system][2], including additional steps you must take to be able to # access your file system using a DNS alias. # # The system response shows the DNS aliases that Amazon FSx is # attempting to associate with the file system. Use the API operation to # monitor the status of the aliases Amazon FSx is associating with the # file system. # # # # [1]: https://docs.aws.amazon.com/fsx/latest/WindowsGuide/managing-dns-aliases.html # [2]: https://docs.aws.amazon.com/fsx/latest/WindowsGuide/walkthrough05-file-system-custom-CNAME.html # # @option params [String] :client_request_token # (Optional) An idempotency token for resource creation, in a string of # up to 64 ASCII characters. This token is automatically filled on your # behalf when you use the AWS Command Line Interface (AWS CLI) or an AWS # SDK. # # **A suitable default value is auto-generated.** You should normally # not need to pass this option.** # # @option params [required, String] :file_system_id # Specifies the file system with which you want to associate one or more # DNS aliases. # # @option params [required, Array] :aliases # An array of one or more DNS alias names to associate with the file # system. The alias name has to comply with the following formatting # requirements: # # * Formatted as a fully-qualified domain name (FQDN), # hostname.domain , for example, # `accounting.corp.example.com`. # # * Can contain alphanumeric characters and the hyphen (-). # # * Cannot start or end with a hyphen. # # * Can start with a numeric. # # For DNS alias names, Amazon FSx stores alphabetic characters as # lowercase letters (a-z), regardless of how you specify them: as # uppercase letters, lowercase letters, or the corresponding letters in # escape codes. # # @return [Types::AssociateFileSystemAliasesResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods: # # * {Types::AssociateFileSystemAliasesResponse#aliases #aliases} => Array<Types::Alias> # # @example Request syntax with placeholder values # # resp = client.associate_file_system_aliases({ # client_request_token: "ClientRequestToken", # file_system_id: "FileSystemId", # required # aliases: ["AlternateDNSName"], # required # }) # # @example Response structure # # resp.aliases #=> Array # resp.aliases[0].name #=> String # resp.aliases[0].lifecycle #=> String, one of "AVAILABLE", "CREATING", "DELETING", "CREATE_FAILED", "DELETE_FAILED" # # @see http://docs.aws.amazon.com/goto/WebAPI/fsx-2018-03-01/AssociateFileSystemAliases AWS API Documentation # # @overload associate_file_system_aliases(params = {}) # @param [Hash] params ({}) def associate_file_system_aliases(params = {}, options = {}) req = build_request(:associate_file_system_aliases, params) req.send_request(options) end # Cancels an existing Amazon FSx for Lustre data repository task if that # task is in either the `PENDING` or `EXECUTING` state. When you cancel # a task, Amazon FSx does the following. # # * Any files that FSx has already exported are not reverted. # # * FSx continues to export any files that are "in-flight" when the # cancel operation is received. # # * FSx does not export any files that have not yet been exported. # # @option params [required, String] :task_id # Specifies the data repository task to cancel. # # @return [Types::CancelDataRepositoryTaskResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods: # # * {Types::CancelDataRepositoryTaskResponse#lifecycle #lifecycle} => String # * {Types::CancelDataRepositoryTaskResponse#task_id #task_id} => String # # @example Request syntax with placeholder values # # resp = client.cancel_data_repository_task({ # task_id: "TaskId", # required # }) # # @example Response structure # # resp.lifecycle #=> String, one of "PENDING", "EXECUTING", "FAILED", "SUCCEEDED", "CANCELED", "CANCELING" # resp.task_id #=> String # # @see http://docs.aws.amazon.com/goto/WebAPI/fsx-2018-03-01/CancelDataRepositoryTask AWS API Documentation # # @overload cancel_data_repository_task(params = {}) # @param [Hash] params ({}) def cancel_data_repository_task(params = {}, options = {}) req = build_request(:cancel_data_repository_task, params) req.send_request(options) end # Copies an existing backup within the same AWS account to another # Region (cross-Region copy) or within the same Region (in-Region copy). # You can have up to five backup copy requests in progress to a single # destination Region per account. # # You can use cross-Region backup copies for cross-region disaster # recovery. You periodically take backups and copy them to another # Region so that in the event of a disaster in the primary Region, you # can restore from backup and recover availability quickly in the other # Region. You can make cross-Region copies only within your AWS # partition. # # You can also use backup copies to clone your file data set to another # Region or within the same Region. # # You can use the `SourceRegion` parameter to specify the AWS Region # from which the backup will be copied. For example, if you make the # call from the `us-west-1` Region and want to copy a backup from the # `us-east-2` Region, you specify `us-east-2` in the `SourceRegion` # parameter to make a cross-Region copy. If you don't specify a Region, # the backup copy is created in the same Region where the request is # sent from (in-Region copy). # # For more information on creating backup copies, see [ Copying # backups][1] in the *Amazon FSx for Windows User Guide* and [Copying # backups][2] in the *Amazon FSx for Lustre User Guide*. # # # # [1]: https://docs.aws.amazon.com/fsx/latest/WindowsGuide/using-backups.html#copy-backups # [2]: https://docs.aws.amazon.com/fsx/latest/LustreGuide/using-backups-fsx.html#copy-backups # # @option params [String] :client_request_token # (Optional) An idempotency token for resource creation, in a string of # up to 64 ASCII characters. This token is automatically filled on your # behalf when you use the AWS Command Line Interface (AWS CLI) or an AWS # SDK. # # **A suitable default value is auto-generated.** You should normally # not need to pass this option.** # # @option params [required, String] :source_backup_id # The ID of the source backup. Specifies the ID of the backup that is # being copied. # # @option params [String] :source_region # The source AWS Region of the backup. Specifies the AWS Region from # which the backup is being copied. The source and destination Regions # must be in the same AWS partition. If you don't specify a Region, it # defaults to the Region where the request is sent from (in-Region # copy). # # @option params [String] :kms_key_id # The ID of the AWS Key Management Service (AWS KMS) key used to encrypt # the file system's data for Amazon FSx for Windows File Server file # systems and Amazon FSx for Lustre `PERSISTENT_1` file systems at rest. # In either case, if not specified, the Amazon FSx managed key is used. # The Amazon FSx for Lustre `SCRATCH_1` and `SCRATCH_2` file systems are # always encrypted at rest using Amazon FSx managed keys. For more # information, see [Encrypt][1] in the *AWS Key Management Service API # Reference*. # # # # [1]: https://docs.aws.amazon.com/kms/latest/APIReference/API_Encrypt.html # # @option params [Boolean] :copy_tags # A boolean flag indicating whether tags from the source backup should # be copied to the backup copy. This value defaults to false. # # If you set `CopyTags` to true and the source backup has existing tags, # you can use the `Tags` parameter to create new tags, provided that the # sum of the source backup tags and the new tags doesn't exceed 50. # Both sets of tags are merged. If there are tag conflicts (for example, # two tags with the same key but different values), the tags created # with the `Tags` parameter take precedence. # # @option params [Array] :tags # A list of `Tag` values, with a maximum of 50 elements. # # @return [Types::CopyBackupResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods: # # * {Types::CopyBackupResponse#backup #backup} => Types::Backup # # # @example Example: To copy a backup # # # This operation copies an Amazon FSx backup. # # resp = client.copy_backup({ # source_backup_id: "backup-03e3c82e0183b7b6b", # source_region: "us-east-2", # }) # # resp.to_h outputs the following: # { # backup: { # backup_id: "backup-0a3364eded1014b28", # creation_time: Time.parse(1617954808.068), # file_system: { # file_system_id: "fs-0498eed5fe91001ec", # file_system_type: "LUSTRE", # lustre_configuration: { # automatic_backup_retention_days: 0, # deployment_type: "PERSISTENT_1", # per_unit_storage_throughput: 50, # weekly_maintenance_start_time: "1:05:00", # }, # resource_arn: "arn:aws:fsx:us-east-1:012345678912:file-system/fs-0f5179e395f597e66", # storage_capacity: 2400, # storage_type: "SSD", # }, # kms_key_id: "arn:aws:fsx:us-east-1:012345678912:key/d1234e22-543a-12b7-a98f-e12c2b54001a", # lifecycle: "COPYING", # owner_id: "123456789012", # resource_arn: "arn:aws:fsx:us-east-1:012345678912:backup/backup-0a3364eded1014b28", # tags: [ # { # key: "Name", # value: "MyBackup", # }, # ], # type: "USER_INITIATED", # }, # } # # @example Request syntax with placeholder values # # resp = client.copy_backup({ # client_request_token: "ClientRequestToken", # source_backup_id: "SourceBackupId", # required # source_region: "Region", # kms_key_id: "KmsKeyId", # copy_tags: false, # tags: [ # { # key: "TagKey", # required # value: "TagValue", # required # }, # ], # }) # # @example Response structure # # resp.backup.backup_id #=> String # resp.backup.lifecycle #=> String, one of "AVAILABLE", "CREATING", "TRANSFERRING", "DELETED", "FAILED", "PENDING", "COPYING" # resp.backup.failure_details.message #=> String # resp.backup.type #=> String, one of "AUTOMATIC", "USER_INITIATED", "AWS_BACKUP" # resp.backup.progress_percent #=> Integer # resp.backup.creation_time #=> Time # resp.backup.kms_key_id #=> String # resp.backup.resource_arn #=> String # resp.backup.tags #=> Array # resp.backup.tags[0].key #=> String # resp.backup.tags[0].value #=> String # resp.backup.file_system.owner_id #=> String # resp.backup.file_system.creation_time #=> Time # resp.backup.file_system.file_system_id #=> String # resp.backup.file_system.file_system_type #=> String, one of "WINDOWS", "LUSTRE" # resp.backup.file_system.lifecycle #=> String, one of "AVAILABLE", "CREATING", "FAILED", "DELETING", "MISCONFIGURED", "UPDATING" # resp.backup.file_system.failure_details.message #=> String # resp.backup.file_system.storage_capacity #=> Integer # resp.backup.file_system.storage_type #=> String, one of "SSD", "HDD" # resp.backup.file_system.vpc_id #=> String # resp.backup.file_system.subnet_ids #=> Array # resp.backup.file_system.subnet_ids[0] #=> String # resp.backup.file_system.network_interface_ids #=> Array # resp.backup.file_system.network_interface_ids[0] #=> String # resp.backup.file_system.dns_name #=> String # resp.backup.file_system.kms_key_id #=> String # resp.backup.file_system.resource_arn #=> String # resp.backup.file_system.tags #=> Array # resp.backup.file_system.tags[0].key #=> String # resp.backup.file_system.tags[0].value #=> String # resp.backup.file_system.windows_configuration.active_directory_id #=> String # resp.backup.file_system.windows_configuration.self_managed_active_directory_configuration.domain_name #=> String # resp.backup.file_system.windows_configuration.self_managed_active_directory_configuration.organizational_unit_distinguished_name #=> String # resp.backup.file_system.windows_configuration.self_managed_active_directory_configuration.file_system_administrators_group #=> String # resp.backup.file_system.windows_configuration.self_managed_active_directory_configuration.user_name #=> String # resp.backup.file_system.windows_configuration.self_managed_active_directory_configuration.dns_ips #=> Array # resp.backup.file_system.windows_configuration.self_managed_active_directory_configuration.dns_ips[0] #=> String # resp.backup.file_system.windows_configuration.deployment_type #=> String, one of "MULTI_AZ_1", "SINGLE_AZ_1", "SINGLE_AZ_2" # resp.backup.file_system.windows_configuration.remote_administration_endpoint #=> String # resp.backup.file_system.windows_configuration.preferred_subnet_id #=> String # resp.backup.file_system.windows_configuration.preferred_file_server_ip #=> String # resp.backup.file_system.windows_configuration.throughput_capacity #=> Integer # resp.backup.file_system.windows_configuration.maintenance_operations_in_progress #=> Array # resp.backup.file_system.windows_configuration.maintenance_operations_in_progress[0] #=> String, one of "PATCHING", "BACKING_UP" # resp.backup.file_system.windows_configuration.weekly_maintenance_start_time #=> String # resp.backup.file_system.windows_configuration.daily_automatic_backup_start_time #=> String # resp.backup.file_system.windows_configuration.automatic_backup_retention_days #=> Integer # resp.backup.file_system.windows_configuration.copy_tags_to_backups #=> Boolean # resp.backup.file_system.windows_configuration.aliases #=> Array # resp.backup.file_system.windows_configuration.aliases[0].name #=> String # resp.backup.file_system.windows_configuration.aliases[0].lifecycle #=> String, one of "AVAILABLE", "CREATING", "DELETING", "CREATE_FAILED", "DELETE_FAILED" # resp.backup.file_system.windows_configuration.audit_log_configuration.file_access_audit_log_level #=> String, one of "DISABLED", "SUCCESS_ONLY", "FAILURE_ONLY", "SUCCESS_AND_FAILURE" # resp.backup.file_system.windows_configuration.audit_log_configuration.file_share_access_audit_log_level #=> String, one of "DISABLED", "SUCCESS_ONLY", "FAILURE_ONLY", "SUCCESS_AND_FAILURE" # resp.backup.file_system.windows_configuration.audit_log_configuration.audit_log_destination #=> String # resp.backup.file_system.lustre_configuration.weekly_maintenance_start_time #=> String # resp.backup.file_system.lustre_configuration.data_repository_configuration.lifecycle #=> String, one of "CREATING", "AVAILABLE", "MISCONFIGURED", "UPDATING", "DELETING" # resp.backup.file_system.lustre_configuration.data_repository_configuration.import_path #=> String # resp.backup.file_system.lustre_configuration.data_repository_configuration.export_path #=> String # resp.backup.file_system.lustre_configuration.data_repository_configuration.imported_file_chunk_size #=> Integer # resp.backup.file_system.lustre_configuration.data_repository_configuration.auto_import_policy #=> String, one of "NONE", "NEW", "NEW_CHANGED" # resp.backup.file_system.lustre_configuration.data_repository_configuration.failure_details.message #=> String # resp.backup.file_system.lustre_configuration.deployment_type #=> String, one of "SCRATCH_1", "SCRATCH_2", "PERSISTENT_1" # resp.backup.file_system.lustre_configuration.per_unit_storage_throughput #=> Integer # resp.backup.file_system.lustre_configuration.mount_name #=> String # resp.backup.file_system.lustre_configuration.daily_automatic_backup_start_time #=> String # resp.backup.file_system.lustre_configuration.automatic_backup_retention_days #=> Integer # resp.backup.file_system.lustre_configuration.copy_tags_to_backups #=> Boolean # resp.backup.file_system.lustre_configuration.drive_cache_type #=> String, one of "NONE", "READ" # resp.backup.file_system.lustre_configuration.data_compression_type #=> String, one of "NONE", "LZ4" # resp.backup.file_system.administrative_actions #=> Array # resp.backup.file_system.administrative_actions[0].administrative_action_type #=> String, one of "FILE_SYSTEM_UPDATE", "STORAGE_OPTIMIZATION", "FILE_SYSTEM_ALIAS_ASSOCIATION", "FILE_SYSTEM_ALIAS_DISASSOCIATION" # resp.backup.file_system.administrative_actions[0].progress_percent #=> Integer # resp.backup.file_system.administrative_actions[0].request_time #=> Time # resp.backup.file_system.administrative_actions[0].status #=> String, one of "FAILED", "IN_PROGRESS", "PENDING", "COMPLETED", "UPDATED_OPTIMIZING" # resp.backup.file_system.administrative_actions[0].target_file_system_values #=> Types::FileSystem # resp.backup.file_system.administrative_actions[0].failure_details.message #=> String # resp.backup.directory_information.domain_name #=> String # resp.backup.directory_information.active_directory_id #=> String # resp.backup.directory_information.resource_arn #=> String # resp.backup.owner_id #=> String # resp.backup.source_backup_id #=> String # resp.backup.source_backup_region #=> String # # @see http://docs.aws.amazon.com/goto/WebAPI/fsx-2018-03-01/CopyBackup AWS API Documentation # # @overload copy_backup(params = {}) # @param [Hash] params ({}) def copy_backup(params = {}, options = {}) req = build_request(:copy_backup, params) req.send_request(options) end # Creates a backup of an existing Amazon FSx file system. Creating # regular backups for your file system is a best practice, enabling you # to restore a file system from a backup if an issue arises with the # original file system. # # For Amazon FSx for Lustre file systems, you can create a backup only # for file systems with the following configuration: # # * a Persistent deployment type # # * is *not* linked to a data respository. # # For more information about backing up Amazon FSx for Lustre file # systems, see [Working with FSx for Lustre backups][1]. # # For more information about backing up Amazon FSx for Windows file # systems, see [Working with FSx for Windows backups][2]. # # If a backup with the specified client request token exists, and the # parameters match, this operation returns the description of the # existing backup. If a backup specified client request token exists, # and the parameters don't match, this operation returns # `IncompatibleParameterError`. If a backup with the specified client # request token doesn't exist, `CreateBackup` does the following: # # * Creates a new Amazon FSx backup with an assigned ID, and an initial # lifecycle state of `CREATING`. # # * Returns the description of the backup. # # By using the idempotent operation, you can retry a `CreateBackup` # operation without the risk of creating an extra backup. This approach # can be useful when an initial call fails in a way that makes it # unclear whether a backup was created. If you use the same client # request token and the initial call created a backup, the operation # returns a successful result because all the parameters are the same. # # The `CreateBackup` operation returns while the backup's lifecycle # state is still `CREATING`. You can check the backup creation status by # calling the DescribeBackups operation, which returns the backup state # along with other information. # # # # [1]: https://docs.aws.amazon.com/fsx/latest/LustreGuide/using-backups-fsx.html # [2]: https://docs.aws.amazon.com/fsx/latest/WindowsGuide/using-backups.html # # @option params [required, String] :file_system_id # The ID of the file system to back up. # # @option params [String] :client_request_token # (Optional) A string of up to 64 ASCII characters that Amazon FSx uses # to ensure idempotent creation. This string is automatically filled on # your behalf when you use the AWS Command Line Interface (AWS CLI) or # an AWS SDK. # # **A suitable default value is auto-generated.** You should normally # not need to pass this option.** # # @option params [Array] :tags # (Optional) The tags to apply to the backup at backup creation. The key # value of the `Name` tag appears in the console as the backup name. If # you have set `CopyTagsToBackups` to true, and you specify one or more # tags using the `CreateBackup` action, no existing file system tags are # copied from the file system to the backup. # # @return [Types::CreateBackupResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods: # # * {Types::CreateBackupResponse#backup #backup} => Types::Backup # # # @example Example: To create a new backup # # # This operation creates a new backup. # # resp = client.create_backup({ # file_system_id: "fs-0498eed5fe91001ec", # tags: [ # { # key: "Name", # value: "MyBackup", # }, # ], # }) # # resp.to_h outputs the following: # { # backup: { # backup_id: "backup-03e3c82e0183b7b6b", # creation_time: Time.parse("1481841524.0"), # file_system: { # file_system_id: "fs-0498eed5fe91001ec", # owner_id: "012345678912", # storage_capacity: 300, # windows_configuration: { # active_directory_id: "d-1234abcd12", # automatic_backup_retention_days: 30, # daily_automatic_backup_start_time: "05:00", # weekly_maintenance_start_time: "1:05:00", # }, # }, # lifecycle: "CREATING", # progress_percent: 0, # resource_arn: "arn:aws:fsx:us-east-1:012345678912:backup/backup-03e3c82e0183b7b6b", # tags: [ # { # key: "Name", # value: "MyBackup", # }, # ], # type: "USER_INITIATED", # }, # } # # @example Request syntax with placeholder values # # resp = client.create_backup({ # file_system_id: "FileSystemId", # required # client_request_token: "ClientRequestToken", # tags: [ # { # key: "TagKey", # required # value: "TagValue", # required # }, # ], # }) # # @example Response structure # # resp.backup.backup_id #=> String # resp.backup.lifecycle #=> String, one of "AVAILABLE", "CREATING", "TRANSFERRING", "DELETED", "FAILED", "PENDING", "COPYING" # resp.backup.failure_details.message #=> String # resp.backup.type #=> String, one of "AUTOMATIC", "USER_INITIATED", "AWS_BACKUP" # resp.backup.progress_percent #=> Integer # resp.backup.creation_time #=> Time # resp.backup.kms_key_id #=> String # resp.backup.resource_arn #=> String # resp.backup.tags #=> Array # resp.backup.tags[0].key #=> String # resp.backup.tags[0].value #=> String # resp.backup.file_system.owner_id #=> String # resp.backup.file_system.creation_time #=> Time # resp.backup.file_system.file_system_id #=> String # resp.backup.file_system.file_system_type #=> String, one of "WINDOWS", "LUSTRE" # resp.backup.file_system.lifecycle #=> String, one of "AVAILABLE", "CREATING", "FAILED", "DELETING", "MISCONFIGURED", "UPDATING" # resp.backup.file_system.failure_details.message #=> String # resp.backup.file_system.storage_capacity #=> Integer # resp.backup.file_system.storage_type #=> String, one of "SSD", "HDD" # resp.backup.file_system.vpc_id #=> String # resp.backup.file_system.subnet_ids #=> Array # resp.backup.file_system.subnet_ids[0] #=> String # resp.backup.file_system.network_interface_ids #=> Array # resp.backup.file_system.network_interface_ids[0] #=> String # resp.backup.file_system.dns_name #=> String # resp.backup.file_system.kms_key_id #=> String # resp.backup.file_system.resource_arn #=> String # resp.backup.file_system.tags #=> Array # resp.backup.file_system.tags[0].key #=> String # resp.backup.file_system.tags[0].value #=> String # resp.backup.file_system.windows_configuration.active_directory_id #=> String # resp.backup.file_system.windows_configuration.self_managed_active_directory_configuration.domain_name #=> String # resp.backup.file_system.windows_configuration.self_managed_active_directory_configuration.organizational_unit_distinguished_name #=> String # resp.backup.file_system.windows_configuration.self_managed_active_directory_configuration.file_system_administrators_group #=> String # resp.backup.file_system.windows_configuration.self_managed_active_directory_configuration.user_name #=> String # resp.backup.file_system.windows_configuration.self_managed_active_directory_configuration.dns_ips #=> Array # resp.backup.file_system.windows_configuration.self_managed_active_directory_configuration.dns_ips[0] #=> String # resp.backup.file_system.windows_configuration.deployment_type #=> String, one of "MULTI_AZ_1", "SINGLE_AZ_1", "SINGLE_AZ_2" # resp.backup.file_system.windows_configuration.remote_administration_endpoint #=> String # resp.backup.file_system.windows_configuration.preferred_subnet_id #=> String # resp.backup.file_system.windows_configuration.preferred_file_server_ip #=> String # resp.backup.file_system.windows_configuration.throughput_capacity #=> Integer # resp.backup.file_system.windows_configuration.maintenance_operations_in_progress #=> Array # resp.backup.file_system.windows_configuration.maintenance_operations_in_progress[0] #=> String, one of "PATCHING", "BACKING_UP" # resp.backup.file_system.windows_configuration.weekly_maintenance_start_time #=> String # resp.backup.file_system.windows_configuration.daily_automatic_backup_start_time #=> String # resp.backup.file_system.windows_configuration.automatic_backup_retention_days #=> Integer # resp.backup.file_system.windows_configuration.copy_tags_to_backups #=> Boolean # resp.backup.file_system.windows_configuration.aliases #=> Array # resp.backup.file_system.windows_configuration.aliases[0].name #=> String # resp.backup.file_system.windows_configuration.aliases[0].lifecycle #=> String, one of "AVAILABLE", "CREATING", "DELETING", "CREATE_FAILED", "DELETE_FAILED" # resp.backup.file_system.windows_configuration.audit_log_configuration.file_access_audit_log_level #=> String, one of "DISABLED", "SUCCESS_ONLY", "FAILURE_ONLY", "SUCCESS_AND_FAILURE" # resp.backup.file_system.windows_configuration.audit_log_configuration.file_share_access_audit_log_level #=> String, one of "DISABLED", "SUCCESS_ONLY", "FAILURE_ONLY", "SUCCESS_AND_FAILURE" # resp.backup.file_system.windows_configuration.audit_log_configuration.audit_log_destination #=> String # resp.backup.file_system.lustre_configuration.weekly_maintenance_start_time #=> String # resp.backup.file_system.lustre_configuration.data_repository_configuration.lifecycle #=> String, one of "CREATING", "AVAILABLE", "MISCONFIGURED", "UPDATING", "DELETING" # resp.backup.file_system.lustre_configuration.data_repository_configuration.import_path #=> String # resp.backup.file_system.lustre_configuration.data_repository_configuration.export_path #=> String # resp.backup.file_system.lustre_configuration.data_repository_configuration.imported_file_chunk_size #=> Integer # resp.backup.file_system.lustre_configuration.data_repository_configuration.auto_import_policy #=> String, one of "NONE", "NEW", "NEW_CHANGED" # resp.backup.file_system.lustre_configuration.data_repository_configuration.failure_details.message #=> String # resp.backup.file_system.lustre_configuration.deployment_type #=> String, one of "SCRATCH_1", "SCRATCH_2", "PERSISTENT_1" # resp.backup.file_system.lustre_configuration.per_unit_storage_throughput #=> Integer # resp.backup.file_system.lustre_configuration.mount_name #=> String # resp.backup.file_system.lustre_configuration.daily_automatic_backup_start_time #=> String # resp.backup.file_system.lustre_configuration.automatic_backup_retention_days #=> Integer # resp.backup.file_system.lustre_configuration.copy_tags_to_backups #=> Boolean # resp.backup.file_system.lustre_configuration.drive_cache_type #=> String, one of "NONE", "READ" # resp.backup.file_system.lustre_configuration.data_compression_type #=> String, one of "NONE", "LZ4" # resp.backup.file_system.administrative_actions #=> Array # resp.backup.file_system.administrative_actions[0].administrative_action_type #=> String, one of "FILE_SYSTEM_UPDATE", "STORAGE_OPTIMIZATION", "FILE_SYSTEM_ALIAS_ASSOCIATION", "FILE_SYSTEM_ALIAS_DISASSOCIATION" # resp.backup.file_system.administrative_actions[0].progress_percent #=> Integer # resp.backup.file_system.administrative_actions[0].request_time #=> Time # resp.backup.file_system.administrative_actions[0].status #=> String, one of "FAILED", "IN_PROGRESS", "PENDING", "COMPLETED", "UPDATED_OPTIMIZING" # resp.backup.file_system.administrative_actions[0].target_file_system_values #=> Types::FileSystem # resp.backup.file_system.administrative_actions[0].failure_details.message #=> String # resp.backup.directory_information.domain_name #=> String # resp.backup.directory_information.active_directory_id #=> String # resp.backup.directory_information.resource_arn #=> String # resp.backup.owner_id #=> String # resp.backup.source_backup_id #=> String # resp.backup.source_backup_region #=> String # # @see http://docs.aws.amazon.com/goto/WebAPI/fsx-2018-03-01/CreateBackup AWS API Documentation # # @overload create_backup(params = {}) # @param [Hash] params ({}) def create_backup(params = {}, options = {}) req = build_request(:create_backup, params) req.send_request(options) end # Creates an Amazon FSx for Lustre data repository task. You use data # repository tasks to perform bulk operations between your Amazon FSx # file system and its linked data repository. An example of a data # repository task is exporting any data and metadata changes, including # POSIX metadata, to files, directories, and symbolic links (symlinks) # from your FSx file system to its linked data repository. A # `CreateDataRepositoryTask` operation will fail if a data repository is # not linked to the FSx file system. To learn more about data repository # tasks, see [Data Repository Tasks][1]. To learn more about linking a # data repository to your file system, see [Linking your file system to # an S3 bucket][2]. # # # # [1]: https://docs.aws.amazon.com/fsx/latest/LustreGuide/data-repository-tasks.html # [2]: https://docs.aws.amazon.com/fsx/latest/LustreGuide/create-fs-linked-data-repo.html # # @option params [required, String] :type # Specifies the type of data repository task to create. # # @option params [Array] :paths # (Optional) The path or paths on the Amazon FSx file system to use when # the data repository task is processed. The default path is the file # system root directory. The paths you provide need to be relative to # the mount point of the file system. If the mount point is `/mnt/fsx` # and `/mnt/fsx/path1` is a directory or file on the file system you # want to export, then the path to provide is `path1`. If a path that # you provide isn't valid, the task fails. # # @option params [required, String] :file_system_id # The globally unique ID of the file system, assigned by Amazon FSx. # # @option params [required, Types::CompletionReport] :report # Defines whether or not Amazon FSx provides a CompletionReport once the # task has completed. A CompletionReport provides a detailed report on # the files that Amazon FSx processed that meet the criteria specified # by the `Scope` parameter. For more information, see [Working with Task # Completion Reports][1]. # # # # [1]: https://docs.aws.amazon.com/fsx/latest/LustreGuide/task-completion-report.html # # @option params [String] :client_request_token # (Optional) An idempotency token for resource creation, in a string of # up to 64 ASCII characters. This token is automatically filled on your # behalf when you use the AWS Command Line Interface (AWS CLI) or an AWS # SDK. # # **A suitable default value is auto-generated.** You should normally # not need to pass this option.** # # @option params [Array] :tags # A list of `Tag` values, with a maximum of 50 elements. # # @return [Types::CreateDataRepositoryTaskResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods: # # * {Types::CreateDataRepositoryTaskResponse#data_repository_task #data_repository_task} => Types::DataRepositoryTask # # @example Request syntax with placeholder values # # resp = client.create_data_repository_task({ # type: "EXPORT_TO_REPOSITORY", # required, accepts EXPORT_TO_REPOSITORY # paths: ["DataRepositoryTaskPath"], # file_system_id: "FileSystemId", # required # report: { # required # enabled: false, # required # path: "ArchivePath", # format: "REPORT_CSV_20191124", # accepts REPORT_CSV_20191124 # scope: "FAILED_FILES_ONLY", # accepts FAILED_FILES_ONLY # }, # client_request_token: "ClientRequestToken", # tags: [ # { # key: "TagKey", # required # value: "TagValue", # required # }, # ], # }) # # @example Response structure # # resp.data_repository_task.task_id #=> String # resp.data_repository_task.lifecycle #=> String, one of "PENDING", "EXECUTING", "FAILED", "SUCCEEDED", "CANCELED", "CANCELING" # resp.data_repository_task.type #=> String, one of "EXPORT_TO_REPOSITORY" # resp.data_repository_task.creation_time #=> Time # resp.data_repository_task.start_time #=> Time # resp.data_repository_task.end_time #=> Time # resp.data_repository_task.resource_arn #=> String # resp.data_repository_task.tags #=> Array # resp.data_repository_task.tags[0].key #=> String # resp.data_repository_task.tags[0].value #=> String # resp.data_repository_task.file_system_id #=> String # resp.data_repository_task.paths #=> Array # resp.data_repository_task.paths[0] #=> String # resp.data_repository_task.failure_details.message #=> String # resp.data_repository_task.status.total_count #=> Integer # resp.data_repository_task.status.succeeded_count #=> Integer # resp.data_repository_task.status.failed_count #=> Integer # resp.data_repository_task.status.last_updated_time #=> Time # resp.data_repository_task.report.enabled #=> Boolean # resp.data_repository_task.report.path #=> String # resp.data_repository_task.report.format #=> String, one of "REPORT_CSV_20191124" # resp.data_repository_task.report.scope #=> String, one of "FAILED_FILES_ONLY" # # @see http://docs.aws.amazon.com/goto/WebAPI/fsx-2018-03-01/CreateDataRepositoryTask AWS API Documentation # # @overload create_data_repository_task(params = {}) # @param [Hash] params ({}) def create_data_repository_task(params = {}, options = {}) req = build_request(:create_data_repository_task, params) req.send_request(options) end # Creates a new, empty Amazon FSx file system. # # If a file system with the specified client request token exists and # the parameters match, `CreateFileSystem` returns the description of # the existing file system. If a file system specified client request # token exists and the parameters don't match, this call returns # `IncompatibleParameterError`. If a file system with the specified # client request token doesn't exist, `CreateFileSystem` does the # following: # # * Creates a new, empty Amazon FSx file system with an assigned ID, and # an initial lifecycle state of `CREATING`. # # * Returns the description of the file system. # # This operation requires a client request token in the request that # Amazon FSx uses to ensure idempotent creation. This means that calling # the operation multiple times with the same client request token has no # effect. By using the idempotent operation, you can retry a # `CreateFileSystem` operation without the risk of creating an extra # file system. This approach can be useful when an initial call fails in # a way that makes it unclear whether a file system was created. # Examples are if a transport level timeout occurred, or your connection # was reset. If you use the same client request token and the initial # call created a file system, the client receives success as long as the # parameters are the same. # # The `CreateFileSystem` call returns while the file system's lifecycle # state is still `CREATING`. You can check the file-system creation # status by calling the DescribeFileSystems operation, which returns the # file system state along with other information. # # # # @option params [String] :client_request_token # A string of up to 64 ASCII characters that Amazon FSx uses to ensure # idempotent creation. This string is automatically filled on your # behalf when you use the AWS Command Line Interface (AWS CLI) or an AWS # SDK. # # **A suitable default value is auto-generated.** You should normally # not need to pass this option.** # # @option params [required, String] :file_system_type # The type of Amazon FSx file system to create, either `WINDOWS` or # `LUSTRE`. # # @option params [required, Integer] :storage_capacity # Sets the storage capacity of the file system that you're creating. # # For Lustre file systems: # # * For `SCRATCH_2` and `PERSISTENT_1 SSD` deployment types, valid # values are 1200 GiB, 2400 GiB, and increments of 2400 GiB. # # * For `PERSISTENT HDD` file systems, valid values are increments of # 6000 GiB for 12 MB/s/TiB file systems and increments of 1800 GiB for # 40 MB/s/TiB file systems. # # * For `SCRATCH_1` deployment type, valid values are 1200 GiB, 2400 # GiB, and increments of 3600 GiB. # # For Windows file systems: # # * If `StorageType=SSD`, valid values are 32 GiB - 65,536 GiB (64 TiB). # # * If `StorageType=HDD`, valid values are 2000 GiB - 65,536 GiB (64 # TiB). # # @option params [String] :storage_type # Sets the storage type for the file system you're creating. Valid # values are `SSD` and `HDD`. # # * Set to `SSD` to use solid state drive storage. SSD is supported on # all Windows and Lustre deployment types. # # * Set to `HDD` to use hard disk drive storage. HDD is supported on # `SINGLE_AZ_2` and `MULTI_AZ_1` Windows file system deployment types, # and on `PERSISTENT` Lustre file system deployment types. # # Default value is `SSD`. For more information, see [ Storage Type # Options][1] in the *Amazon FSx for Windows User Guide* and [Multiple # Storage Options][2] in the *Amazon FSx for Lustre User Guide*. # # # # [1]: https://docs.aws.amazon.com/fsx/latest/WindowsGuide/optimize-fsx-costs.html#storage-type-options # [2]: https://docs.aws.amazon.com/fsx/latest/LustreGuide/what-is.html#storage-options # # @option params [required, Array] :subnet_ids # Specifies the IDs of the subnets that the file system will be # accessible from. For Windows `MULTI_AZ_1` file system deployment # types, provide exactly two subnet IDs, one for the preferred file # server and one for the standby file server. You specify one of these # subnets as the preferred subnet using the `WindowsConfiguration > # PreferredSubnetID` property. For more information, see [ Availability # and durability: Single-AZ and Multi-AZ file systems][1]. # # For Windows `SINGLE_AZ_1` and `SINGLE_AZ_2` file system deployment # types and Lustre file systems, provide exactly one subnet ID. The file # server is launched in that subnet's Availability Zone. # # # # [1]: https://docs.aws.amazon.com/fsx/latest/WindowsGuide/high-availability-multiAZ.html # # @option params [Array] :security_group_ids # A list of IDs specifying the security groups to apply to all network # interfaces created for file system access. This list isn't returned # in later requests to describe the file system. # # @option params [Array] :tags # The tags to apply to the file system being created. The key value of # the `Name` tag appears in the console as the file system name. # # @option params [String] :kms_key_id # The ID of the AWS Key Management Service (AWS KMS) key used to encrypt # the file system's data for Amazon FSx for Windows File Server file # systems and Amazon FSx for Lustre `PERSISTENT_1` file systems at rest. # In either case, if not specified, the Amazon FSx managed key is used. # The Amazon FSx for Lustre `SCRATCH_1` and `SCRATCH_2` file systems are # always encrypted at rest using Amazon FSx managed keys. For more # information, see [Encrypt][1] in the *AWS Key Management Service API # Reference*. # # # # [1]: https://docs.aws.amazon.com/kms/latest/APIReference/API_Encrypt.html # # @option params [Types::CreateFileSystemWindowsConfiguration] :windows_configuration # The Microsoft Windows configuration for the file system being created. # # @option params [Types::CreateFileSystemLustreConfiguration] :lustre_configuration # The Lustre configuration for the file system being created. # # @return [Types::CreateFileSystemResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods: # # * {Types::CreateFileSystemResponse#file_system #file_system} => Types::FileSystem # # # @example Example: To create a new file system # # # This operation creates a new Amazon FSx for Windows File Server file system. # # resp = client.create_file_system({ # client_request_token: "a8ca07e4-61ec-4399-99f4-19853801bcd5", # file_system_type: "WINDOWS", # kms_key_id: "arn:aws:kms:us-east-1:012345678912:key/1111abcd-2222-3333-4444-55556666eeff", # security_group_ids: [ # "sg-edcd9784", # ], # storage_capacity: 3200, # storage_type: "HDD", # subnet_ids: [ # "subnet-1234abcd", # ], # tags: [ # { # key: "Name", # value: "MyFileSystem", # }, # ], # windows_configuration: { # active_directory_id: "d-1234abcd12", # aliases: [ # "accounting.corp.example.com", # ], # automatic_backup_retention_days: 30, # daily_automatic_backup_start_time: "05:00", # throughput_capacity: 32, # weekly_maintenance_start_time: "1:05:00", # }, # }) # # resp.to_h outputs the following: # { # file_system: { # creation_time: Time.parse("1481841524.0"), # dns_name: "fs-0123456789abcdef0.fsx.com", # file_system_id: "fs-0123456789abcdef0", # kms_key_id: "arn:aws:kms:us-east-1:012345678912:key/1111abcd-2222-3333-4444-55556666eeff", # lifecycle: "CREATING", # owner_id: "012345678912", # resource_arn: "arn:aws:fsx:us-east-1:012345678912:file-system/fs-0123456789abcdef0", # storage_capacity: 3200, # storage_type: "HDD", # subnet_ids: [ # "subnet-1234abcd", # ], # tags: [ # { # key: "Name", # value: "MyFileSystem", # }, # ], # vpc_id: "vpc-ab1234cd", # windows_configuration: { # active_directory_id: "d-1234abcd12", # aliases: [ # { # lifecycle: "CREATING", # name: "accounting.corp.example.com", # }, # ], # automatic_backup_retention_days: 30, # daily_automatic_backup_start_time: "05:00", # throughput_capacity: 32, # weekly_maintenance_start_time: "1:05:00", # }, # }, # } # # @example Request syntax with placeholder values # # resp = client.create_file_system({ # client_request_token: "ClientRequestToken", # file_system_type: "WINDOWS", # required, accepts WINDOWS, LUSTRE # storage_capacity: 1, # required # storage_type: "SSD", # accepts SSD, HDD # subnet_ids: ["SubnetId"], # required # security_group_ids: ["SecurityGroupId"], # tags: [ # { # key: "TagKey", # required # value: "TagValue", # required # }, # ], # kms_key_id: "KmsKeyId", # windows_configuration: { # active_directory_id: "DirectoryId", # self_managed_active_directory_configuration: { # domain_name: "ActiveDirectoryFullyQualifiedName", # required # organizational_unit_distinguished_name: "OrganizationalUnitDistinguishedName", # file_system_administrators_group: "FileSystemAdministratorsGroupName", # user_name: "DirectoryUserName", # required # password: "DirectoryPassword", # required # dns_ips: ["IpAddress"], # required # }, # deployment_type: "MULTI_AZ_1", # accepts MULTI_AZ_1, SINGLE_AZ_1, SINGLE_AZ_2 # preferred_subnet_id: "SubnetId", # throughput_capacity: 1, # required # weekly_maintenance_start_time: "WeeklyTime", # daily_automatic_backup_start_time: "DailyTime", # automatic_backup_retention_days: 1, # copy_tags_to_backups: false, # aliases: ["AlternateDNSName"], # audit_log_configuration: { # file_access_audit_log_level: "DISABLED", # required, accepts DISABLED, SUCCESS_ONLY, FAILURE_ONLY, SUCCESS_AND_FAILURE # file_share_access_audit_log_level: "DISABLED", # required, accepts DISABLED, SUCCESS_ONLY, FAILURE_ONLY, SUCCESS_AND_FAILURE # audit_log_destination: "GeneralARN", # }, # }, # lustre_configuration: { # weekly_maintenance_start_time: "WeeklyTime", # import_path: "ArchivePath", # export_path: "ArchivePath", # imported_file_chunk_size: 1, # deployment_type: "SCRATCH_1", # accepts SCRATCH_1, SCRATCH_2, PERSISTENT_1 # auto_import_policy: "NONE", # accepts NONE, NEW, NEW_CHANGED # per_unit_storage_throughput: 1, # daily_automatic_backup_start_time: "DailyTime", # automatic_backup_retention_days: 1, # copy_tags_to_backups: false, # drive_cache_type: "NONE", # accepts NONE, READ # data_compression_type: "NONE", # accepts NONE, LZ4 # }, # }) # # @example Response structure # # resp.file_system.owner_id #=> String # resp.file_system.creation_time #=> Time # resp.file_system.file_system_id #=> String # resp.file_system.file_system_type #=> String, one of "WINDOWS", "LUSTRE" # resp.file_system.lifecycle #=> String, one of "AVAILABLE", "CREATING", "FAILED", "DELETING", "MISCONFIGURED", "UPDATING" # resp.file_system.failure_details.message #=> String # resp.file_system.storage_capacity #=> Integer # resp.file_system.storage_type #=> String, one of "SSD", "HDD" # resp.file_system.vpc_id #=> String # resp.file_system.subnet_ids #=> Array # resp.file_system.subnet_ids[0] #=> String # resp.file_system.network_interface_ids #=> Array # resp.file_system.network_interface_ids[0] #=> String # resp.file_system.dns_name #=> String # resp.file_system.kms_key_id #=> String # resp.file_system.resource_arn #=> String # resp.file_system.tags #=> Array # resp.file_system.tags[0].key #=> String # resp.file_system.tags[0].value #=> String # resp.file_system.windows_configuration.active_directory_id #=> String # resp.file_system.windows_configuration.self_managed_active_directory_configuration.domain_name #=> String # resp.file_system.windows_configuration.self_managed_active_directory_configuration.organizational_unit_distinguished_name #=> String # resp.file_system.windows_configuration.self_managed_active_directory_configuration.file_system_administrators_group #=> String # resp.file_system.windows_configuration.self_managed_active_directory_configuration.user_name #=> String # resp.file_system.windows_configuration.self_managed_active_directory_configuration.dns_ips #=> Array # resp.file_system.windows_configuration.self_managed_active_directory_configuration.dns_ips[0] #=> String # resp.file_system.windows_configuration.deployment_type #=> String, one of "MULTI_AZ_1", "SINGLE_AZ_1", "SINGLE_AZ_2" # resp.file_system.windows_configuration.remote_administration_endpoint #=> String # resp.file_system.windows_configuration.preferred_subnet_id #=> String # resp.file_system.windows_configuration.preferred_file_server_ip #=> String # resp.file_system.windows_configuration.throughput_capacity #=> Integer # resp.file_system.windows_configuration.maintenance_operations_in_progress #=> Array # resp.file_system.windows_configuration.maintenance_operations_in_progress[0] #=> String, one of "PATCHING", "BACKING_UP" # resp.file_system.windows_configuration.weekly_maintenance_start_time #=> String # resp.file_system.windows_configuration.daily_automatic_backup_start_time #=> String # resp.file_system.windows_configuration.automatic_backup_retention_days #=> Integer # resp.file_system.windows_configuration.copy_tags_to_backups #=> Boolean # resp.file_system.windows_configuration.aliases #=> Array # resp.file_system.windows_configuration.aliases[0].name #=> String # resp.file_system.windows_configuration.aliases[0].lifecycle #=> String, one of "AVAILABLE", "CREATING", "DELETING", "CREATE_FAILED", "DELETE_FAILED" # resp.file_system.windows_configuration.audit_log_configuration.file_access_audit_log_level #=> String, one of "DISABLED", "SUCCESS_ONLY", "FAILURE_ONLY", "SUCCESS_AND_FAILURE" # resp.file_system.windows_configuration.audit_log_configuration.file_share_access_audit_log_level #=> String, one of "DISABLED", "SUCCESS_ONLY", "FAILURE_ONLY", "SUCCESS_AND_FAILURE" # resp.file_system.windows_configuration.audit_log_configuration.audit_log_destination #=> String # resp.file_system.lustre_configuration.weekly_maintenance_start_time #=> String # resp.file_system.lustre_configuration.data_repository_configuration.lifecycle #=> String, one of "CREATING", "AVAILABLE", "MISCONFIGURED", "UPDATING", "DELETING" # resp.file_system.lustre_configuration.data_repository_configuration.import_path #=> String # resp.file_system.lustre_configuration.data_repository_configuration.export_path #=> String # resp.file_system.lustre_configuration.data_repository_configuration.imported_file_chunk_size #=> Integer # resp.file_system.lustre_configuration.data_repository_configuration.auto_import_policy #=> String, one of "NONE", "NEW", "NEW_CHANGED" # resp.file_system.lustre_configuration.data_repository_configuration.failure_details.message #=> String # resp.file_system.lustre_configuration.deployment_type #=> String, one of "SCRATCH_1", "SCRATCH_2", "PERSISTENT_1" # resp.file_system.lustre_configuration.per_unit_storage_throughput #=> Integer # resp.file_system.lustre_configuration.mount_name #=> String # resp.file_system.lustre_configuration.daily_automatic_backup_start_time #=> String # resp.file_system.lustre_configuration.automatic_backup_retention_days #=> Integer # resp.file_system.lustre_configuration.copy_tags_to_backups #=> Boolean # resp.file_system.lustre_configuration.drive_cache_type #=> String, one of "NONE", "READ" # resp.file_system.lustre_configuration.data_compression_type #=> String, one of "NONE", "LZ4" # resp.file_system.administrative_actions #=> Array # resp.file_system.administrative_actions[0].administrative_action_type #=> String, one of "FILE_SYSTEM_UPDATE", "STORAGE_OPTIMIZATION", "FILE_SYSTEM_ALIAS_ASSOCIATION", "FILE_SYSTEM_ALIAS_DISASSOCIATION" # resp.file_system.administrative_actions[0].progress_percent #=> Integer # resp.file_system.administrative_actions[0].request_time #=> Time # resp.file_system.administrative_actions[0].status #=> String, one of "FAILED", "IN_PROGRESS", "PENDING", "COMPLETED", "UPDATED_OPTIMIZING" # resp.file_system.administrative_actions[0].target_file_system_values #=> Types::FileSystem # resp.file_system.administrative_actions[0].failure_details.message #=> String # # @see http://docs.aws.amazon.com/goto/WebAPI/fsx-2018-03-01/CreateFileSystem AWS API Documentation # # @overload create_file_system(params = {}) # @param [Hash] params ({}) def create_file_system(params = {}, options = {}) req = build_request(:create_file_system, params) req.send_request(options) end # Creates a new Amazon FSx file system from an existing Amazon FSx # backup. # # If a file system with the specified client request token exists and # the parameters match, this operation returns the description of the # file system. If a client request token specified by the file system # exists and the parameters don't match, this call returns # `IncompatibleParameterError`. If a file system with the specified # client request token doesn't exist, this operation does the # following: # # * Creates a new Amazon FSx file system from backup with an assigned # ID, and an initial lifecycle state of `CREATING`. # # * Returns the description of the file system. # # Parameters like Active Directory, default share name, automatic # backup, and backup settings default to the parameters of the file # system that was backed up, unless overridden. You can explicitly # supply other settings. # # By using the idempotent operation, you can retry a # `CreateFileSystemFromBackup` call without the risk of creating an # extra file system. This approach can be useful when an initial call # fails in a way that makes it unclear whether a file system was # created. Examples are if a transport level timeout occurred, or your # connection was reset. If you use the same client request token and the # initial call created a file system, the client receives success as # long as the parameters are the same. # # The `CreateFileSystemFromBackup` call returns while the file system's # lifecycle state is still `CREATING`. You can check the file-system # creation status by calling the DescribeFileSystems operation, which # returns the file system state along with other information. # # # # @option params [required, String] :backup_id # The ID of the source backup. Specifies the backup you are copying. # # @option params [String] :client_request_token # A string of up to 64 ASCII characters that Amazon FSx uses to ensure # idempotent creation. This string is automatically filled on your # behalf when you use the AWS Command Line Interface (AWS CLI) or an AWS # SDK. # # **A suitable default value is auto-generated.** You should normally # not need to pass this option.** # # @option params [required, Array] :subnet_ids # Specifies the IDs of the subnets that the file system will be # accessible from. For Windows `MULTI_AZ_1` file system deployment # types, provide exactly two subnet IDs, one for the preferred file # server and one for the standby file server. You specify one of these # subnets as the preferred subnet using the `WindowsConfiguration > # PreferredSubnetID` property. # # For Windows `SINGLE_AZ_1` and `SINGLE_AZ_2` deployment types and # Lustre file systems, provide exactly one subnet ID. The file server is # launched in that subnet's Availability Zone. # # @option params [Array] :security_group_ids # A list of IDs for the security groups that apply to the specified # network interfaces created for file system access. These security # groups apply to all network interfaces. This value isn't returned in # later DescribeFileSystem requests. # # @option params [Array] :tags # The tags to be applied to the file system at file system creation. The # key value of the `Name` tag appears in the console as the file system # name. # # @option params [Types::CreateFileSystemWindowsConfiguration] :windows_configuration # The configuration for this Microsoft Windows file system. # # @option params [Types::CreateFileSystemLustreConfiguration] :lustre_configuration # The Lustre configuration for the file system being created. # # @option params [String] :storage_type # Sets the storage type for the Windows file system you're creating # from a backup. Valid values are `SSD` and `HDD`. # # * Set to `SSD` to use solid state drive storage. Supported on all # Windows deployment types. # # * Set to `HDD` to use hard disk drive storage. Supported on # `SINGLE_AZ_2` and `MULTI_AZ_1` Windows file system deployment types. # # Default value is `SSD`. # # HDD and SSD storage types have different minimum storage capacity # requirements. A restored file system's storage capacity is tied to # the file system that was backed up. You can create a file system that # uses HDD storage from a backup of a file system that used SSD storage # only if the original SSD file system had a storage capacity of at # least 2000 GiB. # # # # @option params [String] :kms_key_id # The ID of the AWS Key Management Service (AWS KMS) key used to encrypt # the file system's data for Amazon FSx for Windows File Server file # systems and Amazon FSx for Lustre `PERSISTENT_1` file systems at rest. # In either case, if not specified, the Amazon FSx managed key is used. # The Amazon FSx for Lustre `SCRATCH_1` and `SCRATCH_2` file systems are # always encrypted at rest using Amazon FSx managed keys. For more # information, see [Encrypt][1] in the *AWS Key Management Service API # Reference*. # # # # [1]: https://docs.aws.amazon.com/kms/latest/APIReference/API_Encrypt.html # # @return [Types::CreateFileSystemFromBackupResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods: # # * {Types::CreateFileSystemFromBackupResponse#file_system #file_system} => Types::FileSystem # # # @example Example: To create a new file system from backup # # # This operation creates a new file system from backup. # # resp = client.create_file_system_from_backup({ # backup_id: "backup-03e3c82e0183b7b6b", # client_request_token: "f4c94ed7-238d-4c46-93db-48cd62ec33b7", # security_group_ids: [ # "sg-edcd9784", # ], # subnet_ids: [ # "subnet-1234abcd", # ], # tags: [ # { # key: "Name", # value: "MyFileSystem", # }, # ], # windows_configuration: { # throughput_capacity: 8, # }, # }) # # resp.to_h outputs the following: # { # file_system: { # creation_time: Time.parse("1481841524.0"), # dns_name: "fs-0498eed5fe91001ec.fsx.com", # file_system_id: "fs-0498eed5fe91001ec", # kms_key_id: "arn:aws:kms:us-east-1:012345678912:key/0ff3ea8d-130e-4133-877f-93908b6fdbd6", # lifecycle: "CREATING", # owner_id: "012345678912", # resource_arn: "arn:aws:fsx:us-east-1:012345678912:file-system/fs-0498eed5fe91001ec", # storage_capacity: 300, # subnet_ids: [ # "subnet-1234abcd", # ], # tags: [ # { # key: "Name", # value: "MyFileSystem", # }, # ], # vpc_id: "vpc-ab1234cd", # windows_configuration: { # active_directory_id: "d-1234abcd12", # automatic_backup_retention_days: 30, # daily_automatic_backup_start_time: "05:00", # throughput_capacity: 8, # weekly_maintenance_start_time: "1:05:00", # }, # }, # } # # @example Request syntax with placeholder values # # resp = client.create_file_system_from_backup({ # backup_id: "BackupId", # required # client_request_token: "ClientRequestToken", # subnet_ids: ["SubnetId"], # required # security_group_ids: ["SecurityGroupId"], # tags: [ # { # key: "TagKey", # required # value: "TagValue", # required # }, # ], # windows_configuration: { # active_directory_id: "DirectoryId", # self_managed_active_directory_configuration: { # domain_name: "ActiveDirectoryFullyQualifiedName", # required # organizational_unit_distinguished_name: "OrganizationalUnitDistinguishedName", # file_system_administrators_group: "FileSystemAdministratorsGroupName", # user_name: "DirectoryUserName", # required # password: "DirectoryPassword", # required # dns_ips: ["IpAddress"], # required # }, # deployment_type: "MULTI_AZ_1", # accepts MULTI_AZ_1, SINGLE_AZ_1, SINGLE_AZ_2 # preferred_subnet_id: "SubnetId", # throughput_capacity: 1, # required # weekly_maintenance_start_time: "WeeklyTime", # daily_automatic_backup_start_time: "DailyTime", # automatic_backup_retention_days: 1, # copy_tags_to_backups: false, # aliases: ["AlternateDNSName"], # audit_log_configuration: { # file_access_audit_log_level: "DISABLED", # required, accepts DISABLED, SUCCESS_ONLY, FAILURE_ONLY, SUCCESS_AND_FAILURE # file_share_access_audit_log_level: "DISABLED", # required, accepts DISABLED, SUCCESS_ONLY, FAILURE_ONLY, SUCCESS_AND_FAILURE # audit_log_destination: "GeneralARN", # }, # }, # lustre_configuration: { # weekly_maintenance_start_time: "WeeklyTime", # import_path: "ArchivePath", # export_path: "ArchivePath", # imported_file_chunk_size: 1, # deployment_type: "SCRATCH_1", # accepts SCRATCH_1, SCRATCH_2, PERSISTENT_1 # auto_import_policy: "NONE", # accepts NONE, NEW, NEW_CHANGED # per_unit_storage_throughput: 1, # daily_automatic_backup_start_time: "DailyTime", # automatic_backup_retention_days: 1, # copy_tags_to_backups: false, # drive_cache_type: "NONE", # accepts NONE, READ # data_compression_type: "NONE", # accepts NONE, LZ4 # }, # storage_type: "SSD", # accepts SSD, HDD # kms_key_id: "KmsKeyId", # }) # # @example Response structure # # resp.file_system.owner_id #=> String # resp.file_system.creation_time #=> Time # resp.file_system.file_system_id #=> String # resp.file_system.file_system_type #=> String, one of "WINDOWS", "LUSTRE" # resp.file_system.lifecycle #=> String, one of "AVAILABLE", "CREATING", "FAILED", "DELETING", "MISCONFIGURED", "UPDATING" # resp.file_system.failure_details.message #=> String # resp.file_system.storage_capacity #=> Integer # resp.file_system.storage_type #=> String, one of "SSD", "HDD" # resp.file_system.vpc_id #=> String # resp.file_system.subnet_ids #=> Array # resp.file_system.subnet_ids[0] #=> String # resp.file_system.network_interface_ids #=> Array # resp.file_system.network_interface_ids[0] #=> String # resp.file_system.dns_name #=> String # resp.file_system.kms_key_id #=> String # resp.file_system.resource_arn #=> String # resp.file_system.tags #=> Array # resp.file_system.tags[0].key #=> String # resp.file_system.tags[0].value #=> String # resp.file_system.windows_configuration.active_directory_id #=> String # resp.file_system.windows_configuration.self_managed_active_directory_configuration.domain_name #=> String # resp.file_system.windows_configuration.self_managed_active_directory_configuration.organizational_unit_distinguished_name #=> String # resp.file_system.windows_configuration.self_managed_active_directory_configuration.file_system_administrators_group #=> String # resp.file_system.windows_configuration.self_managed_active_directory_configuration.user_name #=> String # resp.file_system.windows_configuration.self_managed_active_directory_configuration.dns_ips #=> Array # resp.file_system.windows_configuration.self_managed_active_directory_configuration.dns_ips[0] #=> String # resp.file_system.windows_configuration.deployment_type #=> String, one of "MULTI_AZ_1", "SINGLE_AZ_1", "SINGLE_AZ_2" # resp.file_system.windows_configuration.remote_administration_endpoint #=> String # resp.file_system.windows_configuration.preferred_subnet_id #=> String # resp.file_system.windows_configuration.preferred_file_server_ip #=> String # resp.file_system.windows_configuration.throughput_capacity #=> Integer # resp.file_system.windows_configuration.maintenance_operations_in_progress #=> Array # resp.file_system.windows_configuration.maintenance_operations_in_progress[0] #=> String, one of "PATCHING", "BACKING_UP" # resp.file_system.windows_configuration.weekly_maintenance_start_time #=> String # resp.file_system.windows_configuration.daily_automatic_backup_start_time #=> String # resp.file_system.windows_configuration.automatic_backup_retention_days #=> Integer # resp.file_system.windows_configuration.copy_tags_to_backups #=> Boolean # resp.file_system.windows_configuration.aliases #=> Array # resp.file_system.windows_configuration.aliases[0].name #=> String # resp.file_system.windows_configuration.aliases[0].lifecycle #=> String, one of "AVAILABLE", "CREATING", "DELETING", "CREATE_FAILED", "DELETE_FAILED" # resp.file_system.windows_configuration.audit_log_configuration.file_access_audit_log_level #=> String, one of "DISABLED", "SUCCESS_ONLY", "FAILURE_ONLY", "SUCCESS_AND_FAILURE" # resp.file_system.windows_configuration.audit_log_configuration.file_share_access_audit_log_level #=> String, one of "DISABLED", "SUCCESS_ONLY", "FAILURE_ONLY", "SUCCESS_AND_FAILURE" # resp.file_system.windows_configuration.audit_log_configuration.audit_log_destination #=> String # resp.file_system.lustre_configuration.weekly_maintenance_start_time #=> String # resp.file_system.lustre_configuration.data_repository_configuration.lifecycle #=> String, one of "CREATING", "AVAILABLE", "MISCONFIGURED", "UPDATING", "DELETING" # resp.file_system.lustre_configuration.data_repository_configuration.import_path #=> String # resp.file_system.lustre_configuration.data_repository_configuration.export_path #=> String # resp.file_system.lustre_configuration.data_repository_configuration.imported_file_chunk_size #=> Integer # resp.file_system.lustre_configuration.data_repository_configuration.auto_import_policy #=> String, one of "NONE", "NEW", "NEW_CHANGED" # resp.file_system.lustre_configuration.data_repository_configuration.failure_details.message #=> String # resp.file_system.lustre_configuration.deployment_type #=> String, one of "SCRATCH_1", "SCRATCH_2", "PERSISTENT_1" # resp.file_system.lustre_configuration.per_unit_storage_throughput #=> Integer # resp.file_system.lustre_configuration.mount_name #=> String # resp.file_system.lustre_configuration.daily_automatic_backup_start_time #=> String # resp.file_system.lustre_configuration.automatic_backup_retention_days #=> Integer # resp.file_system.lustre_configuration.copy_tags_to_backups #=> Boolean # resp.file_system.lustre_configuration.drive_cache_type #=> String, one of "NONE", "READ" # resp.file_system.lustre_configuration.data_compression_type #=> String, one of "NONE", "LZ4" # resp.file_system.administrative_actions #=> Array # resp.file_system.administrative_actions[0].administrative_action_type #=> String, one of "FILE_SYSTEM_UPDATE", "STORAGE_OPTIMIZATION", "FILE_SYSTEM_ALIAS_ASSOCIATION", "FILE_SYSTEM_ALIAS_DISASSOCIATION" # resp.file_system.administrative_actions[0].progress_percent #=> Integer # resp.file_system.administrative_actions[0].request_time #=> Time # resp.file_system.administrative_actions[0].status #=> String, one of "FAILED", "IN_PROGRESS", "PENDING", "COMPLETED", "UPDATED_OPTIMIZING" # resp.file_system.administrative_actions[0].target_file_system_values #=> Types::FileSystem # resp.file_system.administrative_actions[0].failure_details.message #=> String # # @see http://docs.aws.amazon.com/goto/WebAPI/fsx-2018-03-01/CreateFileSystemFromBackup AWS API Documentation # # @overload create_file_system_from_backup(params = {}) # @param [Hash] params ({}) def create_file_system_from_backup(params = {}, options = {}) req = build_request(:create_file_system_from_backup, params) req.send_request(options) end # Deletes an Amazon FSx backup, deleting its contents. After deletion, # the backup no longer exists, and its data is gone. # # The `DeleteBackup` call returns instantly. The backup will not show up # in later `DescribeBackups` calls. # # The data in a deleted backup is also deleted and can't be recovered # by any means. # # @option params [required, String] :backup_id # The ID of the backup you want to delete. # # @option params [String] :client_request_token # A string of up to 64 ASCII characters that Amazon FSx uses to ensure # idempotent deletion. This is automatically filled on your behalf when # using the AWS CLI or SDK. # # **A suitable default value is auto-generated.** You should normally # not need to pass this option.** # # @return [Types::DeleteBackupResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods: # # * {Types::DeleteBackupResponse#backup_id #backup_id} => String # * {Types::DeleteBackupResponse#lifecycle #lifecycle} => String # # # @example Example: To delete a backup # # # This operation deletes an Amazon FSx file system backup. # # resp = client.delete_backup({ # backup_id: "backup-03e3c82e0183b7b6b", # }) # # resp.to_h outputs the following: # { # backup_id: "backup-03e3c82e0183b7b6b", # lifecycle: "DELETED", # } # # @example Request syntax with placeholder values # # resp = client.delete_backup({ # backup_id: "BackupId", # required # client_request_token: "ClientRequestToken", # }) # # @example Response structure # # resp.backup_id #=> String # resp.lifecycle #=> String, one of "AVAILABLE", "CREATING", "TRANSFERRING", "DELETED", "FAILED", "PENDING", "COPYING" # # @see http://docs.aws.amazon.com/goto/WebAPI/fsx-2018-03-01/DeleteBackup AWS API Documentation # # @overload delete_backup(params = {}) # @param [Hash] params ({}) def delete_backup(params = {}, options = {}) req = build_request(:delete_backup, params) req.send_request(options) end # Deletes a file system, deleting its contents. After deletion, the file # system no longer exists, and its data is gone. Any existing automatic # backups will also be deleted. # # By default, when you delete an Amazon FSx for Windows File Server file # system, a final backup is created upon deletion. This final backup is # not subject to the file system's retention policy, and must be # manually deleted. # # The `DeleteFileSystem` action returns while the file system has the # `DELETING` status. You can check the file system deletion status by # calling the DescribeFileSystems action, which returns a list of file # systems in your account. If you pass the file system ID for a deleted # file system, the DescribeFileSystems returns a `FileSystemNotFound` # error. # # Deleting an Amazon FSx for Lustre file system will fail with a 400 # BadRequest if a data repository task is in a `PENDING` or `EXECUTING` # state. # # # # The data in a deleted file system is also deleted and can't be # recovered by any means. # # @option params [required, String] :file_system_id # The ID of the file system you want to delete. # # @option params [String] :client_request_token # A string of up to 64 ASCII characters that Amazon FSx uses to ensure # idempotent deletion. This is automatically filled on your behalf when # using the AWS CLI or SDK. # # **A suitable default value is auto-generated.** You should normally # not need to pass this option.** # # @option params [Types::DeleteFileSystemWindowsConfiguration] :windows_configuration # The configuration object for the Microsoft Windows file system used in # the `DeleteFileSystem` operation. # # @option params [Types::DeleteFileSystemLustreConfiguration] :lustre_configuration # The configuration object for the Amazon FSx for Lustre file system # being deleted in the `DeleteFileSystem` operation. # # @return [Types::DeleteFileSystemResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods: # # * {Types::DeleteFileSystemResponse#file_system_id #file_system_id} => String # * {Types::DeleteFileSystemResponse#lifecycle #lifecycle} => String # * {Types::DeleteFileSystemResponse#windows_response #windows_response} => Types::DeleteFileSystemWindowsResponse # * {Types::DeleteFileSystemResponse#lustre_response #lustre_response} => Types::DeleteFileSystemLustreResponse # # # @example Example: To delete a file system # # # This operation deletes an Amazon FSx file system. # # resp = client.delete_file_system({ # file_system_id: "fs-0498eed5fe91001ec", # }) # # resp.to_h outputs the following: # { # file_system_id: "fs-0498eed5fe91001ec", # lifecycle: "DELETING", # } # # @example Request syntax with placeholder values # # resp = client.delete_file_system({ # file_system_id: "FileSystemId", # required # client_request_token: "ClientRequestToken", # windows_configuration: { # skip_final_backup: false, # final_backup_tags: [ # { # key: "TagKey", # required # value: "TagValue", # required # }, # ], # }, # lustre_configuration: { # skip_final_backup: false, # final_backup_tags: [ # { # key: "TagKey", # required # value: "TagValue", # required # }, # ], # }, # }) # # @example Response structure # # resp.file_system_id #=> String # resp.lifecycle #=> String, one of "AVAILABLE", "CREATING", "FAILED", "DELETING", "MISCONFIGURED", "UPDATING" # resp.windows_response.final_backup_id #=> String # resp.windows_response.final_backup_tags #=> Array # resp.windows_response.final_backup_tags[0].key #=> String # resp.windows_response.final_backup_tags[0].value #=> String # resp.lustre_response.final_backup_id #=> String # resp.lustre_response.final_backup_tags #=> Array # resp.lustre_response.final_backup_tags[0].key #=> String # resp.lustre_response.final_backup_tags[0].value #=> String # # @see http://docs.aws.amazon.com/goto/WebAPI/fsx-2018-03-01/DeleteFileSystem AWS API Documentation # # @overload delete_file_system(params = {}) # @param [Hash] params ({}) def delete_file_system(params = {}, options = {}) req = build_request(:delete_file_system, params) req.send_request(options) end # Returns the description of specific Amazon FSx backups, if a # `BackupIds` value is provided for that backup. Otherwise, it returns # all backups owned by your AWS account in the AWS Region of the # endpoint that you're calling. # # When retrieving all backups, you can optionally specify the # `MaxResults` parameter to limit the number of backups in a response. # If more backups remain, Amazon FSx returns a `NextToken` value in the # response. In this case, send a later request with the `NextToken` # request parameter set to the value of `NextToken` from the last # response. # # This action is used in an iterative process to retrieve a list of your # backups. `DescribeBackups` is called first without a `NextToken`value. # Then the action continues to be called with the `NextToken` parameter # set to the value of the last `NextToken` value until a response has no # `NextToken`. # # When using this action, keep the following in mind: # # * The implementation might return fewer than `MaxResults` file system # descriptions while still including a `NextToken` value. # # * The order of backups returned in the response of one # `DescribeBackups` call and the order of backups returned across the # responses of a multi-call iteration is unspecified. # # @option params [Array] :backup_ids # IDs of the backups you want to retrieve (String). This overrides any # filters. If any IDs are not found, BackupNotFound will be thrown. # # @option params [Array] :filters # Filters structure. Supported names are file-system-id and backup-type. # # @option params [Integer] :max_results # Maximum number of backups to return in the response (integer). This # parameter value must be greater than 0. The number of items that # Amazon FSx returns is the minimum of the `MaxResults` parameter # specified in the request and the service's internal maximum number of # items per page. # # @option params [String] :next_token # Opaque pagination token returned from a previous `DescribeBackups` # operation (String). If a token present, the action continues the list # from where the returning call left off. # # @return [Types::DescribeBackupsResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods: # # * {Types::DescribeBackupsResponse#backups #backups} => Array<Types::Backup> # * {Types::DescribeBackupsResponse#next_token #next_token} => String # # The returned {Seahorse::Client::Response response} is a pageable response and is Enumerable. For details on usage see {Aws::PageableResponse PageableResponse}. # # # @example Example: To describe Amazon FSx backups # # # This operation describes all of the Amazon FSx backups in an account. # # resp = client.describe_backups({ # }) # # resp.to_h outputs the following: # { # backups: [ # { # backup_id: "backup-03e3c82e0183b7b6b", # creation_time: Time.parse("1481841524.0"), # file_system: { # file_system_id: "fs-0498eed5fe91001ec", # owner_id: "012345678912", # storage_capacity: 300, # windows_configuration: { # active_directory_id: "d-1234abcd12", # automatic_backup_retention_days: 30, # daily_automatic_backup_start_time: "05:00", # weekly_maintenance_start_time: "1:05:00", # }, # }, # lifecycle: "AVAILABLE", # resource_arn: "arn:aws:fsx:us-east-1:012345678912:backup/backup-03e3c82e0183b7b6b", # tags: [ # { # key: "Name", # value: "MyBackup", # }, # ], # type: "USER_INITIATED", # }, # ], # } # # @example Request syntax with placeholder values # # resp = client.describe_backups({ # backup_ids: ["BackupId"], # filters: [ # { # name: "file-system-id", # accepts file-system-id, backup-type, file-system-type # values: ["FilterValue"], # }, # ], # max_results: 1, # next_token: "NextToken", # }) # # @example Response structure # # resp.backups #=> Array # resp.backups[0].backup_id #=> String # resp.backups[0].lifecycle #=> String, one of "AVAILABLE", "CREATING", "TRANSFERRING", "DELETED", "FAILED", "PENDING", "COPYING" # resp.backups[0].failure_details.message #=> String # resp.backups[0].type #=> String, one of "AUTOMATIC", "USER_INITIATED", "AWS_BACKUP" # resp.backups[0].progress_percent #=> Integer # resp.backups[0].creation_time #=> Time # resp.backups[0].kms_key_id #=> String # resp.backups[0].resource_arn #=> String # resp.backups[0].tags #=> Array # resp.backups[0].tags[0].key #=> String # resp.backups[0].tags[0].value #=> String # resp.backups[0].file_system.owner_id #=> String # resp.backups[0].file_system.creation_time #=> Time # resp.backups[0].file_system.file_system_id #=> String # resp.backups[0].file_system.file_system_type #=> String, one of "WINDOWS", "LUSTRE" # resp.backups[0].file_system.lifecycle #=> String, one of "AVAILABLE", "CREATING", "FAILED", "DELETING", "MISCONFIGURED", "UPDATING" # resp.backups[0].file_system.failure_details.message #=> String # resp.backups[0].file_system.storage_capacity #=> Integer # resp.backups[0].file_system.storage_type #=> String, one of "SSD", "HDD" # resp.backups[0].file_system.vpc_id #=> String # resp.backups[0].file_system.subnet_ids #=> Array # resp.backups[0].file_system.subnet_ids[0] #=> String # resp.backups[0].file_system.network_interface_ids #=> Array # resp.backups[0].file_system.network_interface_ids[0] #=> String # resp.backups[0].file_system.dns_name #=> String # resp.backups[0].file_system.kms_key_id #=> String # resp.backups[0].file_system.resource_arn #=> String # resp.backups[0].file_system.tags #=> Array # resp.backups[0].file_system.tags[0].key #=> String # resp.backups[0].file_system.tags[0].value #=> String # resp.backups[0].file_system.windows_configuration.active_directory_id #=> String # resp.backups[0].file_system.windows_configuration.self_managed_active_directory_configuration.domain_name #=> String # resp.backups[0].file_system.windows_configuration.self_managed_active_directory_configuration.organizational_unit_distinguished_name #=> String # resp.backups[0].file_system.windows_configuration.self_managed_active_directory_configuration.file_system_administrators_group #=> String # resp.backups[0].file_system.windows_configuration.self_managed_active_directory_configuration.user_name #=> String # resp.backups[0].file_system.windows_configuration.self_managed_active_directory_configuration.dns_ips #=> Array # resp.backups[0].file_system.windows_configuration.self_managed_active_directory_configuration.dns_ips[0] #=> String # resp.backups[0].file_system.windows_configuration.deployment_type #=> String, one of "MULTI_AZ_1", "SINGLE_AZ_1", "SINGLE_AZ_2" # resp.backups[0].file_system.windows_configuration.remote_administration_endpoint #=> String # resp.backups[0].file_system.windows_configuration.preferred_subnet_id #=> String # resp.backups[0].file_system.windows_configuration.preferred_file_server_ip #=> String # resp.backups[0].file_system.windows_configuration.throughput_capacity #=> Integer # resp.backups[0].file_system.windows_configuration.maintenance_operations_in_progress #=> Array # resp.backups[0].file_system.windows_configuration.maintenance_operations_in_progress[0] #=> String, one of "PATCHING", "BACKING_UP" # resp.backups[0].file_system.windows_configuration.weekly_maintenance_start_time #=> String # resp.backups[0].file_system.windows_configuration.daily_automatic_backup_start_time #=> String # resp.backups[0].file_system.windows_configuration.automatic_backup_retention_days #=> Integer # resp.backups[0].file_system.windows_configuration.copy_tags_to_backups #=> Boolean # resp.backups[0].file_system.windows_configuration.aliases #=> Array # resp.backups[0].file_system.windows_configuration.aliases[0].name #=> String # resp.backups[0].file_system.windows_configuration.aliases[0].lifecycle #=> String, one of "AVAILABLE", "CREATING", "DELETING", "CREATE_FAILED", "DELETE_FAILED" # resp.backups[0].file_system.windows_configuration.audit_log_configuration.file_access_audit_log_level #=> String, one of "DISABLED", "SUCCESS_ONLY", "FAILURE_ONLY", "SUCCESS_AND_FAILURE" # resp.backups[0].file_system.windows_configuration.audit_log_configuration.file_share_access_audit_log_level #=> String, one of "DISABLED", "SUCCESS_ONLY", "FAILURE_ONLY", "SUCCESS_AND_FAILURE" # resp.backups[0].file_system.windows_configuration.audit_log_configuration.audit_log_destination #=> String # resp.backups[0].file_system.lustre_configuration.weekly_maintenance_start_time #=> String # resp.backups[0].file_system.lustre_configuration.data_repository_configuration.lifecycle #=> String, one of "CREATING", "AVAILABLE", "MISCONFIGURED", "UPDATING", "DELETING" # resp.backups[0].file_system.lustre_configuration.data_repository_configuration.import_path #=> String # resp.backups[0].file_system.lustre_configuration.data_repository_configuration.export_path #=> String # resp.backups[0].file_system.lustre_configuration.data_repository_configuration.imported_file_chunk_size #=> Integer # resp.backups[0].file_system.lustre_configuration.data_repository_configuration.auto_import_policy #=> String, one of "NONE", "NEW", "NEW_CHANGED" # resp.backups[0].file_system.lustre_configuration.data_repository_configuration.failure_details.message #=> String # resp.backups[0].file_system.lustre_configuration.deployment_type #=> String, one of "SCRATCH_1", "SCRATCH_2", "PERSISTENT_1" # resp.backups[0].file_system.lustre_configuration.per_unit_storage_throughput #=> Integer # resp.backups[0].file_system.lustre_configuration.mount_name #=> String # resp.backups[0].file_system.lustre_configuration.daily_automatic_backup_start_time #=> String # resp.backups[0].file_system.lustre_configuration.automatic_backup_retention_days #=> Integer # resp.backups[0].file_system.lustre_configuration.copy_tags_to_backups #=> Boolean # resp.backups[0].file_system.lustre_configuration.drive_cache_type #=> String, one of "NONE", "READ" # resp.backups[0].file_system.lustre_configuration.data_compression_type #=> String, one of "NONE", "LZ4" # resp.backups[0].file_system.administrative_actions #=> Array # resp.backups[0].file_system.administrative_actions[0].administrative_action_type #=> String, one of "FILE_SYSTEM_UPDATE", "STORAGE_OPTIMIZATION", "FILE_SYSTEM_ALIAS_ASSOCIATION", "FILE_SYSTEM_ALIAS_DISASSOCIATION" # resp.backups[0].file_system.administrative_actions[0].progress_percent #=> Integer # resp.backups[0].file_system.administrative_actions[0].request_time #=> Time # resp.backups[0].file_system.administrative_actions[0].status #=> String, one of "FAILED", "IN_PROGRESS", "PENDING", "COMPLETED", "UPDATED_OPTIMIZING" # resp.backups[0].file_system.administrative_actions[0].target_file_system_values #=> Types::FileSystem # resp.backups[0].file_system.administrative_actions[0].failure_details.message #=> String # resp.backups[0].directory_information.domain_name #=> String # resp.backups[0].directory_information.active_directory_id #=> String # resp.backups[0].directory_information.resource_arn #=> String # resp.backups[0].owner_id #=> String # resp.backups[0].source_backup_id #=> String # resp.backups[0].source_backup_region #=> String # resp.next_token #=> String # # @see http://docs.aws.amazon.com/goto/WebAPI/fsx-2018-03-01/DescribeBackups AWS API Documentation # # @overload describe_backups(params = {}) # @param [Hash] params ({}) def describe_backups(params = {}, options = {}) req = build_request(:describe_backups, params) req.send_request(options) end # Returns the description of specific Amazon FSx for Lustre data # repository tasks, if one or more `TaskIds` values are provided in the # request, or if filters are used in the request. You can use filters to # narrow the response to include just tasks for specific file systems, # or tasks in a specific lifecycle state. Otherwise, it returns all data # repository tasks owned by your AWS account in the AWS Region of the # endpoint that you're calling. # # When retrieving all tasks, you can paginate the response by using the # optional `MaxResults` parameter to limit the number of tasks returned # in a response. If more tasks remain, Amazon FSx returns a `NextToken` # value in the response. In this case, send a later request with the # `NextToken` request parameter set to the value of `NextToken` from the # last response. # # @option params [Array] :task_ids # (Optional) IDs of the tasks whose descriptions you want to retrieve # (String). # # @option params [Array] :filters # (Optional) You can use filters to narrow the # `DescribeDataRepositoryTasks` response to include just tasks for # specific file systems, or tasks in a specific lifecycle state. # # @option params [Integer] :max_results # The maximum number of resources to return in the response. This value # must be an integer greater than zero. # # @option params [String] :next_token # (Optional) Opaque pagination token returned from a previous operation # (String). If present, this token indicates from what point you can # continue processing the request, where the previous `NextToken` value # left off. # # @return [Types::DescribeDataRepositoryTasksResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods: # # * {Types::DescribeDataRepositoryTasksResponse#data_repository_tasks #data_repository_tasks} => Array<Types::DataRepositoryTask> # * {Types::DescribeDataRepositoryTasksResponse#next_token #next_token} => String # # The returned {Seahorse::Client::Response response} is a pageable response and is Enumerable. For details on usage see {Aws::PageableResponse PageableResponse}. # # @example Request syntax with placeholder values # # resp = client.describe_data_repository_tasks({ # task_ids: ["TaskId"], # filters: [ # { # name: "file-system-id", # accepts file-system-id, task-lifecycle # values: ["DataRepositoryTaskFilterValue"], # }, # ], # max_results: 1, # next_token: "NextToken", # }) # # @example Response structure # # resp.data_repository_tasks #=> Array # resp.data_repository_tasks[0].task_id #=> String # resp.data_repository_tasks[0].lifecycle #=> String, one of "PENDING", "EXECUTING", "FAILED", "SUCCEEDED", "CANCELED", "CANCELING" # resp.data_repository_tasks[0].type #=> String, one of "EXPORT_TO_REPOSITORY" # resp.data_repository_tasks[0].creation_time #=> Time # resp.data_repository_tasks[0].start_time #=> Time # resp.data_repository_tasks[0].end_time #=> Time # resp.data_repository_tasks[0].resource_arn #=> String # resp.data_repository_tasks[0].tags #=> Array # resp.data_repository_tasks[0].tags[0].key #=> String # resp.data_repository_tasks[0].tags[0].value #=> String # resp.data_repository_tasks[0].file_system_id #=> String # resp.data_repository_tasks[0].paths #=> Array # resp.data_repository_tasks[0].paths[0] #=> String # resp.data_repository_tasks[0].failure_details.message #=> String # resp.data_repository_tasks[0].status.total_count #=> Integer # resp.data_repository_tasks[0].status.succeeded_count #=> Integer # resp.data_repository_tasks[0].status.failed_count #=> Integer # resp.data_repository_tasks[0].status.last_updated_time #=> Time # resp.data_repository_tasks[0].report.enabled #=> Boolean # resp.data_repository_tasks[0].report.path #=> String # resp.data_repository_tasks[0].report.format #=> String, one of "REPORT_CSV_20191124" # resp.data_repository_tasks[0].report.scope #=> String, one of "FAILED_FILES_ONLY" # resp.next_token #=> String # # @see http://docs.aws.amazon.com/goto/WebAPI/fsx-2018-03-01/DescribeDataRepositoryTasks AWS API Documentation # # @overload describe_data_repository_tasks(params = {}) # @param [Hash] params ({}) def describe_data_repository_tasks(params = {}, options = {}) req = build_request(:describe_data_repository_tasks, params) req.send_request(options) end # Returns the DNS aliases that are associated with the specified Amazon # FSx for Windows File Server file system. A history of all DNS aliases # that have been associated with and disassociated from the file system # is available in the list of AdministrativeAction provided in the # DescribeFileSystems operation response. # # @option params [String] :client_request_token # (Optional) An idempotency token for resource creation, in a string of # up to 64 ASCII characters. This token is automatically filled on your # behalf when you use the AWS Command Line Interface (AWS CLI) or an AWS # SDK. # # **A suitable default value is auto-generated.** You should normally # not need to pass this option.** # # @option params [required, String] :file_system_id # The ID of the file system to return the associated DNS aliases for # (String). # # @option params [Integer] :max_results # Maximum number of DNS aliases to return in the response (integer). # This parameter value must be greater than 0. The number of items that # Amazon FSx returns is the minimum of the `MaxResults` parameter # specified in the request and the service's internal maximum number of # items per page. # # @option params [String] :next_token # Opaque pagination token returned from a previous # `DescribeFileSystemAliases` operation (String). If a token is included # in the request, the action continues the list from where the previous # returning call left off. # # @return [Types::DescribeFileSystemAliasesResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods: # # * {Types::DescribeFileSystemAliasesResponse#aliases #aliases} => Array<Types::Alias> # * {Types::DescribeFileSystemAliasesResponse#next_token #next_token} => String # # The returned {Seahorse::Client::Response response} is a pageable response and is Enumerable. For details on usage see {Aws::PageableResponse PageableResponse}. # # @example Request syntax with placeholder values # # resp = client.describe_file_system_aliases({ # client_request_token: "ClientRequestToken", # file_system_id: "FileSystemId", # required # max_results: 1, # next_token: "NextToken", # }) # # @example Response structure # # resp.aliases #=> Array # resp.aliases[0].name #=> String # resp.aliases[0].lifecycle #=> String, one of "AVAILABLE", "CREATING", "DELETING", "CREATE_FAILED", "DELETE_FAILED" # resp.next_token #=> String # # @see http://docs.aws.amazon.com/goto/WebAPI/fsx-2018-03-01/DescribeFileSystemAliases AWS API Documentation # # @overload describe_file_system_aliases(params = {}) # @param [Hash] params ({}) def describe_file_system_aliases(params = {}, options = {}) req = build_request(:describe_file_system_aliases, params) req.send_request(options) end # Returns the description of specific Amazon FSx file systems, if a # `FileSystemIds` value is provided for that file system. Otherwise, it # returns descriptions of all file systems owned by your AWS account in # the AWS Region of the endpoint that you're calling. # # When retrieving all file system descriptions, you can optionally # specify the `MaxResults` parameter to limit the number of descriptions # in a response. If more file system descriptions remain, Amazon FSx # returns a `NextToken` value in the response. In this case, send a # later request with the `NextToken` request parameter set to the value # of `NextToken` from the last response. # # This action is used in an iterative process to retrieve a list of your # file system descriptions. `DescribeFileSystems` is called first # without a `NextToken`value. Then the action continues to be called # with the `NextToken` parameter set to the value of the last # `NextToken` value until a response has no `NextToken`. # # When using this action, keep the following in mind: # # * The implementation might return fewer than `MaxResults` file system # descriptions while still including a `NextToken` value. # # * The order of file systems returned in the response of one # `DescribeFileSystems` call and the order of file systems returned # across the responses of a multicall iteration is unspecified. # # @option params [Array] :file_system_ids # IDs of the file systems whose descriptions you want to retrieve # (String). # # @option params [Integer] :max_results # Maximum number of file systems to return in the response (integer). # This parameter value must be greater than 0. The number of items that # Amazon FSx returns is the minimum of the `MaxResults` parameter # specified in the request and the service's internal maximum number of # items per page. # # @option params [String] :next_token # Opaque pagination token returned from a previous `DescribeFileSystems` # operation (String). If a token present, the action continues the list # from where the returning call left off. # # @return [Types::DescribeFileSystemsResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods: # # * {Types::DescribeFileSystemsResponse#file_systems #file_systems} => Array<Types::FileSystem> # * {Types::DescribeFileSystemsResponse#next_token #next_token} => String # # The returned {Seahorse::Client::Response response} is a pageable response and is Enumerable. For details on usage see {Aws::PageableResponse PageableResponse}. # # # @example Example: To describe an Amazon FSx file system # # # This operation describes all of the Amazon FSx file systems in an account. # # resp = client.describe_file_systems({ # }) # # resp.to_h outputs the following: # { # file_systems: [ # { # creation_time: Time.parse("1481841524.0"), # dns_name: "fs-0498eed5fe91001ec.fsx.com", # file_system_id: "fs-0498eed5fe91001ec", # kms_key_id: "arn:aws:kms:us-east-1:012345678912:key/0ff3ea8d-130e-4133-877f-93908b6fdbd6", # lifecycle: "AVAILABLE", # network_interface_ids: [ # "eni-abcd1234", # ], # owner_id: "012345678912", # resource_arn: "arn:aws:fsx:us-east-1:012345678912:file-system/fs-0498eed5fe91001ec", # storage_capacity: 300, # subnet_ids: [ # "subnet-1234abcd", # ], # tags: [ # { # key: "Name", # value: "MyFileSystem", # }, # ], # vpc_id: "vpc-ab1234cd", # windows_configuration: { # active_directory_id: "d-1234abcd12", # automatic_backup_retention_days: 30, # daily_automatic_backup_start_time: "05:00", # throughput_capacity: 8, # weekly_maintenance_start_time: "1:05:00", # }, # }, # ], # } # # @example Request syntax with placeholder values # # resp = client.describe_file_systems({ # file_system_ids: ["FileSystemId"], # max_results: 1, # next_token: "NextToken", # }) # # @example Response structure # # resp.file_systems #=> Array # resp.file_systems[0].owner_id #=> String # resp.file_systems[0].creation_time #=> Time # resp.file_systems[0].file_system_id #=> String # resp.file_systems[0].file_system_type #=> String, one of "WINDOWS", "LUSTRE" # resp.file_systems[0].lifecycle #=> String, one of "AVAILABLE", "CREATING", "FAILED", "DELETING", "MISCONFIGURED", "UPDATING" # resp.file_systems[0].failure_details.message #=> String # resp.file_systems[0].storage_capacity #=> Integer # resp.file_systems[0].storage_type #=> String, one of "SSD", "HDD" # resp.file_systems[0].vpc_id #=> String # resp.file_systems[0].subnet_ids #=> Array # resp.file_systems[0].subnet_ids[0] #=> String # resp.file_systems[0].network_interface_ids #=> Array # resp.file_systems[0].network_interface_ids[0] #=> String # resp.file_systems[0].dns_name #=> String # resp.file_systems[0].kms_key_id #=> String # resp.file_systems[0].resource_arn #=> String # resp.file_systems[0].tags #=> Array # resp.file_systems[0].tags[0].key #=> String # resp.file_systems[0].tags[0].value #=> String # resp.file_systems[0].windows_configuration.active_directory_id #=> String # resp.file_systems[0].windows_configuration.self_managed_active_directory_configuration.domain_name #=> String # resp.file_systems[0].windows_configuration.self_managed_active_directory_configuration.organizational_unit_distinguished_name #=> String # resp.file_systems[0].windows_configuration.self_managed_active_directory_configuration.file_system_administrators_group #=> String # resp.file_systems[0].windows_configuration.self_managed_active_directory_configuration.user_name #=> String # resp.file_systems[0].windows_configuration.self_managed_active_directory_configuration.dns_ips #=> Array # resp.file_systems[0].windows_configuration.self_managed_active_directory_configuration.dns_ips[0] #=> String # resp.file_systems[0].windows_configuration.deployment_type #=> String, one of "MULTI_AZ_1", "SINGLE_AZ_1", "SINGLE_AZ_2" # resp.file_systems[0].windows_configuration.remote_administration_endpoint #=> String # resp.file_systems[0].windows_configuration.preferred_subnet_id #=> String # resp.file_systems[0].windows_configuration.preferred_file_server_ip #=> String # resp.file_systems[0].windows_configuration.throughput_capacity #=> Integer # resp.file_systems[0].windows_configuration.maintenance_operations_in_progress #=> Array # resp.file_systems[0].windows_configuration.maintenance_operations_in_progress[0] #=> String, one of "PATCHING", "BACKING_UP" # resp.file_systems[0].windows_configuration.weekly_maintenance_start_time #=> String # resp.file_systems[0].windows_configuration.daily_automatic_backup_start_time #=> String # resp.file_systems[0].windows_configuration.automatic_backup_retention_days #=> Integer # resp.file_systems[0].windows_configuration.copy_tags_to_backups #=> Boolean # resp.file_systems[0].windows_configuration.aliases #=> Array # resp.file_systems[0].windows_configuration.aliases[0].name #=> String # resp.file_systems[0].windows_configuration.aliases[0].lifecycle #=> String, one of "AVAILABLE", "CREATING", "DELETING", "CREATE_FAILED", "DELETE_FAILED" # resp.file_systems[0].windows_configuration.audit_log_configuration.file_access_audit_log_level #=> String, one of "DISABLED", "SUCCESS_ONLY", "FAILURE_ONLY", "SUCCESS_AND_FAILURE" # resp.file_systems[0].windows_configuration.audit_log_configuration.file_share_access_audit_log_level #=> String, one of "DISABLED", "SUCCESS_ONLY", "FAILURE_ONLY", "SUCCESS_AND_FAILURE" # resp.file_systems[0].windows_configuration.audit_log_configuration.audit_log_destination #=> String # resp.file_systems[0].lustre_configuration.weekly_maintenance_start_time #=> String # resp.file_systems[0].lustre_configuration.data_repository_configuration.lifecycle #=> String, one of "CREATING", "AVAILABLE", "MISCONFIGURED", "UPDATING", "DELETING" # resp.file_systems[0].lustre_configuration.data_repository_configuration.import_path #=> String # resp.file_systems[0].lustre_configuration.data_repository_configuration.export_path #=> String # resp.file_systems[0].lustre_configuration.data_repository_configuration.imported_file_chunk_size #=> Integer # resp.file_systems[0].lustre_configuration.data_repository_configuration.auto_import_policy #=> String, one of "NONE", "NEW", "NEW_CHANGED" # resp.file_systems[0].lustre_configuration.data_repository_configuration.failure_details.message #=> String # resp.file_systems[0].lustre_configuration.deployment_type #=> String, one of "SCRATCH_1", "SCRATCH_2", "PERSISTENT_1" # resp.file_systems[0].lustre_configuration.per_unit_storage_throughput #=> Integer # resp.file_systems[0].lustre_configuration.mount_name #=> String # resp.file_systems[0].lustre_configuration.daily_automatic_backup_start_time #=> String # resp.file_systems[0].lustre_configuration.automatic_backup_retention_days #=> Integer # resp.file_systems[0].lustre_configuration.copy_tags_to_backups #=> Boolean # resp.file_systems[0].lustre_configuration.drive_cache_type #=> String, one of "NONE", "READ" # resp.file_systems[0].lustre_configuration.data_compression_type #=> String, one of "NONE", "LZ4" # resp.file_systems[0].administrative_actions #=> Array # resp.file_systems[0].administrative_actions[0].administrative_action_type #=> String, one of "FILE_SYSTEM_UPDATE", "STORAGE_OPTIMIZATION", "FILE_SYSTEM_ALIAS_ASSOCIATION", "FILE_SYSTEM_ALIAS_DISASSOCIATION" # resp.file_systems[0].administrative_actions[0].progress_percent #=> Integer # resp.file_systems[0].administrative_actions[0].request_time #=> Time # resp.file_systems[0].administrative_actions[0].status #=> String, one of "FAILED", "IN_PROGRESS", "PENDING", "COMPLETED", "UPDATED_OPTIMIZING" # resp.file_systems[0].administrative_actions[0].target_file_system_values #=> Types::FileSystem # resp.file_systems[0].administrative_actions[0].failure_details.message #=> String # resp.next_token #=> String # # @see http://docs.aws.amazon.com/goto/WebAPI/fsx-2018-03-01/DescribeFileSystems AWS API Documentation # # @overload describe_file_systems(params = {}) # @param [Hash] params ({}) def describe_file_systems(params = {}, options = {}) req = build_request(:describe_file_systems, params) req.send_request(options) end # Use this action to disassociate, or remove, one or more Domain Name # Service (DNS) aliases from an Amazon FSx for Windows File Server file # system. If you attempt to disassociate a DNS alias that is not # associated with the file system, Amazon FSx responds with a 400 Bad # Request. For more information, see [Working with DNS Aliases][1]. # # The system generated response showing the DNS aliases that Amazon FSx # is attempting to disassociate from the file system. Use the API # operation to monitor the status of the aliases Amazon FSx is # disassociating with the file system. # # # # [1]: https://docs.aws.amazon.com/fsx/latest/WindowsGuide/managing-dns-aliases.html # # @option params [String] :client_request_token # (Optional) An idempotency token for resource creation, in a string of # up to 64 ASCII characters. This token is automatically filled on your # behalf when you use the AWS Command Line Interface (AWS CLI) or an AWS # SDK. # # **A suitable default value is auto-generated.** You should normally # not need to pass this option.** # # @option params [required, String] :file_system_id # Specifies the file system from which to disassociate the DNS aliases. # # @option params [required, Array] :aliases # An array of one or more DNS alias names to disassociate, or remove, # from the file system. # # @return [Types::DisassociateFileSystemAliasesResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods: # # * {Types::DisassociateFileSystemAliasesResponse#aliases #aliases} => Array<Types::Alias> # # @example Request syntax with placeholder values # # resp = client.disassociate_file_system_aliases({ # client_request_token: "ClientRequestToken", # file_system_id: "FileSystemId", # required # aliases: ["AlternateDNSName"], # required # }) # # @example Response structure # # resp.aliases #=> Array # resp.aliases[0].name #=> String # resp.aliases[0].lifecycle #=> String, one of "AVAILABLE", "CREATING", "DELETING", "CREATE_FAILED", "DELETE_FAILED" # # @see http://docs.aws.amazon.com/goto/WebAPI/fsx-2018-03-01/DisassociateFileSystemAliases AWS API Documentation # # @overload disassociate_file_system_aliases(params = {}) # @param [Hash] params ({}) def disassociate_file_system_aliases(params = {}, options = {}) req = build_request(:disassociate_file_system_aliases, params) req.send_request(options) end # Lists tags for an Amazon FSx file systems and backups in the case of # Amazon FSx for Windows File Server. # # When retrieving all tags, you can optionally specify the `MaxResults` # parameter to limit the number of tags in a response. If more tags # remain, Amazon FSx returns a `NextToken` value in the response. In # this case, send a later request with the `NextToken` request parameter # set to the value of `NextToken` from the last response. # # This action is used in an iterative process to retrieve a list of your # tags. `ListTagsForResource` is called first without a # `NextToken`value. Then the action continues to be called with the # `NextToken` parameter set to the value of the last `NextToken` value # until a response has no `NextToken`. # # When using this action, keep the following in mind: # # * The implementation might return fewer than `MaxResults` file system # descriptions while still including a `NextToken` value. # # * The order of tags returned in the response of one # `ListTagsForResource` call and the order of tags returned across the # responses of a multi-call iteration is unspecified. # # @option params [required, String] :resource_arn # The ARN of the Amazon FSx resource that will have its tags listed. # # @option params [Integer] :max_results # Maximum number of tags to return in the response (integer). This # parameter value must be greater than 0. The number of items that # Amazon FSx returns is the minimum of the `MaxResults` parameter # specified in the request and the service's internal maximum number of # items per page. # # @option params [String] :next_token # Opaque pagination token returned from a previous `ListTagsForResource` # operation (String). If a token present, the action continues the list # from where the returning call left off. # # @return [Types::ListTagsForResourceResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods: # # * {Types::ListTagsForResourceResponse#tags #tags} => Array<Types::Tag> # * {Types::ListTagsForResourceResponse#next_token #next_token} => String # # # @example Example: To list tags for a resource # # # This operation lists tags for an Amazon FSx resource. # # resp = client.list_tags_for_resource({ # resource_arn: "arn:aws:fsx:us-east-1:012345678912:file-system/fs-0498eed5fe91001ec", # }) # # resp.to_h outputs the following: # { # tags: [ # { # key: "Name", # value: "MyFileSystem", # }, # ], # } # # @example Request syntax with placeholder values # # resp = client.list_tags_for_resource({ # resource_arn: "ResourceARN", # required # max_results: 1, # next_token: "NextToken", # }) # # @example Response structure # # resp.tags #=> Array # resp.tags[0].key #=> String # resp.tags[0].value #=> String # resp.next_token #=> String # # @see http://docs.aws.amazon.com/goto/WebAPI/fsx-2018-03-01/ListTagsForResource AWS API Documentation # # @overload list_tags_for_resource(params = {}) # @param [Hash] params ({}) def list_tags_for_resource(params = {}, options = {}) req = build_request(:list_tags_for_resource, params) req.send_request(options) end # Tags an Amazon FSx resource. # # @option params [required, String] :resource_arn # The Amazon Resource Name (ARN) of the Amazon FSx resource that you # want to tag. # # @option params [required, Array] :tags # A list of tags for the resource. If a tag with a given key already # exists, the value is replaced by the one specified in this parameter. # # @return [Struct] Returns an empty {Seahorse::Client::Response response}. # # # @example Example: To tag a resource # # # This operation tags an Amazon FSx resource. # # resp = client.tag_resource({ # resource_arn: "arn:aws:fsx:us-east-1:012345678912:file-system/fs-0498eed5fe91001ec", # tags: [ # { # key: "Name", # value: "MyFileSystem", # }, # ], # }) # # @example Request syntax with placeholder values # # resp = client.tag_resource({ # resource_arn: "ResourceARN", # required # tags: [ # required # { # key: "TagKey", # required # value: "TagValue", # required # }, # ], # }) # # @see http://docs.aws.amazon.com/goto/WebAPI/fsx-2018-03-01/TagResource AWS API Documentation # # @overload tag_resource(params = {}) # @param [Hash] params ({}) def tag_resource(params = {}, options = {}) req = build_request(:tag_resource, params) req.send_request(options) end # This action removes a tag from an Amazon FSx resource. # # @option params [required, String] :resource_arn # The ARN of the Amazon FSx resource to untag. # # @option params [required, Array] :tag_keys # A list of keys of tags on the resource to untag. In case the tag key # doesn't exist, the call will still succeed to be idempotent. # # @return [Struct] Returns an empty {Seahorse::Client::Response response}. # # # @example Example: To untag a resource # # # This operation untags an Amazon FSx resource. # # resp = client.untag_resource({ # resource_arn: "arn:aws:fsx:us-east-1:012345678912:file-system/fs-0498eed5fe91001ec", # tag_keys: [ # "Name", # ], # }) # # @example Request syntax with placeholder values # # resp = client.untag_resource({ # resource_arn: "ResourceARN", # required # tag_keys: ["TagKey"], # required # }) # # @see http://docs.aws.amazon.com/goto/WebAPI/fsx-2018-03-01/UntagResource AWS API Documentation # # @overload untag_resource(params = {}) # @param [Hash] params ({}) def untag_resource(params = {}, options = {}) req = build_request(:untag_resource, params) req.send_request(options) end # Use this operation to update the configuration of an existing Amazon # FSx file system. You can update multiple properties in a single # request. # # For Amazon FSx for Windows File Server file systems, you can update # the following properties: # # * AuditLogConfiguration # # * AutomaticBackupRetentionDays # # * DailyAutomaticBackupStartTime # # * SelfManagedActiveDirectoryConfiguration # # * StorageCapacity # # * ThroughputCapacity # # * WeeklyMaintenanceStartTime # # For Amazon FSx for Lustre file systems, you can update the following # properties: # # * AutoImportPolicy # # * AutomaticBackupRetentionDays # # * DailyAutomaticBackupStartTime # # * DataCompressionType # # * StorageCapacity # # * WeeklyMaintenanceStartTime # # @option params [required, String] :file_system_id # Identifies the file system that you are updating. # # @option params [String] :client_request_token # A string of up to 64 ASCII characters that Amazon FSx uses to ensure # idempotent updates. This string is automatically filled on your behalf # when you use the AWS Command Line Interface (AWS CLI) or an AWS SDK. # # **A suitable default value is auto-generated.** You should normally # not need to pass this option.** # # @option params [Integer] :storage_capacity # Use this parameter to increase the storage capacity of an Amazon FSx # file system. Specifies the storage capacity target value, GiB, to # increase the storage capacity for the file system that you're # updating. You cannot make a storage capacity increase request if there # is an existing storage capacity increase request in progress. # # For Windows file systems, the storage capacity target value must be at # least 10 percent (%) greater than the current storage capacity value. # In order to increase storage capacity, the file system must have at # least 16 MB/s of throughput capacity. # # For Lustre file systems, the storage capacity target value can be the # following: # # * For `SCRATCH_2` and `PERSISTENT_1 SSD` deployment types, valid # values are in multiples of 2400 GiB. The value must be greater than # the current storage capacity. # # * For `PERSISTENT HDD` file systems, valid values are multiples of # 6000 GiB for 12 MB/s/TiB file systems and multiples of 1800 GiB for # 40 MB/s/TiB file systems. The values must be greater than the # current storage capacity. # # * For `SCRATCH_1` file systems, you cannot increase the storage # capacity. # # For more information, see [Managing storage capacity][1] in the # *Amazon FSx for Windows File Server User Guide* and [Managing storage # and throughput capacity][2] in the *Amazon FSx for Lustre User Guide*. # # # # [1]: https://docs.aws.amazon.com/fsx/latest/WindowsGuide/managing-storage-capacity.html # [2]: https://docs.aws.amazon.com/fsx/latest/LustreGuide/managing-storage-capacity.html # # @option params [Types::UpdateFileSystemWindowsConfiguration] :windows_configuration # The configuration updates for an Amazon FSx for Windows File Server # file system. # # @option params [Types::UpdateFileSystemLustreConfiguration] :lustre_configuration # The configuration object for Amazon FSx for Lustre file systems used # in the `UpdateFileSystem` operation. # # @return [Types::UpdateFileSystemResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods: # # * {Types::UpdateFileSystemResponse#file_system #file_system} => Types::FileSystem # # # @example Example: To update an existing file system # # # This operation updates an existing file system. # # resp = client.update_file_system({ # file_system_id: "fs-0498eed5fe91001ec", # windows_configuration: { # automatic_backup_retention_days: 10, # daily_automatic_backup_start_time: "06:00", # weekly_maintenance_start_time: "3:06:00", # }, # }) # # resp.to_h outputs the following: # { # file_system: { # creation_time: Time.parse("1481841524.0"), # dns_name: "fs-0498eed5fe91001ec.fsx.com", # file_system_id: "fs-0498eed5fe91001ec", # kms_key_id: "arn:aws:kms:us-east-1:012345678912:key/0ff3ea8d-130e-4133-877f-93908b6fdbd6", # lifecycle: "AVAILABLE", # owner_id: "012345678912", # resource_arn: "arn:aws:fsx:us-east-1:012345678912:file-system/fs-0498eed5fe91001ec", # storage_capacity: 300, # subnet_ids: [ # "subnet-1234abcd", # ], # tags: [ # { # key: "Name", # value: "MyFileSystem", # }, # ], # vpc_id: "vpc-ab1234cd", # windows_configuration: { # automatic_backup_retention_days: 10, # daily_automatic_backup_start_time: "06:00", # throughput_capacity: 8, # weekly_maintenance_start_time: "3:06:00", # }, # }, # } # # @example Request syntax with placeholder values # # resp = client.update_file_system({ # file_system_id: "FileSystemId", # required # client_request_token: "ClientRequestToken", # storage_capacity: 1, # windows_configuration: { # weekly_maintenance_start_time: "WeeklyTime", # daily_automatic_backup_start_time: "DailyTime", # automatic_backup_retention_days: 1, # throughput_capacity: 1, # self_managed_active_directory_configuration: { # user_name: "DirectoryUserName", # password: "DirectoryPassword", # dns_ips: ["IpAddress"], # }, # audit_log_configuration: { # file_access_audit_log_level: "DISABLED", # required, accepts DISABLED, SUCCESS_ONLY, FAILURE_ONLY, SUCCESS_AND_FAILURE # file_share_access_audit_log_level: "DISABLED", # required, accepts DISABLED, SUCCESS_ONLY, FAILURE_ONLY, SUCCESS_AND_FAILURE # audit_log_destination: "GeneralARN", # }, # }, # lustre_configuration: { # weekly_maintenance_start_time: "WeeklyTime", # daily_automatic_backup_start_time: "DailyTime", # automatic_backup_retention_days: 1, # auto_import_policy: "NONE", # accepts NONE, NEW, NEW_CHANGED # data_compression_type: "NONE", # accepts NONE, LZ4 # }, # }) # # @example Response structure # # resp.file_system.owner_id #=> String # resp.file_system.creation_time #=> Time # resp.file_system.file_system_id #=> String # resp.file_system.file_system_type #=> String, one of "WINDOWS", "LUSTRE" # resp.file_system.lifecycle #=> String, one of "AVAILABLE", "CREATING", "FAILED", "DELETING", "MISCONFIGURED", "UPDATING" # resp.file_system.failure_details.message #=> String # resp.file_system.storage_capacity #=> Integer # resp.file_system.storage_type #=> String, one of "SSD", "HDD" # resp.file_system.vpc_id #=> String # resp.file_system.subnet_ids #=> Array # resp.file_system.subnet_ids[0] #=> String # resp.file_system.network_interface_ids #=> Array # resp.file_system.network_interface_ids[0] #=> String # resp.file_system.dns_name #=> String # resp.file_system.kms_key_id #=> String # resp.file_system.resource_arn #=> String # resp.file_system.tags #=> Array # resp.file_system.tags[0].key #=> String # resp.file_system.tags[0].value #=> String # resp.file_system.windows_configuration.active_directory_id #=> String # resp.file_system.windows_configuration.self_managed_active_directory_configuration.domain_name #=> String # resp.file_system.windows_configuration.self_managed_active_directory_configuration.organizational_unit_distinguished_name #=> String # resp.file_system.windows_configuration.self_managed_active_directory_configuration.file_system_administrators_group #=> String # resp.file_system.windows_configuration.self_managed_active_directory_configuration.user_name #=> String # resp.file_system.windows_configuration.self_managed_active_directory_configuration.dns_ips #=> Array # resp.file_system.windows_configuration.self_managed_active_directory_configuration.dns_ips[0] #=> String # resp.file_system.windows_configuration.deployment_type #=> String, one of "MULTI_AZ_1", "SINGLE_AZ_1", "SINGLE_AZ_2" # resp.file_system.windows_configuration.remote_administration_endpoint #=> String # resp.file_system.windows_configuration.preferred_subnet_id #=> String # resp.file_system.windows_configuration.preferred_file_server_ip #=> String # resp.file_system.windows_configuration.throughput_capacity #=> Integer # resp.file_system.windows_configuration.maintenance_operations_in_progress #=> Array # resp.file_system.windows_configuration.maintenance_operations_in_progress[0] #=> String, one of "PATCHING", "BACKING_UP" # resp.file_system.windows_configuration.weekly_maintenance_start_time #=> String # resp.file_system.windows_configuration.daily_automatic_backup_start_time #=> String # resp.file_system.windows_configuration.automatic_backup_retention_days #=> Integer # resp.file_system.windows_configuration.copy_tags_to_backups #=> Boolean # resp.file_system.windows_configuration.aliases #=> Array # resp.file_system.windows_configuration.aliases[0].name #=> String # resp.file_system.windows_configuration.aliases[0].lifecycle #=> String, one of "AVAILABLE", "CREATING", "DELETING", "CREATE_FAILED", "DELETE_FAILED" # resp.file_system.windows_configuration.audit_log_configuration.file_access_audit_log_level #=> String, one of "DISABLED", "SUCCESS_ONLY", "FAILURE_ONLY", "SUCCESS_AND_FAILURE" # resp.file_system.windows_configuration.audit_log_configuration.file_share_access_audit_log_level #=> String, one of "DISABLED", "SUCCESS_ONLY", "FAILURE_ONLY", "SUCCESS_AND_FAILURE" # resp.file_system.windows_configuration.audit_log_configuration.audit_log_destination #=> String # resp.file_system.lustre_configuration.weekly_maintenance_start_time #=> String # resp.file_system.lustre_configuration.data_repository_configuration.lifecycle #=> String, one of "CREATING", "AVAILABLE", "MISCONFIGURED", "UPDATING", "DELETING" # resp.file_system.lustre_configuration.data_repository_configuration.import_path #=> String # resp.file_system.lustre_configuration.data_repository_configuration.export_path #=> String # resp.file_system.lustre_configuration.data_repository_configuration.imported_file_chunk_size #=> Integer # resp.file_system.lustre_configuration.data_repository_configuration.auto_import_policy #=> String, one of "NONE", "NEW", "NEW_CHANGED" # resp.file_system.lustre_configuration.data_repository_configuration.failure_details.message #=> String # resp.file_system.lustre_configuration.deployment_type #=> String, one of "SCRATCH_1", "SCRATCH_2", "PERSISTENT_1" # resp.file_system.lustre_configuration.per_unit_storage_throughput #=> Integer # resp.file_system.lustre_configuration.mount_name #=> String # resp.file_system.lustre_configuration.daily_automatic_backup_start_time #=> String # resp.file_system.lustre_configuration.automatic_backup_retention_days #=> Integer # resp.file_system.lustre_configuration.copy_tags_to_backups #=> Boolean # resp.file_system.lustre_configuration.drive_cache_type #=> String, one of "NONE", "READ" # resp.file_system.lustre_configuration.data_compression_type #=> String, one of "NONE", "LZ4" # resp.file_system.administrative_actions #=> Array # resp.file_system.administrative_actions[0].administrative_action_type #=> String, one of "FILE_SYSTEM_UPDATE", "STORAGE_OPTIMIZATION", "FILE_SYSTEM_ALIAS_ASSOCIATION", "FILE_SYSTEM_ALIAS_DISASSOCIATION" # resp.file_system.administrative_actions[0].progress_percent #=> Integer # resp.file_system.administrative_actions[0].request_time #=> Time # resp.file_system.administrative_actions[0].status #=> String, one of "FAILED", "IN_PROGRESS", "PENDING", "COMPLETED", "UPDATED_OPTIMIZING" # resp.file_system.administrative_actions[0].target_file_system_values #=> Types::FileSystem # resp.file_system.administrative_actions[0].failure_details.message #=> String # # @see http://docs.aws.amazon.com/goto/WebAPI/fsx-2018-03-01/UpdateFileSystem AWS API Documentation # # @overload update_file_system(params = {}) # @param [Hash] params ({}) def update_file_system(params = {}, options = {}) req = build_request(:update_file_system, params) req.send_request(options) end # @!endgroup # @param params ({}) # @api private def build_request(operation_name, params = {}) handlers = @handlers.for(operation_name) context = Seahorse::Client::RequestContext.new( operation_name: operation_name, operation: config.api.operation(operation_name), client: self, params: params, config: config) context[:gem_name] = 'aws-sdk-fsx' context[:gem_version] = '1.39.0' Seahorse::Client::Request.new(handlers, context) end # @api private # @deprecated def waiter_names [] end class << self # @api private attr_reader :identifier # @api private def errors_module Errors end end end end