# frozen_string_literal: true # WARNING ABOUT GENERATED CODE # # This file is generated. See the contributing guide for more information: # https://github.com/aws/aws-sdk-ruby/blob/master/CONTRIBUTING.md # # WARNING ABOUT GENERATED CODE require 'seahorse/client/plugins/content_length.rb' require 'aws-sdk-core/plugins/credentials_configuration.rb' require 'aws-sdk-core/plugins/logging.rb' require 'aws-sdk-core/plugins/param_converter.rb' require 'aws-sdk-core/plugins/param_validator.rb' require 'aws-sdk-core/plugins/user_agent.rb' require 'aws-sdk-core/plugins/helpful_socket_errors.rb' require 'aws-sdk-core/plugins/retry_errors.rb' require 'aws-sdk-core/plugins/global_configuration.rb' require 'aws-sdk-core/plugins/regional_endpoint.rb' require 'aws-sdk-core/plugins/endpoint_discovery.rb' require 'aws-sdk-core/plugins/endpoint_pattern.rb' require 'aws-sdk-core/plugins/response_paging.rb' require 'aws-sdk-core/plugins/stub_responses.rb' require 'aws-sdk-core/plugins/idempotency_token.rb' require 'aws-sdk-core/plugins/jsonvalue_converter.rb' require 'aws-sdk-core/plugins/client_metrics_plugin.rb' require 'aws-sdk-core/plugins/client_metrics_send_plugin.rb' require 'aws-sdk-core/plugins/transfer_encoding.rb' require 'aws-sdk-core/plugins/http_checksum.rb' require 'aws-sdk-core/plugins/signature_v4.rb' require 'aws-sdk-core/plugins/protocols/json_rpc.rb' Aws::Plugins::GlobalConfiguration.add_identifier(:fsx) module Aws::FSx # An API client for FSx. To construct a client, you need to configure a `:region` and `:credentials`. # # client = Aws::FSx::Client.new( # region: region_name, # credentials: credentials, # # ... # ) # # For details on configuring region and credentials see # the [developer guide](/sdk-for-ruby/v3/developer-guide/setup-config.html). # # See {#initialize} for a full list of supported configuration options. class Client < Seahorse::Client::Base include Aws::ClientStubs @identifier = :fsx set_api(ClientApi::API) add_plugin(Seahorse::Client::Plugins::ContentLength) add_plugin(Aws::Plugins::CredentialsConfiguration) add_plugin(Aws::Plugins::Logging) add_plugin(Aws::Plugins::ParamConverter) add_plugin(Aws::Plugins::ParamValidator) add_plugin(Aws::Plugins::UserAgent) add_plugin(Aws::Plugins::HelpfulSocketErrors) add_plugin(Aws::Plugins::RetryErrors) add_plugin(Aws::Plugins::GlobalConfiguration) add_plugin(Aws::Plugins::RegionalEndpoint) add_plugin(Aws::Plugins::EndpointDiscovery) add_plugin(Aws::Plugins::EndpointPattern) add_plugin(Aws::Plugins::ResponsePaging) add_plugin(Aws::Plugins::StubResponses) add_plugin(Aws::Plugins::IdempotencyToken) add_plugin(Aws::Plugins::JsonvalueConverter) add_plugin(Aws::Plugins::ClientMetricsPlugin) add_plugin(Aws::Plugins::ClientMetricsSendPlugin) add_plugin(Aws::Plugins::TransferEncoding) add_plugin(Aws::Plugins::HttpChecksum) add_plugin(Aws::Plugins::SignatureV4) add_plugin(Aws::Plugins::Protocols::JsonRpc) # @overload initialize(options) # @param [Hash] options # @option options [required, Aws::CredentialProvider] :credentials # Your AWS credentials. This can be an instance of any one of the # following classes: # # * `Aws::Credentials` - Used for configuring static, non-refreshing # credentials. # # * `Aws::InstanceProfileCredentials` - Used for loading credentials # from an EC2 IMDS on an EC2 instance. # # * `Aws::SharedCredentials` - Used for loading credentials from a # shared file, such as `~/.aws/config`. # # * `Aws::AssumeRoleCredentials` - Used when you need to assume a role. # # When `:credentials` are not configured directly, the following # locations will be searched for credentials: # # * `Aws.config[:credentials]` # * The `:access_key_id`, `:secret_access_key`, and `:session_token` options. # * ENV['AWS_ACCESS_KEY_ID'], ENV['AWS_SECRET_ACCESS_KEY'] # * `~/.aws/credentials` # * `~/.aws/config` # * EC2 IMDS instance profile - When used by default, the timeouts are # very aggressive. Construct and pass an instance of # `Aws::InstanceProfileCredentails` to enable retries and extended # timeouts. # # @option options [required, String] :region # The AWS region to connect to. The configured `:region` is # used to determine the service `:endpoint`. When not passed, # a default `:region` is searched for in the following locations: # # * `Aws.config[:region]` # * `ENV['AWS_REGION']` # * `ENV['AMAZON_REGION']` # * `ENV['AWS_DEFAULT_REGION']` # * `~/.aws/credentials` # * `~/.aws/config` # # @option options [String] :access_key_id # # @option options [Boolean] :active_endpoint_cache (false) # When set to `true`, a thread polling for endpoints will be running in # the background every 60 secs (default). Defaults to `false`. # # @option options [Boolean] :adaptive_retry_wait_to_fill (true) # Used only in `adaptive` retry mode. When true, the request will sleep # until there is sufficent client side capacity to retry the request. # When false, the request will raise a `RetryCapacityNotAvailableError` and will # not retry instead of sleeping. # # @option options [Boolean] :client_side_monitoring (false) # When `true`, client-side metrics will be collected for all API requests from # this client. # # @option options [String] :client_side_monitoring_client_id ("") # Allows you to provide an identifier for this client which will be attached to # all generated client side metrics. Defaults to an empty string. # # @option options [String] :client_side_monitoring_host ("127.0.0.1") # Allows you to specify the DNS hostname or IPv4 or IPv6 address that the client # side monitoring agent is running on, where client metrics will be published via UDP. # # @option options [Integer] :client_side_monitoring_port (31000) # Required for publishing client metrics. The port that the client side monitoring # agent is running on, where client metrics will be published via UDP. # # @option options [Aws::ClientSideMonitoring::Publisher] :client_side_monitoring_publisher (Aws::ClientSideMonitoring::Publisher) # Allows you to provide a custom client-side monitoring publisher class. By default, # will use the Client Side Monitoring Agent Publisher. # # @option options [Boolean] :convert_params (true) # When `true`, an attempt is made to coerce request parameters into # the required types. # # @option options [Boolean] :correct_clock_skew (true) # Used only in `standard` and adaptive retry modes. Specifies whether to apply # a clock skew correction and retry requests with skewed client clocks. # # @option options [Boolean] :disable_host_prefix_injection (false) # Set to true to disable SDK automatically adding host prefix # to default service endpoint when available. # # @option options [String] :endpoint # The client endpoint is normally constructed from the `:region` # option. You should only configure an `:endpoint` when connecting # to test or custom endpoints. This should be a valid HTTP(S) URI. # # @option options [Integer] :endpoint_cache_max_entries (1000) # Used for the maximum size limit of the LRU cache storing endpoints data # for endpoint discovery enabled operations. Defaults to 1000. # # @option options [Integer] :endpoint_cache_max_threads (10) # Used for the maximum threads in use for polling endpoints to be cached, defaults to 10. # # @option options [Integer] :endpoint_cache_poll_interval (60) # When :endpoint_discovery and :active_endpoint_cache is enabled, # Use this option to config the time interval in seconds for making # requests fetching endpoints information. Defaults to 60 sec. # # @option options [Boolean] :endpoint_discovery (false) # When set to `true`, endpoint discovery will be enabled for operations when available. # # @option options [Aws::Log::Formatter] :log_formatter (Aws::Log::Formatter.default) # The log formatter. # # @option options [Symbol] :log_level (:info) # The log level to send messages to the `:logger` at. # # @option options [Logger] :logger # The Logger instance to send log messages to. If this option # is not set, logging will be disabled. # # @option options [Integer] :max_attempts (3) # An integer representing the maximum number attempts that will be made for # a single request, including the initial attempt. For example, # setting this value to 5 will result in a request being retried up to # 4 times. Used in `standard` and `adaptive` retry modes. # # @option options [String] :profile ("default") # Used when loading credentials from the shared credentials file # at HOME/.aws/credentials. When not specified, 'default' is used. # # @option options [Proc] :retry_backoff # A proc or lambda used for backoff. Defaults to 2**retries * retry_base_delay. # This option is only used in the `legacy` retry mode. # # @option options [Float] :retry_base_delay (0.3) # The base delay in seconds used by the default backoff function. This option # is only used in the `legacy` retry mode. # # @option options [Symbol] :retry_jitter (:none) # A delay randomiser function used by the default backoff function. # Some predefined functions can be referenced by name - :none, :equal, :full, # otherwise a Proc that takes and returns a number. This option is only used # in the `legacy` retry mode. # # @see https://www.awsarchitectureblog.com/2015/03/backoff.html # # @option options [Integer] :retry_limit (3) # The maximum number of times to retry failed requests. Only # ~ 500 level server errors and certain ~ 400 level client errors # are retried. Generally, these are throttling errors, data # checksum errors, networking errors, timeout errors, auth errors, # endpoint discovery, and errors from expired credentials. # This option is only used in the `legacy` retry mode. # # @option options [Integer] :retry_max_delay (0) # The maximum number of seconds to delay between retries (0 for no limit) # used by the default backoff function. This option is only used in the # `legacy` retry mode. # # @option options [String] :retry_mode ("legacy") # Specifies which retry algorithm to use. Values are: # # * `legacy` - The pre-existing retry behavior. This is default value if # no retry mode is provided. # # * `standard` - A standardized set of retry rules across the AWS SDKs. # This includes support for retry quotas, which limit the number of # unsuccessful retries a client can make. # # * `adaptive` - An experimental retry mode that includes all the # functionality of `standard` mode along with automatic client side # throttling. This is a provisional mode that may change behavior # in the future. # # # @option options [String] :secret_access_key # # @option options [String] :session_token # # @option options [Boolean] :simple_json (false) # Disables request parameter conversion, validation, and formatting. # Also disable response data type conversions. This option is useful # when you want to ensure the highest level of performance by # avoiding overhead of walking request parameters and response data # structures. # # When `:simple_json` is enabled, the request parameters hash must # be formatted exactly as the DynamoDB API expects. # # @option options [Boolean] :stub_responses (false) # Causes the client to return stubbed responses. By default # fake responses are generated and returned. You can specify # the response data to return or errors to raise by calling # {ClientStubs#stub_responses}. See {ClientStubs} for more information. # # ** Please note ** When response stubbing is enabled, no HTTP # requests are made, and retries are disabled. # # @option options [Boolean] :validate_params (true) # When `true`, request parameters are validated before # sending the request. # # @option options [URI::HTTP,String] :http_proxy A proxy to send # requests through. Formatted like 'http://proxy.com:123'. # # @option options [Float] :http_open_timeout (15) The number of # seconds to wait when opening a HTTP session before raising a # `Timeout::Error`. # # @option options [Integer] :http_read_timeout (60) The default # number of seconds to wait for response data. This value can # safely be set per-request on the session. # # @option options [Float] :http_idle_timeout (5) The number of # seconds a connection is allowed to sit idle before it is # considered stale. Stale connections are closed and removed # from the pool before making a request. # # @option options [Float] :http_continue_timeout (1) The number of # seconds to wait for a 100-continue response before sending the # request body. This option has no effect unless the request has # "Expect" header set to "100-continue". Defaults to `nil` which # disables this behaviour. This value can safely be set per # request on the session. # # @option options [Boolean] :http_wire_trace (false) When `true`, # HTTP debug output will be sent to the `:logger`. # # @option options [Boolean] :ssl_verify_peer (true) When `true`, # SSL peer certificates are verified when establishing a # connection. # # @option options [String] :ssl_ca_bundle Full path to the SSL # certificate authority bundle file that should be used when # verifying peer certificates. If you do not pass # `:ssl_ca_bundle` or `:ssl_ca_directory` the the system default # will be used if available. # # @option options [String] :ssl_ca_directory Full path of the # directory that contains the unbundled SSL certificate # authority files for verifying peer certificates. If you do # not pass `:ssl_ca_bundle` or `:ssl_ca_directory` the the # system default will be used if available. # def initialize(*args) super end # @!group API Operations # Cancels an existing Amazon FSx for Lustre data repository task if that # task is in either the `PENDING` or `EXECUTING` state. When you cancel # a task, Amazon FSx does the following. # # * Any files that FSx has already exported are not reverted. # # * FSx continues to export any files that are "in-flight" when the # cancel operation is received. # # * FSx does not export any files that have not yet been exported. # # @option params [required, String] :task_id # Specifies the data repository task to cancel. # # @return [Types::CancelDataRepositoryTaskResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods: # # * {Types::CancelDataRepositoryTaskResponse#lifecycle #lifecycle} => String # * {Types::CancelDataRepositoryTaskResponse#task_id #task_id} => String # # @example Request syntax with placeholder values # # resp = client.cancel_data_repository_task({ # task_id: "TaskId", # required # }) # # @example Response structure # # resp.lifecycle #=> String, one of "PENDING", "EXECUTING", "FAILED", "SUCCEEDED", "CANCELED", "CANCELING" # resp.task_id #=> String # # @see http://docs.aws.amazon.com/goto/WebAPI/fsx-2018-03-01/CancelDataRepositoryTask AWS API Documentation # # @overload cancel_data_repository_task(params = {}) # @param [Hash] params ({}) def cancel_data_repository_task(params = {}, options = {}) req = build_request(:cancel_data_repository_task, params) req.send_request(options) end # Creates a backup of an existing Amazon FSx file system. Creating # regular backups for your file system is a best practice, enabling you # to restore a file system from a backup if an issue arises with the # original file system. # # For Amazon FSx for Lustre file systems, you can create a backup only # for file systems with the following configuration: # # * a Persistent deployment type # # * is *not* linked to an Amazon S3 data respository. # # For more information, see # [https://docs.aws.amazon.com/fsx/latest/LustreGuide/lustre-backups.html][1]. # # If a backup with the specified client request token exists, and the # parameters match, this operation returns the description of the # existing backup. If a backup specified client request token exists, # and the parameters don't match, this operation returns # `IncompatibleParameterError`. If a backup with the specified client # request token doesn't exist, `CreateBackup` does the following: # # * Creates a new Amazon FSx backup with an assigned ID, and an initial # lifecycle state of `CREATING`. # # * Returns the description of the backup. # # By using the idempotent operation, you can retry a `CreateBackup` # operation without the risk of creating an extra backup. This approach # can be useful when an initial call fails in a way that makes it # unclear whether a backup was created. If you use the same client # request token and the initial call created a backup, the operation # returns a successful result because all the parameters are the same. # # The `CreateBackup` operation returns while the backup's lifecycle # state is still `CREATING`. You can check the backup creation status by # calling the DescribeBackups operation, which returns the backup state # along with other information. # # # # [1]: https://docs.aws.amazon.com/fsx/latest/LustreGuide/lustre-backups.html # # @option params [required, String] :file_system_id # The ID of the file system to back up. # # @option params [String] :client_request_token # A string of up to 64 ASCII characters that Amazon FSx uses to ensure # idempotent creation. This string is automatically filled on your # behalf when you use the AWS Command Line Interface (AWS CLI) or an AWS # SDK. # # **A suitable default value is auto-generated.** You should normally # not need to pass this option.** # # @option params [Array] :tags # The tags to apply to the backup at backup creation. The key value of # the `Name` tag appears in the console as the backup name. If you have # set `CopyTagsToBackups` to true, and you specify one or more tags # using the `CreateBackup` action, no existing tags on the file system # are copied from the file system to the backup. # # @return [Types::CreateBackupResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods: # # * {Types::CreateBackupResponse#backup #backup} => Types::Backup # # # @example Example: To create a new backup # # # This operation creates a new backup. # # resp = client.create_backup({ # file_system_id: "fs-0498eed5fe91001ec", # tags: [ # { # key: "Name", # value: "MyBackup", # }, # ], # }) # # resp.to_h outputs the following: # { # backup: { # backup_id: "backup-03e3c82e0183b7b6b", # creation_time: Time.parse("1481841524.0"), # file_system: { # file_system_id: "fs-0498eed5fe91001ec", # owner_id: "012345678912", # storage_capacity: 300, # windows_configuration: { # active_directory_id: "d-1234abcd12", # automatic_backup_retention_days: 30, # daily_automatic_backup_start_time: "05:00", # weekly_maintenance_start_time: "1:05:00", # }, # }, # lifecycle: "CREATING", # progress_percent: 0, # resource_arn: "arn:aws:fsx:us-east-1:012345678912:backup/backup-03e3c82e0183b7b6b", # tags: [ # { # key: "Name", # value: "MyBackup", # }, # ], # type: "USER_INITIATED", # }, # } # # @example Request syntax with placeholder values # # resp = client.create_backup({ # file_system_id: "FileSystemId", # required # client_request_token: "ClientRequestToken", # tags: [ # { # key: "TagKey", # value: "TagValue", # }, # ], # }) # # @example Response structure # # resp.backup.backup_id #=> String # resp.backup.lifecycle #=> String, one of "AVAILABLE", "CREATING", "DELETED", "FAILED" # resp.backup.failure_details.message #=> String # resp.backup.type #=> String, one of "AUTOMATIC", "USER_INITIATED" # resp.backup.progress_percent #=> Integer # resp.backup.creation_time #=> Time # resp.backup.kms_key_id #=> String # resp.backup.resource_arn #=> String # resp.backup.tags #=> Array # resp.backup.tags[0].key #=> String # resp.backup.tags[0].value #=> String # resp.backup.file_system.owner_id #=> String # resp.backup.file_system.creation_time #=> Time # resp.backup.file_system.file_system_id #=> String # resp.backup.file_system.file_system_type #=> String, one of "WINDOWS", "LUSTRE" # resp.backup.file_system.lifecycle #=> String, one of "AVAILABLE", "CREATING", "FAILED", "DELETING", "MISCONFIGURED", "UPDATING" # resp.backup.file_system.failure_details.message #=> String # resp.backup.file_system.storage_capacity #=> Integer # resp.backup.file_system.storage_type #=> String, one of "SSD", "HDD" # resp.backup.file_system.vpc_id #=> String # resp.backup.file_system.subnet_ids #=> Array # resp.backup.file_system.subnet_ids[0] #=> String # resp.backup.file_system.network_interface_ids #=> Array # resp.backup.file_system.network_interface_ids[0] #=> String # resp.backup.file_system.dns_name #=> String # resp.backup.file_system.kms_key_id #=> String # resp.backup.file_system.resource_arn #=> String # resp.backup.file_system.tags #=> Array # resp.backup.file_system.tags[0].key #=> String # resp.backup.file_system.tags[0].value #=> String # resp.backup.file_system.windows_configuration.active_directory_id #=> String # resp.backup.file_system.windows_configuration.self_managed_active_directory_configuration.domain_name #=> String # resp.backup.file_system.windows_configuration.self_managed_active_directory_configuration.organizational_unit_distinguished_name #=> String # resp.backup.file_system.windows_configuration.self_managed_active_directory_configuration.file_system_administrators_group #=> String # resp.backup.file_system.windows_configuration.self_managed_active_directory_configuration.user_name #=> String # resp.backup.file_system.windows_configuration.self_managed_active_directory_configuration.dns_ips #=> Array # resp.backup.file_system.windows_configuration.self_managed_active_directory_configuration.dns_ips[0] #=> String # resp.backup.file_system.windows_configuration.deployment_type #=> String, one of "MULTI_AZ_1", "SINGLE_AZ_1", "SINGLE_AZ_2" # resp.backup.file_system.windows_configuration.remote_administration_endpoint #=> String # resp.backup.file_system.windows_configuration.preferred_subnet_id #=> String # resp.backup.file_system.windows_configuration.preferred_file_server_ip #=> String # resp.backup.file_system.windows_configuration.throughput_capacity #=> Integer # resp.backup.file_system.windows_configuration.maintenance_operations_in_progress #=> Array # resp.backup.file_system.windows_configuration.maintenance_operations_in_progress[0] #=> String, one of "PATCHING", "BACKING_UP" # resp.backup.file_system.windows_configuration.weekly_maintenance_start_time #=> String # resp.backup.file_system.windows_configuration.daily_automatic_backup_start_time #=> String # resp.backup.file_system.windows_configuration.automatic_backup_retention_days #=> Integer # resp.backup.file_system.windows_configuration.copy_tags_to_backups #=> Boolean # resp.backup.file_system.lustre_configuration.weekly_maintenance_start_time #=> String # resp.backup.file_system.lustre_configuration.data_repository_configuration.import_path #=> String # resp.backup.file_system.lustre_configuration.data_repository_configuration.export_path #=> String # resp.backup.file_system.lustre_configuration.data_repository_configuration.imported_file_chunk_size #=> Integer # resp.backup.file_system.lustre_configuration.deployment_type #=> String, one of "SCRATCH_1", "SCRATCH_2", "PERSISTENT_1" # resp.backup.file_system.lustre_configuration.per_unit_storage_throughput #=> Integer # resp.backup.file_system.lustre_configuration.mount_name #=> String # resp.backup.file_system.lustre_configuration.daily_automatic_backup_start_time #=> String # resp.backup.file_system.lustre_configuration.automatic_backup_retention_days #=> Integer # resp.backup.file_system.lustre_configuration.copy_tags_to_backups #=> Boolean # resp.backup.file_system.administrative_actions #=> Array # resp.backup.file_system.administrative_actions[0].administrative_action_type #=> String, one of "FILE_SYSTEM_UPDATE", "STORAGE_OPTIMIZATION" # resp.backup.file_system.administrative_actions[0].progress_percent #=> Integer # resp.backup.file_system.administrative_actions[0].request_time #=> Time # resp.backup.file_system.administrative_actions[0].status #=> String, one of "FAILED", "IN_PROGRESS", "PENDING", "COMPLETED", "UPDATED_OPTIMIZING" # resp.backup.file_system.administrative_actions[0].target_file_system_values #=> Types::FileSystem # resp.backup.file_system.administrative_actions[0].failure_details.message #=> String # resp.backup.directory_information.domain_name #=> String # resp.backup.directory_information.active_directory_id #=> String # # @see http://docs.aws.amazon.com/goto/WebAPI/fsx-2018-03-01/CreateBackup AWS API Documentation # # @overload create_backup(params = {}) # @param [Hash] params ({}) def create_backup(params = {}, options = {}) req = build_request(:create_backup, params) req.send_request(options) end # Creates an Amazon FSx for Lustre data repository task. You use data # repository tasks to perform bulk operations between your Amazon FSx # file system and its linked data repository. An example of a data # repository task is exporting any data and metadata changes, including # POSIX metadata, to files, directories, and symbolic links (symlinks) # from your FSx file system to its linked data repository. A # `CreateDataRepositoryTask` operation will fail if a data repository is # not linked to the FSx file system. To learn more about data repository # tasks, see [Using Data Repository Tasks][1]. To learn more about # linking a data repository to your file system, see [Setting the Export # Prefix][2]. # # # # [1]: https://docs.aws.amazon.com/fsx/latest/LustreGuide/data-repository-tasks.html # [2]: https://docs.aws.amazon.com/fsx/latest/LustreGuide/export-data-repository.html#export-prefix # # @option params [required, String] :type # Specifies the type of data repository task to create. # # @option params [Array] :paths # (Optional) The path or paths on the Amazon FSx file system to use when # the data repository task is processed. The default path is the file # system root directory. The paths you provide need to be relative to # the mount point of the file system. If the mount point is `/mnt/fsx` # and `/mnt/fsx/path1` is a directory or file on the file system you # want to export, then the path to provide is `path1`. If a path that # you provide isn't valid, the task fails. # # @option params [required, String] :file_system_id # The globally unique ID of the file system, assigned by Amazon FSx. # # @option params [required, Types::CompletionReport] :report # Defines whether or not Amazon FSx provides a CompletionReport once the # task has completed. A CompletionReport provides a detailed report on # the files that Amazon FSx processed that meet the criteria specified # by the `Scope` parameter. For more information, see [Working with Task # Completion Reports][1]. # # # # [1]: https://docs.aws.amazon.com/fsx/latest/LustreGuide/task-completion-report.html # # @option params [String] :client_request_token # (Optional) An idempotency token for resource creation, in a string of # up to 64 ASCII characters. This token is automatically filled on your # behalf when you use the AWS Command Line Interface (AWS CLI) or an AWS # SDK. # # **A suitable default value is auto-generated.** You should normally # not need to pass this option.** # # @option params [Array] :tags # A list of `Tag` values, with a maximum of 50 elements. # # @return [Types::CreateDataRepositoryTaskResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods: # # * {Types::CreateDataRepositoryTaskResponse#data_repository_task #data_repository_task} => Types::DataRepositoryTask # # @example Request syntax with placeholder values # # resp = client.create_data_repository_task({ # type: "EXPORT_TO_REPOSITORY", # required, accepts EXPORT_TO_REPOSITORY # paths: ["DataRepositoryTaskPath"], # file_system_id: "FileSystemId", # required # report: { # required # enabled: false, # required # path: "ArchivePath", # format: "REPORT_CSV_20191124", # accepts REPORT_CSV_20191124 # scope: "FAILED_FILES_ONLY", # accepts FAILED_FILES_ONLY # }, # client_request_token: "ClientRequestToken", # tags: [ # { # key: "TagKey", # value: "TagValue", # }, # ], # }) # # @example Response structure # # resp.data_repository_task.task_id #=> String # resp.data_repository_task.lifecycle #=> String, one of "PENDING", "EXECUTING", "FAILED", "SUCCEEDED", "CANCELED", "CANCELING" # resp.data_repository_task.type #=> String, one of "EXPORT_TO_REPOSITORY" # resp.data_repository_task.creation_time #=> Time # resp.data_repository_task.start_time #=> Time # resp.data_repository_task.end_time #=> Time # resp.data_repository_task.resource_arn #=> String # resp.data_repository_task.tags #=> Array # resp.data_repository_task.tags[0].key #=> String # resp.data_repository_task.tags[0].value #=> String # resp.data_repository_task.file_system_id #=> String # resp.data_repository_task.paths #=> Array # resp.data_repository_task.paths[0] #=> String # resp.data_repository_task.failure_details.message #=> String # resp.data_repository_task.status.total_count #=> Integer # resp.data_repository_task.status.succeeded_count #=> Integer # resp.data_repository_task.status.failed_count #=> Integer # resp.data_repository_task.status.last_updated_time #=> Time # resp.data_repository_task.report.enabled #=> Boolean # resp.data_repository_task.report.path #=> String # resp.data_repository_task.report.format #=> String, one of "REPORT_CSV_20191124" # resp.data_repository_task.report.scope #=> String, one of "FAILED_FILES_ONLY" # # @see http://docs.aws.amazon.com/goto/WebAPI/fsx-2018-03-01/CreateDataRepositoryTask AWS API Documentation # # @overload create_data_repository_task(params = {}) # @param [Hash] params ({}) def create_data_repository_task(params = {}, options = {}) req = build_request(:create_data_repository_task, params) req.send_request(options) end # Creates a new, empty Amazon FSx file system. # # If a file system with the specified client request token exists and # the parameters match, `CreateFileSystem` returns the description of # the existing file system. If a file system specified client request # token exists and the parameters don't match, this call returns # `IncompatibleParameterError`. If a file system with the specified # client request token doesn't exist, `CreateFileSystem` does the # following: # # * Creates a new, empty Amazon FSx file system with an assigned ID, and # an initial lifecycle state of `CREATING`. # # * Returns the description of the file system. # # This operation requires a client request token in the request that # Amazon FSx uses to ensure idempotent creation. This means that calling # the operation multiple times with the same client request token has no # effect. By using the idempotent operation, you can retry a # `CreateFileSystem` operation without the risk of creating an extra # file system. This approach can be useful when an initial call fails in # a way that makes it unclear whether a file system was created. # Examples are if a transport level timeout occurred, or your connection # was reset. If you use the same client request token and the initial # call created a file system, the client receives success as long as the # parameters are the same. # # The `CreateFileSystem` call returns while the file system's lifecycle # state is still `CREATING`. You can check the file-system creation # status by calling the DescribeFileSystems operation, which returns the # file system state along with other information. # # # # @option params [String] :client_request_token # A string of up to 64 ASCII characters that Amazon FSx uses to ensure # idempotent creation. This string is automatically filled on your # behalf when you use the AWS Command Line Interface (AWS CLI) or an AWS # SDK. # # **A suitable default value is auto-generated.** You should normally # not need to pass this option.** # # @option params [required, String] :file_system_type # The type of Amazon FSx file system to create, either `WINDOWS` or # `LUSTRE`. # # @option params [required, Integer] :storage_capacity # Sets the storage capacity of the file system that you're creating. # # For Lustre file systems: # # * For `SCRATCH_2` and `PERSISTENT_1` deployment types, valid values # are 1.2, 2.4, and increments of 2.4 TiB. # # * For `SCRATCH_1` deployment type, valid values are 1.2, 2.4, and # increments of 3.6 TiB. # # For Windows file systems: # # * If `StorageType=SSD`, valid values are 32 GiB - 65,536 GiB (64 TiB). # # * If `StorageType=HDD`, valid values are 2000 GiB - 65,536 GiB (64 # TiB). # # @option params [String] :storage_type # Sets the storage type for the Amazon FSx for Windows file system # you're creating. Valid values are `SSD` and `HDD`. # # * Set to `SSD` to use solid state drive storage. SSD is supported on # all Windows deployment types. # # * Set to `HDD` to use hard disk drive storage. HDD is supported on # `SINGLE_AZ_2` and `MULTI_AZ_1` Windows file system deployment types. # # Default value is `SSD`. For more information, see [ Storage Type # Options][1] in the *Amazon FSx for Windows User Guide*. # # # # [1]: https://docs.aws.amazon.com/fsx/latest/WindowsGuide/optimize-fsx-costs.html#storage-type-options # # @option params [required, Array] :subnet_ids # Specifies the IDs of the subnets that the file system will be # accessible from. For Windows `MULTI_AZ_1` file system deployment # types, provide exactly two subnet IDs, one for the preferred file # server and one for the standby file server. You specify one of these # subnets as the preferred subnet using the `WindowsConfiguration > # PreferredSubnetID` property. # # For Windows `SINGLE_AZ_1` and `SINGLE_AZ_2` file system deployment # types and Lustre file systems, provide exactly one subnet ID. The file # server is launched in that subnet's Availability Zone. # # @option params [Array] :security_group_ids # A list of IDs specifying the security groups to apply to all network # interfaces created for file system access. This list isn't returned # in later requests to describe the file system. # # @option params [Array] :tags # The tags to apply to the file system being created. The key value of # the `Name` tag appears in the console as the file system name. # # @option params [String] :kms_key_id # The ID of the AWS Key Management Service (AWS KMS) key used to encrypt # the file system's data for Amazon FSx for Windows File Server file # systems and Amazon FSx for Lustre `PERSISTENT_1` file systems at rest. # In either case, if not specified, the Amazon FSx managed key is used. # The Amazon FSx for Lustre `SCRATCH_1` and `SCRATCH_2` file systems are # always encrypted at rest using Amazon FSx managed keys. For more # information, see [Encrypt][1] in the *AWS Key Management Service API # Reference*. # # # # [1]: https://docs.aws.amazon.com/kms/latest/APIReference/API_Encrypt.html # # @option params [Types::CreateFileSystemWindowsConfiguration] :windows_configuration # The Microsoft Windows configuration for the file system being created. # # @option params [Types::CreateFileSystemLustreConfiguration] :lustre_configuration # The Lustre configuration for the file system being created. # # @return [Types::CreateFileSystemResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods: # # * {Types::CreateFileSystemResponse#file_system #file_system} => Types::FileSystem # # # @example Example: To create a new file system # # # This operation creates a new file system. # # resp = client.create_file_system({ # client_request_token: "a8ca07e4-61ec-4399-99f4-19853801bcd5", # file_system_type: "WINDOWS", # kms_key_id: "arn:aws:kms:us-east-1:012345678912:key/0ff3ea8d-130e-4133-877f-93908b6fdbd6", # security_group_ids: [ # "sg-edcd9784", # ], # storage_capacity: 300, # subnet_ids: [ # "subnet-1234abcd", # ], # tags: [ # { # key: "Name", # value: "MyFileSystem", # }, # ], # windows_configuration: { # active_directory_id: "d-1234abcd12", # automatic_backup_retention_days: 30, # daily_automatic_backup_start_time: "05:00", # throughput_capacity: 8, # weekly_maintenance_start_time: "1:05:00", # }, # }) # # resp.to_h outputs the following: # { # file_system: { # creation_time: Time.parse("1481841524.0"), # dns_name: "fs-0498eed5fe91001ec.fsx.com", # file_system_id: "fs-0498eed5fe91001ec", # kms_key_id: "arn:aws:kms:us-east-1:012345678912:key/0ff3ea8d-130e-4133-877f-93908b6fdbd6", # lifecycle: "CREATING", # owner_id: "012345678912", # resource_arn: "arn:aws:fsx:us-east-1:012345678912:file-system/fs-0498eed5fe91001ec", # storage_capacity: 300, # subnet_ids: [ # "subnet-1234abcd", # ], # tags: [ # { # key: "Name", # value: "MyFileSystem", # }, # ], # vpc_id: "vpc-ab1234cd", # windows_configuration: { # active_directory_id: "d-1234abcd12", # automatic_backup_retention_days: 30, # daily_automatic_backup_start_time: "05:00", # throughput_capacity: 8, # weekly_maintenance_start_time: "1:05:00", # }, # }, # } # # @example Request syntax with placeholder values # # resp = client.create_file_system({ # client_request_token: "ClientRequestToken", # file_system_type: "WINDOWS", # required, accepts WINDOWS, LUSTRE # storage_capacity: 1, # required # storage_type: "SSD", # accepts SSD, HDD # subnet_ids: ["SubnetId"], # required # security_group_ids: ["SecurityGroupId"], # tags: [ # { # key: "TagKey", # value: "TagValue", # }, # ], # kms_key_id: "KmsKeyId", # windows_configuration: { # active_directory_id: "DirectoryId", # self_managed_active_directory_configuration: { # domain_name: "ActiveDirectoryFullyQualifiedName", # required # organizational_unit_distinguished_name: "OrganizationalUnitDistinguishedName", # file_system_administrators_group: "FileSystemAdministratorsGroupName", # user_name: "DirectoryUserName", # required # password: "DirectoryPassword", # required # dns_ips: ["IpAddress"], # required # }, # deployment_type: "MULTI_AZ_1", # accepts MULTI_AZ_1, SINGLE_AZ_1, SINGLE_AZ_2 # preferred_subnet_id: "SubnetId", # throughput_capacity: 1, # required # weekly_maintenance_start_time: "WeeklyTime", # daily_automatic_backup_start_time: "DailyTime", # automatic_backup_retention_days: 1, # copy_tags_to_backups: false, # }, # lustre_configuration: { # weekly_maintenance_start_time: "WeeklyTime", # import_path: "ArchivePath", # export_path: "ArchivePath", # imported_file_chunk_size: 1, # deployment_type: "SCRATCH_1", # accepts SCRATCH_1, SCRATCH_2, PERSISTENT_1 # per_unit_storage_throughput: 1, # daily_automatic_backup_start_time: "DailyTime", # automatic_backup_retention_days: 1, # copy_tags_to_backups: false, # }, # }) # # @example Response structure # # resp.file_system.owner_id #=> String # resp.file_system.creation_time #=> Time # resp.file_system.file_system_id #=> String # resp.file_system.file_system_type #=> String, one of "WINDOWS", "LUSTRE" # resp.file_system.lifecycle #=> String, one of "AVAILABLE", "CREATING", "FAILED", "DELETING", "MISCONFIGURED", "UPDATING" # resp.file_system.failure_details.message #=> String # resp.file_system.storage_capacity #=> Integer # resp.file_system.storage_type #=> String, one of "SSD", "HDD" # resp.file_system.vpc_id #=> String # resp.file_system.subnet_ids #=> Array # resp.file_system.subnet_ids[0] #=> String # resp.file_system.network_interface_ids #=> Array # resp.file_system.network_interface_ids[0] #=> String # resp.file_system.dns_name #=> String # resp.file_system.kms_key_id #=> String # resp.file_system.resource_arn #=> String # resp.file_system.tags #=> Array # resp.file_system.tags[0].key #=> String # resp.file_system.tags[0].value #=> String # resp.file_system.windows_configuration.active_directory_id #=> String # resp.file_system.windows_configuration.self_managed_active_directory_configuration.domain_name #=> String # resp.file_system.windows_configuration.self_managed_active_directory_configuration.organizational_unit_distinguished_name #=> String # resp.file_system.windows_configuration.self_managed_active_directory_configuration.file_system_administrators_group #=> String # resp.file_system.windows_configuration.self_managed_active_directory_configuration.user_name #=> String # resp.file_system.windows_configuration.self_managed_active_directory_configuration.dns_ips #=> Array # resp.file_system.windows_configuration.self_managed_active_directory_configuration.dns_ips[0] #=> String # resp.file_system.windows_configuration.deployment_type #=> String, one of "MULTI_AZ_1", "SINGLE_AZ_1", "SINGLE_AZ_2" # resp.file_system.windows_configuration.remote_administration_endpoint #=> String # resp.file_system.windows_configuration.preferred_subnet_id #=> String # resp.file_system.windows_configuration.preferred_file_server_ip #=> String # resp.file_system.windows_configuration.throughput_capacity #=> Integer # resp.file_system.windows_configuration.maintenance_operations_in_progress #=> Array # resp.file_system.windows_configuration.maintenance_operations_in_progress[0] #=> String, one of "PATCHING", "BACKING_UP" # resp.file_system.windows_configuration.weekly_maintenance_start_time #=> String # resp.file_system.windows_configuration.daily_automatic_backup_start_time #=> String # resp.file_system.windows_configuration.automatic_backup_retention_days #=> Integer # resp.file_system.windows_configuration.copy_tags_to_backups #=> Boolean # resp.file_system.lustre_configuration.weekly_maintenance_start_time #=> String # resp.file_system.lustre_configuration.data_repository_configuration.import_path #=> String # resp.file_system.lustre_configuration.data_repository_configuration.export_path #=> String # resp.file_system.lustre_configuration.data_repository_configuration.imported_file_chunk_size #=> Integer # resp.file_system.lustre_configuration.deployment_type #=> String, one of "SCRATCH_1", "SCRATCH_2", "PERSISTENT_1" # resp.file_system.lustre_configuration.per_unit_storage_throughput #=> Integer # resp.file_system.lustre_configuration.mount_name #=> String # resp.file_system.lustre_configuration.daily_automatic_backup_start_time #=> String # resp.file_system.lustre_configuration.automatic_backup_retention_days #=> Integer # resp.file_system.lustre_configuration.copy_tags_to_backups #=> Boolean # resp.file_system.administrative_actions #=> Array # resp.file_system.administrative_actions[0].administrative_action_type #=> String, one of "FILE_SYSTEM_UPDATE", "STORAGE_OPTIMIZATION" # resp.file_system.administrative_actions[0].progress_percent #=> Integer # resp.file_system.administrative_actions[0].request_time #=> Time # resp.file_system.administrative_actions[0].status #=> String, one of "FAILED", "IN_PROGRESS", "PENDING", "COMPLETED", "UPDATED_OPTIMIZING" # resp.file_system.administrative_actions[0].target_file_system_values #=> Types::FileSystem # resp.file_system.administrative_actions[0].failure_details.message #=> String # # @see http://docs.aws.amazon.com/goto/WebAPI/fsx-2018-03-01/CreateFileSystem AWS API Documentation # # @overload create_file_system(params = {}) # @param [Hash] params ({}) def create_file_system(params = {}, options = {}) req = build_request(:create_file_system, params) req.send_request(options) end # Creates a new Amazon FSx file system from an existing Amazon FSx # backup. # # If a file system with the specified client request token exists and # the parameters match, this operation returns the description of the # file system. If a client request token specified by the file system # exists and the parameters don't match, this call returns # `IncompatibleParameterError`. If a file system with the specified # client request token doesn't exist, this operation does the # following: # # * Creates a new Amazon FSx file system from backup with an assigned # ID, and an initial lifecycle state of `CREATING`. # # * Returns the description of the file system. # # Parameters like Active Directory, default share name, automatic # backup, and backup settings default to the parameters of the file # system that was backed up, unless overridden. You can explicitly # supply other settings. # # By using the idempotent operation, you can retry a # `CreateFileSystemFromBackup` call without the risk of creating an # extra file system. This approach can be useful when an initial call # fails in a way that makes it unclear whether a file system was # created. Examples are if a transport level timeout occurred, or your # connection was reset. If you use the same client request token and the # initial call created a file system, the client receives success as # long as the parameters are the same. # # The `CreateFileSystemFromBackup` call returns while the file system's # lifecycle state is still `CREATING`. You can check the file-system # creation status by calling the DescribeFileSystems operation, which # returns the file system state along with other information. # # # # @option params [required, String] :backup_id # The ID of the backup. Specifies the backup to use if you're creating # a file system from an existing backup. # # @option params [String] :client_request_token # A string of up to 64 ASCII characters that Amazon FSx uses to ensure # idempotent creation. This string is automatically filled on your # behalf when you use the AWS Command Line Interface (AWS CLI) or an AWS # SDK. # # **A suitable default value is auto-generated.** You should normally # not need to pass this option.** # # @option params [required, Array] :subnet_ids # Specifies the IDs of the subnets that the file system will be # accessible from. For Windows `MULTI_AZ_1` file system deployment # types, provide exactly two subnet IDs, one for the preferred file # server and one for the standby file server. You specify one of these # subnets as the preferred subnet using the `WindowsConfiguration > # PreferredSubnetID` property. # # For Windows `SINGLE_AZ_1` and `SINGLE_AZ_2` deployment types and # Lustre file systems, provide exactly one subnet ID. The file server is # launched in that subnet's Availability Zone. # # @option params [Array] :security_group_ids # A list of IDs for the security groups that apply to the specified # network interfaces created for file system access. These security # groups apply to all network interfaces. This value isn't returned in # later DescribeFileSystem requests. # # @option params [Array] :tags # The tags to be applied to the file system at file system creation. The # key value of the `Name` tag appears in the console as the file system # name. # # @option params [Types::CreateFileSystemWindowsConfiguration] :windows_configuration # The configuration for this Microsoft Windows file system. # # @option params [Types::CreateFileSystemLustreConfiguration] :lustre_configuration # The Lustre configuration for the file system being created. # # @option params [String] :storage_type # Sets the storage type for the Windows file system you're creating # from a backup. Valid values are `SSD` and `HDD`. # # * Set to `SSD` to use solid state drive storage. Supported on all # Windows deployment types. # # * Set to `HDD` to use hard disk drive storage. Supported on # `SINGLE_AZ_2` and `MULTI_AZ_1` Windows file system deployment types. # # Default value is `SSD`. # # HDD and SSD storage types have different minimum storage capacity # requirements. A restored file system's storage capacity is tied to # the file system that was backed up. You can create a file system that # uses HDD storage from a backup of a file system that used SSD storage # only if the original SSD file system had a storage capacity of at # least 2000 GiB. # # # # @return [Types::CreateFileSystemFromBackupResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods: # # * {Types::CreateFileSystemFromBackupResponse#file_system #file_system} => Types::FileSystem # # # @example Example: To create a new file system from backup # # # This operation creates a new file system from backup. # # resp = client.create_file_system_from_backup({ # backup_id: "backup-03e3c82e0183b7b6b", # client_request_token: "f4c94ed7-238d-4c46-93db-48cd62ec33b7", # security_group_ids: [ # "sg-edcd9784", # ], # subnet_ids: [ # "subnet-1234abcd", # ], # tags: [ # { # key: "Name", # value: "MyFileSystem", # }, # ], # windows_configuration: { # throughput_capacity: 8, # }, # }) # # resp.to_h outputs the following: # { # file_system: { # creation_time: Time.parse("1481841524.0"), # dns_name: "fs-0498eed5fe91001ec.fsx.com", # file_system_id: "fs-0498eed5fe91001ec", # kms_key_id: "arn:aws:kms:us-east-1:012345678912:key/0ff3ea8d-130e-4133-877f-93908b6fdbd6", # lifecycle: "CREATING", # owner_id: "012345678912", # resource_arn: "arn:aws:fsx:us-east-1:012345678912:file-system/fs-0498eed5fe91001ec", # storage_capacity: 300, # subnet_ids: [ # "subnet-1234abcd", # ], # tags: [ # { # key: "Name", # value: "MyFileSystem", # }, # ], # vpc_id: "vpc-ab1234cd", # windows_configuration: { # active_directory_id: "d-1234abcd12", # automatic_backup_retention_days: 30, # daily_automatic_backup_start_time: "05:00", # throughput_capacity: 8, # weekly_maintenance_start_time: "1:05:00", # }, # }, # } # # @example Request syntax with placeholder values # # resp = client.create_file_system_from_backup({ # backup_id: "BackupId", # required # client_request_token: "ClientRequestToken", # subnet_ids: ["SubnetId"], # required # security_group_ids: ["SecurityGroupId"], # tags: [ # { # key: "TagKey", # value: "TagValue", # }, # ], # windows_configuration: { # active_directory_id: "DirectoryId", # self_managed_active_directory_configuration: { # domain_name: "ActiveDirectoryFullyQualifiedName", # required # organizational_unit_distinguished_name: "OrganizationalUnitDistinguishedName", # file_system_administrators_group: "FileSystemAdministratorsGroupName", # user_name: "DirectoryUserName", # required # password: "DirectoryPassword", # required # dns_ips: ["IpAddress"], # required # }, # deployment_type: "MULTI_AZ_1", # accepts MULTI_AZ_1, SINGLE_AZ_1, SINGLE_AZ_2 # preferred_subnet_id: "SubnetId", # throughput_capacity: 1, # required # weekly_maintenance_start_time: "WeeklyTime", # daily_automatic_backup_start_time: "DailyTime", # automatic_backup_retention_days: 1, # copy_tags_to_backups: false, # }, # lustre_configuration: { # weekly_maintenance_start_time: "WeeklyTime", # import_path: "ArchivePath", # export_path: "ArchivePath", # imported_file_chunk_size: 1, # deployment_type: "SCRATCH_1", # accepts SCRATCH_1, SCRATCH_2, PERSISTENT_1 # per_unit_storage_throughput: 1, # daily_automatic_backup_start_time: "DailyTime", # automatic_backup_retention_days: 1, # copy_tags_to_backups: false, # }, # storage_type: "SSD", # accepts SSD, HDD # }) # # @example Response structure # # resp.file_system.owner_id #=> String # resp.file_system.creation_time #=> Time # resp.file_system.file_system_id #=> String # resp.file_system.file_system_type #=> String, one of "WINDOWS", "LUSTRE" # resp.file_system.lifecycle #=> String, one of "AVAILABLE", "CREATING", "FAILED", "DELETING", "MISCONFIGURED", "UPDATING" # resp.file_system.failure_details.message #=> String # resp.file_system.storage_capacity #=> Integer # resp.file_system.storage_type #=> String, one of "SSD", "HDD" # resp.file_system.vpc_id #=> String # resp.file_system.subnet_ids #=> Array # resp.file_system.subnet_ids[0] #=> String # resp.file_system.network_interface_ids #=> Array # resp.file_system.network_interface_ids[0] #=> String # resp.file_system.dns_name #=> String # resp.file_system.kms_key_id #=> String # resp.file_system.resource_arn #=> String # resp.file_system.tags #=> Array # resp.file_system.tags[0].key #=> String # resp.file_system.tags[0].value #=> String # resp.file_system.windows_configuration.active_directory_id #=> String # resp.file_system.windows_configuration.self_managed_active_directory_configuration.domain_name #=> String # resp.file_system.windows_configuration.self_managed_active_directory_configuration.organizational_unit_distinguished_name #=> String # resp.file_system.windows_configuration.self_managed_active_directory_configuration.file_system_administrators_group #=> String # resp.file_system.windows_configuration.self_managed_active_directory_configuration.user_name #=> String # resp.file_system.windows_configuration.self_managed_active_directory_configuration.dns_ips #=> Array # resp.file_system.windows_configuration.self_managed_active_directory_configuration.dns_ips[0] #=> String # resp.file_system.windows_configuration.deployment_type #=> String, one of "MULTI_AZ_1", "SINGLE_AZ_1", "SINGLE_AZ_2" # resp.file_system.windows_configuration.remote_administration_endpoint #=> String # resp.file_system.windows_configuration.preferred_subnet_id #=> String # resp.file_system.windows_configuration.preferred_file_server_ip #=> String # resp.file_system.windows_configuration.throughput_capacity #=> Integer # resp.file_system.windows_configuration.maintenance_operations_in_progress #=> Array # resp.file_system.windows_configuration.maintenance_operations_in_progress[0] #=> String, one of "PATCHING", "BACKING_UP" # resp.file_system.windows_configuration.weekly_maintenance_start_time #=> String # resp.file_system.windows_configuration.daily_automatic_backup_start_time #=> String # resp.file_system.windows_configuration.automatic_backup_retention_days #=> Integer # resp.file_system.windows_configuration.copy_tags_to_backups #=> Boolean # resp.file_system.lustre_configuration.weekly_maintenance_start_time #=> String # resp.file_system.lustre_configuration.data_repository_configuration.import_path #=> String # resp.file_system.lustre_configuration.data_repository_configuration.export_path #=> String # resp.file_system.lustre_configuration.data_repository_configuration.imported_file_chunk_size #=> Integer # resp.file_system.lustre_configuration.deployment_type #=> String, one of "SCRATCH_1", "SCRATCH_2", "PERSISTENT_1" # resp.file_system.lustre_configuration.per_unit_storage_throughput #=> Integer # resp.file_system.lustre_configuration.mount_name #=> String # resp.file_system.lustre_configuration.daily_automatic_backup_start_time #=> String # resp.file_system.lustre_configuration.automatic_backup_retention_days #=> Integer # resp.file_system.lustre_configuration.copy_tags_to_backups #=> Boolean # resp.file_system.administrative_actions #=> Array # resp.file_system.administrative_actions[0].administrative_action_type #=> String, one of "FILE_SYSTEM_UPDATE", "STORAGE_OPTIMIZATION" # resp.file_system.administrative_actions[0].progress_percent #=> Integer # resp.file_system.administrative_actions[0].request_time #=> Time # resp.file_system.administrative_actions[0].status #=> String, one of "FAILED", "IN_PROGRESS", "PENDING", "COMPLETED", "UPDATED_OPTIMIZING" # resp.file_system.administrative_actions[0].target_file_system_values #=> Types::FileSystem # resp.file_system.administrative_actions[0].failure_details.message #=> String # # @see http://docs.aws.amazon.com/goto/WebAPI/fsx-2018-03-01/CreateFileSystemFromBackup AWS API Documentation # # @overload create_file_system_from_backup(params = {}) # @param [Hash] params ({}) def create_file_system_from_backup(params = {}, options = {}) req = build_request(:create_file_system_from_backup, params) req.send_request(options) end # Deletes an Amazon FSx backup, deleting its contents. After deletion, # the backup no longer exists, and its data is gone. # # The `DeleteBackup` call returns instantly. The backup will not show up # in later `DescribeBackups` calls. # # The data in a deleted backup is also deleted and can't be recovered # by any means. # # @option params [required, String] :backup_id # The ID of the backup you want to delete. # # @option params [String] :client_request_token # A string of up to 64 ASCII characters that Amazon FSx uses to ensure # idempotent deletion. This is automatically filled on your behalf when # using the AWS CLI or SDK. # # **A suitable default value is auto-generated.** You should normally # not need to pass this option.** # # @return [Types::DeleteBackupResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods: # # * {Types::DeleteBackupResponse#backup_id #backup_id} => String # * {Types::DeleteBackupResponse#lifecycle #lifecycle} => String # # # @example Example: To delete a backup # # # This operation deletes an Amazon FSx file system backup. # # resp = client.delete_backup({ # backup_id: "backup-03e3c82e0183b7b6b", # }) # # resp.to_h outputs the following: # { # backup_id: "backup-03e3c82e0183b7b6b", # lifecycle: "DELETED", # } # # @example Request syntax with placeholder values # # resp = client.delete_backup({ # backup_id: "BackupId", # required # client_request_token: "ClientRequestToken", # }) # # @example Response structure # # resp.backup_id #=> String # resp.lifecycle #=> String, one of "AVAILABLE", "CREATING", "DELETED", "FAILED" # # @see http://docs.aws.amazon.com/goto/WebAPI/fsx-2018-03-01/DeleteBackup AWS API Documentation # # @overload delete_backup(params = {}) # @param [Hash] params ({}) def delete_backup(params = {}, options = {}) req = build_request(:delete_backup, params) req.send_request(options) end # Deletes a file system, deleting its contents. After deletion, the file # system no longer exists, and its data is gone. Any existing automatic # backups will also be deleted. # # By default, when you delete an Amazon FSx for Windows File Server file # system, a final backup is created upon deletion. This final backup is # not subject to the file system's retention policy, and must be # manually deleted. # # The `DeleteFileSystem` action returns while the file system has the # `DELETING` status. You can check the file system deletion status by # calling the DescribeFileSystems action, which returns a list of file # systems in your account. If you pass the file system ID for a deleted # file system, the DescribeFileSystems returns a `FileSystemNotFound` # error. # # Deleting an Amazon FSx for Lustre file system will fail with a 400 # BadRequest if a data repository task is in a `PENDING` or `EXECUTING` # state. # # # # The data in a deleted file system is also deleted and can't be # recovered by any means. # # @option params [required, String] :file_system_id # The ID of the file system you want to delete. # # @option params [String] :client_request_token # A string of up to 64 ASCII characters that Amazon FSx uses to ensure # idempotent deletion. This is automatically filled on your behalf when # using the AWS CLI or SDK. # # **A suitable default value is auto-generated.** You should normally # not need to pass this option.** # # @option params [Types::DeleteFileSystemWindowsConfiguration] :windows_configuration # The configuration object for the Microsoft Windows file system used in # the `DeleteFileSystem` operation. # # @option params [Types::DeleteFileSystemLustreConfiguration] :lustre_configuration # The configuration object for the Amazon FSx for Lustre file system # being deleted in the `DeleteFileSystem` operation. # # @return [Types::DeleteFileSystemResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods: # # * {Types::DeleteFileSystemResponse#file_system_id #file_system_id} => String # * {Types::DeleteFileSystemResponse#lifecycle #lifecycle} => String # * {Types::DeleteFileSystemResponse#windows_response #windows_response} => Types::DeleteFileSystemWindowsResponse # * {Types::DeleteFileSystemResponse#lustre_response #lustre_response} => Types::DeleteFileSystemLustreResponse # # # @example Example: To delete a file system # # # This operation deletes an Amazon FSx file system. # # resp = client.delete_file_system({ # file_system_id: "fs-0498eed5fe91001ec", # }) # # resp.to_h outputs the following: # { # file_system_id: "fs-0498eed5fe91001ec", # lifecycle: "DELETING", # } # # @example Request syntax with placeholder values # # resp = client.delete_file_system({ # file_system_id: "FileSystemId", # required # client_request_token: "ClientRequestToken", # windows_configuration: { # skip_final_backup: false, # final_backup_tags: [ # { # key: "TagKey", # value: "TagValue", # }, # ], # }, # lustre_configuration: { # skip_final_backup: false, # final_backup_tags: [ # { # key: "TagKey", # value: "TagValue", # }, # ], # }, # }) # # @example Response structure # # resp.file_system_id #=> String # resp.lifecycle #=> String, one of "AVAILABLE", "CREATING", "FAILED", "DELETING", "MISCONFIGURED", "UPDATING" # resp.windows_response.final_backup_id #=> String # resp.windows_response.final_backup_tags #=> Array # resp.windows_response.final_backup_tags[0].key #=> String # resp.windows_response.final_backup_tags[0].value #=> String # resp.lustre_response.final_backup_id #=> String # resp.lustre_response.final_backup_tags #=> Array # resp.lustre_response.final_backup_tags[0].key #=> String # resp.lustre_response.final_backup_tags[0].value #=> String # # @see http://docs.aws.amazon.com/goto/WebAPI/fsx-2018-03-01/DeleteFileSystem AWS API Documentation # # @overload delete_file_system(params = {}) # @param [Hash] params ({}) def delete_file_system(params = {}, options = {}) req = build_request(:delete_file_system, params) req.send_request(options) end # Returns the description of specific Amazon FSx backups, if a # `BackupIds` value is provided for that backup. Otherwise, it returns # all backups owned by your AWS account in the AWS Region of the # endpoint that you're calling. # # When retrieving all backups, you can optionally specify the # `MaxResults` parameter to limit the number of backups in a response. # If more backups remain, Amazon FSx returns a `NextToken` value in the # response. In this case, send a later request with the `NextToken` # request parameter set to the value of `NextToken` from the last # response. # # This action is used in an iterative process to retrieve a list of your # backups. `DescribeBackups` is called first without a `NextToken`value. # Then the action continues to be called with the `NextToken` parameter # set to the value of the last `NextToken` value until a response has no # `NextToken`. # # When using this action, keep the following in mind: # # * The implementation might return fewer than `MaxResults` file system # descriptions while still including a `NextToken` value. # # * The order of backups returned in the response of one # `DescribeBackups` call and the order of backups returned across the # responses of a multi-call iteration is unspecified. # # @option params [Array] :backup_ids # IDs of the backups you want to retrieve (String). This overrides any # filters. If any IDs are not found, BackupNotFound will be thrown. # # @option params [Array] :filters # Filters structure. Supported names are file-system-id and backup-type. # # @option params [Integer] :max_results # Maximum number of backups to return in the response (integer). This # parameter value must be greater than 0. The number of items that # Amazon FSx returns is the minimum of the `MaxResults` parameter # specified in the request and the service's internal maximum number of # items per page. # # @option params [String] :next_token # Opaque pagination token returned from a previous `DescribeBackups` # operation (String). If a token present, the action continues the list # from where the returning call left off. # # @return [Types::DescribeBackupsResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods: # # * {Types::DescribeBackupsResponse#backups #backups} => Array<Types::Backup> # * {Types::DescribeBackupsResponse#next_token #next_token} => String # # The returned {Seahorse::Client::Response response} is a pageable response and is Enumerable. For details on usage see {Aws::PageableResponse PageableResponse}. # # # @example Example: To describe Amazon FSx backups # # # This operation describes all of the Amazon FSx backups in an account. # # resp = client.describe_backups({ # }) # # resp.to_h outputs the following: # { # backups: [ # { # backup_id: "backup-03e3c82e0183b7b6b", # creation_time: Time.parse("1481841524.0"), # file_system: { # file_system_id: "fs-0498eed5fe91001ec", # owner_id: "012345678912", # storage_capacity: 300, # windows_configuration: { # active_directory_id: "d-1234abcd12", # automatic_backup_retention_days: 30, # daily_automatic_backup_start_time: "05:00", # weekly_maintenance_start_time: "1:05:00", # }, # }, # lifecycle: "AVAILABLE", # resource_arn: "arn:aws:fsx:us-east-1:012345678912:backup/backup-03e3c82e0183b7b6b", # tags: [ # { # key: "Name", # value: "MyBackup", # }, # ], # type: "USER_INITIATED", # }, # ], # } # # @example Request syntax with placeholder values # # resp = client.describe_backups({ # backup_ids: ["BackupId"], # filters: [ # { # name: "file-system-id", # accepts file-system-id, backup-type, file-system-type # values: ["FilterValue"], # }, # ], # max_results: 1, # next_token: "NextToken", # }) # # @example Response structure # # resp.backups #=> Array # resp.backups[0].backup_id #=> String # resp.backups[0].lifecycle #=> String, one of "AVAILABLE", "CREATING", "DELETED", "FAILED" # resp.backups[0].failure_details.message #=> String # resp.backups[0].type #=> String, one of "AUTOMATIC", "USER_INITIATED" # resp.backups[0].progress_percent #=> Integer # resp.backups[0].creation_time #=> Time # resp.backups[0].kms_key_id #=> String # resp.backups[0].resource_arn #=> String # resp.backups[0].tags #=> Array # resp.backups[0].tags[0].key #=> String # resp.backups[0].tags[0].value #=> String # resp.backups[0].file_system.owner_id #=> String # resp.backups[0].file_system.creation_time #=> Time # resp.backups[0].file_system.file_system_id #=> String # resp.backups[0].file_system.file_system_type #=> String, one of "WINDOWS", "LUSTRE" # resp.backups[0].file_system.lifecycle #=> String, one of "AVAILABLE", "CREATING", "FAILED", "DELETING", "MISCONFIGURED", "UPDATING" # resp.backups[0].file_system.failure_details.message #=> String # resp.backups[0].file_system.storage_capacity #=> Integer # resp.backups[0].file_system.storage_type #=> String, one of "SSD", "HDD" # resp.backups[0].file_system.vpc_id #=> String # resp.backups[0].file_system.subnet_ids #=> Array # resp.backups[0].file_system.subnet_ids[0] #=> String # resp.backups[0].file_system.network_interface_ids #=> Array # resp.backups[0].file_system.network_interface_ids[0] #=> String # resp.backups[0].file_system.dns_name #=> String # resp.backups[0].file_system.kms_key_id #=> String # resp.backups[0].file_system.resource_arn #=> String # resp.backups[0].file_system.tags #=> Array # resp.backups[0].file_system.tags[0].key #=> String # resp.backups[0].file_system.tags[0].value #=> String # resp.backups[0].file_system.windows_configuration.active_directory_id #=> String # resp.backups[0].file_system.windows_configuration.self_managed_active_directory_configuration.domain_name #=> String # resp.backups[0].file_system.windows_configuration.self_managed_active_directory_configuration.organizational_unit_distinguished_name #=> String # resp.backups[0].file_system.windows_configuration.self_managed_active_directory_configuration.file_system_administrators_group #=> String # resp.backups[0].file_system.windows_configuration.self_managed_active_directory_configuration.user_name #=> String # resp.backups[0].file_system.windows_configuration.self_managed_active_directory_configuration.dns_ips #=> Array # resp.backups[0].file_system.windows_configuration.self_managed_active_directory_configuration.dns_ips[0] #=> String # resp.backups[0].file_system.windows_configuration.deployment_type #=> String, one of "MULTI_AZ_1", "SINGLE_AZ_1", "SINGLE_AZ_2" # resp.backups[0].file_system.windows_configuration.remote_administration_endpoint #=> String # resp.backups[0].file_system.windows_configuration.preferred_subnet_id #=> String # resp.backups[0].file_system.windows_configuration.preferred_file_server_ip #=> String # resp.backups[0].file_system.windows_configuration.throughput_capacity #=> Integer # resp.backups[0].file_system.windows_configuration.maintenance_operations_in_progress #=> Array # resp.backups[0].file_system.windows_configuration.maintenance_operations_in_progress[0] #=> String, one of "PATCHING", "BACKING_UP" # resp.backups[0].file_system.windows_configuration.weekly_maintenance_start_time #=> String # resp.backups[0].file_system.windows_configuration.daily_automatic_backup_start_time #=> String # resp.backups[0].file_system.windows_configuration.automatic_backup_retention_days #=> Integer # resp.backups[0].file_system.windows_configuration.copy_tags_to_backups #=> Boolean # resp.backups[0].file_system.lustre_configuration.weekly_maintenance_start_time #=> String # resp.backups[0].file_system.lustre_configuration.data_repository_configuration.import_path #=> String # resp.backups[0].file_system.lustre_configuration.data_repository_configuration.export_path #=> String # resp.backups[0].file_system.lustre_configuration.data_repository_configuration.imported_file_chunk_size #=> Integer # resp.backups[0].file_system.lustre_configuration.deployment_type #=> String, one of "SCRATCH_1", "SCRATCH_2", "PERSISTENT_1" # resp.backups[0].file_system.lustre_configuration.per_unit_storage_throughput #=> Integer # resp.backups[0].file_system.lustre_configuration.mount_name #=> String # resp.backups[0].file_system.lustre_configuration.daily_automatic_backup_start_time #=> String # resp.backups[0].file_system.lustre_configuration.automatic_backup_retention_days #=> Integer # resp.backups[0].file_system.lustre_configuration.copy_tags_to_backups #=> Boolean # resp.backups[0].file_system.administrative_actions #=> Array # resp.backups[0].file_system.administrative_actions[0].administrative_action_type #=> String, one of "FILE_SYSTEM_UPDATE", "STORAGE_OPTIMIZATION" # resp.backups[0].file_system.administrative_actions[0].progress_percent #=> Integer # resp.backups[0].file_system.administrative_actions[0].request_time #=> Time # resp.backups[0].file_system.administrative_actions[0].status #=> String, one of "FAILED", "IN_PROGRESS", "PENDING", "COMPLETED", "UPDATED_OPTIMIZING" # resp.backups[0].file_system.administrative_actions[0].target_file_system_values #=> Types::FileSystem # resp.backups[0].file_system.administrative_actions[0].failure_details.message #=> String # resp.backups[0].directory_information.domain_name #=> String # resp.backups[0].directory_information.active_directory_id #=> String # resp.next_token #=> String # # @see http://docs.aws.amazon.com/goto/WebAPI/fsx-2018-03-01/DescribeBackups AWS API Documentation # # @overload describe_backups(params = {}) # @param [Hash] params ({}) def describe_backups(params = {}, options = {}) req = build_request(:describe_backups, params) req.send_request(options) end # Returns the description of specific Amazon FSx for Lustre data # repository tasks, if one or more `TaskIds` values are provided in the # request, or if filters are used in the request. You can use filters to # narrow the response to include just tasks for specific file systems, # or tasks in a specific lifecycle state. Otherwise, it returns all data # repository tasks owned by your AWS account in the AWS Region of the # endpoint that you're calling. # # When retrieving all tasks, you can paginate the response by using the # optional `MaxResults` parameter to limit the number of tasks returned # in a response. If more tasks remain, Amazon FSx returns a `NextToken` # value in the response. In this case, send a later request with the # `NextToken` request parameter set to the value of `NextToken` from the # last response. # # @option params [Array] :task_ids # (Optional) IDs of the tasks whose descriptions you want to retrieve # (String). # # @option params [Array] :filters # (Optional) You can use filters to narrow the # `DescribeDataRepositoryTasks` response to include just tasks for # specific file systems, or tasks in a specific lifecycle state. # # @option params [Integer] :max_results # The maximum number of resources to return in the response. This value # must be an integer greater than zero. # # @option params [String] :next_token # (Optional) Opaque pagination token returned from a previous operation # (String). If present, this token indicates from what point you can # continue processing the request, where the previous `NextToken` value # left off. # # @return [Types::DescribeDataRepositoryTasksResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods: # # * {Types::DescribeDataRepositoryTasksResponse#data_repository_tasks #data_repository_tasks} => Array<Types::DataRepositoryTask> # * {Types::DescribeDataRepositoryTasksResponse#next_token #next_token} => String # # The returned {Seahorse::Client::Response response} is a pageable response and is Enumerable. For details on usage see {Aws::PageableResponse PageableResponse}. # # @example Request syntax with placeholder values # # resp = client.describe_data_repository_tasks({ # task_ids: ["TaskId"], # filters: [ # { # name: "file-system-id", # accepts file-system-id, task-lifecycle # values: ["DataRepositoryTaskFilterValue"], # }, # ], # max_results: 1, # next_token: "NextToken", # }) # # @example Response structure # # resp.data_repository_tasks #=> Array # resp.data_repository_tasks[0].task_id #=> String # resp.data_repository_tasks[0].lifecycle #=> String, one of "PENDING", "EXECUTING", "FAILED", "SUCCEEDED", "CANCELED", "CANCELING" # resp.data_repository_tasks[0].type #=> String, one of "EXPORT_TO_REPOSITORY" # resp.data_repository_tasks[0].creation_time #=> Time # resp.data_repository_tasks[0].start_time #=> Time # resp.data_repository_tasks[0].end_time #=> Time # resp.data_repository_tasks[0].resource_arn #=> String # resp.data_repository_tasks[0].tags #=> Array # resp.data_repository_tasks[0].tags[0].key #=> String # resp.data_repository_tasks[0].tags[0].value #=> String # resp.data_repository_tasks[0].file_system_id #=> String # resp.data_repository_tasks[0].paths #=> Array # resp.data_repository_tasks[0].paths[0] #=> String # resp.data_repository_tasks[0].failure_details.message #=> String # resp.data_repository_tasks[0].status.total_count #=> Integer # resp.data_repository_tasks[0].status.succeeded_count #=> Integer # resp.data_repository_tasks[0].status.failed_count #=> Integer # resp.data_repository_tasks[0].status.last_updated_time #=> Time # resp.data_repository_tasks[0].report.enabled #=> Boolean # resp.data_repository_tasks[0].report.path #=> String # resp.data_repository_tasks[0].report.format #=> String, one of "REPORT_CSV_20191124" # resp.data_repository_tasks[0].report.scope #=> String, one of "FAILED_FILES_ONLY" # resp.next_token #=> String # # @see http://docs.aws.amazon.com/goto/WebAPI/fsx-2018-03-01/DescribeDataRepositoryTasks AWS API Documentation # # @overload describe_data_repository_tasks(params = {}) # @param [Hash] params ({}) def describe_data_repository_tasks(params = {}, options = {}) req = build_request(:describe_data_repository_tasks, params) req.send_request(options) end # Returns the description of specific Amazon FSx file systems, if a # `FileSystemIds` value is provided for that file system. Otherwise, it # returns descriptions of all file systems owned by your AWS account in # the AWS Region of the endpoint that you're calling. # # When retrieving all file system descriptions, you can optionally # specify the `MaxResults` parameter to limit the number of descriptions # in a response. If more file system descriptions remain, Amazon FSx # returns a `NextToken` value in the response. In this case, send a # later request with the `NextToken` request parameter set to the value # of `NextToken` from the last response. # # This action is used in an iterative process to retrieve a list of your # file system descriptions. `DescribeFileSystems` is called first # without a `NextToken`value. Then the action continues to be called # with the `NextToken` parameter set to the value of the last # `NextToken` value until a response has no `NextToken`. # # When using this action, keep the following in mind: # # * The implementation might return fewer than `MaxResults` file system # descriptions while still including a `NextToken` value. # # * The order of file systems returned in the response of one # `DescribeFileSystems` call and the order of file systems returned # across the responses of a multicall iteration is unspecified. # # @option params [Array] :file_system_ids # IDs of the file systems whose descriptions you want to retrieve # (String). # # @option params [Integer] :max_results # Maximum number of file systems to return in the response (integer). # This parameter value must be greater than 0. The number of items that # Amazon FSx returns is the minimum of the `MaxResults` parameter # specified in the request and the service's internal maximum number of # items per page. # # @option params [String] :next_token # Opaque pagination token returned from a previous `DescribeFileSystems` # operation (String). If a token present, the action continues the list # from where the returning call left off. # # @return [Types::DescribeFileSystemsResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods: # # * {Types::DescribeFileSystemsResponse#file_systems #file_systems} => Array<Types::FileSystem> # * {Types::DescribeFileSystemsResponse#next_token #next_token} => String # # The returned {Seahorse::Client::Response response} is a pageable response and is Enumerable. For details on usage see {Aws::PageableResponse PageableResponse}. # # # @example Example: To describe an Amazon FSx file system # # # This operation describes all of the Amazon FSx file systems in an account. # # resp = client.describe_file_systems({ # }) # # resp.to_h outputs the following: # { # file_systems: [ # { # creation_time: Time.parse("1481841524.0"), # dns_name: "fs-0498eed5fe91001ec.fsx.com", # file_system_id: "fs-0498eed5fe91001ec", # kms_key_id: "arn:aws:kms:us-east-1:012345678912:key/0ff3ea8d-130e-4133-877f-93908b6fdbd6", # lifecycle: "AVAILABLE", # network_interface_ids: [ # "eni-abcd1234", # ], # owner_id: "012345678912", # resource_arn: "arn:aws:fsx:us-east-1:012345678912:file-system/fs-0498eed5fe91001ec", # storage_capacity: 300, # subnet_ids: [ # "subnet-1234abcd", # ], # tags: [ # { # key: "Name", # value: "MyFileSystem", # }, # ], # vpc_id: "vpc-ab1234cd", # windows_configuration: { # active_directory_id: "d-1234abcd12", # automatic_backup_retention_days: 30, # daily_automatic_backup_start_time: "05:00", # throughput_capacity: 8, # weekly_maintenance_start_time: "1:05:00", # }, # }, # ], # } # # @example Request syntax with placeholder values # # resp = client.describe_file_systems({ # file_system_ids: ["FileSystemId"], # max_results: 1, # next_token: "NextToken", # }) # # @example Response structure # # resp.file_systems #=> Array # resp.file_systems[0].owner_id #=> String # resp.file_systems[0].creation_time #=> Time # resp.file_systems[0].file_system_id #=> String # resp.file_systems[0].file_system_type #=> String, one of "WINDOWS", "LUSTRE" # resp.file_systems[0].lifecycle #=> String, one of "AVAILABLE", "CREATING", "FAILED", "DELETING", "MISCONFIGURED", "UPDATING" # resp.file_systems[0].failure_details.message #=> String # resp.file_systems[0].storage_capacity #=> Integer # resp.file_systems[0].storage_type #=> String, one of "SSD", "HDD" # resp.file_systems[0].vpc_id #=> String # resp.file_systems[0].subnet_ids #=> Array # resp.file_systems[0].subnet_ids[0] #=> String # resp.file_systems[0].network_interface_ids #=> Array # resp.file_systems[0].network_interface_ids[0] #=> String # resp.file_systems[0].dns_name #=> String # resp.file_systems[0].kms_key_id #=> String # resp.file_systems[0].resource_arn #=> String # resp.file_systems[0].tags #=> Array # resp.file_systems[0].tags[0].key #=> String # resp.file_systems[0].tags[0].value #=> String # resp.file_systems[0].windows_configuration.active_directory_id #=> String # resp.file_systems[0].windows_configuration.self_managed_active_directory_configuration.domain_name #=> String # resp.file_systems[0].windows_configuration.self_managed_active_directory_configuration.organizational_unit_distinguished_name #=> String # resp.file_systems[0].windows_configuration.self_managed_active_directory_configuration.file_system_administrators_group #=> String # resp.file_systems[0].windows_configuration.self_managed_active_directory_configuration.user_name #=> String # resp.file_systems[0].windows_configuration.self_managed_active_directory_configuration.dns_ips #=> Array # resp.file_systems[0].windows_configuration.self_managed_active_directory_configuration.dns_ips[0] #=> String # resp.file_systems[0].windows_configuration.deployment_type #=> String, one of "MULTI_AZ_1", "SINGLE_AZ_1", "SINGLE_AZ_2" # resp.file_systems[0].windows_configuration.remote_administration_endpoint #=> String # resp.file_systems[0].windows_configuration.preferred_subnet_id #=> String # resp.file_systems[0].windows_configuration.preferred_file_server_ip #=> String # resp.file_systems[0].windows_configuration.throughput_capacity #=> Integer # resp.file_systems[0].windows_configuration.maintenance_operations_in_progress #=> Array # resp.file_systems[0].windows_configuration.maintenance_operations_in_progress[0] #=> String, one of "PATCHING", "BACKING_UP" # resp.file_systems[0].windows_configuration.weekly_maintenance_start_time #=> String # resp.file_systems[0].windows_configuration.daily_automatic_backup_start_time #=> String # resp.file_systems[0].windows_configuration.automatic_backup_retention_days #=> Integer # resp.file_systems[0].windows_configuration.copy_tags_to_backups #=> Boolean # resp.file_systems[0].lustre_configuration.weekly_maintenance_start_time #=> String # resp.file_systems[0].lustre_configuration.data_repository_configuration.import_path #=> String # resp.file_systems[0].lustre_configuration.data_repository_configuration.export_path #=> String # resp.file_systems[0].lustre_configuration.data_repository_configuration.imported_file_chunk_size #=> Integer # resp.file_systems[0].lustre_configuration.deployment_type #=> String, one of "SCRATCH_1", "SCRATCH_2", "PERSISTENT_1" # resp.file_systems[0].lustre_configuration.per_unit_storage_throughput #=> Integer # resp.file_systems[0].lustre_configuration.mount_name #=> String # resp.file_systems[0].lustre_configuration.daily_automatic_backup_start_time #=> String # resp.file_systems[0].lustre_configuration.automatic_backup_retention_days #=> Integer # resp.file_systems[0].lustre_configuration.copy_tags_to_backups #=> Boolean # resp.file_systems[0].administrative_actions #=> Array # resp.file_systems[0].administrative_actions[0].administrative_action_type #=> String, one of "FILE_SYSTEM_UPDATE", "STORAGE_OPTIMIZATION" # resp.file_systems[0].administrative_actions[0].progress_percent #=> Integer # resp.file_systems[0].administrative_actions[0].request_time #=> Time # resp.file_systems[0].administrative_actions[0].status #=> String, one of "FAILED", "IN_PROGRESS", "PENDING", "COMPLETED", "UPDATED_OPTIMIZING" # resp.file_systems[0].administrative_actions[0].target_file_system_values #=> Types::FileSystem # resp.file_systems[0].administrative_actions[0].failure_details.message #=> String # resp.next_token #=> String # # @see http://docs.aws.amazon.com/goto/WebAPI/fsx-2018-03-01/DescribeFileSystems AWS API Documentation # # @overload describe_file_systems(params = {}) # @param [Hash] params ({}) def describe_file_systems(params = {}, options = {}) req = build_request(:describe_file_systems, params) req.send_request(options) end # Lists tags for an Amazon FSx file systems and backups in the case of # Amazon FSx for Windows File Server. # # When retrieving all tags, you can optionally specify the `MaxResults` # parameter to limit the number of tags in a response. If more tags # remain, Amazon FSx returns a `NextToken` value in the response. In # this case, send a later request with the `NextToken` request parameter # set to the value of `NextToken` from the last response. # # This action is used in an iterative process to retrieve a list of your # tags. `ListTagsForResource` is called first without a # `NextToken`value. Then the action continues to be called with the # `NextToken` parameter set to the value of the last `NextToken` value # until a response has no `NextToken`. # # When using this action, keep the following in mind: # # * The implementation might return fewer than `MaxResults` file system # descriptions while still including a `NextToken` value. # # * The order of tags returned in the response of one # `ListTagsForResource` call and the order of tags returned across the # responses of a multi-call iteration is unspecified. # # @option params [required, String] :resource_arn # The ARN of the Amazon FSx resource that will have its tags listed. # # @option params [Integer] :max_results # Maximum number of tags to return in the response (integer). This # parameter value must be greater than 0. The number of items that # Amazon FSx returns is the minimum of the `MaxResults` parameter # specified in the request and the service's internal maximum number of # items per page. # # @option params [String] :next_token # Opaque pagination token returned from a previous `ListTagsForResource` # operation (String). If a token present, the action continues the list # from where the returning call left off. # # @return [Types::ListTagsForResourceResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods: # # * {Types::ListTagsForResourceResponse#tags #tags} => Array<Types::Tag> # * {Types::ListTagsForResourceResponse#next_token #next_token} => String # # # @example Example: To list tags for a resource # # # This operation lists tags for an Amazon FSx resource. # # resp = client.list_tags_for_resource({ # resource_arn: "arn:aws:fsx:us-east-1:012345678912:file-system/fs-0498eed5fe91001ec", # }) # # resp.to_h outputs the following: # { # tags: [ # { # key: "Name", # value: "MyFileSystem", # }, # ], # } # # @example Request syntax with placeholder values # # resp = client.list_tags_for_resource({ # resource_arn: "ResourceARN", # required # max_results: 1, # next_token: "NextToken", # }) # # @example Response structure # # resp.tags #=> Array # resp.tags[0].key #=> String # resp.tags[0].value #=> String # resp.next_token #=> String # # @see http://docs.aws.amazon.com/goto/WebAPI/fsx-2018-03-01/ListTagsForResource AWS API Documentation # # @overload list_tags_for_resource(params = {}) # @param [Hash] params ({}) def list_tags_for_resource(params = {}, options = {}) req = build_request(:list_tags_for_resource, params) req.send_request(options) end # Tags an Amazon FSx resource. # # @option params [required, String] :resource_arn # The Amazon Resource Name (ARN) of the Amazon FSx resource that you # want to tag. # # @option params [required, Array] :tags # A list of tags for the resource. If a tag with a given key already # exists, the value is replaced by the one specified in this parameter. # # @return [Struct] Returns an empty {Seahorse::Client::Response response}. # # # @example Example: To tag a resource # # # This operation tags an Amazon FSx resource. # # resp = client.tag_resource({ # resource_arn: "arn:aws:fsx:us-east-1:012345678912:file-system/fs-0498eed5fe91001ec", # tags: [ # { # key: "Name", # value: "MyFileSystem", # }, # ], # }) # # @example Request syntax with placeholder values # # resp = client.tag_resource({ # resource_arn: "ResourceARN", # required # tags: [ # required # { # key: "TagKey", # value: "TagValue", # }, # ], # }) # # @see http://docs.aws.amazon.com/goto/WebAPI/fsx-2018-03-01/TagResource AWS API Documentation # # @overload tag_resource(params = {}) # @param [Hash] params ({}) def tag_resource(params = {}, options = {}) req = build_request(:tag_resource, params) req.send_request(options) end # This action removes a tag from an Amazon FSx resource. # # @option params [required, String] :resource_arn # The ARN of the Amazon FSx resource to untag. # # @option params [required, Array] :tag_keys # A list of keys of tags on the resource to untag. In case the tag key # doesn't exist, the call will still succeed to be idempotent. # # @return [Struct] Returns an empty {Seahorse::Client::Response response}. # # # @example Example: To untag a resource # # # This operation untags an Amazon FSx resource. # # resp = client.untag_resource({ # resource_arn: "arn:aws:fsx:us-east-1:012345678912:file-system/fs-0498eed5fe91001ec", # tag_keys: [ # "Name", # ], # }) # # @example Request syntax with placeholder values # # resp = client.untag_resource({ # resource_arn: "ResourceARN", # required # tag_keys: ["TagKey"], # required # }) # # @see http://docs.aws.amazon.com/goto/WebAPI/fsx-2018-03-01/UntagResource AWS API Documentation # # @overload untag_resource(params = {}) # @param [Hash] params ({}) def untag_resource(params = {}, options = {}) req = build_request(:untag_resource, params) req.send_request(options) end # Use this operation to update the configuration of an existing Amazon # FSx file system. For an Amazon FSx for Lustre file system, you can # update only the WeeklyMaintenanceStartTime. For an Amazon for Windows # File Server file system, you can update the following properties: # # * AutomaticBackupRetentionDays # # * DailyAutomaticBackupStartTime # # * SelfManagedActiveDirectoryConfiguration # # * StorageCapacity # # * ThroughputCapacity # # * WeeklyMaintenanceStartTime # # You can update multiple properties in a single request. # # @option params [required, String] :file_system_id # Identifies the file system that you are updating. # # @option params [String] :client_request_token # A string of up to 64 ASCII characters that Amazon FSx uses to ensure # idempotent updates. This string is automatically filled on your behalf # when you use the AWS Command Line Interface (AWS CLI) or an AWS SDK. # # **A suitable default value is auto-generated.** You should normally # not need to pass this option.** # # @option params [Integer] :storage_capacity # Use this parameter to increase the storage capacity of an Amazon FSx # for Windows File Server file system. Specifies the storage capacity # target value, GiB, for the file system you're updating. The storage # capacity target value must be at least 10 percent (%) greater than the # current storage capacity value. In order to increase storage capacity, # the file system needs to have at least 16 MB/s of throughput capacity. # You cannot make a storage capacity increase request if there is an # existing storage capacity increase request in progress. For more # information, see [Managing Storage Capacity][1]. # # # # [1]: https://docs.aws.amazon.com/fsx/latest/WindowsGuide/managing-storage-capacity.html # # @option params [Types::UpdateFileSystemWindowsConfiguration] :windows_configuration # The configuration updates for an Amazon FSx for Windows File Server # file system. # # @option params [Types::UpdateFileSystemLustreConfiguration] :lustre_configuration # The configuration object for Amazon FSx for Lustre file systems used # in the `UpdateFileSystem` operation. # # @return [Types::UpdateFileSystemResponse] Returns a {Seahorse::Client::Response response} object which responds to the following methods: # # * {Types::UpdateFileSystemResponse#file_system #file_system} => Types::FileSystem # # # @example Example: To update an existing file system # # # This operation updates an existing file system. # # resp = client.update_file_system({ # file_system_id: "fs-0498eed5fe91001ec", # windows_configuration: { # automatic_backup_retention_days: 10, # daily_automatic_backup_start_time: "06:00", # weekly_maintenance_start_time: "3:06:00", # }, # }) # # resp.to_h outputs the following: # { # file_system: { # creation_time: Time.parse("1481841524.0"), # dns_name: "fs-0498eed5fe91001ec.fsx.com", # file_system_id: "fs-0498eed5fe91001ec", # kms_key_id: "arn:aws:kms:us-east-1:012345678912:key/0ff3ea8d-130e-4133-877f-93908b6fdbd6", # lifecycle: "AVAILABLE", # owner_id: "012345678912", # resource_arn: "arn:aws:fsx:us-east-1:012345678912:file-system/fs-0498eed5fe91001ec", # storage_capacity: 300, # subnet_ids: [ # "subnet-1234abcd", # ], # tags: [ # { # key: "Name", # value: "MyFileSystem", # }, # ], # vpc_id: "vpc-ab1234cd", # windows_configuration: { # automatic_backup_retention_days: 10, # daily_automatic_backup_start_time: "06:00", # throughput_capacity: 8, # weekly_maintenance_start_time: "3:06:00", # }, # }, # } # # @example Request syntax with placeholder values # # resp = client.update_file_system({ # file_system_id: "FileSystemId", # required # client_request_token: "ClientRequestToken", # storage_capacity: 1, # windows_configuration: { # weekly_maintenance_start_time: "WeeklyTime", # daily_automatic_backup_start_time: "DailyTime", # automatic_backup_retention_days: 1, # throughput_capacity: 1, # self_managed_active_directory_configuration: { # user_name: "DirectoryUserName", # password: "DirectoryPassword", # dns_ips: ["IpAddress"], # }, # }, # lustre_configuration: { # weekly_maintenance_start_time: "WeeklyTime", # daily_automatic_backup_start_time: "DailyTime", # automatic_backup_retention_days: 1, # }, # }) # # @example Response structure # # resp.file_system.owner_id #=> String # resp.file_system.creation_time #=> Time # resp.file_system.file_system_id #=> String # resp.file_system.file_system_type #=> String, one of "WINDOWS", "LUSTRE" # resp.file_system.lifecycle #=> String, one of "AVAILABLE", "CREATING", "FAILED", "DELETING", "MISCONFIGURED", "UPDATING" # resp.file_system.failure_details.message #=> String # resp.file_system.storage_capacity #=> Integer # resp.file_system.storage_type #=> String, one of "SSD", "HDD" # resp.file_system.vpc_id #=> String # resp.file_system.subnet_ids #=> Array # resp.file_system.subnet_ids[0] #=> String # resp.file_system.network_interface_ids #=> Array # resp.file_system.network_interface_ids[0] #=> String # resp.file_system.dns_name #=> String # resp.file_system.kms_key_id #=> String # resp.file_system.resource_arn #=> String # resp.file_system.tags #=> Array # resp.file_system.tags[0].key #=> String # resp.file_system.tags[0].value #=> String # resp.file_system.windows_configuration.active_directory_id #=> String # resp.file_system.windows_configuration.self_managed_active_directory_configuration.domain_name #=> String # resp.file_system.windows_configuration.self_managed_active_directory_configuration.organizational_unit_distinguished_name #=> String # resp.file_system.windows_configuration.self_managed_active_directory_configuration.file_system_administrators_group #=> String # resp.file_system.windows_configuration.self_managed_active_directory_configuration.user_name #=> String # resp.file_system.windows_configuration.self_managed_active_directory_configuration.dns_ips #=> Array # resp.file_system.windows_configuration.self_managed_active_directory_configuration.dns_ips[0] #=> String # resp.file_system.windows_configuration.deployment_type #=> String, one of "MULTI_AZ_1", "SINGLE_AZ_1", "SINGLE_AZ_2" # resp.file_system.windows_configuration.remote_administration_endpoint #=> String # resp.file_system.windows_configuration.preferred_subnet_id #=> String # resp.file_system.windows_configuration.preferred_file_server_ip #=> String # resp.file_system.windows_configuration.throughput_capacity #=> Integer # resp.file_system.windows_configuration.maintenance_operations_in_progress #=> Array # resp.file_system.windows_configuration.maintenance_operations_in_progress[0] #=> String, one of "PATCHING", "BACKING_UP" # resp.file_system.windows_configuration.weekly_maintenance_start_time #=> String # resp.file_system.windows_configuration.daily_automatic_backup_start_time #=> String # resp.file_system.windows_configuration.automatic_backup_retention_days #=> Integer # resp.file_system.windows_configuration.copy_tags_to_backups #=> Boolean # resp.file_system.lustre_configuration.weekly_maintenance_start_time #=> String # resp.file_system.lustre_configuration.data_repository_configuration.import_path #=> String # resp.file_system.lustre_configuration.data_repository_configuration.export_path #=> String # resp.file_system.lustre_configuration.data_repository_configuration.imported_file_chunk_size #=> Integer # resp.file_system.lustre_configuration.deployment_type #=> String, one of "SCRATCH_1", "SCRATCH_2", "PERSISTENT_1" # resp.file_system.lustre_configuration.per_unit_storage_throughput #=> Integer # resp.file_system.lustre_configuration.mount_name #=> String # resp.file_system.lustre_configuration.daily_automatic_backup_start_time #=> String # resp.file_system.lustre_configuration.automatic_backup_retention_days #=> Integer # resp.file_system.lustre_configuration.copy_tags_to_backups #=> Boolean # resp.file_system.administrative_actions #=> Array # resp.file_system.administrative_actions[0].administrative_action_type #=> String, one of "FILE_SYSTEM_UPDATE", "STORAGE_OPTIMIZATION" # resp.file_system.administrative_actions[0].progress_percent #=> Integer # resp.file_system.administrative_actions[0].request_time #=> Time # resp.file_system.administrative_actions[0].status #=> String, one of "FAILED", "IN_PROGRESS", "PENDING", "COMPLETED", "UPDATED_OPTIMIZING" # resp.file_system.administrative_actions[0].target_file_system_values #=> Types::FileSystem # resp.file_system.administrative_actions[0].failure_details.message #=> String # # @see http://docs.aws.amazon.com/goto/WebAPI/fsx-2018-03-01/UpdateFileSystem AWS API Documentation # # @overload update_file_system(params = {}) # @param [Hash] params ({}) def update_file_system(params = {}, options = {}) req = build_request(:update_file_system, params) req.send_request(options) end # @!endgroup # @param params ({}) # @api private def build_request(operation_name, params = {}) handlers = @handlers.for(operation_name) context = Seahorse::Client::RequestContext.new( operation_name: operation_name, operation: config.api.operation(operation_name), client: self, params: params, config: config) context[:gem_name] = 'aws-sdk-fsx' context[:gem_version] = '1.23.0' Seahorse::Client::Request.new(handlers, context) end # @api private # @deprecated def waiter_names [] end class << self # @api private attr_reader :identifier # @api private def errors_module Errors end end end end