lib/google/apis/dataplex_v1/classes.rb in google-apis-dataplex_v1-0.19.0 vs lib/google/apis/dataplex_v1/classes.rb in google-apis-dataplex_v1-0.20.0

- old
+ new

@@ -469,12 +469,12 @@ # Optional. Cron schedule (https://en.wikipedia.org/wiki/Cron) for running # discovery periodically. Successive discovery runs must be scheduled at least # 60 minutes apart. The default value is to run discovery every 60 minutes. To # explicitly set a timezone to the cron tab, apply a prefix in the cron tab: " # CRON_TZ=$`IANA_TIME_ZONE`" or TZ=$`IANA_TIME_ZONE`". The $`IANA_TIME_ZONE` may - # only be a valid string from IANA time zone database. For example, "CRON_TZ= - # America/New_York 1 * * * *", or "TZ=America/New_York 1 * * * *". + # only be a valid string from IANA time zone database. For example, CRON_TZ= + # America/New_York 1 * * * *, or TZ=America/New_York 1 * * * *. # Corresponds to the JSON property `schedule` # @return [String] attr_accessor :schedule def initialize(**args) @@ -906,17 +906,16 @@ # DataProfileResult defines the output of DataProfileScan. Each field of the # table will have field type specific profile result. class GoogleCloudDataplexV1DataProfileResult include Google::Apis::Core::Hashable - # Profile information describing the structure and layout of the data and - # contains the profile info. + # Contains name, type, mode and field type specific profile information. # Corresponds to the JSON property `profile` # @return [Google::Apis::DataplexV1::GoogleCloudDataplexV1DataProfileResultProfile] attr_accessor :profile - # The count of all rows in the sampled data. Return 0, if zero rows. + # The count of rows scanned. # Corresponds to the JSON property `rowCount` # @return [Fixnum] attr_accessor :row_count # The data scanned during processing (e.g. in incremental DataScan) @@ -934,16 +933,15 @@ @row_count = args[:row_count] if args.key?(:row_count) @scanned_data = args[:scanned_data] if args.key?(:scanned_data) end end - # Profile information describing the structure and layout of the data and - # contains the profile info. + # Contains name, type, mode and field type specific profile information. class GoogleCloudDataplexV1DataProfileResultProfile include Google::Apis::Core::Hashable - # The sequence of fields describing data in table entities. + # List of fields with structural and profile information for each field. # Corresponds to the JSON property `fields` # @return [Array<Google::Apis::DataplexV1::GoogleCloudDataplexV1DataProfileResultProfileField>] attr_accessor :fields def initialize(**args) @@ -954,26 +952,27 @@ def update!(**args) @fields = args[:fields] if args.key?(:fields) end end - # Represents a column field within a table schema. + # A field within a table. class GoogleCloudDataplexV1DataProfileResultProfileField include Google::Apis::Core::Hashable - # The mode of the field. Its value will be: REQUIRED, if it is a required field. - # NULLABLE, if it is an optional field. REPEATED, if it is a repeated field. + # The mode of the field. Possible values include: REQUIRED, if it is a required + # field. NULLABLE, if it is an optional field. REPEATED, if it is a repeated + # field. # Corresponds to the JSON property `mode` # @return [String] attr_accessor :mode # The name of the field. # Corresponds to the JSON property `name` # @return [String] attr_accessor :name - # ProfileInfo defines the profile information for each schema field type. + # The profile information for each field type. # Corresponds to the JSON property `profile` # @return [Google::Apis::DataplexV1::GoogleCloudDataplexV1DataProfileResultProfileFieldProfileInfo] attr_accessor :profile # The field data type. Possible values include: STRING BYTE INT64 INT32 INT16 @@ -993,43 +992,44 @@ @profile = args[:profile] if args.key?(:profile) @type = args[:type] if args.key?(:type) end end - # ProfileInfo defines the profile information for each schema field type. + # The profile information for each field type. class GoogleCloudDataplexV1DataProfileResultProfileFieldProfileInfo include Google::Apis::Core::Hashable - # The ratio of rows that are distinct against the rows in the sampled data. + # Ratio of rows with distinct values against total scanned rows. Not available + # for complex non-groupable field type RECORD and fields with REPEATABLE mode. # Corresponds to the JSON property `distinctRatio` # @return [Float] attr_accessor :distinct_ratio - # DoubleFieldInfo defines output for any double type field. + # The profile information for a double type field. # Corresponds to the JSON property `doubleProfile` # @return [Google::Apis::DataplexV1::GoogleCloudDataplexV1DataProfileResultProfileFieldProfileInfoDoubleFieldInfo] attr_accessor :double_profile - # IntegerFieldInfo defines output for any integer type field. + # The profile information for an integer type field. # Corresponds to the JSON property `integerProfile` # @return [Google::Apis::DataplexV1::GoogleCloudDataplexV1DataProfileResultProfileFieldProfileInfoIntegerFieldInfo] attr_accessor :integer_profile - # The ratio of null rows against the rows in the sampled data. + # Ratio of rows with null value against total scanned rows. # Corresponds to the JSON property `nullRatio` # @return [Float] attr_accessor :null_ratio - # StringFieldInfo defines output info for any string type field. + # The profile information for a string type field. # Corresponds to the JSON property `stringProfile` # @return [Google::Apis::DataplexV1::GoogleCloudDataplexV1DataProfileResultProfileFieldProfileInfoStringFieldInfo] attr_accessor :string_profile - # The array of top N values of the field in the sampled data. Currently N is set - # as 10 or equal to distinct values in the field, whichever is smaller. This - # will be optional for complex non-groupable data-types such as JSON, ARRAY, - # JSON, STRUCT. + # The list of top N non-null values and number of times they occur in the + # scanned data. N is 10 or equal to the number of distinct values in the field, + # whichever is smaller. Not available for complex non-groupable field type + # RECORD and fields with REPEATABLE mode. # Corresponds to the JSON property `topNValues` # @return [Array<Google::Apis::DataplexV1::GoogleCloudDataplexV1DataProfileResultProfileFieldProfileInfoTopNValue>] attr_accessor :top_n_values def initialize(**args) @@ -1045,47 +1045,45 @@ @string_profile = args[:string_profile] if args.key?(:string_profile) @top_n_values = args[:top_n_values] if args.key?(:top_n_values) end end - # DoubleFieldInfo defines output for any double type field. + # The profile information for a double type field. class GoogleCloudDataplexV1DataProfileResultProfileFieldProfileInfoDoubleFieldInfo include Google::Apis::Core::Hashable - # The average of non-null values of double field in the sampled data. Return NaN, - # if the field has a NaN. Optional if zero non-null rows. + # Average of non-null values in the scanned data. NaN, if the field has a NaN. # Corresponds to the JSON property `average` # @return [Float] attr_accessor :average - # The maximum value of a double field in the sampled data. Return NaN, if the - # field has a NaN. Optional if zero non-null rows. + # Maximum of non-null values in the scanned data. NaN, if the field has a NaN. # Corresponds to the JSON property `max` # @return [Float] attr_accessor :max - # The minimum value of a double field in the sampled data. Return NaN, if the - # field has a NaN. Optional if zero non-null rows. + # Minimum of non-null values in the scanned data. NaN, if the field has a NaN. # Corresponds to the JSON property `min` # @return [Float] attr_accessor :min - # A quartile divide the numebr of data points into four parts, or quarters, of + # A quartile divides the number of data points into four parts, or quarters, of # more-or-less equal size. Three main quartiles used are: The first quartile (Q1) # splits off the lowest 25% of data from the highest 75%. It is also known as # the lower or 25th empirical quartile, as 25% of the data is below this point. # The second quartile (Q2) is the median of a data set. So, 50% of the data lies # below this point. The third quartile (Q3) splits off the highest 25% of data # from the lowest 75%. It is known as the upper or 75th empirical quartile, as - # 75% of the data lies below this point. So, here the quartiles is provided as - # an ordered list of quartile values, occurring in order Q1, median, Q3. + # 75% of the data lies below this point. Here, the quartiles is provided as an + # ordered list of quartile values for the scanned data, occurring in order Q1, + # median, Q3. # Corresponds to the JSON property `quartiles` # @return [Array<Float>] attr_accessor :quartiles - # The standard deviation of non-null of double field in the sampled data. Return - # NaN, if the field has a NaN. Optional if zero non-null rows. + # Standard deviation of non-null values in the scanned data. NaN, if the field + # has a NaN. # Corresponds to the JSON property `standardDeviation` # @return [Float] attr_accessor :standard_deviation def initialize(**args) @@ -1100,47 +1098,45 @@ @quartiles = args[:quartiles] if args.key?(:quartiles) @standard_deviation = args[:standard_deviation] if args.key?(:standard_deviation) end end - # IntegerFieldInfo defines output for any integer type field. + # The profile information for an integer type field. class GoogleCloudDataplexV1DataProfileResultProfileFieldProfileInfoIntegerFieldInfo include Google::Apis::Core::Hashable - # The average of non-null values of integer field in the sampled data. Return - # NaN, if the field has a NaN. Optional if zero non-null rows. + # Average of non-null values in the scanned data. NaN, if the field has a NaN. # Corresponds to the JSON property `average` # @return [Float] attr_accessor :average - # The maximum value of an integer field in the sampled data. Return NaN, if the - # field has a NaN. Optional if zero non-null rows. + # Maximum of non-null values in the scanned data. NaN, if the field has a NaN. # Corresponds to the JSON property `max` # @return [Fixnum] attr_accessor :max - # The minimum value of an integer field in the sampled data. Return NaN, if the - # field has a NaN. Optional if zero non-null rows. + # Minimum of non-null values in the scanned data. NaN, if the field has a NaN. # Corresponds to the JSON property `min` # @return [Fixnum] attr_accessor :min - # A quartile divide the number of data points into four parts, or quarters, of + # A quartile divides the number of data points into four parts, or quarters, of # more-or-less equal size. Three main quartiles used are: The first quartile (Q1) # splits off the lowest 25% of data from the highest 75%. It is also known as # the lower or 25th empirical quartile, as 25% of the data is below this point. # The second quartile (Q2) is the median of a data set. So, 50% of the data lies # below this point. The third quartile (Q3) splits off the highest 25% of data # from the lowest 75%. It is known as the upper or 75th empirical quartile, as - # 75% of the data lies below this point. So, here the quartiles is provided as - # an ordered list of quartile values, occurring in order Q1, median, Q3. + # 75% of the data lies below this point. Here, the quartiles is provided as an + # ordered list of quartile values for the scanned data, occurring in order Q1, + # median, Q3. # Corresponds to the JSON property `quartiles` # @return [Array<Fixnum>] attr_accessor :quartiles - # The standard deviation of non-null of integer field in the sampled data. - # Return NaN, if the field has a NaN. Optional if zero non-null rows. + # Standard deviation of non-null values in the scanned data. NaN, if the field + # has a NaN. # Corresponds to the JSON property `standardDeviation` # @return [Float] attr_accessor :standard_deviation def initialize(**args) @@ -1155,28 +1151,25 @@ @quartiles = args[:quartiles] if args.key?(:quartiles) @standard_deviation = args[:standard_deviation] if args.key?(:standard_deviation) end end - # StringFieldInfo defines output info for any string type field. + # The profile information for a string type field. class GoogleCloudDataplexV1DataProfileResultProfileFieldProfileInfoStringFieldInfo include Google::Apis::Core::Hashable - # The average length of a string field in the sampled data. Optional if zero non- - # null rows. + # Average length of non-null values in the scanned data. # Corresponds to the JSON property `averageLength` # @return [Float] attr_accessor :average_length - # The maximum length of a string field in the sampled data. Optional if zero non- - # null rows. + # Maximum length of non-null values in the scanned data. # Corresponds to the JSON property `maxLength` # @return [Fixnum] attr_accessor :max_length - # The minimum length of the string field in the sampled data. Optional if zero - # non-null rows. + # Minimum length of non-null values in the scanned data. # Corresponds to the JSON property `minLength` # @return [Fixnum] attr_accessor :min_length def initialize(**args) @@ -1189,20 +1182,20 @@ @max_length = args[:max_length] if args.key?(:max_length) @min_length = args[:min_length] if args.key?(:min_length) end end - # The TopNValue defines the structure of output of top N values of a field. + # Top N non-null values in the scanned data. class GoogleCloudDataplexV1DataProfileResultProfileFieldProfileInfoTopNValue include Google::Apis::Core::Hashable - # The frequency count of the corresponding value in the field. + # Count of the corresponding value in the scanned data. # Corresponds to the JSON property `count` # @return [Fixnum] attr_accessor :count - # The value is the string value of the actual value from the field. + # String value of a top N non-null value. # Corresponds to the JSON property `value` # @return [String] attr_accessor :value def initialize(**args) @@ -1227,12 +1220,12 @@ # Update properties of this object def update!(**args) end end - # DataQualityDimensionResult provides a more detailed, per-dimension level view - # of the results. + # DataQualityDimensionResult provides a more detailed, per-dimension view of the + # results. class GoogleCloudDataplexV1DataQualityDimensionResult include Google::Apis::Core::Hashable # Whether the dimension passed or failed. # Corresponds to the JSON property `passed` @@ -1252,11 +1245,11 @@ # The output of a DataQualityScan. class GoogleCloudDataplexV1DataQualityResult include Google::Apis::Core::Hashable - # A list of results at the dimension-level. + # A list of results at the dimension level. # Corresponds to the JSON property `dimensions` # @return [Array<Google::Apis::DataplexV1::GoogleCloudDataplexV1DataQualityDimensionResult>] attr_accessor :dimensions # Overall data quality result -- true if all rules passed. @@ -1302,19 +1295,19 @@ # Corresponds to the JSON property `column` # @return [String] attr_accessor :column # Required. The dimension a rule belongs to. Results are also aggregated at the - # dimension-level. Supported dimensions are "COMPLETENESS", "ACCURACY", " + # dimension level. Supported dimensions are "COMPLETENESS", "ACCURACY", " # CONSISTENCY", "VALIDITY", "UNIQUENESS", "INTEGRITY" # Corresponds to the JSON property `dimension` # @return [String] attr_accessor :dimension # Optional. Rows with null values will automatically fail a rule, unless # ignore_null is true. In that case, such null rows are trivially considered - # passing. Only applicable to ColumnMap rules. + # passing.Only applicable to ColumnMap rules. # Corresponds to the JSON property `ignoreNull` # @return [Boolean] attr_accessor :ignore_null alias_method :ignore_null?, :ignore_null @@ -1331,13 +1324,13 @@ # Evaluates whether each column value matches a specified regex. # Corresponds to the JSON property `regexExpectation` # @return [Google::Apis::DataplexV1::GoogleCloudDataplexV1DataQualityRuleRegexExpectation] attr_accessor :regex_expectation - # Evaluates whether each row passes the specified condition. The SQL expression - # needs to use BigQuery standard SQL syntax and should produce a boolean per row - # as the result. Example: col1 >= 0 AND col2 < 10 + # Evaluates whether each row passes the specified condition.The SQL expression + # needs to use BigQuery standard SQL syntax and should produce a boolean value + # per row as the result.Example: col1 >= 0 AND col2 < 10 # Corresponds to the JSON property `rowConditionExpectation` # @return [Google::Apis::DataplexV1::GoogleCloudDataplexV1DataQualityRuleRowConditionExpectation] attr_accessor :row_condition_expectation # Evaluates whether each column value is contained by a specified set. @@ -1349,19 +1342,19 @@ # range. # Corresponds to the JSON property `statisticRangeExpectation` # @return [Google::Apis::DataplexV1::GoogleCloudDataplexV1DataQualityRuleStatisticRangeExpectation] attr_accessor :statistic_range_expectation - # Evaluates whether the provided expression is true. The SQL expression needs to + # Evaluates whether the provided expression is true.The SQL expression needs to # use BigQuery standard SQL syntax and should produce a scalar boolean result. # Example: MIN(col1) >= 0 # Corresponds to the JSON property `tableConditionExpectation` # @return [Google::Apis::DataplexV1::GoogleCloudDataplexV1DataQualityRuleTableConditionExpectation] attr_accessor :table_condition_expectation # Optional. The minimum ratio of passing_rows / total_rows required to pass this - # rule, with a range of 0.0, 1.00 indicates default value (i.e. 1.0) + # rule, with a range of 0.0, 1.0.0 indicates default value (i.e. 1.0). # Corresponds to the JSON property `threshold` # @return [Float] attr_accessor :threshold # Evaluates whether the column has duplicates. @@ -1418,19 +1411,19 @@ # Corresponds to the JSON property `minValue` # @return [String] attr_accessor :min_value # Optional. Whether each value needs to be strictly lesser than ('<') the - # maximum, or if equality is allowed. Only relevant if a max_value has been + # maximum, or if equality is allowed.Only relevant if a max_value has been # defined. Default = false. # Corresponds to the JSON property `strictMaxEnabled` # @return [Boolean] attr_accessor :strict_max_enabled alias_method :strict_max_enabled?, :strict_max_enabled # Optional. Whether each value needs to be strictly greater than ('>') the - # minimum, or if equality is allowed. Only relevant if a min_value has been + # minimum, or if equality is allowed.Only relevant if a min_value has been # defined. Default = false. # Corresponds to the JSON property `strictMinEnabled` # @return [Boolean] attr_accessor :strict_min_enabled alias_method :strict_min_enabled?, :strict_min_enabled @@ -1450,11 +1443,11 @@ # Evaluates whether each column value matches a specified regex. class GoogleCloudDataplexV1DataQualityRuleRegexExpectation include Google::Apis::Core::Hashable - # + # A regular expression the column value is expected to match. # Corresponds to the JSON property `regex` # @return [String] attr_accessor :regex def initialize(**args) @@ -1465,19 +1458,18 @@ def update!(**args) @regex = args[:regex] if args.key?(:regex) end end - # DataQualityRuleResult provides a more detailed, per-rule level view of the - # results. + # DataQualityRuleResult provides a more detailed, per-rule view of the results. class GoogleCloudDataplexV1DataQualityRuleResult include Google::Apis::Core::Hashable # The number of rows a rule was evaluated against. This field is only valid for - # ColumnMap type rules. Evaluated count can be configured to either (1) include - # all rows (default) - with null rows automatically failing rule evaluation OR ( - # 2) exclude null rows from the evaluated_count, by setting ignore_nulls = true + # ColumnMap type rules.Evaluated count can be configured to either include all + # rows (default) - with null rows automatically failing rule evaluation, or + # exclude null rows from the evaluated_count, by setting ignore_nulls = true. # Corresponds to the JSON property `evaluatedCount` # @return [Fixnum] attr_accessor :evaluated_count # The query to find rows that did not pass this rule. Only applies to ColumnMap @@ -1528,17 +1520,17 @@ @passed_count = args[:passed_count] if args.key?(:passed_count) @rule = args[:rule] if args.key?(:rule) end end - # Evaluates whether each row passes the specified condition. The SQL expression - # needs to use BigQuery standard SQL syntax and should produce a boolean per row - # as the result. Example: col1 >= 0 AND col2 < 10 + # Evaluates whether each row passes the specified condition.The SQL expression + # needs to use BigQuery standard SQL syntax and should produce a boolean value + # per row as the result.Example: col1 >= 0 AND col2 < 10 class GoogleCloudDataplexV1DataQualityRuleRowConditionExpectation include Google::Apis::Core::Hashable - # + # The SQL expression. # Corresponds to the JSON property `sqlExpression` # @return [String] attr_accessor :sql_expression def initialize(**args) @@ -1553,11 +1545,11 @@ # Evaluates whether each column value is contained by a specified set. class GoogleCloudDataplexV1DataQualityRuleSetExpectation include Google::Apis::Core::Hashable - # + # Expected values for the column value. # Corresponds to the JSON property `values` # @return [Array<String>] attr_accessor :values def initialize(**args) @@ -1591,19 +1583,19 @@ # Corresponds to the JSON property `statistic` # @return [String] attr_accessor :statistic # Whether column statistic needs to be strictly lesser than ('<') the maximum, - # or if equality is allowed. Only relevant if a max_value has been defined. + # or if equality is allowed.Only relevant if a max_value has been defined. # Default = false. # Corresponds to the JSON property `strictMaxEnabled` # @return [Boolean] attr_accessor :strict_max_enabled alias_method :strict_max_enabled?, :strict_max_enabled # Whether column statistic needs to be strictly greater than ('>') the minimum, - # or if equality is allowed. Only relevant if a min_value has been defined. + # or if equality is allowed.Only relevant if a min_value has been defined. # Default = false. # Corresponds to the JSON property `strictMinEnabled` # @return [Boolean] attr_accessor :strict_min_enabled alias_method :strict_min_enabled?, :strict_min_enabled @@ -1620,17 +1612,17 @@ @strict_max_enabled = args[:strict_max_enabled] if args.key?(:strict_max_enabled) @strict_min_enabled = args[:strict_min_enabled] if args.key?(:strict_min_enabled) end end - # Evaluates whether the provided expression is true. The SQL expression needs to + # Evaluates whether the provided expression is true.The SQL expression needs to # use BigQuery standard SQL syntax and should produce a scalar boolean result. # Example: MIN(col1) >= 0 class GoogleCloudDataplexV1DataQualityRuleTableConditionExpectation include Google::Apis::Core::Hashable - # + # The SQL expression. # Corresponds to the JSON property `sqlExpression` # @return [String] attr_accessor :sql_expression def initialize(**args) @@ -1675,12 +1667,12 @@ @rules = args[:rules] if args.key?(:rules) end end # Represents a user-visible job which provides the insights for the related data - # source. For examples: - Data Quality: generates queries based on the rules and - # run against the data to get data quality check results. - Data Profile: + # source.For example: Data Quality: generates queries based on the rules and + # runs against the data to get data quality check results. Data Profile: # analyzes the data in table(s) and generates insights about the structure, # content and relationships (such as null percent, cardinality, min/max/mean, # etc). class GoogleCloudDataplexV1DataScan include Google::Apis::Core::Hashable @@ -1714,16 +1706,16 @@ # DataQualityScan related setting. # Corresponds to the JSON property `dataQualitySpec` # @return [Google::Apis::DataplexV1::GoogleCloudDataplexV1DataQualitySpec] attr_accessor :data_quality_spec - # Optional. Description of the scan. * Must be between 1-1024 characters. + # Optional. Description of the scan. Must be between 1-1024 characters. # Corresponds to the JSON property `description` # @return [String] attr_accessor :description - # Optional. User friendly display name. * Must be between 1-256 characters. + # Optional. User friendly display name. Must be between 1-256 characters. # Corresponds to the JSON property `displayName` # @return [String] attr_accessor :display_name # DataScan execution settings. @@ -1740,13 +1732,12 @@ # Corresponds to the JSON property `labels` # @return [Hash<String,String>] attr_accessor :labels # Output only. The relative resource name of the scan, of the form: projects/` - # project`/locations/`location_id`/dataScans/`datascan_id`. where `project` - # refers to a project_id or project_number and location_id refers to a GCP - # region. + # project`/locations/`location_id`/dataScans/`datascan_id`, where project refers + # to a project_id or project_number and location_id refers to a GCP region. # Corresponds to the JSON property `name` # @return [String] attr_accessor :name # Output only. Current state of the DataScan. @@ -1936,12 +1927,13 @@ # DataScan execution settings. class GoogleCloudDataplexV1DataScanExecutionSpec include Google::Apis::Core::Hashable - # Immutable. The unnested field (Date or Timestamp) that contains values that - # monotonically increase over time. + # Immutable. The unnested field (of type Date or Timestamp) that contains values + # which monotonically increase over time.If not specified, a data scan will run + # for all data in the table. # Corresponds to the JSON property `field` # @return [String] attr_accessor :field # DataScan scheduling and trigger settings. @@ -1983,11 +1975,11 @@ @latest_job_end_time = args[:latest_job_end_time] if args.key?(:latest_job_end_time) @latest_job_start_time = args[:latest_job_start_time] if args.key?(:latest_job_start_time) end end - # A DataScanJob represents an instance of a data scan. + # A DataScanJob represents an instance of DataScan execution. class GoogleCloudDataplexV1DataScanJob include Google::Apis::Core::Hashable # DataProfileResult defines the output of DataProfileScan. Each field of the # table will have field type specific profile result. @@ -2020,11 +2012,11 @@ # @return [String] attr_accessor :message # Output only. The relative resource name of the DataScanJob, of the form: # projects/`project`/locations/`location_id`/dataScans/`datascan_id`/jobs/` - # job_id`. where `project` refers to a project_id or project_number and + # job_id`, where project refers to a project_id or project_number and # location_id refers to a GCP region. # Corresponds to the JSON property `name` # @return [String] attr_accessor :name @@ -2070,13 +2062,13 @@ # The data source for DataScan. class GoogleCloudDataplexV1DataSource include Google::Apis::Core::Hashable - # Immutable. The dataplex entity that contains the data for DataScan, of the - # form: projects/`project_number`/locations/`location_id`/lakes/`lake_id`/zones/` - # zone_id`/entities/`entity_id`. + # Immutable. The Dataplex entity that represents the data source (e.g. BigQuery + # table) for DataScan, of the form: projects/`project_number`/locations/` + # location_id`/lakes/`lake_id`/zones/`zone_id`/entities/`entity_id`. # Corresponds to the JSON property `entity` # @return [String] attr_accessor :entity def initialize(**args) @@ -2363,10 +2355,16 @@ # Required. Immutable. The type of entity. # Corresponds to the JSON property `type` # @return [String] attr_accessor :type + # Output only. System generated unique ID for the Entity. This ID will be + # different if the Entity is deleted and re-created with the same name. + # Corresponds to the JSON property `uid` + # @return [String] + attr_accessor :uid + # Output only. The time when the entity was last updated. # Corresponds to the JSON property `updateTime` # @return [String] attr_accessor :update_time @@ -2390,10 +2388,11 @@ @id = args[:id] if args.key?(:id) @name = args[:name] if args.key?(:name) @schema = args[:schema] if args.key?(:schema) @system = args[:system] if args.key?(:system) @type = args[:type] if args.key?(:type) + @uid = args[:uid] if args.key?(:uid) @update_time = args[:update_time] if args.key?(:update_time) end end # Provides compatibility information for various metadata stores. @@ -3087,11 +3086,11 @@ # List DataScanJobs response. class GoogleCloudDataplexV1ListDataScanJobsResponse include Google::Apis::Core::Hashable - # DataScanJobs (metadata only) under a given dataScan. + # DataScanJobs (BASIC view only) under a given dataScan. # Corresponds to the JSON property `dataScanJobs` # @return [Array<Google::Apis::DataplexV1::GoogleCloudDataplexV1DataScanJob>] attr_accessor :data_scan_jobs # Token to retrieve the next page of results, or empty if there are no more @@ -3113,11 +3112,11 @@ # List dataScans response. class GoogleCloudDataplexV1ListDataScansResponse include Google::Apis::Core::Hashable - # DataScans (metadata only) under the given parent location. + # DataScans (BASIC view only) under the given parent location. # Corresponds to the JSON property `dataScans` # @return [Array<Google::Apis::DataplexV1::GoogleCloudDataplexV1DataScan>] attr_accessor :data_scans # Token to retrieve the next page of results, or empty if there are no more @@ -3480,11 +3479,11 @@ # Run DataScan Response. class GoogleCloudDataplexV1RunDataScanResponse include Google::Apis::Core::Hashable - # A DataScanJob represents an instance of a data scan. + # A DataScanJob represents an instance of DataScan execution. # Corresponds to the JSON property `job` # @return [Google::Apis::DataplexV1::GoogleCloudDataplexV1DataScanJob] attr_accessor :job def initialize(**args) @@ -3550,22 +3549,22 @@ # A data range denoted by a pair of start/end values of a field. class GoogleCloudDataplexV1ScannedDataIncrementalField include Google::Apis::Core::Hashable - # Value that marks the end of the range + # Value that marks the end of the range. # Corresponds to the JSON property `end` # @return [String] attr_accessor :end - # The field that contains values which monotonically increases over time (e.g. - # timestamp). + # The field that contains values which monotonically increases over time (e.g. a + # timestamp column). # Corresponds to the JSON property `field` # @return [String] attr_accessor :field - # Value that marks the start of the range + # Value that marks the start of the range. # Corresponds to the JSON property `start` # @return [String] attr_accessor :start def initialize(**args) @@ -4456,12 +4455,12 @@ # Optional. Cron schedule (https://en.wikipedia.org/wiki/Cron) for running tasks # periodically. To explicitly set a timezone to the cron tab, apply a prefix in # the cron tab: "CRON_TZ=$`IANA_TIME_ZONE`" or "TZ=$`IANA_TIME_ZONE`". The $` # IANA_TIME_ZONE` may only be a valid string from IANA time zone database. For - # example, "CRON_TZ=America/New_York 1 * * * *", or "TZ=America/New_York 1 * * * - # *". This field is required for RECURRING tasks. + # example, CRON_TZ=America/New_York 1 * * * *, or TZ=America/New_York 1 * * * *. + # This field is required for RECURRING tasks. # Corresponds to the JSON property `schedule` # @return [String] attr_accessor :schedule # Optional. The first run of the task will be after this time. If not specified, @@ -4492,11 +4491,11 @@ # DataScan scheduling and trigger settings. class GoogleCloudDataplexV1Trigger include Google::Apis::Core::Hashable - # The scan runs one-time via RunDataScan API. + # The scan runs once via RunDataScan API. # Corresponds to the JSON property `onDemand` # @return [Google::Apis::DataplexV1::GoogleCloudDataplexV1TriggerOnDemand] attr_accessor :on_demand # The scan is scheduled to run periodically. @@ -4513,11 +4512,11 @@ @on_demand = args[:on_demand] if args.key?(:on_demand) @schedule = args[:schedule] if args.key?(:schedule) end end - # The scan runs one-time via RunDataScan API. + # The scan runs once via RunDataScan API. class GoogleCloudDataplexV1TriggerOnDemand include Google::Apis::Core::Hashable def initialize(**args) update!(**args) @@ -4530,16 +4529,17 @@ # The scan is scheduled to run periodically. class GoogleCloudDataplexV1TriggerSchedule include Google::Apis::Core::Hashable - # Required. Cron schedule (https://en.wikipedia.org/wiki/Cron) for running scans - # periodically. To explicitly set a timezone to the cron tab, apply a prefix in + # Required. Cron (https://en.wikipedia.org/wiki/Cron) schedule for running scans + # periodically.To explicitly set a timezone in the cron tab, apply a prefix in # the cron tab: "CRON_TZ=$`IANA_TIME_ZONE`" or "TZ=$`IANA_TIME_ZONE`". The $` - # IANA_TIME_ZONE` may only be a valid string from IANA time zone database. For - # example, "CRON_TZ=America/New_York 1 * * * *", or "TZ=America/New_York 1 * * * - # *". This field is required for Schedule scans. + # IANA_TIME_ZONE` may only be a valid string from IANA time zone database ( + # wikipedia (https://en.wikipedia.org/wiki/List_of_tz_database_time_zones#List)). + # For example, CRON_TZ=America/New_York 1 * * * *, or TZ=America/New_York 1 * * + # * *.This field is required for Schedule scans. # Corresponds to the JSON property `cron` # @return [String] attr_accessor :cron def initialize(**args) @@ -4682,11 +4682,11 @@ # Optional. Cron schedule (https://en.wikipedia.org/wiki/Cron) for running # discovery periodically. Successive discovery runs must be scheduled at least # 60 minutes apart. The default value is to run discovery every 60 minutes. To # explicitly set a timezone to the cron tab, apply a prefix in the cron tab: " # CRON_TZ=$`IANA_TIME_ZONE`" or TZ=$`IANA_TIME_ZONE`". The $`IANA_TIME_ZONE` may - # only be a valid string from IANA time zone database. For example, "CRON_TZ= - # America/New_York 1 * * * *", or "TZ=America/New_York 1 * * * *". + # only be a valid string from IANA time zone database. For example, CRON_TZ= + # America/New_York 1 * * * *, or TZ=America/New_York 1 * * * *. # Corresponds to the JSON property `schedule` # @return [String] attr_accessor :schedule def initialize(**args)