diff --git a/.codegen/_openapi_sha b/.codegen/_openapi_sha index d227bcf74..2cc1f0bd8 100755 --- a/.codegen/_openapi_sha +++ b/.codegen/_openapi_sha @@ -1 +1 @@ -539df954f01ca41273fbe025ced73b4b6a8fda7e \ No newline at end of file +c3a1161e432b5db5ec2669b0a855be7cca3a81d6 \ No newline at end of file diff --git a/.gitattributes b/.gitattributes index 44139fc3e..9fad2a72b 100755 --- a/.gitattributes +++ b/.gitattributes @@ -598,6 +598,7 @@ databricks-sdk-java/src/main/java/com/databricks/sdk/service/catalog/TableType.j databricks-sdk-java/src/main/java/com/databricks/sdk/service/catalog/TablesAPI.java linguist-generated=true databricks-sdk-java/src/main/java/com/databricks/sdk/service/catalog/TablesImpl.java linguist-generated=true databricks-sdk-java/src/main/java/com/databricks/sdk/service/catalog/TablesService.java linguist-generated=true +databricks-sdk-java/src/main/java/com/databricks/sdk/service/catalog/TagAssignmentSourceType.java linguist-generated=true databricks-sdk-java/src/main/java/com/databricks/sdk/service/catalog/TagKeyValue.java linguist-generated=true databricks-sdk-java/src/main/java/com/databricks/sdk/service/catalog/TemporaryCredentials.java linguist-generated=true databricks-sdk-java/src/main/java/com/databricks/sdk/service/catalog/TemporaryPathCredentialsAPI.java linguist-generated=true @@ -1480,6 +1481,7 @@ databricks-sdk-java/src/main/java/com/databricks/sdk/service/jobs/Source.java li databricks-sdk-java/src/main/java/com/databricks/sdk/service/jobs/SparkJarTask.java linguist-generated=true databricks-sdk-java/src/main/java/com/databricks/sdk/service/jobs/SparkPythonTask.java linguist-generated=true databricks-sdk-java/src/main/java/com/databricks/sdk/service/jobs/SparkSubmitTask.java linguist-generated=true +databricks-sdk-java/src/main/java/com/databricks/sdk/service/jobs/SparseCheckout.java linguist-generated=true databricks-sdk-java/src/main/java/com/databricks/sdk/service/jobs/SqlAlertOutput.java linguist-generated=true databricks-sdk-java/src/main/java/com/databricks/sdk/service/jobs/SqlAlertState.java linguist-generated=true databricks-sdk-java/src/main/java/com/databricks/sdk/service/jobs/SqlDashboardOutput.java linguist-generated=true diff --git a/NEXT_CHANGELOG.md b/NEXT_CHANGELOG.md old mode 100644 new mode 100755 index 2d6c4e6e3..c4fd1f031 --- a/NEXT_CHANGELOG.md +++ b/NEXT_CHANGELOG.md @@ -13,3 +13,9 @@ ### Internal Changes ### API Changes +* Add `sourceType`, `updateTime` and `updatedBy` fields for `com.databricks.sdk.service.catalog.EntityTagAssignment`. +* Add `skipNotify` field for `com.databricks.sdk.service.dashboards.Subscription`. +* Add `sparseCheckout` field for `com.databricks.sdk.service.jobs.GitSource`. +* Add `disableAutoOptimization`, `maxRetries`, `minRetryIntervalMillis` and `retryOnTimeout` fields for `com.databricks.sdk.service.jobs.RunTask`. +* Add `disableAutoOptimization`, `maxRetries`, `minRetryIntervalMillis` and `retryOnTimeout` fields for `com.databricks.sdk.service.jobs.SubmitTask`. +* Add `EDGEGRID_AKAMAI` enum value for `com.databricks.sdk.service.catalog.CredentialType`. \ No newline at end of file diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/catalog/CredentialType.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/catalog/CredentialType.java index 6cf2586e2..7d6a6aca3 100755 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/catalog/CredentialType.java +++ b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/catalog/CredentialType.java @@ -4,11 +4,12 @@ import com.databricks.sdk.support.Generated; -/** Next Id: 15 */ +/** Next Id: 16 */ @Generated public enum CredentialType { ANY_STATIC_CREDENTIAL, BEARER_TOKEN, + EDGEGRID_AKAMAI, OAUTH_ACCESS_TOKEN, OAUTH_M2M, OAUTH_MTLS, diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/catalog/EntityTagAssignment.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/catalog/EntityTagAssignment.java index 95c3c464e..1ac630031 100755 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/catalog/EntityTagAssignment.java +++ b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/catalog/EntityTagAssignment.java @@ -5,6 +5,7 @@ import com.databricks.sdk.support.Generated; import com.databricks.sdk.support.ToStringer; import com.fasterxml.jackson.annotation.JsonProperty; +import com.google.protobuf.Timestamp; import java.util.Objects; /** Represents a tag assignment to an entity */ @@ -21,6 +22,10 @@ public class EntityTagAssignment { @JsonProperty("entity_type") private String entityType; + /** The source type of the tag assignment, e.g., user-assigned or system-assigned */ + @JsonProperty("source_type") + private TagAssignmentSourceType sourceType; + /** The key of the tag */ @JsonProperty("tag_key") private String tagKey; @@ -29,6 +34,14 @@ public class EntityTagAssignment { @JsonProperty("tag_value") private String tagValue; + /** The timestamp when the tag assignment was last updated */ + @JsonProperty("update_time") + private Timestamp updateTime; + + /** The user or principal who updated the tag assignment */ + @JsonProperty("updated_by") + private String updatedBy; + public EntityTagAssignment setEntityName(String entityName) { this.entityName = entityName; return this; @@ -47,6 +60,15 @@ public String getEntityType() { return entityType; } + public EntityTagAssignment setSourceType(TagAssignmentSourceType sourceType) { + this.sourceType = sourceType; + return this; + } + + public TagAssignmentSourceType getSourceType() { + return sourceType; + } + public EntityTagAssignment setTagKey(String tagKey) { this.tagKey = tagKey; return this; @@ -65,6 +87,24 @@ public String getTagValue() { return tagValue; } + public EntityTagAssignment setUpdateTime(Timestamp updateTime) { + this.updateTime = updateTime; + return this; + } + + public Timestamp getUpdateTime() { + return updateTime; + } + + public EntityTagAssignment setUpdatedBy(String updatedBy) { + this.updatedBy = updatedBy; + return this; + } + + public String getUpdatedBy() { + return updatedBy; + } + @Override public boolean equals(Object o) { if (this == o) return true; @@ -72,13 +112,17 @@ public boolean equals(Object o) { EntityTagAssignment that = (EntityTagAssignment) o; return Objects.equals(entityName, that.entityName) && Objects.equals(entityType, that.entityType) + && Objects.equals(sourceType, that.sourceType) && Objects.equals(tagKey, that.tagKey) - && Objects.equals(tagValue, that.tagValue); + && Objects.equals(tagValue, that.tagValue) + && Objects.equals(updateTime, that.updateTime) + && Objects.equals(updatedBy, that.updatedBy); } @Override public int hashCode() { - return Objects.hash(entityName, entityType, tagKey, tagValue); + return Objects.hash( + entityName, entityType, sourceType, tagKey, tagValue, updateTime, updatedBy); } @Override @@ -86,8 +130,11 @@ public String toString() { return new ToStringer(EntityTagAssignment.class) .add("entityName", entityName) .add("entityType", entityType) + .add("sourceType", sourceType) .add("tagKey", tagKey) .add("tagValue", tagValue) + .add("updateTime", updateTime) + .add("updatedBy", updatedBy) .toString(); } } diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/catalog/SecurableKind.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/catalog/SecurableKind.java index a21a9b821..175a780e1 100755 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/catalog/SecurableKind.java +++ b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/catalog/SecurableKind.java @@ -4,7 +4,7 @@ import com.databricks.sdk.support.Generated; -/** Latest kind: CONNECTION_OKTA_SYSTEM_LOGS_SSWS_TOKEN = 295; Next id: 296 */ +/** Latest kind: CONNECTION_AKAMAI_WAF_EDGEGRID = 296; Next id: 297 */ @Generated public enum SecurableKind { TABLE_DB_STORAGE, diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/catalog/TagAssignmentSourceType.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/catalog/TagAssignmentSourceType.java new file mode 100755 index 000000000..ee09c7b59 --- /dev/null +++ b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/catalog/TagAssignmentSourceType.java @@ -0,0 +1,11 @@ +// Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT. + +package com.databricks.sdk.service.catalog; + +import com.databricks.sdk.support.Generated; + +/** Enum representing the source type of a tag assignment */ +@Generated +public enum TagAssignmentSourceType { + TAG_ASSIGNMENT_SOURCE_TYPE_SYSTEM_DATA_CLASSIFICATION, +} diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/dashboards/Subscription.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/dashboards/Subscription.java index c271bee40..3a8978fcc 100755 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/dashboards/Subscription.java +++ b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/dashboards/Subscription.java @@ -35,6 +35,14 @@ public class Subscription { @JsonProperty("schedule_id") private String scheduleId; + /** + * Controls whether notifications are sent to the subscriber for scheduled dashboard refreshes. If + * not defined, defaults to false in the backend to match the current behavior (refresh and + * notify) + */ + @JsonProperty("skip_notify") + private Boolean skipNotify; + /** Subscriber details for users and destinations to be added as subscribers to the schedule. */ @JsonProperty("subscriber") private Subscriber subscriber; @@ -92,6 +100,15 @@ public String getScheduleId() { return scheduleId; } + public Subscription setSkipNotify(Boolean skipNotify) { + this.skipNotify = skipNotify; + return this; + } + + public Boolean getSkipNotify() { + return skipNotify; + } + public Subscription setSubscriber(Subscriber subscriber) { this.subscriber = subscriber; return this; @@ -129,6 +146,7 @@ public boolean equals(Object o) { && Objects.equals(dashboardId, that.dashboardId) && Objects.equals(etag, that.etag) && Objects.equals(scheduleId, that.scheduleId) + && Objects.equals(skipNotify, that.skipNotify) && Objects.equals(subscriber, that.subscriber) && Objects.equals(subscriptionId, that.subscriptionId) && Objects.equals(updateTime, that.updateTime); @@ -142,6 +160,7 @@ public int hashCode() { dashboardId, etag, scheduleId, + skipNotify, subscriber, subscriptionId, updateTime); @@ -155,6 +174,7 @@ public String toString() { .add("dashboardId", dashboardId) .add("etag", etag) .add("scheduleId", scheduleId) + .add("skipNotify", skipNotify) .add("subscriber", subscriber) .add("subscriptionId", subscriptionId) .add("updateTime", updateTime) diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/jobs/GitSource.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/jobs/GitSource.java index 88c0dae70..27bb17bee 100755 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/jobs/GitSource.java +++ b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/jobs/GitSource.java @@ -61,6 +61,10 @@ public class GitSource { @JsonProperty("job_source") private JobSource jobSource; + /** */ + @JsonProperty("sparse_checkout") + private SparseCheckout sparseCheckout; + public GitSource setGitBranch(String gitBranch) { this.gitBranch = gitBranch; return this; @@ -124,6 +128,15 @@ public JobSource getJobSource() { return jobSource; } + public GitSource setSparseCheckout(SparseCheckout sparseCheckout) { + this.sparseCheckout = sparseCheckout; + return this; + } + + public SparseCheckout getSparseCheckout() { + return sparseCheckout; + } + @Override public boolean equals(Object o) { if (this == o) return true; @@ -135,12 +148,14 @@ public boolean equals(Object o) { && Objects.equals(gitSnapshot, that.gitSnapshot) && Objects.equals(gitTag, that.gitTag) && Objects.equals(gitUrl, that.gitUrl) - && Objects.equals(jobSource, that.jobSource); + && Objects.equals(jobSource, that.jobSource) + && Objects.equals(sparseCheckout, that.sparseCheckout); } @Override public int hashCode() { - return Objects.hash(gitBranch, gitCommit, gitProvider, gitSnapshot, gitTag, gitUrl, jobSource); + return Objects.hash( + gitBranch, gitCommit, gitProvider, gitSnapshot, gitTag, gitUrl, jobSource, sparseCheckout); } @Override @@ -153,6 +168,7 @@ public String toString() { .add("gitTag", gitTag) .add("gitUrl", gitUrl) .add("jobSource", jobSource) + .add("sparseCheckout", sparseCheckout) .toString(); } } diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/jobs/RunTask.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/jobs/RunTask.java index 480e343b7..a60ca38ae 100755 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/jobs/RunTask.java +++ b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/jobs/RunTask.java @@ -88,6 +88,10 @@ public class RunTask { @JsonProperty("description") private String description; + /** An option to disable auto optimization in serverless */ + @JsonProperty("disable_auto_optimization") + private Boolean disableAutoOptimization; + /** * The actual performance target used by the serverless run during execution. This can differ from * the client-set performance target on the request depending on whether the performance mode is @@ -175,6 +179,22 @@ public class RunTask { @JsonProperty("libraries") private Collection libraries; + /** + * An optional maximum number of times to retry an unsuccessful run. A run is considered to be + * unsuccessful if it completes with the `FAILED` result_state or `INTERNAL_ERROR` + * `life_cycle_state`. The value `-1` means to retry indefinitely and the value `0` means to never + * retry. + */ + @JsonProperty("max_retries") + private Long maxRetries; + + /** + * An optional minimal interval in milliseconds between the start of the failed run and the + * subsequent retry run. The default behavior is that unsuccessful runs are immediately retried. + */ + @JsonProperty("min_retry_interval_millis") + private Long minRetryIntervalMillis; + /** If new_cluster, a description of a new cluster that is created for each run. */ @JsonProperty("new_cluster") private com.databricks.sdk.service.compute.ClusterSpec newCluster; @@ -215,6 +235,13 @@ public class RunTask { @JsonProperty("resolved_values") private ResolvedValues resolvedValues; + /** + * An optional policy to specify whether to retry a job when it times out. The default behavior is + * to not retry on timeout. + */ + @JsonProperty("retry_on_timeout") + private Boolean retryOnTimeout; + /** The time in milliseconds it took the job run and all of its repairs to finish. */ @JsonProperty("run_duration") private Long runDuration; @@ -416,6 +443,15 @@ public String getDescription() { return description; } + public RunTask setDisableAutoOptimization(Boolean disableAutoOptimization) { + this.disableAutoOptimization = disableAutoOptimization; + return this; + } + + public Boolean getDisableAutoOptimization() { + return disableAutoOptimization; + } + public RunTask setEffectivePerformanceTarget(PerformanceTarget effectivePerformanceTarget) { this.effectivePerformanceTarget = effectivePerformanceTarget; return this; @@ -515,6 +551,24 @@ public Collection getLibraries() { return libraries; } + public RunTask setMaxRetries(Long maxRetries) { + this.maxRetries = maxRetries; + return this; + } + + public Long getMaxRetries() { + return maxRetries; + } + + public RunTask setMinRetryIntervalMillis(Long minRetryIntervalMillis) { + this.minRetryIntervalMillis = minRetryIntervalMillis; + return this; + } + + public Long getMinRetryIntervalMillis() { + return minRetryIntervalMillis; + } + public RunTask setNewCluster(com.databricks.sdk.service.compute.ClusterSpec newCluster) { this.newCluster = newCluster; return this; @@ -587,6 +641,15 @@ public ResolvedValues getResolvedValues() { return resolvedValues; } + public RunTask setRetryOnTimeout(Boolean retryOnTimeout) { + this.retryOnTimeout = retryOnTimeout; + return this; + } + + public Boolean getRetryOnTimeout() { + return retryOnTimeout; + } + public RunTask setRunDuration(Long runDuration) { this.runDuration = runDuration; return this; @@ -748,6 +811,7 @@ public boolean equals(Object o) { && Objects.equals(dbtTask, that.dbtTask) && Objects.equals(dependsOn, that.dependsOn) && Objects.equals(description, that.description) + && Objects.equals(disableAutoOptimization, that.disableAutoOptimization) && Objects.equals(effectivePerformanceTarget, that.effectivePerformanceTarget) && Objects.equals(emailNotifications, that.emailNotifications) && Objects.equals(endTime, that.endTime) @@ -759,6 +823,8 @@ public boolean equals(Object o) { && Objects.equals(gitSource, that.gitSource) && Objects.equals(jobClusterKey, that.jobClusterKey) && Objects.equals(libraries, that.libraries) + && Objects.equals(maxRetries, that.maxRetries) + && Objects.equals(minRetryIntervalMillis, that.minRetryIntervalMillis) && Objects.equals(newCluster, that.newCluster) && Objects.equals(notebookTask, that.notebookTask) && Objects.equals(notificationSettings, that.notificationSettings) @@ -767,6 +833,7 @@ public boolean equals(Object o) { && Objects.equals(pythonWheelTask, that.pythonWheelTask) && Objects.equals(queueDuration, that.queueDuration) && Objects.equals(resolvedValues, that.resolvedValues) + && Objects.equals(retryOnTimeout, that.retryOnTimeout) && Objects.equals(runDuration, that.runDuration) && Objects.equals(runId, that.runId) && Objects.equals(runIf, that.runIf) @@ -800,6 +867,7 @@ public int hashCode() { dbtTask, dependsOn, description, + disableAutoOptimization, effectivePerformanceTarget, emailNotifications, endTime, @@ -811,6 +879,8 @@ public int hashCode() { gitSource, jobClusterKey, libraries, + maxRetries, + minRetryIntervalMillis, newCluster, notebookTask, notificationSettings, @@ -819,6 +889,7 @@ public int hashCode() { pythonWheelTask, queueDuration, resolvedValues, + retryOnTimeout, runDuration, runId, runIf, @@ -852,6 +923,7 @@ public String toString() { .add("dbtTask", dbtTask) .add("dependsOn", dependsOn) .add("description", description) + .add("disableAutoOptimization", disableAutoOptimization) .add("effectivePerformanceTarget", effectivePerformanceTarget) .add("emailNotifications", emailNotifications) .add("endTime", endTime) @@ -863,6 +935,8 @@ public String toString() { .add("gitSource", gitSource) .add("jobClusterKey", jobClusterKey) .add("libraries", libraries) + .add("maxRetries", maxRetries) + .add("minRetryIntervalMillis", minRetryIntervalMillis) .add("newCluster", newCluster) .add("notebookTask", notebookTask) .add("notificationSettings", notificationSettings) @@ -871,6 +945,7 @@ public String toString() { .add("pythonWheelTask", pythonWheelTask) .add("queueDuration", queueDuration) .add("resolvedValues", resolvedValues) + .add("retryOnTimeout", retryOnTimeout) .add("runDuration", runDuration) .add("runId", runId) .add("runIf", runIf) diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/jobs/SparseCheckout.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/jobs/SparseCheckout.java new file mode 100755 index 000000000..193e1083d --- /dev/null +++ b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/jobs/SparseCheckout.java @@ -0,0 +1,43 @@ +// Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT. + +package com.databricks.sdk.service.jobs; + +import com.databricks.sdk.support.Generated; +import com.databricks.sdk.support.ToStringer; +import com.fasterxml.jackson.annotation.JsonProperty; +import java.util.Collection; +import java.util.Objects; + +@Generated +public class SparseCheckout { + /** List of patterns to include for sparse checkout. */ + @JsonProperty("patterns") + private Collection patterns; + + public SparseCheckout setPatterns(Collection patterns) { + this.patterns = patterns; + return this; + } + + public Collection getPatterns() { + return patterns; + } + + @Override + public boolean equals(Object o) { + if (this == o) return true; + if (o == null || getClass() != o.getClass()) return false; + SparseCheckout that = (SparseCheckout) o; + return Objects.equals(patterns, that.patterns); + } + + @Override + public int hashCode() { + return Objects.hash(patterns); + } + + @Override + public String toString() { + return new ToStringer(SparseCheckout.class).add("patterns", patterns).toString(); + } +} diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/jobs/SubmitTask.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/jobs/SubmitTask.java index 2f9339f6e..8dc713ec8 100755 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/jobs/SubmitTask.java +++ b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/jobs/SubmitTask.java @@ -61,6 +61,10 @@ public class SubmitTask { @JsonProperty("description") private String description; + /** An option to disable auto optimization in serverless */ + @JsonProperty("disable_auto_optimization") + private Boolean disableAutoOptimization; + /** * An optional set of email addresses notified when the task run begins or completes. The default * behavior is to not send any emails. @@ -105,6 +109,22 @@ public class SubmitTask { @JsonProperty("libraries") private Collection libraries; + /** + * An optional maximum number of times to retry an unsuccessful run. A run is considered to be + * unsuccessful if it completes with the `FAILED` result_state or `INTERNAL_ERROR` + * `life_cycle_state`. The value `-1` means to retry indefinitely and the value `0` means to never + * retry. + */ + @JsonProperty("max_retries") + private Long maxRetries; + + /** + * An optional minimal interval in milliseconds between the start of the failed run and the + * subsequent retry run. The default behavior is that unsuccessful runs are immediately retried. + */ + @JsonProperty("min_retry_interval_millis") + private Long minRetryIntervalMillis; + /** If new_cluster, a description of a new cluster that is created for each run. */ @JsonProperty("new_cluster") private com.databricks.sdk.service.compute.ClusterSpec newCluster; @@ -137,6 +157,13 @@ public class SubmitTask { @JsonProperty("python_wheel_task") private PythonWheelTask pythonWheelTask; + /** + * An optional policy to specify whether to retry a job when it times out. The default behavior is + * to not retry on timeout. + */ + @JsonProperty("retry_on_timeout") + private Boolean retryOnTimeout; + /** * An optional value indicating the condition that determines whether the task should be run once * its dependencies have been completed. When omitted, defaults to `ALL_SUCCESS`. See @@ -273,6 +300,15 @@ public String getDescription() { return description; } + public SubmitTask setDisableAutoOptimization(Boolean disableAutoOptimization) { + this.disableAutoOptimization = disableAutoOptimization; + return this; + } + + public Boolean getDisableAutoOptimization() { + return disableAutoOptimization; + } + public SubmitTask setEmailNotifications(JobEmailNotifications emailNotifications) { this.emailNotifications = emailNotifications; return this; @@ -336,6 +372,24 @@ public Collection getLibraries() { return libraries; } + public SubmitTask setMaxRetries(Long maxRetries) { + this.maxRetries = maxRetries; + return this; + } + + public Long getMaxRetries() { + return maxRetries; + } + + public SubmitTask setMinRetryIntervalMillis(Long minRetryIntervalMillis) { + this.minRetryIntervalMillis = minRetryIntervalMillis; + return this; + } + + public Long getMinRetryIntervalMillis() { + return minRetryIntervalMillis; + } + public SubmitTask setNewCluster(com.databricks.sdk.service.compute.ClusterSpec newCluster) { this.newCluster = newCluster; return this; @@ -390,6 +444,15 @@ public PythonWheelTask getPythonWheelTask() { return pythonWheelTask; } + public SubmitTask setRetryOnTimeout(Boolean retryOnTimeout) { + this.retryOnTimeout = retryOnTimeout; + return this; + } + + public Boolean getRetryOnTimeout() { + return retryOnTimeout; + } + public SubmitTask setRunIf(RunIf runIf) { this.runIf = runIf; return this; @@ -485,6 +548,7 @@ public boolean equals(Object o) { && Objects.equals(dbtTask, that.dbtTask) && Objects.equals(dependsOn, that.dependsOn) && Objects.equals(description, that.description) + && Objects.equals(disableAutoOptimization, that.disableAutoOptimization) && Objects.equals(emailNotifications, that.emailNotifications) && Objects.equals(environmentKey, that.environmentKey) && Objects.equals(existingClusterId, that.existingClusterId) @@ -492,12 +556,15 @@ public boolean equals(Object o) { && Objects.equals(genAiComputeTask, that.genAiComputeTask) && Objects.equals(health, that.health) && Objects.equals(libraries, that.libraries) + && Objects.equals(maxRetries, that.maxRetries) + && Objects.equals(minRetryIntervalMillis, that.minRetryIntervalMillis) && Objects.equals(newCluster, that.newCluster) && Objects.equals(notebookTask, that.notebookTask) && Objects.equals(notificationSettings, that.notificationSettings) && Objects.equals(pipelineTask, that.pipelineTask) && Objects.equals(powerBiTask, that.powerBiTask) && Objects.equals(pythonWheelTask, that.pythonWheelTask) + && Objects.equals(retryOnTimeout, that.retryOnTimeout) && Objects.equals(runIf, that.runIf) && Objects.equals(runJobTask, that.runJobTask) && Objects.equals(sparkJarTask, that.sparkJarTask) @@ -521,6 +588,7 @@ public int hashCode() { dbtTask, dependsOn, description, + disableAutoOptimization, emailNotifications, environmentKey, existingClusterId, @@ -528,12 +596,15 @@ public int hashCode() { genAiComputeTask, health, libraries, + maxRetries, + minRetryIntervalMillis, newCluster, notebookTask, notificationSettings, pipelineTask, powerBiTask, pythonWheelTask, + retryOnTimeout, runIf, runJobTask, sparkJarTask, @@ -557,6 +628,7 @@ public String toString() { .add("dbtTask", dbtTask) .add("dependsOn", dependsOn) .add("description", description) + .add("disableAutoOptimization", disableAutoOptimization) .add("emailNotifications", emailNotifications) .add("environmentKey", environmentKey) .add("existingClusterId", existingClusterId) @@ -564,12 +636,15 @@ public String toString() { .add("genAiComputeTask", genAiComputeTask) .add("health", health) .add("libraries", libraries) + .add("maxRetries", maxRetries) + .add("minRetryIntervalMillis", minRetryIntervalMillis) .add("newCluster", newCluster) .add("notebookTask", notebookTask) .add("notificationSettings", notificationSettings) .add("pipelineTask", pipelineTask) .add("powerBiTask", powerBiTask) .add("pythonWheelTask", pythonWheelTask) + .add("retryOnTimeout", retryOnTimeout) .add("runIf", runIf) .add("runJobTask", runJobTask) .add("sparkJarTask", sparkJarTask) diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/provisioning/CreateWorkspaceRequest.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/provisioning/CreateWorkspaceRequest.java index 4051768bc..40e3fd63c 100755 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/provisioning/CreateWorkspaceRequest.java +++ b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/provisioning/CreateWorkspaceRequest.java @@ -14,7 +14,10 @@ public class CreateWorkspaceRequest { @JsonProperty("aws_region") private String awsRegion; - /** The cloud name. This field always has the value `gcp`. */ + /** + * DEPRECATED: This field is being ignored by the server and will be removed in the future. The + * cloud name. This field always has the value `gcp`. + */ @JsonProperty("cloud") private String cloud; diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/sql/CreateWarehouseRequest.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/sql/CreateWarehouseRequest.java index 2c29ba7cb..355151317 100755 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/sql/CreateWarehouseRequest.java +++ b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/sql/CreateWarehouseRequest.java @@ -31,7 +31,7 @@ public class CreateWarehouseRequest { * queries, please tune max_num_clusters. * *

Supported values: - 2X-Small - X-Small - Small - Medium - Large - X-Large - 2X-Large - - * 3X-Large - 4X-Large + * 3X-Large - 4X-Large - 5X-Large */ @JsonProperty("cluster_size") private String clusterSize; diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/sql/EditWarehouseRequest.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/sql/EditWarehouseRequest.java index 251b1555b..3343356c6 100755 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/sql/EditWarehouseRequest.java +++ b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/sql/EditWarehouseRequest.java @@ -36,7 +36,7 @@ public class EditWarehouseRequest { * queries, please tune max_num_clusters. * *

Supported values: - 2X-Small - X-Small - Small - Medium - Large - X-Large - 2X-Large - - * 3X-Large - 4X-Large + * 3X-Large - 4X-Large - 5X-Large */ @JsonProperty("cluster_size") private String clusterSize; diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/sql/EndpointInfo.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/sql/EndpointInfo.java index 03ae34840..4132cc7dd 100755 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/sql/EndpointInfo.java +++ b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/sql/EndpointInfo.java @@ -30,7 +30,7 @@ public class EndpointInfo { * queries, please tune max_num_clusters. * *

Supported values: - 2X-Small - X-Small - Small - Medium - Large - X-Large - 2X-Large - - * 3X-Large - 4X-Large + * 3X-Large - 4X-Large - 5X-Large */ @JsonProperty("cluster_size") private String clusterSize; diff --git a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/sql/GetWarehouseResponse.java b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/sql/GetWarehouseResponse.java index 6a789a483..f4a691556 100755 --- a/databricks-sdk-java/src/main/java/com/databricks/sdk/service/sql/GetWarehouseResponse.java +++ b/databricks-sdk-java/src/main/java/com/databricks/sdk/service/sql/GetWarehouseResponse.java @@ -30,7 +30,7 @@ public class GetWarehouseResponse { * queries, please tune max_num_clusters. * *

Supported values: - 2X-Small - X-Small - Small - Medium - Large - X-Large - 2X-Large - - * 3X-Large - 4X-Large + * 3X-Large - 4X-Large - 5X-Large */ @JsonProperty("cluster_size") private String clusterSize;