Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
3 changes: 2 additions & 1 deletion bundle/internal/tf/schema/config.go
Original file line number Diff line number Diff line change
Expand Up @@ -12,16 +12,17 @@ type Config struct {
AzureTenantId string `json:"azure_tenant_id,omitempty"`
AzureUseMsi bool `json:"azure_use_msi,omitempty"`
AzureWorkspaceResourceId string `json:"azure_workspace_resource_id,omitempty"`
DatabricksCliPath string `json:"databricks_cli_path,omitempty"`
ClientId string `json:"client_id,omitempty"`
ClientSecret string `json:"client_secret,omitempty"`
ConfigFile string `json:"config_file,omitempty"`
DatabricksCliPath string `json:"databricks_cli_path,omitempty"`
DebugHeaders bool `json:"debug_headers,omitempty"`
DebugTruncateBytes int `json:"debug_truncate_bytes,omitempty"`
GoogleCredentials string `json:"google_credentials,omitempty"`
GoogleServiceAccount string `json:"google_service_account,omitempty"`
Host string `json:"host,omitempty"`
HttpTimeoutSeconds int `json:"http_timeout_seconds,omitempty"`
MetadataServiceUrl string `json:"metadata_service_url,omitempty"`
Password string `json:"password,omitempty"`
Profile string `json:"profile,omitempty"`
RateLimit int `json:"rate_limit,omitempty"`
Expand Down
15 changes: 10 additions & 5 deletions bundle/internal/tf/schema/data_source_cluster.go
Original file line number Diff line number Diff line change
Expand Up @@ -120,12 +120,17 @@ type DataSourceClusterClusterInfoInitScriptsS3 struct {
Region string `json:"region,omitempty"`
}

type DataSourceClusterClusterInfoInitScriptsWorkspace struct {
Destination string `json:"destination,omitempty"`
}

type DataSourceClusterClusterInfoInitScripts struct {
Abfss *DataSourceClusterClusterInfoInitScriptsAbfss `json:"abfss,omitempty"`
Dbfs *DataSourceClusterClusterInfoInitScriptsDbfs `json:"dbfs,omitempty"`
File *DataSourceClusterClusterInfoInitScriptsFile `json:"file,omitempty"`
Gcs *DataSourceClusterClusterInfoInitScriptsGcs `json:"gcs,omitempty"`
S3 *DataSourceClusterClusterInfoInitScriptsS3 `json:"s3,omitempty"`
Abfss *DataSourceClusterClusterInfoInitScriptsAbfss `json:"abfss,omitempty"`
Dbfs *DataSourceClusterClusterInfoInitScriptsDbfs `json:"dbfs,omitempty"`
File *DataSourceClusterClusterInfoInitScriptsFile `json:"file,omitempty"`
Gcs *DataSourceClusterClusterInfoInitScriptsGcs `json:"gcs,omitempty"`
S3 *DataSourceClusterClusterInfoInitScriptsS3 `json:"s3,omitempty"`
Workspace *DataSourceClusterClusterInfoInitScriptsWorkspace `json:"workspace,omitempty"`
}

type DataSourceClusterClusterInfoTerminationReason struct {
Expand Down
2 changes: 1 addition & 1 deletion bundle/internal/tf/schema/data_source_cluster_policy.go
Original file line number Diff line number Diff line change
Expand Up @@ -6,5 +6,5 @@ type DataSourceClusterPolicy struct {
Definition string `json:"definition,omitempty"`
Id string `json:"id,omitempty"`
MaxClustersPerUser int `json:"max_clusters_per_user,omitempty"`
Name string `json:"name"`
Name string `json:"name,omitempty"`
}
75 changes: 60 additions & 15 deletions bundle/internal/tf/schema/data_source_job.go
Original file line number Diff line number Diff line change
Expand Up @@ -127,12 +127,17 @@ type DataSourceJobJobSettingsSettingsJobClusterNewClusterInitScriptsS3 struct {
Region string `json:"region,omitempty"`
}

type DataSourceJobJobSettingsSettingsJobClusterNewClusterInitScriptsWorkspace struct {
Destination string `json:"destination,omitempty"`
}

type DataSourceJobJobSettingsSettingsJobClusterNewClusterInitScripts struct {
Abfss *DataSourceJobJobSettingsSettingsJobClusterNewClusterInitScriptsAbfss `json:"abfss,omitempty"`
Dbfs *DataSourceJobJobSettingsSettingsJobClusterNewClusterInitScriptsDbfs `json:"dbfs,omitempty"`
File *DataSourceJobJobSettingsSettingsJobClusterNewClusterInitScriptsFile `json:"file,omitempty"`
Gcs *DataSourceJobJobSettingsSettingsJobClusterNewClusterInitScriptsGcs `json:"gcs,omitempty"`
S3 *DataSourceJobJobSettingsSettingsJobClusterNewClusterInitScriptsS3 `json:"s3,omitempty"`
Abfss *DataSourceJobJobSettingsSettingsJobClusterNewClusterInitScriptsAbfss `json:"abfss,omitempty"`
Dbfs *DataSourceJobJobSettingsSettingsJobClusterNewClusterInitScriptsDbfs `json:"dbfs,omitempty"`
File *DataSourceJobJobSettingsSettingsJobClusterNewClusterInitScriptsFile `json:"file,omitempty"`
Gcs *DataSourceJobJobSettingsSettingsJobClusterNewClusterInitScriptsGcs `json:"gcs,omitempty"`
S3 *DataSourceJobJobSettingsSettingsJobClusterNewClusterInitScriptsS3 `json:"s3,omitempty"`
Workspace *DataSourceJobJobSettingsSettingsJobClusterNewClusterInitScriptsWorkspace `json:"workspace,omitempty"`
}

type DataSourceJobJobSettingsSettingsJobClusterNewClusterWorkloadTypeClients struct {
Expand Down Expand Up @@ -303,12 +308,17 @@ type DataSourceJobJobSettingsSettingsNewClusterInitScriptsS3 struct {
Region string `json:"region,omitempty"`
}

type DataSourceJobJobSettingsSettingsNewClusterInitScriptsWorkspace struct {
Destination string `json:"destination,omitempty"`
}

type DataSourceJobJobSettingsSettingsNewClusterInitScripts struct {
Abfss *DataSourceJobJobSettingsSettingsNewClusterInitScriptsAbfss `json:"abfss,omitempty"`
Dbfs *DataSourceJobJobSettingsSettingsNewClusterInitScriptsDbfs `json:"dbfs,omitempty"`
File *DataSourceJobJobSettingsSettingsNewClusterInitScriptsFile `json:"file,omitempty"`
Gcs *DataSourceJobJobSettingsSettingsNewClusterInitScriptsGcs `json:"gcs,omitempty"`
S3 *DataSourceJobJobSettingsSettingsNewClusterInitScriptsS3 `json:"s3,omitempty"`
Abfss *DataSourceJobJobSettingsSettingsNewClusterInitScriptsAbfss `json:"abfss,omitempty"`
Dbfs *DataSourceJobJobSettingsSettingsNewClusterInitScriptsDbfs `json:"dbfs,omitempty"`
File *DataSourceJobJobSettingsSettingsNewClusterInitScriptsFile `json:"file,omitempty"`
Gcs *DataSourceJobJobSettingsSettingsNewClusterInitScriptsGcs `json:"gcs,omitempty"`
S3 *DataSourceJobJobSettingsSettingsNewClusterInitScriptsS3 `json:"s3,omitempty"`
Workspace *DataSourceJobJobSettingsSettingsNewClusterInitScriptsWorkspace `json:"workspace,omitempty"`
}

type DataSourceJobJobSettingsSettingsNewClusterWorkloadTypeClients struct {
Expand Down Expand Up @@ -359,6 +369,11 @@ type DataSourceJobJobSettingsSettingsNotebookTask struct {
Source string `json:"source,omitempty"`
}

type DataSourceJobJobSettingsSettingsNotificationSettings struct {
NoAlertForCanceledRuns bool `json:"no_alert_for_canceled_runs,omitempty"`
NoAlertForSkippedRuns bool `json:"no_alert_for_skipped_runs,omitempty"`
}

type DataSourceJobJobSettingsSettingsPipelineTask struct {
PipelineId string `json:"pipeline_id"`
}
Expand All @@ -370,6 +385,9 @@ type DataSourceJobJobSettingsSettingsPythonWheelTask struct {
Parameters []string `json:"parameters,omitempty"`
}

type DataSourceJobJobSettingsSettingsQueue struct {
}

type DataSourceJobJobSettingsSettingsSchedule struct {
PauseStatus string `json:"pause_status,omitempty"`
QuartzCronExpression string `json:"quartz_cron_expression"`
Expand All @@ -385,6 +403,7 @@ type DataSourceJobJobSettingsSettingsSparkJarTask struct {
type DataSourceJobJobSettingsSettingsSparkPythonTask struct {
Parameters []string `json:"parameters,omitempty"`
PythonFile string `json:"python_file"`
Source string `json:"source,omitempty"`
}

type DataSourceJobJobSettingsSettingsSparkSubmitTask struct {
Expand Down Expand Up @@ -533,12 +552,17 @@ type DataSourceJobJobSettingsSettingsTaskNewClusterInitScriptsS3 struct {
Region string `json:"region,omitempty"`
}

type DataSourceJobJobSettingsSettingsTaskNewClusterInitScriptsWorkspace struct {
Destination string `json:"destination,omitempty"`
}

type DataSourceJobJobSettingsSettingsTaskNewClusterInitScripts struct {
Abfss *DataSourceJobJobSettingsSettingsTaskNewClusterInitScriptsAbfss `json:"abfss,omitempty"`
Dbfs *DataSourceJobJobSettingsSettingsTaskNewClusterInitScriptsDbfs `json:"dbfs,omitempty"`
File *DataSourceJobJobSettingsSettingsTaskNewClusterInitScriptsFile `json:"file,omitempty"`
Gcs *DataSourceJobJobSettingsSettingsTaskNewClusterInitScriptsGcs `json:"gcs,omitempty"`
S3 *DataSourceJobJobSettingsSettingsTaskNewClusterInitScriptsS3 `json:"s3,omitempty"`
Abfss *DataSourceJobJobSettingsSettingsTaskNewClusterInitScriptsAbfss `json:"abfss,omitempty"`
Dbfs *DataSourceJobJobSettingsSettingsTaskNewClusterInitScriptsDbfs `json:"dbfs,omitempty"`
File *DataSourceJobJobSettingsSettingsTaskNewClusterInitScriptsFile `json:"file,omitempty"`
Gcs *DataSourceJobJobSettingsSettingsTaskNewClusterInitScriptsGcs `json:"gcs,omitempty"`
S3 *DataSourceJobJobSettingsSettingsTaskNewClusterInitScriptsS3 `json:"s3,omitempty"`
Workspace *DataSourceJobJobSettingsSettingsTaskNewClusterInitScriptsWorkspace `json:"workspace,omitempty"`
}

type DataSourceJobJobSettingsSettingsTaskNewClusterWorkloadTypeClients struct {
Expand Down Expand Up @@ -609,6 +633,7 @@ type DataSourceJobJobSettingsSettingsTaskSparkJarTask struct {
type DataSourceJobJobSettingsSettingsTaskSparkPythonTask struct {
Parameters []string `json:"parameters,omitempty"`
PythonFile string `json:"python_file"`
Source string `json:"source,omitempty"`
}

type DataSourceJobJobSettingsSettingsTaskSparkSubmitTask struct {
Expand All @@ -623,6 +648,10 @@ type DataSourceJobJobSettingsSettingsTaskSqlTaskDashboard struct {
DashboardId string `json:"dashboard_id"`
}

type DataSourceJobJobSettingsSettingsTaskSqlTaskFile struct {
Path string `json:"path"`
}

type DataSourceJobJobSettingsSettingsTaskSqlTaskQuery struct {
QueryId string `json:"query_id"`
}
Expand All @@ -632,6 +661,7 @@ type DataSourceJobJobSettingsSettingsTaskSqlTask struct {
WarehouseId string `json:"warehouse_id,omitempty"`
Alert *DataSourceJobJobSettingsSettingsTaskSqlTaskAlert `json:"alert,omitempty"`
Dashboard *DataSourceJobJobSettingsSettingsTaskSqlTaskDashboard `json:"dashboard,omitempty"`
File *DataSourceJobJobSettingsSettingsTaskSqlTaskFile `json:"file,omitempty"`
Query *DataSourceJobJobSettingsSettingsTaskSqlTaskQuery `json:"query,omitempty"`
}

Expand All @@ -642,6 +672,7 @@ type DataSourceJobJobSettingsSettingsTask struct {
MaxRetries int `json:"max_retries,omitempty"`
MinRetryIntervalMillis int `json:"min_retry_interval_millis,omitempty"`
RetryOnTimeout bool `json:"retry_on_timeout,omitempty"`
RunIf string `json:"run_if,omitempty"`
TaskKey string `json:"task_key,omitempty"`
TimeoutSeconds int `json:"timeout_seconds,omitempty"`
DbtTask *DataSourceJobJobSettingsSettingsTaskDbtTask `json:"dbt_task,omitempty"`
Expand All @@ -658,6 +689,17 @@ type DataSourceJobJobSettingsSettingsTask struct {
SqlTask *DataSourceJobJobSettingsSettingsTaskSqlTask `json:"sql_task,omitempty"`
}

type DataSourceJobJobSettingsSettingsTriggerFileArrival struct {
MinTimeBetweenTriggerSeconds int `json:"min_time_between_trigger_seconds,omitempty"`
Url string `json:"url"`
WaitAfterLastChangeSeconds int `json:"wait_after_last_change_seconds,omitempty"`
}

type DataSourceJobJobSettingsSettingsTrigger struct {
PauseStatus string `json:"pause_status,omitempty"`
FileArrival *DataSourceJobJobSettingsSettingsTriggerFileArrival `json:"file_arrival,omitempty"`
}

type DataSourceJobJobSettingsSettingsWebhookNotificationsOnFailure struct {
Id string `json:"id"`
}
Expand Down Expand Up @@ -694,13 +736,16 @@ type DataSourceJobJobSettingsSettings struct {
Library []DataSourceJobJobSettingsSettingsLibrary `json:"library,omitempty"`
NewCluster *DataSourceJobJobSettingsSettingsNewCluster `json:"new_cluster,omitempty"`
NotebookTask *DataSourceJobJobSettingsSettingsNotebookTask `json:"notebook_task,omitempty"`
NotificationSettings *DataSourceJobJobSettingsSettingsNotificationSettings `json:"notification_settings,omitempty"`
PipelineTask *DataSourceJobJobSettingsSettingsPipelineTask `json:"pipeline_task,omitempty"`
PythonWheelTask *DataSourceJobJobSettingsSettingsPythonWheelTask `json:"python_wheel_task,omitempty"`
Queue *DataSourceJobJobSettingsSettingsQueue `json:"queue,omitempty"`
Schedule *DataSourceJobJobSettingsSettingsSchedule `json:"schedule,omitempty"`
SparkJarTask *DataSourceJobJobSettingsSettingsSparkJarTask `json:"spark_jar_task,omitempty"`
SparkPythonTask *DataSourceJobJobSettingsSettingsSparkPythonTask `json:"spark_python_task,omitempty"`
SparkSubmitTask *DataSourceJobJobSettingsSettingsSparkSubmitTask `json:"spark_submit_task,omitempty"`
Task []DataSourceJobJobSettingsSettingsTask `json:"task,omitempty"`
Trigger *DataSourceJobJobSettingsSettingsTrigger `json:"trigger,omitempty"`
WebhookNotifications *DataSourceJobJobSettingsSettingsWebhookNotifications `json:"webhook_notifications,omitempty"`
}

Expand Down
9 changes: 9 additions & 0 deletions bundle/internal/tf/schema/data_source_pipelines.go
Original file line number Diff line number Diff line change
@@ -0,0 +1,9 @@
// Generated from Databricks Terraform provider schema. DO NOT EDIT.

package schema

type DataSourcePipelines struct {
Id string `json:"id,omitempty"`
Ids []string `json:"ids,omitempty"`
PipelineName string `json:"pipeline_name,omitempty"`
}
2 changes: 2 additions & 0 deletions bundle/internal/tf/schema/data_sources.go
Original file line number Diff line number Diff line change
Expand Up @@ -23,6 +23,7 @@ type DataSources struct {
NodeType map[string]*DataSourceNodeType `json:"databricks_node_type,omitempty"`
Notebook map[string]*DataSourceNotebook `json:"databricks_notebook,omitempty"`
NotebookPaths map[string]*DataSourceNotebookPaths `json:"databricks_notebook_paths,omitempty"`
Pipelines map[string]*DataSourcePipelines `json:"databricks_pipelines,omitempty"`
Schemas map[string]*DataSourceSchemas `json:"databricks_schemas,omitempty"`
ServicePrincipal map[string]*DataSourceServicePrincipal `json:"databricks_service_principal,omitempty"`
ServicePrincipals map[string]*DataSourceServicePrincipals `json:"databricks_service_principals,omitempty"`
Expand Down Expand Up @@ -59,6 +60,7 @@ func NewDataSources() *DataSources {
NodeType: make(map[string]*DataSourceNodeType),
Notebook: make(map[string]*DataSourceNotebook),
NotebookPaths: make(map[string]*DataSourceNotebookPaths),
Pipelines: make(map[string]*DataSourcePipelines),
Schemas: make(map[string]*DataSourceSchemas),
ServicePrincipal: make(map[string]*DataSourceServicePrincipal),
ServicePrincipals: make(map[string]*DataSourceServicePrincipals),
Expand Down
15 changes: 10 additions & 5 deletions bundle/internal/tf/schema/resource_cluster.go
Original file line number Diff line number Diff line change
Expand Up @@ -98,12 +98,17 @@ type ResourceClusterInitScriptsS3 struct {
Region string `json:"region,omitempty"`
}

type ResourceClusterInitScriptsWorkspace struct {
Destination string `json:"destination,omitempty"`
}

type ResourceClusterInitScripts struct {
Abfss *ResourceClusterInitScriptsAbfss `json:"abfss,omitempty"`
Dbfs *ResourceClusterInitScriptsDbfs `json:"dbfs,omitempty"`
File *ResourceClusterInitScriptsFile `json:"file,omitempty"`
Gcs *ResourceClusterInitScriptsGcs `json:"gcs,omitempty"`
S3 *ResourceClusterInitScriptsS3 `json:"s3,omitempty"`
Abfss *ResourceClusterInitScriptsAbfss `json:"abfss,omitempty"`
Dbfs *ResourceClusterInitScriptsDbfs `json:"dbfs,omitempty"`
File *ResourceClusterInitScriptsFile `json:"file,omitempty"`
Gcs *ResourceClusterInitScriptsGcs `json:"gcs,omitempty"`
S3 *ResourceClusterInitScriptsS3 `json:"s3,omitempty"`
Workspace *ResourceClusterInitScriptsWorkspace `json:"workspace,omitempty"`
}

type ResourceClusterLibraryCran struct {
Expand Down
13 changes: 8 additions & 5 deletions bundle/internal/tf/schema/resource_cluster_policy.go
Original file line number Diff line number Diff line change
Expand Up @@ -3,9 +3,12 @@
package schema

type ResourceClusterPolicy struct {
Definition string `json:"definition"`
Id string `json:"id,omitempty"`
MaxClustersPerUser int `json:"max_clusters_per_user,omitempty"`
Name string `json:"name"`
PolicyId string `json:"policy_id,omitempty"`
Definition string `json:"definition,omitempty"`
Description string `json:"description,omitempty"`
Id string `json:"id,omitempty"`
MaxClustersPerUser int `json:"max_clusters_per_user,omitempty"`
Name string `json:"name"`
PolicyFamilyDefinitionOverrides string `json:"policy_family_definition_overrides,omitempty"`
PolicyFamilyId string `json:"policy_family_id,omitempty"`
PolicyId string `json:"policy_id,omitempty"`
}
Loading