263 lines
9.2 KiB
Go
263 lines
9.2 KiB
Go
package mongodbflex
|
|
|
|
import (
|
|
"context"
|
|
"fmt"
|
|
"net/http"
|
|
|
|
"github.com/stackitcloud/terraform-provider-stackit/stackit/internal/conversion"
|
|
mongodbflexUtils "github.com/stackitcloud/terraform-provider-stackit/stackit/internal/services/mongodbflex/utils"
|
|
|
|
"github.com/hashicorp/terraform-plugin-framework/datasource"
|
|
"github.com/hashicorp/terraform-plugin-framework/schema/validator"
|
|
"github.com/hashicorp/terraform-plugin-framework/types/basetypes"
|
|
"github.com/hashicorp/terraform-plugin-log/tflog"
|
|
"github.com/stackitcloud/terraform-provider-stackit/stackit/internal/core"
|
|
"github.com/stackitcloud/terraform-provider-stackit/stackit/internal/utils"
|
|
"github.com/stackitcloud/terraform-provider-stackit/stackit/internal/validate"
|
|
|
|
"github.com/hashicorp/terraform-plugin-framework/datasource/schema"
|
|
"github.com/hashicorp/terraform-plugin-framework/types"
|
|
"github.com/stackitcloud/stackit-sdk-go/services/mongodbflex"
|
|
)
|
|
|
|
// Ensure the implementation satisfies the expected interfaces.
|
|
var (
|
|
_ datasource.DataSource = &instanceDataSource{}
|
|
)
|
|
|
|
// NewInstanceDataSource is a helper function to simplify the provider implementation.
|
|
func NewInstanceDataSource() datasource.DataSource {
|
|
return &instanceDataSource{}
|
|
}
|
|
|
|
// instanceDataSource is the data source implementation.
|
|
type instanceDataSource struct {
|
|
client *mongodbflex.APIClient
|
|
providerData core.ProviderData
|
|
}
|
|
|
|
// Metadata returns the data source type name.
|
|
func (d *instanceDataSource) Metadata(_ context.Context, req datasource.MetadataRequest, resp *datasource.MetadataResponse) {
|
|
resp.TypeName = req.ProviderTypeName + "_mongodbflex_instance"
|
|
}
|
|
|
|
// Configure adds the provider configured client to the data source.
|
|
func (d *instanceDataSource) Configure(ctx context.Context, req datasource.ConfigureRequest, resp *datasource.ConfigureResponse) {
|
|
var ok bool
|
|
d.providerData, ok = conversion.ParseProviderData(ctx, req.ProviderData, &resp.Diagnostics)
|
|
if !ok {
|
|
return
|
|
}
|
|
|
|
apiClient := mongodbflexUtils.ConfigureClient(ctx, &d.providerData, &resp.Diagnostics)
|
|
if resp.Diagnostics.HasError() {
|
|
return
|
|
}
|
|
d.client = apiClient
|
|
tflog.Info(ctx, "MongoDB Flex instance client configured")
|
|
}
|
|
|
|
// Schema defines the schema for the data source.
|
|
func (d *instanceDataSource) Schema(_ context.Context, _ datasource.SchemaRequest, resp *datasource.SchemaResponse) {
|
|
descriptions := map[string]string{
|
|
"main": "MongoDB Flex instance data source schema. Must have a `region` specified in the provider configuration.",
|
|
"id": "Terraform's internal data source ID. It is structured as \"`project_id`,`region`,`instance_id`\".",
|
|
"instance_id": "ID of the MongoDB Flex instance.",
|
|
"project_id": "STACKIT project ID to which the instance is associated.",
|
|
"name": "Instance name.",
|
|
"acl": "The Access Control List (ACL) for the MongoDB Flex instance.",
|
|
"backup_schedule": `The backup schedule. Should follow the cron scheduling system format (e.g. "0 0 * * *").`,
|
|
"options": "Custom parameters for the MongoDB Flex instance.",
|
|
"type": "Type of the MongoDB Flex instance.",
|
|
"snapshot_retention_days": "The number of days that continuous backups (controlled via the `backup_schedule`) will be retained.",
|
|
"daily_snapshot_retention_days": "The number of days that daily backups will be retained.",
|
|
"weekly_snapshot_retention_weeks": "The number of weeks that weekly backups will be retained.",
|
|
"monthly_snapshot_retention_months": "The number of months that monthly backups will be retained.",
|
|
"point_in_time_window_hours": "The number of hours back in time the point-in-time recovery feature will be able to recover.",
|
|
"region": "The resource region. If not defined, the provider region is used.",
|
|
}
|
|
|
|
resp.Schema = schema.Schema{
|
|
Description: descriptions["main"],
|
|
Attributes: map[string]schema.Attribute{
|
|
"id": schema.StringAttribute{
|
|
Description: descriptions["id"],
|
|
Computed: true,
|
|
},
|
|
"instance_id": schema.StringAttribute{
|
|
Description: descriptions["instance_id"],
|
|
Required: true,
|
|
Validators: []validator.String{
|
|
validate.UUID(),
|
|
validate.NoSeparator(),
|
|
},
|
|
},
|
|
"project_id": schema.StringAttribute{
|
|
Description: descriptions["project_id"],
|
|
Required: true,
|
|
Validators: []validator.String{
|
|
validate.UUID(),
|
|
validate.NoSeparator(),
|
|
},
|
|
},
|
|
"name": schema.StringAttribute{
|
|
Description: descriptions["name"],
|
|
Computed: true,
|
|
},
|
|
"acl": schema.ListAttribute{
|
|
Description: descriptions["acl"],
|
|
ElementType: types.StringType,
|
|
Computed: true,
|
|
},
|
|
"backup_schedule": schema.StringAttribute{
|
|
Description: descriptions["backup_schedule"],
|
|
Computed: true,
|
|
},
|
|
"flavor": schema.SingleNestedAttribute{
|
|
Computed: true,
|
|
Attributes: map[string]schema.Attribute{
|
|
"id": schema.StringAttribute{
|
|
Computed: true,
|
|
},
|
|
"description": schema.StringAttribute{
|
|
Computed: true,
|
|
},
|
|
"cpu": schema.Int64Attribute{
|
|
Computed: true,
|
|
},
|
|
"ram": schema.Int64Attribute{
|
|
Computed: true,
|
|
},
|
|
},
|
|
},
|
|
"replicas": schema.Int64Attribute{
|
|
Computed: true,
|
|
},
|
|
"storage": schema.SingleNestedAttribute{
|
|
Computed: true,
|
|
Attributes: map[string]schema.Attribute{
|
|
"class": schema.StringAttribute{
|
|
Computed: true,
|
|
},
|
|
"size": schema.Int64Attribute{
|
|
Computed: true,
|
|
},
|
|
},
|
|
},
|
|
"version": schema.StringAttribute{
|
|
Computed: true,
|
|
},
|
|
"options": schema.SingleNestedAttribute{
|
|
Description: descriptions["options"],
|
|
Computed: true,
|
|
Attributes: map[string]schema.Attribute{
|
|
"type": schema.StringAttribute{
|
|
Description: descriptions["type"],
|
|
Computed: true,
|
|
},
|
|
"snapshot_retention_days": schema.Int64Attribute{
|
|
Description: descriptions["snapshot_retention_days"],
|
|
Computed: true,
|
|
},
|
|
"daily_snapshot_retention_days": schema.Int64Attribute{
|
|
Description: descriptions["daily_snapshot_retention_days"],
|
|
Computed: true,
|
|
},
|
|
"weekly_snapshot_retention_weeks": schema.Int64Attribute{
|
|
Description: descriptions["weekly_snapshot_retention_weeks"],
|
|
Computed: true,
|
|
},
|
|
"monthly_snapshot_retention_months": schema.Int64Attribute{
|
|
Description: descriptions["monthly_snapshot_retention_months"],
|
|
Computed: true,
|
|
},
|
|
"point_in_time_window_hours": schema.Int64Attribute{
|
|
Description: descriptions["point_in_time_window_hours"],
|
|
Computed: true,
|
|
},
|
|
},
|
|
},
|
|
"region": schema.StringAttribute{
|
|
Optional: true,
|
|
// must be computed to allow for storing the override value from the provider
|
|
Computed: true,
|
|
Description: descriptions["region"],
|
|
},
|
|
},
|
|
}
|
|
}
|
|
|
|
// Read refreshes the Terraform state with the latest data.
|
|
func (d *instanceDataSource) Read(ctx context.Context, req datasource.ReadRequest, resp *datasource.ReadResponse) { // nolint:gocritic // function signature required by Terraform
|
|
var model Model
|
|
diags := req.Config.Get(ctx, &model)
|
|
resp.Diagnostics.Append(diags...)
|
|
if resp.Diagnostics.HasError() {
|
|
return
|
|
}
|
|
|
|
ctx = core.InitProviderContext(ctx)
|
|
|
|
projectId := model.ProjectId.ValueString()
|
|
region := d.providerData.GetRegionWithOverride(model.Region)
|
|
instanceId := model.InstanceId.ValueString()
|
|
ctx = tflog.SetField(ctx, "project_id", projectId)
|
|
ctx = tflog.SetField(ctx, "region", region)
|
|
ctx = tflog.SetField(ctx, "instance_id", instanceId)
|
|
instanceResp, err := d.client.GetInstance(ctx, projectId, instanceId, region).Execute()
|
|
if err != nil {
|
|
utils.LogError(
|
|
ctx,
|
|
&resp.Diagnostics,
|
|
err,
|
|
"Reading instance",
|
|
fmt.Sprintf("Instance with ID %q does not exist in project %q.", instanceId, projectId),
|
|
map[int]string{
|
|
http.StatusForbidden: fmt.Sprintf("Project with ID %q not found or forbidden access", projectId),
|
|
},
|
|
)
|
|
resp.State.RemoveResource(ctx)
|
|
return
|
|
}
|
|
|
|
ctx = core.LogResponse(ctx)
|
|
|
|
var flavor = &flavorModel{}
|
|
if !(model.Flavor.IsNull() || model.Flavor.IsUnknown()) {
|
|
diags = model.Flavor.As(ctx, flavor, basetypes.ObjectAsOptions{})
|
|
resp.Diagnostics.Append(diags...)
|
|
if resp.Diagnostics.HasError() {
|
|
return
|
|
}
|
|
}
|
|
var storage = &storageModel{}
|
|
if !(model.Storage.IsNull() || model.Storage.IsUnknown()) {
|
|
diags = model.Storage.As(ctx, storage, basetypes.ObjectAsOptions{})
|
|
resp.Diagnostics.Append(diags...)
|
|
if resp.Diagnostics.HasError() {
|
|
return
|
|
}
|
|
}
|
|
var options = &optionsModel{}
|
|
if !(model.Options.IsNull() || model.Options.IsUnknown()) {
|
|
diags = model.Options.As(ctx, options, basetypes.ObjectAsOptions{})
|
|
resp.Diagnostics.Append(diags...)
|
|
if resp.Diagnostics.HasError() {
|
|
return
|
|
}
|
|
}
|
|
|
|
err = mapFields(ctx, instanceResp, &model, flavor, storage, options, region)
|
|
if err != nil {
|
|
core.LogAndAddError(ctx, &resp.Diagnostics, "Error reading instance", fmt.Sprintf("Processing API payload: %v", err))
|
|
return
|
|
}
|
|
// Set refreshed state
|
|
diags = resp.State.Set(ctx, model)
|
|
resp.Diagnostics.Append(diags...)
|
|
if resp.Diagnostics.HasError() {
|
|
return
|
|
}
|
|
tflog.Info(ctx, "MongoDB Flex instance read")
|
|
}
|