package postgresflexalpha import ( "context" "fmt" "math" "net/http" "github.com/hashicorp/terraform-plugin-framework/datasource/schema" "github.com/hashicorp/terraform-plugin-framework/diag" postgresflex "tf-provider.git.onstackit.cloud/stackit-dev-tools/terraform-provider-stackitprivatepreview/pkg_gen/postgresflexalpha" "tf-provider.git.onstackit.cloud/stackit-dev-tools/terraform-provider-stackitprivatepreview/stackit/internal/conversion" postgresflexalpha "tf-provider.git.onstackit.cloud/stackit-dev-tools/terraform-provider-stackitprivatepreview/stackit/internal/services/postgresflexalpha/user/datasources_gen" postgresflexUtils "tf-provider.git.onstackit.cloud/stackit-dev-tools/terraform-provider-stackitprivatepreview/stackit/internal/services/postgresflexalpha/utils" "github.com/hashicorp/terraform-plugin-framework/datasource" "github.com/hashicorp/terraform-plugin-framework/types" "github.com/hashicorp/terraform-plugin-log/tflog" "tf-provider.git.onstackit.cloud/stackit-dev-tools/terraform-provider-stackitprivatepreview/stackit/internal/core" "tf-provider.git.onstackit.cloud/stackit-dev-tools/terraform-provider-stackitprivatepreview/stackit/internal/utils" ) // Ensure the implementation satisfies the expected interfaces. var ( _ datasource.DataSource = &userDataSource{} ) // NewUserDataSource is a helper function to simplify the provider implementation. func NewUserDataSource() datasource.DataSource { return &userDataSource{} } // dataSourceModel maps the data source schema data. type dataSourceModel struct { postgresflexalpha.UserModel TerraformID types.String `tfsdk:"id"` } // userDataSource is the data source implementation. type userDataSource struct { client *postgresflex.APIClient providerData core.ProviderData } // Metadata returns the data source type name. func (r *userDataSource) Metadata( _ context.Context, req datasource.MetadataRequest, resp *datasource.MetadataResponse, ) { resp.TypeName = req.ProviderTypeName + "_postgresflexalpha_user" } // Configure adds the provider configured client to the data source. func (r *userDataSource) Configure( ctx context.Context, req datasource.ConfigureRequest, resp *datasource.ConfigureResponse, ) { var ok bool r.providerData, ok = conversion.ParseProviderData(ctx, req.ProviderData, &resp.Diagnostics) if !ok { return } apiClient := postgresflexUtils.ConfigureClient(ctx, &r.providerData, &resp.Diagnostics) if resp.Diagnostics.HasError() { return } r.client = apiClient tflog.Info(ctx, "Postgres Flex user client configured") } // Schema defines the schema for the data source. func (r *userDataSource) Schema(ctx context.Context, _ datasource.SchemaRequest, resp *datasource.SchemaResponse) { s := postgresflexalpha.UserDataSourceSchema(ctx) s.Attributes["id"] = schema.StringAttribute{ Description: "Terraform's internal resource ID. It is structured as \\\"`project_id`,`region`,`instance_id`," + "`user_id`\\\".\",", Optional: true, Computed: true, } resp.Schema = s } // Read refreshes the Terraform state with the latest data. func (r *userDataSource) Read( ctx context.Context, req datasource.ReadRequest, resp *datasource.ReadResponse, ) { // nolint:gocritic // function signature required by Terraform var model dataSourceModel diags := req.Config.Get(ctx, &model) resp.Diagnostics.Append(diags...) if resp.Diagnostics.HasError() { return } ctx = core.InitProviderContext(ctx) projectId := model.ProjectId.ValueString() instanceId := model.InstanceId.ValueString() userId64 := model.UserId.ValueInt64() if userId64 > math.MaxInt32 { core.LogAndAddError(ctx, &resp.Diagnostics, "Error in type conversion", "int value too large (userId)") return } userId := int32(userId64) region := r.providerData.GetRegionWithOverride(model.Region) ctx = tflog.SetField(ctx, "project_id", projectId) ctx = tflog.SetField(ctx, "instance_id", instanceId) ctx = tflog.SetField(ctx, "region", region) ctx = tflog.SetField(ctx, "user_id", userId) recordSetResp, err := r.client.GetUserRequest(ctx, projectId, region, instanceId, userId).Execute() if err != nil { handleReadError(ctx, &diags, err, projectId, instanceId, userId) resp.State.RemoveResource(ctx) return } ctx = core.LogResponse(ctx) // Map response body to schema and populate Computed attribute values err = mapDataSourceFields(recordSetResp, &model, region) if err != nil { core.LogAndAddError( ctx, &resp.Diagnostics, "Error reading user", fmt.Sprintf("Processing API payload: %v", err), ) return } // Set refreshed state diags = resp.State.Set(ctx, model) resp.Diagnostics.Append(diags...) if resp.Diagnostics.HasError() { return } tflog.Info(ctx, "Postgres Flex user read") } // handleReadError centralizes API error handling for the Read operation. func handleReadError( ctx context.Context, diags *diag.Diagnostics, err error, projectId, instanceId string, userId int32, ) { utils.LogError( ctx, diags, err, "Reading user", fmt.Sprintf( "User with ID %q or instance with ID %q does not exist in project %q.", userId, instanceId, projectId, ), map[int]string{ http.StatusBadRequest: fmt.Sprintf( "Invalid user request parameters for project %q and instance %q.", projectId, instanceId, ), http.StatusNotFound: fmt.Sprintf( "User, instance %q, or project %q or user %q not found.", instanceId, projectId, userId, ), http.StatusForbidden: fmt.Sprintf("Forbidden access to project %q.", projectId), }, ) }