Skip to content

Commit 36085a8

Browse files
committed
formatting
1 parent 9aa894e commit 36085a8

File tree

5 files changed

+39
-39
lines changed

5 files changed

+39
-39
lines changed

terraform/README.adoc

Lines changed: 4 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -18,7 +18,8 @@ export CONFLUENT_CLOUD_API_SECRET<API SECRET>
1818

1919
== Execute Terraform Manifests
2020

21-
The terraform manifests require the Confluent Cloud organization ID in order to provision infrastructure. This can be found in the Confluent Cloud console in the "Organization Settings" and exported to an environment variable:
21+
The terraform manifests require the Confluent Cloud organization ID in order to provision infrastructure.
22+
This can be found in the Confluent Cloud console in the "Organization Settings" and exported to an environment variable:
2223

2324
```bash
2425
export TF_VAR_org_id=<ORG ID VALUE FROM CONSOLE>
@@ -42,7 +43,8 @@ Once completed, verify the infrastructure is created in the Confluent Cloud cons
4243

4344
== Using Infrastructure in the Workshop
4445

45-
In the Table API exercises, we'll need API keys and secrets to connect Flink to Confluent Cloud. This `terraform output` command will create a file with those parameters:
46+
In the Table API exercises, we will need API keys and secrets to connect Flink to Confluent Cloud.
47+
This `terraform output` command will create a file with those parameters:
4648

4749
```bash
4850
terraform output -json \

terraform/flink-compute.tf

Lines changed: 9 additions & 11 deletions
Original file line numberDiff line numberDiff line change
@@ -1,18 +1,16 @@
1-
2-
31
resource "confluent_flink_compute_pool" "main_flink_pool" {
4-
display_name = "main_flink_pool"
5-
cloud = var.cloud_provider
6-
region = var.cloud_region
7-
max_cfu = 5
2+
display_name = "main_flink_pool"
3+
cloud = var.cloud_provider
4+
region = var.cloud_region
5+
max_cfu = 5
86
environment {
97
id = confluent_environment.cc_env.id
108
}
119
}
1210

1311
data "confluent_flink_region" "main_flink_region" {
14-
cloud = var.cloud_provider
15-
region = var.cloud_region
12+
cloud = var.cloud_provider
13+
region = var.cloud_region
1614
}
1715

1816
resource "confluent_service_account" "flink_developer" {
@@ -25,23 +23,23 @@ resource "confluent_role_binding" "fd_flink_developer" {
2523
role_name = "FlinkDeveloper"
2624
crn_pattern = confluent_environment.cc_env.resource_name
2725

28-
depends_on = [ confluent_flink_compute_pool.main_flink_pool]
26+
depends_on = [confluent_flink_compute_pool.main_flink_pool]
2927
}
3028

3129
resource "confluent_role_binding" "fd_kafka_write" {
3230
principal = "User:${confluent_service_account.flink_developer.id}"
3331
role_name = "DeveloperWrite"
3432
crn_pattern = "${confluent_kafka_cluster.kafka_cluster.rbac_crn}/kafka=${confluent_kafka_cluster.kafka_cluster.id}/topic=*"
3533

36-
depends_on = [ confluent_kafka_cluster.kafka_cluster]
34+
depends_on = [confluent_kafka_cluster.kafka_cluster]
3735
}
3836

3937
resource "confluent_role_binding" "fd_kafka_read" {
4038
principal = "User:${confluent_service_account.flink_developer.id}"
4139
role_name = "DeveloperRead"
4240
crn_pattern = "${confluent_kafka_cluster.kafka_cluster.rbac_crn}/kafka=${confluent_kafka_cluster.kafka_cluster.id}/topic=*"
4341

44-
depends_on = [ confluent_kafka_cluster.kafka_cluster]
42+
depends_on = [confluent_kafka_cluster.kafka_cluster]
4543
}
4644

4745
resource "confluent_role_binding" "fd_schema_registry_write" {

terraform/output.tf

Lines changed: 11 additions & 11 deletions
Original file line numberDiff line numberDiff line change
@@ -7,45 +7,45 @@ output "kafka_boostrap_servers" {
77
}
88

99
output "cloud" {
10-
value = var.cloud_provider
10+
value = var.cloud_provider
1111
}
1212

1313
output "region" {
14-
value = var.cloud_region
14+
value = var.cloud_region
1515
}
1616

1717
output "organization-id" {
18-
value = replace(var.org_id, "\"", "")
18+
value = replace(var.org_id, "\"", "")
1919
}
2020

2121
output "environment-id" {
22-
value = confluent_environment.cc_env.id
22+
value = confluent_environment.cc_env.id
2323
}
2424

2525
output "compute-pool-id" {
26-
value = confluent_flink_compute_pool.main_flink_pool.id
26+
value = confluent_flink_compute_pool.main_flink_pool.id
2727
}
2828

2929
output "kafka_sasl_jaas_config" {
30-
value = "org.apache.kafka.common.security.plain.PlainLoginModule required username='${confluent_api_key.kafka_developer_kafka_api_key.id}' password='${nonsensitive(confluent_api_key.kafka_developer_kafka_api_key.secret)}';"
30+
value = "org.apache.kafka.common.security.plain.PlainLoginModule required username='${confluent_api_key.kafka_developer_kafka_api_key.id}' password='${nonsensitive(confluent_api_key.kafka_developer_kafka_api_key.secret)}';"
3131
}
3232

3333
output "registry_url" {
34-
value = data.confluent_schema_registry_cluster.advanced.rest_endpoint
34+
value = data.confluent_schema_registry_cluster.advanced.rest_endpoint
3535
}
3636

3737
output "registry_key" {
38-
value = confluent_api_key.sr_manager_kafka_api_key.id
38+
value = confluent_api_key.sr_manager_kafka_api_key.id
3939
}
4040

4141
output "registry_secret" {
42-
value = nonsensitive(confluent_api_key.sr_manager_kafka_api_key.secret)
42+
value = nonsensitive(confluent_api_key.sr_manager_kafka_api_key.secret)
4343
}
4444

4545
output "flink-api-key" {
46-
value = confluent_api_key.flink_developer_api_key.id
46+
value = confluent_api_key.flink_developer_api_key.id
4747
}
4848

4949
output "flink-api-secret" {
50-
value = nonsensitive(confluent_api_key.flink_developer_api_key.secret)
50+
value = nonsensitive(confluent_api_key.flink_developer_api_key.secret)
5151
}

terraform/schema-registry.tf

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -19,7 +19,7 @@ resource "confluent_role_binding" "sr_manager_data_steward" {
1919
role_name = "DataSteward"
2020
crn_pattern = confluent_environment.cc_env.resource_name
2121

22-
depends_on = [ data.confluent_schema_registry_cluster.advanced ]
22+
depends_on = [data.confluent_schema_registry_cluster.advanced]
2323
}
2424

2525
resource "confluent_api_key" "sr_manager_kafka_api_key" {

terraform/variables.tf

Lines changed: 14 additions & 14 deletions
Original file line numberDiff line numberDiff line change
@@ -1,34 +1,34 @@
11
variable "org_id" {
2-
type = string
2+
type = string
33
description = "CC Organization ID"
44
}
55

66
variable "cloud_provider" {
7-
type = string
7+
type = string
88
description = "cloud provider for Confluent Cloud"
9-
default = "AWS"
9+
default = "AWS"
1010
}
1111

1212
variable "cloud_region" {
13-
type = string
13+
type = string
1414
description = "cloud provider region"
15-
default = "us-east-2"
15+
default = "us-east-2"
1616
}
1717

1818
variable "cc_env_name" {
19-
type = string
20-
description = "CC Environment Name"
21-
default = "java_flink_workshop"
19+
type = string
20+
description = "CC Environment Name"
21+
default = "java_flink_workshop"
2222
}
2323

2424
variable "cc_default_kafka_cluster_name" {
25-
type = string
26-
description = "Kafka Default Cluster Name"
27-
default = "workshop"
25+
type = string
26+
description = "Kafka Default Cluster Name"
27+
default = "workshop"
2828
}
2929

3030
variable "cc_default_flink_compute_pool_name" {
31-
type = string
32-
description = "Default Flink Compute Pool Name"
33-
default = "flink-compute-pool"
31+
type = string
32+
description = "Default Flink Compute Pool Name"
33+
default = "flink-compute-pool"
3434
}

0 commit comments

Comments
 (0)