This repository has been archived by the owner on Aug 27, 2024. It is now read-only.
-
Notifications
You must be signed in to change notification settings - Fork 0
/
variables.tf
147 lines (124 loc) · 3.81 KB
/
variables.tf
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
variable "name" {
description = "Will be prefixed to all resource names. Use to easily identify the resources created."
type = string
default = "loader"
}
variable "project_id" {
description = "The GCP project ID."
type = string
}
variable "region" {
description = "The name of the region to deploy within."
type = string
}
variable "zone" {
description = "The zone in which to deploy the instances."
}
variable "network" {
description = "The name of the network to deploy within."
type = string
}
variable "subnetwork" {
description = "The name of the sub-network to deploy within; if populated will override the 'network' setting."
type = string
default = ""
}
variable "machine_type" {
description = "The machine type to use"
type = string
default = "e2-micro"
}
variable "associate_public_ip_address" {
description = "Whether to assign a public ip address to this instance; if false this instance must be behind a Cloud NAT to connect to the internet"
type = bool
default = true
}
variable "ssh_ip_allowlist" {
description = "The list of CIDR ranges to allow SSH traffic from"
type = list(any)
default = [""]
}
variable "ssh_block_project_keys" {
description = "Whether to block project wide SSH keys"
type = bool
default = true
}
variable "ssh_key_pairs" {
description = "The list of SSH key-pairs to add to the servers"
default = []
type = list(object({
user_name = string
public_key = string
}))
}
variable "ubuntu_20_04_source_image" {
description = "The source image to use which must be based of of Ubuntu 20.04; by default the latest community version is used"
default = ""
type = string
}
variable "labels" {
description = "The labels to append to this resource"
default = {}
type = map(string)
}
variable "gcp_logs_enabled" {
description = "Whether application logs should be reported to GCP Logging"
default = true
type = bool
}
variable "images" {
description = <<EOH
The docker images with version tag to deploy on Compute Engine's instances. See here for details:
https://docs.snowplowanalytics.com/docs/pipeline-components-and-applications/loaders-storage-targets/bigquery-loader/
The default is to launch all three apps: 'Stream Loader', 'Mutator' and 'Repeater'.
EOH
type = list(string)
default = [
"snowplow/snowplow-bigquery-streamloader:1.3.2",
"snowplow/snowplow-bigquery-repeater:1.3.2",
"snowplow/snowplow-bigquery-mutator:1.3.2"
]
}
variable "enriched_topic_id" {
description = "The pubsub topic to read enriched messages from."
type = string
}
variable "tags" {
description = "The tags to apply to the created resources."
type = list(string)
default = []
}
variable "telemetry_enabled" {
description = "Whether or not to send telemetry information back to Snowplow Analytics Ltd"
type = bool
default = true
}
variable "user_provided_id" {
description = "An optional unique identifier to identify the telemetry events emitted by this stack"
type = string
default = ""
}
variable "dataset_config" {
description = "The dataset in which to load the Snowplow events. Created by default."
type = object({
name = string
create = bool
})
default = {
name = "snowplow"
create = true
}
}
variable "table_config" {
description = "The table in which to load the Snowplow events. Created by default."
type = object({
name = string
load_timestamp_column = string
load_timestamp_column_partition = string
})
default = {
name = "events"
load_timestamp_column = "load_tstamp"
load_timestamp_column_partition = null
}
}