Skip to content

Commit

Permalink
Namespace substitution vars (#408)
Browse files Browse the repository at this point in the history
closes #404
  • Loading branch information
sujuka99 authored Jan 11, 2024
1 parent 3a86e04 commit 8f5af65
Show file tree
Hide file tree
Showing 51 changed files with 285 additions and 265 deletions.
16 changes: 8 additions & 8 deletions docs/docs/resources/pipeline-components/helm-app.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -3,16 +3,16 @@
name: helm-app # required
# Pipeline prefix that will prefix every component name. If you wish to not
# have any prefix you can specify an empty string.
prefix: ${pipeline_name}-
prefix: ${pipeline.name}-
from: # Must not be null
topics: # read from topic
${pipeline_name}-input-topic:
${pipeline.name}-input-topic:
type: input # Implied when role is NOT specified
${pipeline_name}-extra-topic:
${pipeline.name}-extra-topic:
role: topic-role # Implies `type` to be extra
${pipeline_name}-input-pattern-topic:
${pipeline.name}-input-pattern-topic:
type: pattern # Implied to be an input pattern if `role` is undefined
${pipeline_name}-extra-pattern-topic:
${pipeline.name}-extra-pattern-topic:
type: pattern # Implied to be an extra pattern if `role` is defined
role: some-role
components: # read from specific component
Expand All @@ -28,11 +28,11 @@
# Topic(s) into which the component will write output
to:
topics:
${pipeline_name}-output-topic:
${pipeline.name}-output-topic:
type: output # Implied when role is NOT specified
${pipeline_name}-extra-topic:
${pipeline.name}-extra-topic:
role: topic-role # Implies `type` to be extra; Will throw an error if `type` is defined
${pipeline_name}-error-topic:
${pipeline.name}-error-topic:
type: error
# Currently KPOps supports Avro and JSON schemas.
key_schema: key-schema # must implement SchemaProvider to use
Expand Down
20 changes: 10 additions & 10 deletions docs/docs/resources/pipeline-components/kafka-app.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -4,16 +4,16 @@
name: kafka-app # required
# Pipeline prefix that will prefix every component name. If you wish to not
# have any prefix you can specify an empty string.
prefix: ${pipeline_name}-
prefix: ${pipeline.name}-
from: # Must not be null
topics: # read from topic
${pipeline_name}-input-topic:
${pipeline.name}-input-topic:
type: input # Implied when role is NOT specified
${pipeline_name}-extra-topic:
${pipeline.name}-extra-topic:
role: topic-role # Implies `type` to be extra
${pipeline_name}-input-pattern-topic:
${pipeline.name}-input-pattern-topic:
type: pattern # Implied to be an input pattern if `role` is undefined
${pipeline_name}-extra-pattern-topic:
${pipeline.name}-extra-pattern-topic:
type: pattern # Implied to be an extra pattern if `role` is defined
role: some-role
components: # read from specific component
Expand All @@ -29,11 +29,11 @@
# Topic(s) into which the component will write output
to:
topics:
${pipeline_name}-output-topic:
${pipeline.name}-output-topic:
type: output # Implied when role is NOT specified
${pipeline_name}-extra-topic:
${pipeline.name}-extra-topic:
role: topic-role # Implies `type` to be extra; Will throw an error if `type` is defined
${pipeline_name}-error-topic:
${pipeline.name}-error-topic:
type: error
# Currently KPOps supports Avro and JSON schemas.
key_schema: key-schema # must implement SchemaProvider to use
Expand All @@ -49,8 +49,8 @@
# add the key-value pairs they need.
app: # required
streams: # required
brokers: ${kafka_brokers} # required
schemaRegistryUrl: ${schema_registry.url}
brokers: ${config.kafka_brokers} # required
schemaRegistryUrl: ${config.schema_registry.url}
nameOverride: override-with-this-name # kafka-app-specific
imageTag: "1.0.0" # Example values that are shared between streams-app and producer-app
# Helm repository configuration (optional)
Expand Down
16 changes: 8 additions & 8 deletions docs/docs/resources/pipeline-components/kafka-connector.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -2,16 +2,16 @@
name: kafka-connector # required
# Pipeline prefix that will prefix every component name. If you wish to not
# have any prefix you can specify an empty string.
prefix: ${pipeline_name}-
prefix: ${pipeline.name}-
from: # Must not be null
topics: # read from topic
${pipeline_name}-input-topic:
${pipeline.name}-input-topic:
type: input # Implied when role is NOT specified
${pipeline_name}-extra-topic:
${pipeline.name}-extra-topic:
role: topic-role # Implies `type` to be extra
${pipeline_name}-input-pattern-topic:
${pipeline.name}-input-pattern-topic:
type: pattern # Implied to be an input pattern if `role` is undefined
${pipeline_name}-extra-pattern-topic:
${pipeline.name}-extra-pattern-topic:
type: pattern # Implied to be an extra pattern if `role` is defined
role: some-role
components: # read from specific component
Expand All @@ -27,11 +27,11 @@
# Topic(s) into which the component will write output
to:
topics:
${pipeline_name}-output-topic:
${pipeline.name}-output-topic:
type: output # Implied when role is NOT specified
${pipeline_name}-extra-topic:
${pipeline.name}-extra-topic:
role: topic-role # Implies `type` to be extra; Will throw an error if `type` is defined
${pipeline_name}-error-topic:
${pipeline.name}-error-topic:
type: error
# Currently KPOps supports Avro and JSON schemas.
key_schema: key-schema # must implement SchemaProvider to use
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -3,16 +3,16 @@
name: kafka-sink-connector # required
# Pipeline prefix that will prefix every component name. If you wish to not
# have any prefix you can specify an empty string.
prefix: ${pipeline_name}-
prefix: ${pipeline.name}-
from: # Must not be null
topics: # read from topic
${pipeline_name}-input-topic:
${pipeline.name}-input-topic:
type: input # Implied when role is NOT specified
${pipeline_name}-extra-topic:
${pipeline.name}-extra-topic:
role: topic-role # Implies `type` to be extra
${pipeline_name}-input-pattern-topic:
${pipeline.name}-input-pattern-topic:
type: pattern # Implied to be an input pattern if `role` is undefined
${pipeline_name}-extra-pattern-topic:
${pipeline.name}-extra-pattern-topic:
type: pattern # Implied to be an extra pattern if `role` is defined
role: some-role
components: # read from specific component
Expand All @@ -28,11 +28,11 @@
# Topic(s) into which the component will write output
to:
topics:
${pipeline_name}-output-topic:
${pipeline.name}-output-topic:
type: output # Implied when role is NOT specified
${pipeline_name}-extra-topic:
${pipeline.name}-extra-topic:
role: topic-role # Implies `type` to be extra; Will throw an error if `type` is defined
${pipeline_name}-error-topic:
${pipeline.name}-error-topic:
type: error
# Currently KPOps supports Avro and JSON schemas.
key_schema: key-schema # must implement SchemaProvider to use
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -3,17 +3,17 @@
name: kafka-source-connector # required
# Pipeline prefix that will prefix every component name. If you wish to not
# have any prefix you can specify an empty string.
prefix: ${pipeline_name}-
prefix: ${pipeline.name}-
# The source connector has no `from` section
# from:
# Topic(s) into which the component will write output
to:
topics:
${pipeline_name}-output-topic:
${pipeline.name}-output-topic:
type: output # Implied when role is NOT specified
${pipeline_name}-extra-topic:
${pipeline.name}-extra-topic:
role: topic-role # Implies `type` to be extra; Will throw an error if `type` is defined
${pipeline_name}-error-topic:
${pipeline.name}-error-topic:
type: error
# Currently KPOps supports Avro and JSON schemas.
key_schema: key-schema # must implement SchemaProvider to use
Expand Down
16 changes: 8 additions & 8 deletions docs/docs/resources/pipeline-components/kubernetes-app.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -3,16 +3,16 @@
name: kubernetes-app # required
# Pipeline prefix that will prefix every component name. If you wish to not
# have any prefix you can specify an empty string.
prefix: ${pipeline_name}-
prefix: ${pipeline.name}-
from: # Must not be null
topics: # read from topic
${pipeline_name}-input-topic:
${pipeline.name}-input-topic:
type: input # Implied when role is NOT specified
${pipeline_name}-extra-topic:
${pipeline.name}-extra-topic:
role: topic-role # Implies `type` to be extra
${pipeline_name}-input-pattern-topic:
${pipeline.name}-input-pattern-topic:
type: pattern # Implied to be an input pattern if `role` is undefined
${pipeline_name}-extra-pattern-topic:
${pipeline.name}-extra-pattern-topic:
type: pattern # Implied to be an extra pattern if `role` is defined
role: some-role
components: # read from specific component
Expand All @@ -28,11 +28,11 @@
# Topic(s) into which the component will write output
to:
topics:
${pipeline_name}-output-topic:
${pipeline.name}-output-topic:
type: output # Implied when role is NOT specified
${pipeline_name}-extra-topic:
${pipeline.name}-extra-topic:
role: topic-role # Implies `type` to be extra; Will throw an error if `type` is defined
${pipeline_name}-error-topic:
${pipeline.name}-error-topic:
type: error
# Currently KPOps supports Avro and JSON schemas.
key_schema: key-schema # must implement SchemaProvider to use
Expand Down
Loading

0 comments on commit 8f5af65

Please sign in to comment.