Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Upgrade to Camel 4.4.2 #1620

Merged
merged 2 commits into from
May 6, 2024
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
The table of contents is too big for display.
Diff view
Diff view
  •  
  •  
  •  
Original file line number Diff line number Diff line change
Expand Up @@ -24,7 +24,7 @@
<parent>
<groupId>org.apache.camel.kafkaconnector</groupId>
<artifactId>archetypes</artifactId>
<version>4.1.0-SNAPSHOT</version>
<version>4.4.2-SNAPSHOT</version>
</parent>

<groupId>org.apache.camel.kafkaconnector.archetypes</groupId>
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -24,7 +24,7 @@
<parent>
<groupId>org.apache.camel.kafkaconnector</groupId>
<artifactId>archetypes</artifactId>
<version>4.1.0-SNAPSHOT</version>
<version>4.4.2-SNAPSHOT</version>
</parent>

<groupId>org.apache.camel.kafkaconnector.archetypes</groupId>
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -24,7 +24,7 @@
<parent>
<groupId>org.apache.camel.kafkaconnector</groupId>
<artifactId>archetypes</artifactId>
<version>4.1.0-SNAPSHOT</version>
<version>4.4.2-SNAPSHOT</version>
</parent>

<groupId>org.apache.camel.kafkaconnector.archetypes</groupId>
Expand Down
2 changes: 1 addition & 1 deletion archetypes/pom.xml
Original file line number Diff line number Diff line change
Expand Up @@ -22,7 +22,7 @@
<parent>
<groupId>org.apache.camel.kafkaconnector</groupId>
<artifactId>parent</artifactId>
<version>4.1.0-SNAPSHOT</version>
<version>4.4.2-SNAPSHOT</version>
<relativePath>../parent/pom.xml</relativePath>
</parent>

Expand Down
2 changes: 1 addition & 1 deletion buildingtools/pom.xml
Original file line number Diff line number Diff line change
Expand Up @@ -23,7 +23,7 @@
<parent>
<groupId>org.apache.camel.kafkaconnector</groupId>
<artifactId>parent</artifactId>
<version>4.1.0-SNAPSHOT</version>
<version>4.4.2-SNAPSHOT</version>
<relativePath>../parent/pom.xml</relativePath>
</parent>

Expand Down
2 changes: 1 addition & 1 deletion camel-kafka-connector-catalog/pom.xml
Original file line number Diff line number Diff line change
Expand Up @@ -23,7 +23,7 @@
<parent>
<groupId>org.apache.camel.kafkaconnector</groupId>
<artifactId>parent</artifactId>
<version>4.1.0-SNAPSHOT</version>
<version>4.4.2-SNAPSHOT</version>
<relativePath>../parent/pom.xml</relativePath>
</parent>

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,7 @@
"groupId": "org.apache.camel.kafkaconnector",
"id": "camel-aws-cloudtrail-source-source",
"type": "source",
"version": "4.1.0-SNAPSHOT",
"version": "4.4.2-SNAPSHOT",
"description": "Receive data from an AWS Cloudtrail.\n\nThe basic authentication method for the Cloudtrail service is to specify an access key and a secret key. These parameters are optional because the Kamelet provides a default credentials provider."
},
"properties": {
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,7 @@
"groupId": "org.apache.camel.kafkaconnector",
"id": "camel-aws-cloudwatch-sink-sink",
"type": "sink",
"version": "4.1.0-SNAPSHOT",
"version": "4.4.2-SNAPSHOT",
"description": "Send data to Amazon CloudWatch metrics.\n\nThe basic authentication method for the AWS CloudWatch metrics service is to specify an access key and a secret key. These parameters are optional because the Kamelet provides a default credentials provider.\n\nIf you use the default credentials provider, the CloudWatch client loads the credentials through this provider and doesn't use the basic authentication method.\n\nYou can set the following properties in the header:\n\n`metric-name` \/ `ce-metricname` for the metric name.\n`metric-value` \/ `ce-metricvalue` for the metric value.\n`metric-unit` \/ `ce-metricunit` for the metric unit.\n`metric-timestamp` \/ `ce-metrictimestamp` for the metric timestamp.\n`metric-dimension-name` \/ `ce-metricdimensionname` for the dimension name.\n`metric-dimension-value` \/ `ce-metricdimensionvalue` for the dimension value."
},
"properties": {
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,7 @@
"groupId": "org.apache.camel.kafkaconnector",
"id": "camel-aws-ddb-sink-sink",
"type": "sink",
"version": "4.1.0-SNAPSHOT",
"version": "4.4.2-SNAPSHOT",
"description": "Send data to Amazon DynamoDB. The sent data inserts, updates, or deletes an item on the specified AWS DynamoDB table.\n\nThe basic authentication method for the AWS DynamoDB service is to specify an access key and a secret key. These parameters are optional because the Kamelet provides a default credentials provider.\n\nIf you use the default credentials provider, the DynamoDB client loads the credentials through this provider and doesn't use the basic authentication method.\n\nThis Kamelet expects a JSON-formatted body and it must include the primary key values that define the DynamoDB item. The mapping between the JSON fields and table attribute values is done by key. For example, for '{\"username\":\"oscerd\", \"city\":\"Rome\"}' input, the Kamelet inserts or update an item in the specified AWS DynamoDB table and sets the values for the 'username' and 'city' attributes. \n\nFor PutItem operation the Json body defines all item attributes.\nFor DeleteItem operation the Json body defines only the primary key attributes that identify the item to delete.\nFor UpdateItem operation the Json body defines both key attributes to identify the item to be updated and all item attributes tht get updated on the item.\n\nThe given Json body can use \"operation\", \"key\" and \"item\" as top level properties that will be mapped to respective attribute value maps.\n{\n \"operation\": \"PutItem\"\n \"key\": {},\n \"item\": {}\n}"
},
"properties": {
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,7 @@
"groupId": "org.apache.camel.kafkaconnector",
"id": "camel-aws-ddb-streams-source-source",
"type": "source",
"version": "4.1.0-SNAPSHOT",
"version": "4.4.2-SNAPSHOT",
"description": "Receive events from Amazon DynamoDB Streams.\n\nThe basic authentication method for the AWS DynamoDB Streams service is to specify an access key and a secret key. These parameters are optional because the Kamelet provides a default credentials provider.\n\nIf you use the default credentials provider, the DynamoDB Streams client loads the credentials through this provider and doesn't use the basic authentication method."
},
"properties": {
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,7 @@
"groupId": "org.apache.camel.kafkaconnector",
"id": "camel-aws-ec2-sink-sink",
"type": "sink",
"version": "4.1.0-SNAPSHOT",
"version": "4.4.2-SNAPSHOT",
"description": "Check the status of EC2 instances\n\nAccess Key\/Secret Key are the basic method for authenticating to the AWS EC2 Service. These parameters are optional because the Kamelet provides the 'useDefaultCredentialsProvider'.\n\nWhen using a default Credentials Provider the EC2 client will load the credentials through this provider and won't use the static credential. This is reason for not having the access key and secret key as mandatory parameter for this Kamelet.\n\nThe Kamelet expects the following headers to be set:\n\n- `instanceIds` \/ `ce-instanceids`: as a comma separated list of EC2 instance ids."
},
"properties": {
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,7 @@
"groupId": "org.apache.camel.kafkaconnector",
"id": "camel-aws-eventbridge-sink-sink",
"type": "sink",
"version": "4.1.0-SNAPSHOT",
"version": "4.4.2-SNAPSHOT",
"description": "Send events to an Amazon Eventbridge Eventbus.\n\nThe basic authentication method for the Eventbridge service is to specify an access key and a secret key. These parameters are optional because the Kamelet provides a default credentials provider.\n\nIf you use the default credentials provider, the Eventbridge client loads the credentials through this provider and doesn't use the basic authentication method.\n\nIn the headers, you need to set the `resources-arn` \/ `ce-resources-arn` property to specify the ARN of resources related to the event.\n\nIn the headers, you need to set the `detail-type` \/ `ce-detail-type` property to specify the detail type related to the event.\n\nIn the headers, you need to set the `event-source` \/ `ce-event-source` property to specify the event source related to the event.\n\nIf you do not set the property in the header, the Kamelet uses the exchange ID for the file name."
},
"properties": {
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,7 @@
"groupId": "org.apache.camel.kafkaconnector",
"id": "camel-aws-kinesis-firehose-sink-sink",
"type": "sink",
"version": "4.1.0-SNAPSHOT",
"version": "4.4.2-SNAPSHOT",
"description": "Send message to an AWS Kinesis Firehose Stream\n\nAccess Key\/Secret Key are the basic method for authenticating to the AWS Kinesis Firehose Service. These parameters are optional because the Kamelet provides the 'useDefaultCredentialsProvider'.\n\nWhen using a default Credentials Provider the Kinesis Firehose client will load the credentials through this provider and won't use the static credential. This is reason for not having the access key and secret key as mandatory parameter for this Kamelet."
},
"properties": {
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,7 @@
"groupId": "org.apache.camel.kafkaconnector",
"id": "camel-aws-kinesis-sink-sink",
"type": "sink",
"version": "4.1.0-SNAPSHOT",
"version": "4.4.2-SNAPSHOT",
"description": "Send data to AWS Kinesis.\n\nThe basic authentication method for the Kinesis service is to specify an access key and a secret key. These parameters are optional because the Kamelet provides a default credentials provider.\n\nIf you use the default credentials provider, the Kinesis client loads the credentials through this provider and doesn't use the basic authentication method.\n\nIn the header, you can optionally set the `file` \/ `ce-partition` property to set the Kinesis partition key.\n\nIf you do not set the property in the header, the Kamelet uses the exchange ID for the partition key.\n\nYou can also set the `sequence-number` \/ `ce-sequencenumber` property in the header to specify the Sequence number."
},
"properties": {
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,7 @@
"groupId": "org.apache.camel.kafkaconnector",
"id": "camel-aws-kinesis-source-source",
"type": "source",
"version": "4.1.0-SNAPSHOT",
"version": "4.4.2-SNAPSHOT",
"description": "Receive data from AWS Kinesis.\n\nThe basic authentication method for the Kinesis service is to specify an access key and a secret key. These parameters are optional because the Kamelet provides a default credentials provider.\n\nIf you use the default credentials provider, the Kinesis client loads the credentials through this provider and doesn't use the basic authentication method.\n\nTwo headers will be duplicated with different names for clarity at sink level, CamelAwsKinesisPartitionKey will be duplicated into aws.kinesis.partition.key and CamelAwsKinesisSequenceNumber will be duplicated in aws.kinesis.sequence.number"
},
"properties": {
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,7 @@
"groupId": "org.apache.camel.kafkaconnector",
"id": "camel-aws-lambda-sink-sink",
"type": "sink",
"version": "4.1.0-SNAPSHOT",
"version": "4.4.2-SNAPSHOT",
"description": "Send a payload to an AWS Lambda function.\n\nThe basic authentication method for the Lambda service is to specify an access key and a secret key. These parameters are optional because the Kamelet provides a default credentials provider.\n\nIf you use the default credentials provider, the Lambda client loads the credentials through this provider and doesn't use the basic authentication method."
},
"properties": {
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,7 @@
"groupId": "org.apache.camel.kafkaconnector",
"id": "camel-aws-redshift-sink-sink",
"type": "sink",
"version": "4.1.0-SNAPSHOT",
"version": "4.4.2-SNAPSHOT",
"description": "Send data to an AWS Redshift Database.\n\nThis Kamelet expects a JSON-formatted body. Use key:value pairs to map the JSON fields and parameters. For example, here is a query:\n\n'INSERT INTO accounts (username,city) VALUES (:#username,:#city)'\n\nHere is example input for the example query:\n\n'{ \"username\":\"oscerd\", \"city\":\"Rome\"}'"
},
"properties": {
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,7 @@
"groupId": "org.apache.camel.kafkaconnector",
"id": "camel-aws-redshift-source-source",
"type": "source",
"version": "4.1.0-SNAPSHOT",
"version": "4.4.2-SNAPSHOT",
"description": "Query data from an AWS RedShift Database."
},
"properties": {
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,7 @@
"groupId": "org.apache.camel.kafkaconnector",
"id": "camel-aws-s3-cdc-source-source",
"type": "source",
"version": "4.1.0-SNAPSHOT",
"version": "4.4.2-SNAPSHOT",
"description": "Receive data from AWS SQS subscribed to Eventbridge Bus reporting events related to an S3 bucket or multiple buckets.\nAccess Key\/Secret Key are the basic method for authenticating to the AWS SQS Service.\nTo use this Kamelet you'll need to set up Eventbridge on your bucket and subscribe Eventbridge bus to an SQS Queue.\nFor doing this you'll need to enable Evenbridge notification on your bucket and creating a rule on Eventbridge console related to all the events on S3 bucket and pointing to the SQS Queue specified as parameter in this Kamelet."
},
"properties": {
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,7 @@
"groupId": "org.apache.camel.kafkaconnector",
"id": "camel-aws-s3-sink-sink",
"type": "sink",
"version": "4.1.0-SNAPSHOT",
"version": "4.4.2-SNAPSHOT",
"description": "Upload data to an Amazon S3 Bucket.\n\nThe basic authentication method for the S3 service is to specify an access key and a secret key. These parameters are optional because the Kamelet provides a default credentials provider.\n\nIf you use the default credentials provider, the S3 client loads the credentials through this provider and doesn't use the basic authentication method.\n\nIn the header, you can optionally set the `file` \/ `ce-partition` property to specify the name of the file to upload.\n\nIf you do not set the property in the header, the Kamelet uses the exchange ID for the file name."
},
"properties": {
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,7 @@
"groupId": "org.apache.camel.kafkaconnector",
"id": "camel-aws-s3-source-source",
"type": "source",
"version": "4.1.0-SNAPSHOT",
"version": "4.4.2-SNAPSHOT",
"description": "Receive data from an Amazon S3 Bucket.\n\nThe basic authentication method for the S3 service is to specify an access key and a secret key. These parameters are optional because the Kamelet provides a default credentials provider.\n\nIf you use the default credentials provider, the S3 client loads the credentials through this provider and doesn't use the basic authentication method.\n\nTwo headers will be duplicated with different names for clarity at sink level, CamelAwsS3Key will be duplicated into aws.s3.key and CamelAwsS3BucketName will be duplicated in aws.s3.bucket.name"
},
"properties": {
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,7 @@
"groupId": "org.apache.camel.kafkaconnector",
"id": "camel-aws-s3-streaming-upload-sink-sink",
"type": "sink",
"version": "4.1.0-SNAPSHOT",
"version": "4.4.2-SNAPSHOT",
"description": "Upload data to AWS S3 in streaming upload mode.\n\nAccess Key\/Secret Key are the basic method for authenticating to the AWS S3 Service. These parameters are optional because the Kamelet provides the 'useDefaultCredentialsProvider'.\n\nWhen using a default Credentials Provider the S3 client will load the credentials through this provider and won't use the static credential. This is reason for not having the access key and secret key as mandatory parameter for this Kamelet."
},
"properties": {
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,7 @@
"groupId": "org.apache.camel.kafkaconnector",
"id": "camel-aws-secrets-manager-sink-sink",
"type": "sink",
"version": "4.1.0-SNAPSHOT",
"version": "4.4.2-SNAPSHOT",
"description": "Create a secret in AWS Secrets Manager.\n\nAccess Key\/Secret Key are the basic method for authenticating to the AWS Secrets Manager Service. These parameters are optional because the Kamelet provides the 'useDefaultCredentialsProvider'.\n\nWhen using a default Credentials Provider the Secrets Manager client will load the credentials through this provider and won't use the static credential. This is reason for not having the access key and secret key as mandatory parameter for this Kamelet.\n\nThe Kamelet expects the following headers to be set:\n\n- `secret-name` \/ `ce-secret-name`: as the secret name to create.\n\nIf the header won't be set the exchange ID will be used as secret name."
},
"properties": {
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,7 @@
"groupId": "org.apache.camel.kafkaconnector",
"id": "camel-aws-ses-sink-sink",
"type": "sink",
"version": "4.1.0-SNAPSHOT",
"version": "4.4.2-SNAPSHOT",
"description": "Send email through the Amazon Simple Email Service (SES).\n\nThe basic authentication method for the SES service is to specify an access key and a secret key. These parameters are optional because the Kamelet provides a default credentials provider.\n\nIf you use the default credentials provider, the SES client loads the credentials through this provider and doesn't use the basic authentication method.\n\nIn the header, you can optionally set the following properties:\n\n- `subject` \/ `ce-subject` property to specify the email subject. If you do not set this property in the header, the Kamelet uses the exchange ID for the email subject.\n\n- `to` \/ `ce-to` to provide a comma-separated list of email addresses.\n\n- `reply-to-addresses` \/ `ce-reply-to-addresses`to provider a comma-separated list of reply-to email addresses."
},
"properties": {
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,7 @@
"groupId": "org.apache.camel.kafkaconnector",
"id": "camel-aws-sns-fifo-sink-sink",
"type": "sink",
"version": "4.1.0-SNAPSHOT",
"version": "4.4.2-SNAPSHOT",
"description": "Send message to an AWS SNS FIFO Topic.\n\nAccess Key\/Secret Key are the basic method for authenticating to the AWS SNS Service. These parameters are optional because the Kamelet provides the 'useDefaultCredentialsProvider'.\n\nWhen using a default Credentials Provider the SNS client will load the credentials through this provider and won't use the static credential. This is reason for not having the access key and secret key as mandatory parameter for this Kamelet.\n\nThe Kamelet expects the following headers to be set:\n\n- `subject` \/ `ce-subject`: the subject of the message"
},
"properties": {
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,7 @@
"groupId": "org.apache.camel.kafkaconnector",
"id": "camel-aws-sns-sink-sink",
"type": "sink",
"version": "4.1.0-SNAPSHOT",
"version": "4.4.2-SNAPSHOT",
"description": "Send message to an Amazon Simple Notification Service (SNS) topic.\n\nThe basic authentication method for the SNS service is to specify an access key and a secret key. These parameters are optional because the Kamelet provides a default credentials provider.\n\nIf you use the default credentials provider, the SNS client loads the credentials through this provider and doesn't use the basic authentication method.\n\nIn the header, you can optionally set the `subject` \/ `ce-subject` property to specify the subject of the message."
},
"properties": {
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,7 @@
"groupId": "org.apache.camel.kafkaconnector",
"id": "camel-aws-sqs-batch-sink-sink",
"type": "sink",
"version": "4.1.0-SNAPSHOT",
"version": "4.4.2-SNAPSHOT",
"description": "Send Batch message to an AWS SQS Queue.\n\nAccess Key\/Secret Key are the basic method for authenticating to the AWS SQS Service. These parameters are optional because the Kamelet provides the 'useDefaultCredentialsProvider'.\n\nWhen using a default Credentials Provider the SQS client will load the credentials through this provider and won't use the static credential. This is reason for not having the access key and secret key as mandatory parameter for this Kamelet."
},
"properties": {
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,7 @@
"groupId": "org.apache.camel.kafkaconnector",
"id": "camel-aws-sqs-fifo-sink-sink",
"type": "sink",
"version": "4.1.0-SNAPSHOT",
"version": "4.4.2-SNAPSHOT",
"description": "Send message to an AWS SQS FIFO Queue.\n\nAccess Key\/Secret Key are the basic method for authenticating to the AWS SQS Service. These parameters are optional because the Kamelet provides the 'useDefaultCredentialsProvider'.\n\nWhen using a default Credentials Provider the SQS client will load the credentials through this provider and won't use the static credential. This is reason for not having the access key and secret key as mandatory parameter for this Kamelet."
},
"properties": {
Expand Down
Loading
Loading