chore: fix dependabot alerts (#2770) #24
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
name: "Clustering test" | |
on: | |
push: | |
branches: | |
- main | |
pull_request: | |
branches: [main] | |
release: | |
types: | |
- published | |
permissions: read-all | |
jobs: | |
client-tools: | |
name: Stateless zot with shared reliable storage | |
runs-on: ubuntu-latest | |
# services: | |
# minio: | |
# image: minio/minio:RELEASE.2024-07-16T23-46-41Z | |
# env: | |
# MINIO_ROOT_USER: minioadmin | |
# MINIO_ROOT_PASSWORD: minioadmin | |
# ports: | |
# - 9000:9000 | |
# volumes: | |
# - /tmp/data:/data | |
# options: --name=minio --health-cmd "curl http://localhost:9000/minio/health/live" | |
steps: | |
- uses: actions/checkout@v4 | |
- uses: actions/setup-go@v5 | |
with: | |
cache: false | |
go-version: 1.23.x | |
- name: Install dependencies | |
run: | | |
cd $GITHUB_WORKSPACE | |
go install github.com/swaggo/swag/cmd/[email protected] | |
go mod download | |
sudo apt-get update | |
sudo apt-get -y install rpm uidmap | |
# install skopeo | |
sudo apt-get -y install skopeo | |
# install haproxy | |
sudo apt-get install haproxy | |
- name: Build binaries | |
run: | | |
cd $GITHUB_WORKSPACE | |
make binary | |
make bench | |
- name: Setup minio service | |
run: | | |
docker run -d -p 9000:9000 --name minio \ | |
-e "MINIO_ACCESS_KEY=minioadmin" \ | |
-e "MINIO_SECRET_KEY=minioadmin" \ | |
-v /tmp/data:/data \ | |
-v /tmp/config:/root/.minio \ | |
--health-cmd "curl http://localhost:9000/minio/health/live" \ | |
minio/minio:RELEASE.2024-07-16T23-46-41Z server /data | |
- name: Install py minio | |
run: pip3 install minio | |
- name: Wait for minio to come up | |
run: | | |
curl --connect-timeout 5 \ | |
--max-time 120 \ | |
--retry 12 \ | |
--retry-max-time 120 \ | |
'http://localhost:9000/minio/health/live' | |
- name: Create minio bucket | |
run: | | |
python3 - <<'EOF' | |
from minio import Minio | |
try: | |
minio = Minio( | |
'localhost:9000', | |
access_key='minioadmin', | |
secret_key='minioadmin', | |
secure=False | |
) | |
except Exception as ex: | |
raise | |
minio.make_bucket('zot-storage') | |
print(f'{minio.list_buckets()}') | |
EOF | |
- name: Run haproxy | |
run: | | |
sudo haproxy -d -f examples/cluster/haproxy.cfg -D | |
sleep 10 | |
- name: Prepare configuration files | |
run: | | |
cp test/cluster/config-minio.json test/cluster/config-minio1.json | |
sed -i 's/8081/8081/g' test/cluster/config-minio1.json | |
sed -i 's/\/tmp\/zot/\/tmp\/zot1/g' test/cluster/config-minio1.json | |
cp test/cluster/config-minio.json test/cluster/config-minio2.json | |
sed -i 's/8081/8082/g' test/cluster/config-minio2.json | |
sed -i 's/\/tmp\/zot/\/tmp\/zot2/g' test/cluster/config-minio2.json | |
cp test/cluster/config-minio.json test/cluster/config-minio3.json | |
sed -i 's/8081/8083/g' test/cluster/config-minio3.json | |
sed -i 's/\/tmp\/zot/\/tmp\/zot3/g' test/cluster/config-minio3.json | |
- name: Free up disk space | |
uses: jlumbroso/free-disk-space@main | |
with: | |
# This might remove tools that are actually needed, if set to "true" but frees about 6 GB | |
tool-cache: true | |
# All of these default to true, but feel free to set to "false" if necessary for your workflow | |
android: true | |
dotnet: true | |
haskell: true | |
large-packages: true | |
swap-storage: true | |
- name: Run push-pull tests | |
run: | | |
./bin/zot-linux-amd64 serve test/cluster/config-minio1.json & | |
./bin/zot-linux-amd64 serve test/cluster/config-minio2.json & | |
./bin/zot-linux-amd64 serve test/cluster/config-minio3.json & | |
sleep 20 | |
# run tests | |
skopeo --debug copy --format=oci --dest-tls-verify=false docker://ghcr.io/project-zot/golang:1.20 docker://localhost:8080/golang:1.20 | |
skopeo --debug copy --src-tls-verify=false docker://localhost:8080/golang:1.20 oci:golang:1.20 | |
echo "{\"name\":\"foo\",\"value\":\"bar\"}" > config.json | |
echo "hello world" > artifact.txt | |
oras push --plain-http localhost:8080/hello-artifact:v2 \ | |
--config config.json:application/vnd.acme.rocket.config.v1+json \ | |
artifact.txt:text/plain -d -v | |
rm -f artifact.txt # first delete the file | |
oras pull --plain-http localhost:8080/hello-artifact:v2 -d -v | |
grep -q "hello world" artifact.txt # should print "hello world" | |
if [ $? -ne 0 ]; then \ | |
killall -r zot-*; \ | |
exit 1; \ | |
fi | |
killall -r zot-* | |
env: | |
AWS_ACCESS_KEY_ID: minioadmin | |
AWS_SECRET_ACCESS_KEY: minioadmin | |
- name: Run benchmark with --src-cidr arg | |
run: | | |
./bin/zot-linux-amd64 serve test/cluster/config-minio1.json & | |
./bin/zot-linux-amd64 serve test/cluster/config-minio2.json & | |
./bin/zot-linux-amd64 serve test/cluster/config-minio3.json & | |
sleep 20 | |
# run zb with --src-cidr | |
bin/zb-linux-amd64 -c 10 -n 50 -o ci-cd --src-cidr 127.0.0.0/8 http://localhost:8080 | |
killall -r zot-* | |
# clean zot storage | |
sudo rm -rf /tmp/data/zot-storage/zot | |
env: | |
AWS_ACCESS_KEY_ID: minioadmin | |
AWS_SECRET_ACCESS_KEY: minioadmin | |
- name: Run benchmark with --src-ips arg | |
run: | | |
./bin/zot-linux-amd64 serve test/cluster/config-minio1.json & | |
./bin/zot-linux-amd64 serve test/cluster/config-minio2.json & | |
./bin/zot-linux-amd64 serve test/cluster/config-minio3.json & | |
sleep 20 | |
# run zb with --src-ips | |
bin/zb-linux-amd64 -c 10 -n 50 -o ci-cd --src-ips 127.0.0.2,127.0.0.3,127.0.0.4,127.0.0.5,127.0.0.6,127.0.12.5,127.0.12.6 http://localhost:8080 | |
killall -r zot-* | |
env: | |
AWS_ACCESS_KEY_ID: minioadmin | |
AWS_SECRET_ACCESS_KEY: minioadmin | |
# Download previous benchmark result from cache (if exists) | |
- name: Download previous benchmark data | |
uses: actions/cache@v4 | |
with: | |
path: ./cache | |
key: ${{ runner.os }}-gen1-benchmark-stateless-cluster | |
# Run `github-action-benchmark` action | |
- name: Store benchmark result | |
uses: benchmark-action/[email protected] | |
with: | |
# What benchmark tool the output.txt came from | |
tool: 'customBiggerIsBetter' | |
# Where the output from the benchmark tool is stored | |
output-file-path: ci-cd.json | |
# Where the previous data file is stored | |
external-data-json-path: ./cache/benchmark-data.json | |
# Workflow will fail when an alert happens | |
fail-on-alert: true | |
# Upload the updated cache file for the next job by actions/cache |