Add upload-image-s3 role
This adds a role similar to upload-image-swift to upload dib images to s3 for use by zuul-launcher. Change-Id: Ie802a0221717e6d5d5dcaa91771d07f8d0321889
This commit is contained in:
parent
4b025bbf2f
commit
0c004096c6
@ -5,3 +5,4 @@ Diskimage-Builder Roles
|
|||||||
.. zuul:autorole:: build-diskimage
|
.. zuul:autorole:: build-diskimage
|
||||||
.. zuul:autorole:: convert-diskimage
|
.. zuul:autorole:: convert-diskimage
|
||||||
.. zuul:autorole:: upload-image-swift
|
.. zuul:autorole:: upload-image-swift
|
||||||
|
.. zuul:autorole:: upload-image-s3
|
||||||
|
66
roles/upload-image-s3/README.rst
Normal file
66
roles/upload-image-s3/README.rst
Normal file
@ -0,0 +1,66 @@
|
|||||||
|
Upload a filesystem image to an S3 bucket
|
||||||
|
|
||||||
|
This uploads a filesystem image (for example, one built by diskimage
|
||||||
|
builder) to an S3 bucket. The role returns an artifact to Zuul
|
||||||
|
suitable for use by the zuul-launcher.
|
||||||
|
|
||||||
|
**Role Variables**
|
||||||
|
|
||||||
|
.. zuul:rolevar:: upload_image_s3_endpoint
|
||||||
|
|
||||||
|
The endpoint to use when uploading an image to an s3 compatible
|
||||||
|
service. By default this will be automatically constructed by boto
|
||||||
|
but should be set when working with non-AWS hosted s3 service.
|
||||||
|
|
||||||
|
.. zuul:rolevar:: upload_image_s3_aws_access_key
|
||||||
|
|
||||||
|
AWS access key to use.
|
||||||
|
|
||||||
|
.. zuul:rolevar:: upload_image_s3_aws_secret_key
|
||||||
|
|
||||||
|
AWS secret key for the AWS access key.
|
||||||
|
|
||||||
|
.. zuul:rolevar:: upload_image_s3_bucket
|
||||||
|
|
||||||
|
This role *will not* create buckets which do not already exist.
|
||||||
|
|
||||||
|
Note that you will want to set this to a value that uniquely
|
||||||
|
identifies your Zuul installation.
|
||||||
|
|
||||||
|
The bucket should be dedicated to image uploads so that the
|
||||||
|
"delete_after" option may be safely used.
|
||||||
|
|
||||||
|
.. zuul:rolevar:: upload_image_s3_delete_after
|
||||||
|
:default: 0
|
||||||
|
|
||||||
|
Number of seconds to delete objects after upload. Default is 0
|
||||||
|
(disabled). Each run of the role will attempt to delete any
|
||||||
|
objects in the bucket older than this time.
|
||||||
|
|
||||||
|
It is also recommended to use the AWS console to configure
|
||||||
|
automatic expiration of objects in this bucket.
|
||||||
|
|
||||||
|
.. zuul:rolevar:: upload_image_s3_image_name
|
||||||
|
:default: `{{ build_diskimage_image_name }}`
|
||||||
|
|
||||||
|
The Zuul image name for use by zuul-launcher (e.g., `debian-bookworm`).
|
||||||
|
|
||||||
|
.. zuul:rolevar:: upload_image_s3_format
|
||||||
|
|
||||||
|
The image format (e.g., `qcow2`).
|
||||||
|
|
||||||
|
.. zuul:rolevar:: upload_image_s3_extension
|
||||||
|
:default: `{{ upload_image_s3_format }}`
|
||||||
|
|
||||||
|
The extension to use when uploading (only used in the default
|
||||||
|
values for the following variables.
|
||||||
|
|
||||||
|
.. zuul:rolevar:: upload_image_s3_filename
|
||||||
|
:default: `{{ build_diskimage_image_root }}/{{ build_diskimage_image_name }}.{{ upload_image_s3_extension }}`
|
||||||
|
|
||||||
|
The path of the local file to upload.
|
||||||
|
|
||||||
|
.. zuul:rolevar:: upload_image_s3_name
|
||||||
|
:default: `{{ zuul.build }}-{{ build_diskimage_image_name }}.{{ upload_image_s3_extension }}`
|
||||||
|
|
||||||
|
The object name to use when uploading.
|
5
roles/upload-image-s3/defaults/main.yaml
Normal file
5
roles/upload-image-s3/defaults/main.yaml
Normal file
@ -0,0 +1,5 @@
|
|||||||
|
upload_image_s3_image_name: '{{ build_diskimage_image_name }}'
|
||||||
|
upload_image_s3_delete_after: 0
|
||||||
|
upload_image_s3_filename: '{{ build_diskimage_image_root }}/{{ build_diskimage_image_name }}.{{ upload_image_s3_extension }}'
|
||||||
|
upload_image_s3_name: '{{ zuul.build }}-{{ build_diskimage_image_name }}.{{ upload_image_s3_extension }}'
|
||||||
|
upload_image_s3_extension: '{{ upload_image_s3_format }}'
|
0
roles/upload-image-s3/library/__init__.py
Normal file
0
roles/upload-image-s3/library/__init__.py
Normal file
136
roles/upload-image-s3/library/upload_image_s3.py
Normal file
136
roles/upload-image-s3/library/upload_image_s3.py
Normal file
@ -0,0 +1,136 @@
|
|||||||
|
# Copyright 2014 Rackspace Australia
|
||||||
|
# Copyright 2018 Red Hat, Inc
|
||||||
|
# Copyright 2024-2025 Acme Gating, LLC
|
||||||
|
#
|
||||||
|
# Licensed under the Apache License, Version 2.0 (the "License"); you may
|
||||||
|
# not use this file except in compliance with the License. You may obtain
|
||||||
|
# a copy of the License at
|
||||||
|
#
|
||||||
|
# http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
#
|
||||||
|
# Unless required by applicable law or agreed to in writing, software
|
||||||
|
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
||||||
|
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
||||||
|
# License for the specific language governing permissions and limitations
|
||||||
|
# under the License.
|
||||||
|
|
||||||
|
import argparse
|
||||||
|
import datetime
|
||||||
|
import logging
|
||||||
|
import os
|
||||||
|
import sys
|
||||||
|
import traceback
|
||||||
|
|
||||||
|
import boto3
|
||||||
|
|
||||||
|
from ansible.module_utils.basic import AnsibleModule
|
||||||
|
|
||||||
|
|
||||||
|
def prune(bucket, delete_after):
|
||||||
|
# In case the automatic expiration doesn't work, manually prune old uploads
|
||||||
|
if not delete_after:
|
||||||
|
return
|
||||||
|
target = (datetime.datetime.now(datetime.UTC) -
|
||||||
|
datetime.timedelta(seconds=delete_after))
|
||||||
|
for obj in bucket.objects.all():
|
||||||
|
if obj.last_modified < target:
|
||||||
|
obj.delete()
|
||||||
|
|
||||||
|
|
||||||
|
def run(endpoint, bucket_name, aws_access_key, aws_secret_key,
|
||||||
|
filename, name, delete_after=None):
|
||||||
|
endpoint = endpoint or 'https://s3.amazonaws.com/'
|
||||||
|
s3 = boto3.resource('s3',
|
||||||
|
endpoint_url=endpoint,
|
||||||
|
aws_access_key_id=aws_access_key,
|
||||||
|
aws_secret_access_key=aws_secret_key)
|
||||||
|
bucket = s3.Bucket(bucket_name)
|
||||||
|
|
||||||
|
prune(bucket, delete_after)
|
||||||
|
|
||||||
|
bucket.upload_file(filename, name)
|
||||||
|
|
||||||
|
url = os.path.join(endpoint, bucket_name, name)
|
||||||
|
return url
|
||||||
|
|
||||||
|
|
||||||
|
def ansible_main():
|
||||||
|
module = AnsibleModule(
|
||||||
|
argument_spec=dict(
|
||||||
|
endpoint=dict(type='str'),
|
||||||
|
bucket=dict(required=True, type='str'),
|
||||||
|
filename=dict(required=True, type='path'),
|
||||||
|
name=dict(required=True, type='str'),
|
||||||
|
delete_after=dict(type='int'),
|
||||||
|
aws_access_key=dict(type='str'),
|
||||||
|
aws_secret_key=dict(type='str', no_log=True),
|
||||||
|
)
|
||||||
|
)
|
||||||
|
|
||||||
|
p = module.params
|
||||||
|
|
||||||
|
try:
|
||||||
|
url = run(
|
||||||
|
p.get('endpoint'),
|
||||||
|
p.get('bucket'),
|
||||||
|
p.get('aws_access_key'),
|
||||||
|
p.get('aws_secret_key'),
|
||||||
|
p.get('filename'),
|
||||||
|
p.get('name'),
|
||||||
|
delete_after=p.get('delete_after'),
|
||||||
|
)
|
||||||
|
except Exception:
|
||||||
|
s = "Error uploading to S3"
|
||||||
|
s += "\n" + traceback.format_exc()
|
||||||
|
module.fail_json(
|
||||||
|
changed=False,
|
||||||
|
msg=s)
|
||||||
|
module.exit_json(
|
||||||
|
changed=True,
|
||||||
|
url=url,
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
def cli_main():
|
||||||
|
parser = argparse.ArgumentParser(
|
||||||
|
description="Upload image to S3"
|
||||||
|
)
|
||||||
|
parser.add_argument('--verbose', action='store_true',
|
||||||
|
help='show debug information')
|
||||||
|
parser.add_argument('--endpoint',
|
||||||
|
help='http endpoint of s3 service')
|
||||||
|
parser.add_argument('bucket',
|
||||||
|
help='Name of the bucket to use when uploading')
|
||||||
|
parser.add_argument('filename',
|
||||||
|
help='the file to upload')
|
||||||
|
parser.add_argument('name',
|
||||||
|
help='the object name')
|
||||||
|
parser.add_argument('--delete-after',
|
||||||
|
help='Number of seconds to delete object after '
|
||||||
|
'upload. Default is 3 days (259200 seconds) '
|
||||||
|
'and if set to 0 X-Delete-After will not be set',
|
||||||
|
type=int)
|
||||||
|
|
||||||
|
args = parser.parse_args()
|
||||||
|
|
||||||
|
if args.verbose:
|
||||||
|
logging.basicConfig(level=logging.DEBUG)
|
||||||
|
logging.captureWarnings(True)
|
||||||
|
|
||||||
|
url = run(
|
||||||
|
args.endpoint,
|
||||||
|
args.bucket,
|
||||||
|
None,
|
||||||
|
None,
|
||||||
|
args.filename,
|
||||||
|
args.name,
|
||||||
|
delete_after=args.delete_after,
|
||||||
|
)
|
||||||
|
print(url)
|
||||||
|
|
||||||
|
|
||||||
|
if __name__ == '__main__':
|
||||||
|
if not sys.stdin.isatty():
|
||||||
|
ansible_main()
|
||||||
|
else:
|
||||||
|
cli_main()
|
58
roles/upload-image-s3/tasks/main.yaml
Normal file
58
roles/upload-image-s3/tasks/main.yaml
Normal file
@ -0,0 +1,58 @@
|
|||||||
|
# Run the checksums in the background while we're uploading
|
||||||
|
- name: Get sha256 hash
|
||||||
|
stat:
|
||||||
|
path: '{{ upload_image_s3_filename }}'
|
||||||
|
checksum_algorithm: sha256
|
||||||
|
async: 600
|
||||||
|
poll: 0
|
||||||
|
register: sha256_task
|
||||||
|
|
||||||
|
- name: Get md5 hash
|
||||||
|
stat:
|
||||||
|
path: '{{ upload_image_s3_filename }}'
|
||||||
|
checksum_algorithm: md5
|
||||||
|
async: 600
|
||||||
|
poll: 0
|
||||||
|
register: md5_task
|
||||||
|
|
||||||
|
- name: Upload image to S3
|
||||||
|
no_log: true
|
||||||
|
upload_image_s3:
|
||||||
|
endpoint: "{{ upload_image_s3_endpoint | default(omit) }}"
|
||||||
|
bucket: "{{ upload_image_s3_bucket }}"
|
||||||
|
aws_access_key: "{{ upload_image_s3_aws_access_key }}"
|
||||||
|
aws_secret_key: "{{ upload_image_s3_aws_secret_key }}"
|
||||||
|
filename: '{{ upload_image_s3_filename }}'
|
||||||
|
name: '{{ upload_image_s3_name }}'
|
||||||
|
delete_after: '{{ upload_image_s3_delete_after }}'
|
||||||
|
register: upload_results
|
||||||
|
|
||||||
|
- name: Wait for sha256
|
||||||
|
async_status:
|
||||||
|
jid: "{{ sha256_task.ansible_job_id }}"
|
||||||
|
register: sha256
|
||||||
|
until: sha256.finished
|
||||||
|
retries: 1
|
||||||
|
delay: 10
|
||||||
|
|
||||||
|
- name: Wait for md5
|
||||||
|
async_status:
|
||||||
|
jid: "{{ md5_task.ansible_job_id }}"
|
||||||
|
register: md5
|
||||||
|
until: md5.finished
|
||||||
|
retries: 1
|
||||||
|
delay: 10
|
||||||
|
|
||||||
|
- name: Return artifact to Zuul
|
||||||
|
zuul_return:
|
||||||
|
data:
|
||||||
|
zuul:
|
||||||
|
artifacts:
|
||||||
|
- name: '{{ upload_image_s3_format }} image'
|
||||||
|
url: '{{ upload_results.url }}'
|
||||||
|
metadata:
|
||||||
|
type: 'zuul_image'
|
||||||
|
image_name: '{{ upload_image_s3_image_name }}'
|
||||||
|
format: '{{ upload_image_s3_format }}'
|
||||||
|
sha256: '{{ sha256.stat.checksum }}'
|
||||||
|
md5sum: '{{ md5.stat.checksum }}'
|
89
test-playbooks/upload-image-s3.yaml
Normal file
89
test-playbooks/upload-image-s3.yaml
Normal file
@ -0,0 +1,89 @@
|
|||||||
|
- hosts: all
|
||||||
|
vars:
|
||||||
|
test_aws_access_key: minioadmin
|
||||||
|
test_aws_secret_key: minioadmin
|
||||||
|
test_bucket: zuul
|
||||||
|
test_image: "image"
|
||||||
|
test_filename: "/tmp/testfile"
|
||||||
|
test_objectname_old: "image-old.raw"
|
||||||
|
test_objectname: "image.raw"
|
||||||
|
test_ext: "raw"
|
||||||
|
test_content: "123abc"
|
||||||
|
roles:
|
||||||
|
- ensure-docker
|
||||||
|
- ensure-pip
|
||||||
|
tasks:
|
||||||
|
- name: Install boto
|
||||||
|
pip:
|
||||||
|
extra_args: "--break-system-packages"
|
||||||
|
name:
|
||||||
|
- boto3
|
||||||
|
- botocore
|
||||||
|
|
||||||
|
- name: Start minio server
|
||||||
|
command: >-
|
||||||
|
docker run -d -p 9000:9000
|
||||||
|
-e MINIO_ACCESS_KEY={{ test_aws_access_key }}
|
||||||
|
-e MINIO_SECRET_KEY={{ test_aws_secret_key }}
|
||||||
|
quay.io/minio/minio server /data
|
||||||
|
|
||||||
|
- name: Make sure bucket exists
|
||||||
|
s3_bucket:
|
||||||
|
name: "{{ test_bucket }}"
|
||||||
|
state: present
|
||||||
|
s3_url: 'http://localhost:9000'
|
||||||
|
aws_access_key: "{{ test_aws_access_key }}"
|
||||||
|
aws_secret_key: "{{ test_aws_secret_key }}"
|
||||||
|
|
||||||
|
- name: Add content to tempfile
|
||||||
|
copy:
|
||||||
|
content: "{{ test_content }}"
|
||||||
|
dest: "{{ test_filename }}"
|
||||||
|
|
||||||
|
- name: Upload file to s3
|
||||||
|
include_role:
|
||||||
|
name: upload-image-s3
|
||||||
|
vars:
|
||||||
|
upload_image_s3_endpoint: "http://localhost:9000"
|
||||||
|
upload_image_s3_aws_access_key: "{{ test_aws_access_key }}"
|
||||||
|
upload_image_s3_aws_secret_key: "{{ test_aws_secret_key }}"
|
||||||
|
upload_image_s3_bucket: "{{ test_bucket }}"
|
||||||
|
upload_image_s3_filename: "{{ test_filename }}"
|
||||||
|
upload_image_s3_name: "{{ test_objectname_old }}"
|
||||||
|
upload_image_s3_image_name: "{{ test_image }}"
|
||||||
|
upload_image_s3_format: "{{ test_ext }}"
|
||||||
|
upload_image_s3_delete_after: 1
|
||||||
|
|
||||||
|
- name: Delay for 5 seconds
|
||||||
|
wait_for:
|
||||||
|
timeout: 5
|
||||||
|
|
||||||
|
# This should delete the first object then replace it.
|
||||||
|
- name: Upload file to s3
|
||||||
|
include_role:
|
||||||
|
name: upload-image-s3
|
||||||
|
vars:
|
||||||
|
upload_image_s3_endpoint: "http://localhost:9000"
|
||||||
|
upload_image_s3_aws_access_key: "{{ test_aws_access_key }}"
|
||||||
|
upload_image_s3_aws_secret_key: "{{ test_aws_secret_key }}"
|
||||||
|
upload_image_s3_bucket: "{{ test_bucket }}"
|
||||||
|
upload_image_s3_filename: "{{ test_filename }}"
|
||||||
|
upload_image_s3_name: "{{ test_objectname }}"
|
||||||
|
upload_image_s3_image_name: "{{ test_image }}"
|
||||||
|
upload_image_s3_format: "{{ test_ext }}"
|
||||||
|
upload_image_s3_delete_after: 1
|
||||||
|
|
||||||
|
- name: Download mc
|
||||||
|
get_url:
|
||||||
|
url: https://dl.min.io/client/mc/release/linux-amd64/mc
|
||||||
|
dest: "{{ ansible_user_dir }}/mc"
|
||||||
|
mode: 0755
|
||||||
|
|
||||||
|
- name: Add localhost minio host
|
||||||
|
command: "{{ ansible_user_dir }}/mc config host add local http://localhost:9000 {{ test_aws_access_key }} {{ test_aws_secret_key }}"
|
||||||
|
|
||||||
|
- name: List files in minio bucket
|
||||||
|
command: "{{ ansible_user_dir }}/mc find local/zuul"
|
||||||
|
|
||||||
|
- name: Check for testfile in minio bucket
|
||||||
|
command: "{{ ansible_user_dir }}/mc find local/zuul/{{ test_objectname }}"
|
@ -177,6 +177,14 @@
|
|||||||
- name: ubuntu-noble
|
- name: ubuntu-noble
|
||||||
label: ubuntu-noble
|
label: ubuntu-noble
|
||||||
|
|
||||||
|
- job:
|
||||||
|
name: zuul-jobs-test-upload-image-s3
|
||||||
|
description: Test the upload-image-s3 role
|
||||||
|
files:
|
||||||
|
- roles/upload-image-s3/.*
|
||||||
|
- test-playbooks/upload-image-s3.yaml
|
||||||
|
run: test-playbooks/upload-image-s3.yaml
|
||||||
|
|
||||||
- project:
|
- project:
|
||||||
check:
|
check:
|
||||||
jobs: &id001
|
jobs: &id001
|
||||||
@ -195,6 +203,7 @@
|
|||||||
- zuul-jobs-test-convert-diskimage-ubuntu-focal
|
- zuul-jobs-test-convert-diskimage-ubuntu-focal
|
||||||
- zuul-jobs-test-convert-diskimage-ubuntu-jammy
|
- zuul-jobs-test-convert-diskimage-ubuntu-jammy
|
||||||
- zuul-jobs-test-convert-diskimage-ubuntu-noble
|
- zuul-jobs-test-convert-diskimage-ubuntu-noble
|
||||||
|
- zuul-jobs-test-upload-image-s3
|
||||||
gate:
|
gate:
|
||||||
jobs: *id001
|
jobs: *id001
|
||||||
periodic-weekly:
|
periodic-weekly:
|
||||||
|
Loading…
x
Reference in New Issue
Block a user