debian-koji/tests/test_hub/test_complete_image_build.py
2025-04-15 10:53:50 -04:00

289 lines
11 KiB
Python

import copy
from unittest import mock
import os
import os.path
import shutil
import tempfile
import unittest
import koji
import koji.util
import kojihub
orig_import_archive_internal = kojihub.import_archive_internal
def make_insert_grabber(test):
# test is the test class instance
def grab_insert(insert):
# insert is self for the InsertProcessor instance
# we are replacing execute()
info = [str(insert), insert.data.copy(), insert.rawdata.copy()]
test.inserts.append(info)
return grab_insert
def make_bulk_insert_grabber(test):
# test is the test class instance
def grab_insert(insert, data):
# insert is self for the BulkInsertProcessor instance
# we are replacing _one_insert()
query, params = insert._get_insert(data)
info = [query, copy.copy(params)]
test.inserts.append(info)
return grab_insert
def make_update_grabber(test):
# test is the test class instance
def grab_update(update):
# update is self for the UpdateProcessor instance
# we are replacing execute()
info = [str(update), update.data.copy(), update.rawdata.copy()]
test.updates.append(info)
return grab_update
class TestCompleteImageBuild(unittest.TestCase):
def setUp(self):
self.tempdir = tempfile.mkdtemp()
self.pathinfo = koji.PathInfo(self.tempdir)
mock.patch('koji.pathinfo', new=self.pathinfo).start()
self.hostcalls = kojihub.HostExports()
self.context_db = mock.patch('kojihub.db.context').start()
mock.patch('kojihub.kojihub.Host').start()
self.Task = mock.patch('kojihub.kojihub.Task').start()
self.Task.return_value.assertHost = mock.MagicMock()
self.get_build = mock.patch('kojihub.kojihub.get_build').start()
mock.patch('kojihub.kojihub.get_rpm', new=self.my_get_rpm).start()
self.get_image_build = mock.patch('kojihub.kojihub.get_image_build').start()
mock.patch('kojihub.kojihub.get_archive_type', new=self.my_get_archive_type).start()
mock.patch('kojihub.kojihub.lookup_name', new=self.my_lookup_name).start()
mock.patch.object(kojihub.BuildRoot, 'load', new=self.my_buildroot_load).start()
mock.patch('kojihub.kojihub.import_archive_internal',
new=self.my_import_archive_internal).start()
self._dml = mock.patch('kojihub.kojihub._dml').start()
mock.patch('kojihub.kojihub.build_notification').start()
mock.patch('kojihub.kojihub.assert_policy').start()
mock.patch('kojihub.kojihub.policy_data_from_task').start()
mock.patch('kojihub.kojihub.check_volume_policy',
return_value={'id': 0, 'name': 'DEFAULT'}).start()
self.set_up_callbacks()
self.rpms = {}
self.inserts = []
self.updates = []
mock.patch.object(kojihub.InsertProcessor, 'execute',
new=make_insert_grabber(self)).start()
mock.patch.object(kojihub.BulkInsertProcessor, '_one_insert',
new=make_bulk_insert_grabber(self)).start()
mock.patch.object(kojihub.UpdateProcessor, 'execute',
new=make_update_grabber(self)).start()
mock.patch('kojihub.kojihub.nextval', new=self.my_nextval).start()
self.sequences = {}
def tearDown(self):
mock.patch.stopall()
shutil.rmtree(self.tempdir)
def grab_update(self, update):
# update is self for the UpdateProcessor instance
# we are replacing execute()
info = [str(update), update.data.copy(), update.rawdata.copy()]
self.updates.append(info)
def set_up_files(self, name):
datadir = os.path.join(os.path.dirname(__file__), 'data/image', name)
# load image result data for our test build
data = koji.load_json(datadir + '/data.json')
self.db_expect = koji.load_json(datadir + '/db.json')
for arch in data:
taskdir = koji.pathinfo.task(data[arch]['task_id'])
os.makedirs(taskdir)
filenames = data[arch]['files'] + data[arch]['logs']
for filename in filenames:
path = os.path.join(taskdir, filename)
with open(path, 'wt', encoding='utf-8') as fp:
fp.write('Test file for %s\n%s\n' % (arch, filename))
self.image_data = data
def get_expected_files(self, buildinfo):
data = self.image_data
imgdir = koji.pathinfo.imagebuild(buildinfo)
logdir = koji.pathinfo.build_logs(buildinfo)
paths = []
for arch in data:
for filename in data[arch]['files']:
paths.append(os.path.join(imgdir, filename))
for filename in data[arch]['logs']:
# paths.append(os.path.join(logdir, 'image', arch, filename))
paths.append(os.path.join(logdir, 'image', filename))
# bdir = koji.pathinfo.build(buildinfo)
# paths.append(os.path.join(bdir, 'metadata.json'))
return paths
def my_nextval(self, sequence):
self.sequences.setdefault(sequence, 1000)
self.sequences[sequence] += 1
return self.sequences[sequence]
def my_get_rpm(self, rpminfo, **kw):
key = '%(name)s-%(version)s-%(release)s.%(arch)s' % rpminfo
ret = self.rpms.get(key)
if ret is not None:
return ret
ret = rpminfo.copy()
ret['id'] = len(self.rpms) + 1000
self.rpms[key] = rpminfo
return ret
def my_lookup_name(self, table, info, **kw):
if table == 'btype':
return {
'id': 'BTYPEID:%s' % info,
'name': info,
}
else:
raise Exception("Cannot fake call")
def my_get_archive_type(self, *a, **kw):
return dict.fromkeys(['id', 'name', 'description', 'extensions'], 'ARCHIVETYPE')
@staticmethod
def my_buildroot_load(br, id):
# br is the BuildRoot instance
br.id = id
br.is_standard = True
br.data = {
'br_type': koji.BR_TYPES['STANDARD'],
'id': id,
}
def my_import_archive_internal(self, *a, **kw):
# this is kind of odd, but we need this to fake the archiveinfo
share = {}
old_ip = kojihub.InsertProcessor
def my_ip(table, *a, **kw):
if table == 'archiveinfo':
data = kw['data']
data.setdefault('archive_id', 'ARCHIVE_ID')
share['archiveinfo'] = data
# TODO: need to add id
return old_ip(table, *a, **kw)
def my_ga(archive_id, **kw):
return share['archiveinfo']
with mock.patch('kojihub.kojihub.InsertProcessor', new=my_ip):
with mock.patch('kojihub.kojihub.get_archive', new=my_ga):
return orig_import_archive_internal(*a, **kw)
def set_up_callbacks(self):
new_callbacks = copy.deepcopy(koji.plugin.callbacks)
mock.patch('koji.plugin.callbacks', new=new_callbacks).start()
self.callbacks = []
for cbtype in koji.plugin.callbacks.keys():
koji.plugin.register_callback(cbtype, self.callback)
def callback(self, cbtype, *args, **kwargs):
self.callbacks.append([cbtype, args, kwargs])
def test_complete_image_build(self):
self.set_up_files('import_1')
buildinfo = {
'id': 137,
'task_id': 'TASK_ID',
'name': 'some-image',
'version': '1.2.3.4',
'release': '3',
'epoch': None,
'source': None,
'state': koji.BUILD_STATES['BUILDING'],
'volume_id': 0,
'extra': {},
}
image_info = {'build_id': buildinfo['id']}
self.get_build.return_value = buildinfo
self.get_image_build.return_value = image_info
taskinfo = {'id': 'TASKID', 'method': 'image'}
self.Task.return_value.getInfo.return_value = taskinfo
# run the import call
self.hostcalls.completeImageBuild('TASK_ID', 'BUILD_ID', self.image_data)
# make sure we wrote the files we expect
expected = self.get_expected_files(buildinfo)
files = []
for dirpath, dirnames, filenames in os.walk(self.tempdir + '/packages'):
files.extend([os.path.join(dirpath, fn) for fn in filenames])
self.assertEqual(set(files), set(expected))
# check callbacks
cbtypes = [c[0] for c in self.callbacks]
cb_expect = [
'preImport', # build
'preImport', # archive 1...
'postImport',
'preImport', # archive 2...
'postImport',
'preImport', # archive 3...
'postImport',
'preImport', # archive 4...
'postImport',
'preImport', # archive 5...
'postImport',
'postImport', # build
'preBuildStateChange', # building -> completed
'postBuildStateChange',
]
self.assertEqual(cbtypes, cb_expect)
cb_idx = {}
for c in self.callbacks:
# no callbacks should use *args
self.assertEqual(c[1], ())
cbtype = c[0]
if 'type' in c[2]:
key = "%s:%s" % (cbtype, c[2]['type'])
else:
key = cbtype
cb_idx.setdefault(key, [])
cb_idx[key].append(c[2])
key_expect = [
'postBuildStateChange', 'preBuildStateChange',
'preImport:archive', 'postImport:archive',
'preImport:image', 'postImport:image',
]
self.assertEqual(set(cb_idx.keys()), set(key_expect))
for key in ['preImport:image']:
callbacks = cb_idx[key]
self.assertEqual(len(callbacks), 1)
for cbargs in callbacks:
keys = set(cbargs.keys())
k_expect = set(['type', 'image'])
self.assertEqual(keys, k_expect)
self.assertEqual(cbargs['type'], 'image')
for key in ['postImport:image']:
callbacks = cb_idx[key]
self.assertEqual(len(callbacks), 1)
for cbargs in callbacks:
keys = set(cbargs.keys())
k_expect = set(['type', 'image', 'build', 'fullpath'])
self.assertEqual(keys, k_expect)
self.assertEqual(cbargs['type'], 'image')
self.assertEqual(cbargs['build'], buildinfo)
for key in ['preImport:archive', 'postImport:archive']:
callbacks = cb_idx[key]
self.assertEqual(len(callbacks), 5)
for cbargs in callbacks:
keys = set(cbargs.keys())
k_expect = set(['filepath', 'build_type', 'build', 'fileinfo', 'type', 'archive'])
self.assertEqual(keys, k_expect)
self.assertEqual(cbargs['type'], 'archive')
self.assertEqual(cbargs['build'], buildinfo)
# db operations
# with our other mocks, we should never reach _dml
self._dml.assert_not_called()
data = {'inserts': self.inserts, 'updates': self.updates}
self.assertEqual(data, self.db_expect)