810 lines
31 KiB
Python
Executable file
810 lines
31 KiB
Python
Executable file
#!/usr/bin/python3
|
|
|
|
# Koji daemon that runs in a Windows VM and executes commands associated
|
|
# with a task.
|
|
# Copyright (c) 2010-2014 Red Hat, Inc.
|
|
#
|
|
# Koji is free software; you can redistribute it and/or
|
|
# modify it under the terms of the GNU Lesser General Public
|
|
# License as published by the Free Software Foundation;
|
|
# version 2.1 of the License.
|
|
#
|
|
# This software is distributed in the hope that it will be useful,
|
|
# but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
# Lesser General Public License for more details.
|
|
#
|
|
# You should have received a copy of the GNU Lesser General Public
|
|
# License along with this software; if not, write to the Free Software
|
|
# Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
|
#
|
|
# Authors:
|
|
# Mike Bonnet <mikeb@redhat.com>
|
|
# Jay Greguske <jgregusk@redhat.com>
|
|
#
|
|
# To register this script as a service on Windows 2008 (with Cygwin 1.7.5 installed) run:
|
|
# kojiwind --install
|
|
# in a cygwin shell.
|
|
|
|
import base64
|
|
import glob
|
|
import hashlib
|
|
import logging
|
|
import os
|
|
import re
|
|
import subprocess
|
|
import sys
|
|
import tempfile
|
|
import threading
|
|
import time
|
|
import traceback
|
|
import xmlrpc.client
|
|
# urllib is required by the SCM class which is substituted into this file
|
|
# do not remove the import below
|
|
import urllib # noqa: F401
|
|
import zipfile
|
|
from configparser import RawConfigParser
|
|
from optparse import OptionParser
|
|
|
|
import six # noqa: F401, needed for imported code
|
|
|
|
|
|
MANAGER_PORT = 7000
|
|
|
|
KOJIKAMID = True
|
|
|
|
# INSERT kojikamid dup #
|
|
|
|
|
|
class fakemodule(object):
|
|
pass
|
|
|
|
|
|
# make parts of the above insert accessible as koji.X
|
|
koji = fakemodule()
|
|
koji.GenericError = GenericError # noqa: F821
|
|
koji.BuildError = BuildError # noqa: F821
|
|
koji._open_text_file = _open_text_file # noqa: F821
|
|
|
|
|
|
def encode_int(n):
|
|
"""If n is too large for a 32bit signed, convert it to a string"""
|
|
if n <= 2147483647:
|
|
return n
|
|
# else
|
|
return str(n)
|
|
|
|
|
|
class WindowsBuild(object):
|
|
|
|
LEADING_CHAR = re.compile('^[^A-Za-z_]')
|
|
VAR_CHARS = re.compile('[^A-Za-z0-9_]')
|
|
|
|
def __init__(self, server):
|
|
"""Get task info and setup build directory"""
|
|
self.logger = logging.getLogger('koji.vm')
|
|
self.server = server
|
|
info = server.getTaskInfo()
|
|
self.source_url = info[0]
|
|
self.build_tag = info[1]
|
|
if len(info) > 2:
|
|
self.task_opts = info[2]
|
|
else:
|
|
self.task_opts = {}
|
|
self.workdir = '/tmp/build'
|
|
ensuredir(self.workdir) # noqa: F821
|
|
self.buildreq_dir = os.path.join(self.workdir, 'buildreqs')
|
|
ensuredir(self.buildreq_dir) # noqa: F821
|
|
self.source_dir = None
|
|
self.spec_dir = None
|
|
self.patches_dir = None
|
|
self.buildroot_id = None
|
|
|
|
# we initialize these here for clarity, but they are populated in loadConfig()
|
|
self.name = None
|
|
self.version = None
|
|
self.release = None
|
|
self.epoch = None
|
|
self.description = None
|
|
self.platform = None
|
|
self.preinstalled = []
|
|
self.buildrequires = []
|
|
self.provides = []
|
|
self.shell = None
|
|
self.execute = []
|
|
self.postbuild = []
|
|
self.output = {}
|
|
self.logs = []
|
|
|
|
def checkTools(self):
|
|
"""Is this environment fit to build in, based on the spec file?"""
|
|
errors = []
|
|
for entry in self.preinstalled:
|
|
checkdir = False
|
|
if entry.startswith('/'):
|
|
# Cygwin path
|
|
if entry.endswith('/'):
|
|
# directory
|
|
checkdir = True
|
|
elif entry[1:3] == ':\\':
|
|
# Windows path
|
|
if entry.endswith('\\'):
|
|
# directory
|
|
checkdir = True
|
|
else:
|
|
# Check in the path
|
|
ret, output = run(['/bin/which', entry], log=False)
|
|
output = output.strip()
|
|
if ret:
|
|
errors.append(output)
|
|
else:
|
|
self.logger.info('command %s is available at %s', entry, output)
|
|
continue
|
|
if checkdir:
|
|
if not os.path.isdir(entry):
|
|
errors.append('directory %s does not exist' % entry)
|
|
else:
|
|
self.logger.info('directory %s exists', entry)
|
|
else:
|
|
# file
|
|
if not os.path.isfile(entry):
|
|
errors.append('file %s does not exist' % entry)
|
|
else:
|
|
self.logger.info('file %s exists', entry)
|
|
if errors:
|
|
raise BuildError('error validating build environment: %s' % # noqa: F821
|
|
', '.join(errors))
|
|
|
|
def updateClam(self):
|
|
"""update ClamAV virus definitions"""
|
|
ret, output = run(['/bin/freshclam', '--quiet'])
|
|
if ret:
|
|
raise BuildError('could not update ClamAV database: %s' % output) # noqa: F821
|
|
|
|
def checkEnv(self):
|
|
"""make the environment is fit for building in"""
|
|
for tool in ['/bin/freshclam', '/bin/clamscan', '/bin/patch']:
|
|
if not os.path.isfile(tool):
|
|
raise BuildError('%s is missing from the build environment' % tool) # noqa: F821
|
|
|
|
def zipDir(self, rootdir, filename):
|
|
rootbase = os.path.basename(rootdir)
|
|
roottrim = len(rootdir) - len(rootbase)
|
|
zfo = zipfile.ZipFile(filename, 'w', zipfile.ZIP_DEFLATED)
|
|
for dirpath, dirnames, filenames in os.walk(rootdir):
|
|
for skip in ['CVS', '.svn', '.git']:
|
|
if skip in dirnames:
|
|
dirnames.remove(skip)
|
|
for filename in filenames:
|
|
filepath = os.path.join(dirpath, filename)
|
|
zfo.write(filepath, filepath[roottrim:])
|
|
zfo.close()
|
|
|
|
def checkout(self):
|
|
"""Checkout sources, winspec, and patches, and apply patches"""
|
|
src_scm = SCM(self.source_url) # noqa: F821
|
|
self.source_dir = src_scm.checkout(
|
|
ensuredir(os.path.join(self.workdir, 'source'))) # noqa: F821
|
|
self.zipDir(self.source_dir, os.path.join(self.workdir, 'sources.zip'))
|
|
if 'winspec' in self.task_opts:
|
|
spec_scm = SCM(self.task_opts['winspec']) # noqa: F821
|
|
self.spec_dir = spec_scm.checkout(
|
|
ensuredir(os.path.join(self.workdir, 'spec'))) # noqa: F821
|
|
self.zipDir(self.spec_dir, os.path.join(self.workdir, 'spec.zip'))
|
|
else:
|
|
self.spec_dir = self.source_dir
|
|
if 'patches' in self.task_opts:
|
|
patch_scm = SCM(self.task_opts['patches']) # noqa: F821
|
|
self.patches_dir = patch_scm.checkout(
|
|
ensuredir(os.path.join(self.workdir, 'patches'))) # noqa: F821
|
|
self.zipDir(self.patches_dir, os.path.join(self.workdir, 'patches.zip'))
|
|
self.applyPatches(self.source_dir, self.patches_dir)
|
|
self.virusCheck(self.workdir)
|
|
|
|
def applyPatches(self, sourcedir, patchdir):
|
|
"""Apply patches in patchdir to files in sourcedir)"""
|
|
patches = [patch for patch in os.listdir(patchdir) if
|
|
os.path.isfile(os.path.join(patchdir, patch)) and
|
|
patch.endswith('.patch')]
|
|
if not patches:
|
|
raise BuildError('no patches found at %s' % patchdir) # noqa: F821
|
|
patches.sort()
|
|
for patch in patches:
|
|
cmd = ['/bin/patch', '--verbose', '-d', sourcedir, '-p1', '-i',
|
|
os.path.join(patchdir, patch)]
|
|
run(cmd, fatal=True)
|
|
|
|
def loadConfig(self):
|
|
"""Load build configuration from the spec file."""
|
|
specfiles = [spec for spec in os.listdir(self.spec_dir) if spec.endswith('.ini')]
|
|
if len(specfiles) == 0:
|
|
raise BuildError('No .ini file found') # noqa: F821
|
|
elif len(specfiles) > 1:
|
|
raise BuildError('Multiple .ini files found') # noqa: F821
|
|
|
|
conf = RawConfigParser()
|
|
conf.read(os.path.join(self.spec_dir, specfiles[0]))
|
|
|
|
# [naming] section
|
|
for entry in ('name', 'version', 'release', 'description'):
|
|
setattr(self, entry, conf.get('naming', entry))
|
|
if conf.has_option('naming', 'epoch'):
|
|
self.epoch = conf.get('naming', 'epoch')
|
|
|
|
# [building] section
|
|
self.platform = conf.get('building', 'platform')
|
|
|
|
# preinstalled are paths to files or directories that must exist
|
|
# in the VM for it to execute the build.
|
|
# If the path ends in / or \ it must be a directory, otherwise it must
|
|
# be a file.
|
|
# They may be specified as Cygwin (/cygdrive/c/...) or Windows (C:\...)
|
|
# absolute paths, or without a path in which case it is searched for
|
|
# on the PATH.
|
|
if conf.has_option('building', 'preinstalled'):
|
|
self.preinstalled.extend(
|
|
[e.strip() for e in conf.get('building', 'preinstalled').split('\n') if e])
|
|
|
|
# buildrequires and provides are multi-valued (space-separated)
|
|
for br in conf.get('building', 'buildrequires').split():
|
|
# buildrequires is a space-separated list
|
|
# each item in the list is in the format:
|
|
# pkgname[:opt1:opt2=val2:...]
|
|
# the options are put into a dict
|
|
# if the option has no =val, the value in the dict will be None
|
|
if br:
|
|
br = br.split(':')
|
|
bropts = {}
|
|
for opt in br[1:]:
|
|
if '=' in opt:
|
|
key, val = opt.split('=', 1)
|
|
else:
|
|
key = opt
|
|
val = None
|
|
bropts[key] = val
|
|
self.buildrequires.append((br[0], bropts))
|
|
for prov in conf.get('building', 'provides').split():
|
|
if prov:
|
|
self.provides.append(prov)
|
|
# optionally specify a shell to use (defaults to bash)
|
|
# valid values are: cmd, cmd.exe (alias for cmd), and bash
|
|
if conf.has_option('building', 'shell'):
|
|
self.shell = conf.get('building', 'shell')
|
|
else:
|
|
self.shell = 'bash'
|
|
# execute is multi-valued (newline-separated)
|
|
self.execute.extend([e.strip() for e in conf.get('building', 'execute').split('\n') if e])
|
|
|
|
# postbuild are files or directories that must exist after the build is
|
|
# complete, but are not included in the build output
|
|
# they are specified as paths relative the source directory, and may be
|
|
# in Unix or Windows format
|
|
# each entry may contain shell-style globs, and one or more files
|
|
# matching the glob is considered valid
|
|
if conf.has_option('building', 'postbuild'):
|
|
for entry in conf.get('building', 'postbuild').split('\n'):
|
|
entry = entry.strip()
|
|
if not entry:
|
|
continue
|
|
for var in ('name', 'version', 'release'):
|
|
entry = entry.replace('$' + var, getattr(self, var))
|
|
self.postbuild.append(entry)
|
|
|
|
# [files] section
|
|
for entry in conf.get('files', 'output').split('\n'):
|
|
entry = entry.strip()
|
|
if not entry:
|
|
continue
|
|
tokens = entry.split(':')
|
|
filename = tokens[0]
|
|
for var in ('name', 'version', 'release'):
|
|
filename = filename.replace('$' + var, getattr(self, var))
|
|
metadata = {}
|
|
metadata['platforms'] = tokens[1].split(',')
|
|
if len(tokens) > 2:
|
|
metadata['flags'] = tokens[2].split(',')
|
|
else:
|
|
metadata['flags'] = []
|
|
self.output[filename] = metadata
|
|
self.logs.extend([e.strip() for e in conf.get('files', 'logs').split('\n') if e])
|
|
|
|
def initBuildroot(self):
|
|
"""Create the buildroot object on the hub."""
|
|
repo_id = self.task_opts.get('repo_id')
|
|
if not repo_id:
|
|
raise BuildError('repo_id must be specified') # noqa: F821
|
|
self.buildroot_id = self.server.initBuildroot(repo_id, self.platform)
|
|
|
|
def expireBuildroot(self):
|
|
"""Set the buildroot object to expired on the hub."""
|
|
self.server.expireBuildroot(self.buildroot_id)
|
|
|
|
def fetchFile(self, basedir, buildinfo, fileinfo, brtype):
|
|
"""Download the file from buildreq, at filepath, into the basedir"""
|
|
destpath = os.path.join(basedir, fileinfo['localpath'])
|
|
ensuredir(os.path.dirname(destpath)) # noqa: F821
|
|
if 'checksum_type' in fileinfo:
|
|
checksum_type = CHECKSUM_TYPES[fileinfo['checksum_type']] # noqa: F821
|
|
if checksum_type == 'sha1':
|
|
checksum = hashlib.sha1()
|
|
elif checksum_type == 'sha256':
|
|
checksum = hashlib.sha256()
|
|
elif checksum_type == 'md5':
|
|
checksum = md5_constructor.md5() # noqa: F821
|
|
else:
|
|
raise BuildError('Unknown checksum type %s for %s' % ( # noqa: F821
|
|
checksum_type,
|
|
os.path.basename(fileinfo['localpath'])))
|
|
with koji._open_text_file(destpath, 'wt') as destfile:
|
|
offset = 0
|
|
while True:
|
|
encoded = self.server.getFile(buildinfo, fileinfo, encode_int(offset), 1048576,
|
|
brtype)
|
|
if not encoded:
|
|
break
|
|
data = base64.b64decode(encoded)
|
|
del encoded
|
|
destfile.write(data)
|
|
offset += len(data)
|
|
if 'checksum_type' in fileinfo:
|
|
checksum.update(data)
|
|
# rpms don't have a checksum in the fileinfo, but check it for everything else
|
|
if 'checksum_type' in fileinfo:
|
|
digest = checksum.hexdigest()
|
|
if fileinfo['checksum'] != digest:
|
|
raise BuildError( # noqa: F821
|
|
'checksum validation failed for %s, %s (computed) != %s (provided)' %
|
|
(destpath, digest, fileinfo['checksum']))
|
|
self.logger.info(
|
|
'Retrieved %s (%s bytes, %s: %s)', destpath, offset, checksum_type, digest)
|
|
else:
|
|
self.logger.info('Retrieved %s (%s bytes)', destpath, offset)
|
|
|
|
def fetchBuildReqs(self):
|
|
"""Retrieve buildrequires listed in the spec file"""
|
|
files = []
|
|
rpms = []
|
|
for buildreq, brinfo in self.buildrequires:
|
|
# if no type is specified in the options, default to win
|
|
brtype = brinfo.get('type', 'win')
|
|
buildinfo = self.server.getLatestBuild(self.build_tag, buildreq,
|
|
self.task_opts.get('repo_id'))
|
|
br_dir = os.path.join(self.buildreq_dir, buildreq, brtype)
|
|
ensuredir(br_dir) # noqa: F821
|
|
brinfo['dir'] = br_dir
|
|
brfiles = []
|
|
brinfo['files'] = brfiles
|
|
buildfiles = self.server.getFileList(buildinfo['id'], brtype, brinfo)
|
|
for fileinfo in buildfiles:
|
|
self.fetchFile(br_dir, buildinfo, fileinfo, brtype)
|
|
brfiles.append(fileinfo['localpath'])
|
|
if brtype == 'rpm':
|
|
rpms.append(fileinfo)
|
|
else:
|
|
files.append(fileinfo)
|
|
self.server.updateBuildrootFiles(self.buildroot_id, files, rpms)
|
|
self.virusCheck(self.buildreq_dir)
|
|
|
|
def build(self):
|
|
if self.shell in ('cmd', 'cmd.exe'):
|
|
self.cmdBuild()
|
|
else:
|
|
self.bashBuild()
|
|
# move the zips of the SCM checkouts to their final locations
|
|
for src in ['sources.zip', 'spec.zip', 'patches.zip']:
|
|
srcpath = os.path.join(self.workdir, src)
|
|
if os.path.exists(srcpath):
|
|
dest = '%s-%s-%s-%s' % (self.name, self.version, self.release, src)
|
|
destpath = os.path.join(self.source_dir, dest)
|
|
os.rename(srcpath, destpath)
|
|
self.output[dest] = {'platforms': ['all'],
|
|
'flags': ['src']}
|
|
|
|
def varname(self, name):
|
|
"""
|
|
Convert name to a valid shell variable name.
|
|
Converts leading characters that aren't letters or underscores
|
|
to underscores.
|
|
Converts any other characters that aren't letters, numbers,
|
|
or underscores to underscores.
|
|
"""
|
|
name = self.LEADING_CHAR.sub('_', name)
|
|
name = self.VAR_CHARS.sub('_', name)
|
|
return name
|
|
|
|
def cmdBuild(self):
|
|
"""Do the build: run the execute line(s) with cmd.exe"""
|
|
tmpfd, tmpname = tempfile.mkstemp(prefix='koji-tmp', suffix='.bat',
|
|
dir='/cygdrive/c/Windows/Temp')
|
|
script = koji._open_text_file(tmpfd, 'wt')
|
|
for attr in ['source_dir', 'spec_dir', 'patches_dir']:
|
|
val = getattr(self, attr)
|
|
if val:
|
|
ret, output = run(['/bin/cygpath', '-wa', val], log=False, fatal=True)
|
|
script.write('set %s=%s\r\n' % (attr, output.strip()))
|
|
for buildreq, brinfo in self.buildrequires:
|
|
buildreq = self.varname(buildreq)
|
|
ret, output = run(['/bin/cygpath', '-wa', brinfo['dir']], log=False, fatal=True)
|
|
br_dir = output.strip()
|
|
files = ' '.join(brinfo['files'])
|
|
files.replace('/', '\\')
|
|
if brinfo.get('type'):
|
|
# if the spec file qualifies the buildreq with a type,
|
|
# the env. var is named buildreq_type_{dir,files}
|
|
script.write('set %s_%s_dir=%s\r\n' % (buildreq, brinfo['type'], br_dir))
|
|
script.write('set %s_%s_files=%s\r\n' % (buildreq, brinfo['type'], files))
|
|
else:
|
|
# otherwise it's just buildreq_{dir,files}
|
|
script.write('set %s_dir=%s\r\n' % (buildreq, br_dir))
|
|
script.write('set %s_files=%s\r\n' % (buildreq, files))
|
|
script.write('\r\n')
|
|
script.write('set name=%s\r\n' % self.name)
|
|
script.write('set version=%s\r\n' % self.version)
|
|
script.write('set release=%s\r\n' % self.release)
|
|
for cmd in self.execute:
|
|
script.write(cmd)
|
|
script.write('\r\n')
|
|
script.close()
|
|
cmd = ['cmd.exe', '/C', 'C:\\Windows\\Temp\\' + os.path.basename(tmpname)]
|
|
ret, output = run(cmd, chdir=self.source_dir)
|
|
if ret:
|
|
raise BuildError('build command failed, see build.log for details') # noqa: F821
|
|
|
|
def bashBuild(self):
|
|
"""Do the build: run the execute line(s) with bash"""
|
|
tmpfd, tmpname = tempfile.mkstemp(prefix='koji-tmp.', dir='/tmp')
|
|
script = koji._open_text_file(tmpfd, 'wt')
|
|
script.write("export source_dir='%s'\n" % self.source_dir)
|
|
script.write("export spec_dir='%s'\n" % self.spec_dir)
|
|
if self.patches_dir:
|
|
script.write("export patches_dir='%s'\n" % self.patches_dir)
|
|
for buildreq, brinfo in self.buildrequires:
|
|
buildreq = self.varname(buildreq)
|
|
if brinfo.get('type'):
|
|
script.write("export %s_%s_dir='%s'\n" % (buildreq, brinfo['type'], brinfo['dir']))
|
|
script.write("export %s_%s_files='" % (buildreq, brinfo['type']))
|
|
else:
|
|
script.write("export %s_dir='%s'\n" % (buildreq, brinfo['dir']))
|
|
script.write("export %s_files='" % buildreq)
|
|
for filename in brinfo['files']:
|
|
script.write(filename)
|
|
script.write('\n')
|
|
script.write("'\n\n")
|
|
script.write('export name=%s\n' % self.name)
|
|
script.write('export version=%s\n' % self.version)
|
|
script.write('export release=%s\n' % self.release)
|
|
for cmd in self.execute:
|
|
script.write(cmd)
|
|
script.write('\n')
|
|
script.close()
|
|
cmd = ['/bin/bash', '-e', '-x', tmpname]
|
|
ret, output = run(cmd, chdir=self.source_dir)
|
|
if ret:
|
|
raise BuildError('build command failed, see build.log for details') # noqa: F821
|
|
|
|
def checkBuild(self):
|
|
"""Verify that the build completed successfully."""
|
|
errors = []
|
|
for entry in self.postbuild:
|
|
relpath = entry
|
|
if '\\' in relpath:
|
|
relpath = relpath.replace('\\', '/')
|
|
fullpath = os.path.join(self.source_dir, relpath)
|
|
results = glob.glob(fullpath)
|
|
if fullpath.endswith('/'):
|
|
for result in results:
|
|
if os.path.isdir(result):
|
|
self.logger.info('found directory %s at %s', entry, result)
|
|
break
|
|
else:
|
|
errors.append('directory %s does not exist' % entry)
|
|
else:
|
|
for result in results:
|
|
if os.path.isfile(result):
|
|
self.logger.info('found file %s at %s', entry, result)
|
|
break
|
|
else:
|
|
errors.append('file %s does not exist' % entry)
|
|
self.virusCheck(self.workdir)
|
|
if errors:
|
|
raise BuildError('error validating build output: %s' % # noqa: F821
|
|
', '.join(errors))
|
|
|
|
def virusCheck(self, path):
|
|
"""ensure a path is virus free with ClamAV. path should be absolute"""
|
|
if not path.startswith('/'):
|
|
raise BuildError('Invalid path to scan for viruses: ' + path) # noqa: F821
|
|
run(['/bin/clamscan', '--quiet', '--recursive', path], fatal=True)
|
|
|
|
def gatherResults(self):
|
|
"""Gather information about the output from the build, return it"""
|
|
return {'name': self.name, 'version': self.version, 'release': self.release,
|
|
'epoch': self.epoch,
|
|
'description': self.description, 'platform': self.platform,
|
|
'provides': self.provides,
|
|
'output': self.output, 'logs': self.logs,
|
|
'buildroot_id': self.buildroot_id}
|
|
|
|
def run(self):
|
|
"""Run the entire build process"""
|
|
self.checkEnv()
|
|
self.updateClam()
|
|
self.checkout()
|
|
self.loadConfig()
|
|
self.initBuildroot()
|
|
self.checkTools()
|
|
self.fetchBuildReqs()
|
|
self.build()
|
|
self.checkBuild()
|
|
self.expireBuildroot()
|
|
return self.gatherResults()
|
|
|
|
|
|
def run(cmd, chdir=None, fatal=False, log=True):
|
|
global logfd
|
|
output = ''
|
|
olddir = None
|
|
if chdir:
|
|
olddir = os.getcwd()
|
|
os.chdir(chdir)
|
|
if log:
|
|
logger = logging.getLogger('koji.vm')
|
|
logger.info('$ %s', ' '.join(cmd))
|
|
proc = subprocess.Popen(cmd, stdout=logfd, stderr=subprocess.STDOUT,
|
|
close_fds=True, text=True)
|
|
ret = proc.wait()
|
|
else:
|
|
proc = subprocess.Popen(cmd, stdout=subprocess.PIPE, stderr=subprocess.STDOUT,
|
|
close_fds=True, text=True)
|
|
output, dummy = proc.communicate()
|
|
ret = proc.returncode
|
|
if olddir:
|
|
os.chdir(olddir)
|
|
if ret and fatal:
|
|
msg = 'error running: %s, return code was %s' % (' '.join(cmd), ret)
|
|
if log:
|
|
msg += ', see %s for details' % (os.path.basename(logfd.name))
|
|
else:
|
|
msg += ', output: %s' % output
|
|
raise BuildError(msg) # noqa: F821
|
|
return ret, output
|
|
|
|
|
|
def find_net_info():
|
|
"""
|
|
Find the network gateway configured for this VM.
|
|
"""
|
|
ret, output = run(['ipconfig', '/all'], log=False)
|
|
if ret:
|
|
raise RuntimeError('error running ipconfig, output was: %s' % output)
|
|
macaddr = None
|
|
gateway = None
|
|
for line in output.splitlines():
|
|
line = line.strip()
|
|
# take the first values we find
|
|
if line.startswith('Physical Address'):
|
|
if not macaddr:
|
|
macaddr = line.split()[-1]
|
|
# format it to be consistent with the libvirt MAC address
|
|
macaddr = macaddr.replace('-', ':').lower()
|
|
elif line.startswith('Default Gateway'):
|
|
if not gateway:
|
|
gateway = line.split()[-1]
|
|
|
|
# check that we have valid values
|
|
if macaddr and len(macaddr) != 17:
|
|
macaddr = None
|
|
if gateway and (len(gateway) < 7 or len(gateway) > 15):
|
|
gateway = None
|
|
return macaddr, gateway
|
|
|
|
|
|
def upload_file(server, prefix, path):
|
|
"""upload a single file to the vmd"""
|
|
logger = logging.getLogger('koji.vm')
|
|
destpath = os.path.join(prefix, path)
|
|
fobj = open(destpath, 'rb')
|
|
offset = 0
|
|
sum = hashlib.sha256()
|
|
while True:
|
|
data = fobj.read(131072)
|
|
if not data:
|
|
break
|
|
encoded = base64.b64encode(data).decode()
|
|
server.upload(path, encode_int(offset), encoded)
|
|
offset += len(data)
|
|
sum.update(data)
|
|
fobj.close()
|
|
digest = sum.hexdigest()
|
|
server.verifyChecksum(path, digest, 'sha256')
|
|
logger.info('Uploaded %s (%s bytes, sha256: %s)', destpath, offset, digest)
|
|
|
|
|
|
def get_mgmt_server():
|
|
"""Get a ServerProxy object we can use to retrieve task info"""
|
|
logger = logging.getLogger('koji.vm')
|
|
macaddr, gateway = find_net_info()
|
|
while not (macaddr and gateway):
|
|
# wait for the network connection to come up and get an address
|
|
time.sleep(5)
|
|
macaddr, gateway = find_net_info()
|
|
logger.debug('found MAC address %s, connecting to %s:%s',
|
|
macaddr, gateway, MANAGER_PORT)
|
|
server = xmlrpc.client.ServerProxy('http://%s:%s/' % (gateway, MANAGER_PORT), allow_none=True)
|
|
# we would set a timeout on the socket here, but that is apparently not
|
|
# supported by python/cygwin/Windows
|
|
task_port = server.getPort(macaddr)
|
|
logger.debug('found task-specific port %s', task_port)
|
|
return xmlrpc.client.ServerProxy('http://%s:%s/' % (gateway, task_port), allow_none=True)
|
|
|
|
|
|
def get_options():
|
|
"""handle usage and parse options"""
|
|
usage = """%prog [options]
|
|
Run Koji tasks assigned to a VM.
|
|
Run without any arguments to start this daemon.
|
|
"""
|
|
parser = OptionParser(usage=usage)
|
|
parser.add_option('-d', '--debug', action='store_true', help='Log debug statements')
|
|
parser.add_option('-i', '--install', action='store_true', default=False,
|
|
help='Install this daemon as a service')
|
|
parser.add_option('-u', '--uninstall', action='store_true', default=False,
|
|
help='Uninstall this daemon if it was installed previously as a service')
|
|
(options, args) = parser.parse_args()
|
|
return options
|
|
|
|
|
|
def setup_logging(opts):
|
|
global logfile, logfd
|
|
logger = logging.getLogger('koji.vm')
|
|
level = logging.INFO
|
|
if opts.debug:
|
|
level = logging.DEBUG
|
|
logger.setLevel(level)
|
|
logfd = koji._open_text_file(logfile, 'wt')
|
|
handler = logging.StreamHandler(logfd)
|
|
handler.setLevel(level)
|
|
handler.setFormatter(logging.Formatter('%(asctime)s [%(levelname)s] %(name)s: %(message)s'))
|
|
logger.addHandler(handler)
|
|
return handler
|
|
|
|
|
|
def log_local(msg):
|
|
tb = ''.join(traceback.format_exception(*sys.exc_info()))
|
|
sys.stderr.write('%s: %s\n' % (time.ctime(), msg))
|
|
sys.stderr.write(tb)
|
|
|
|
|
|
def stream_logs(server, handler, builds):
|
|
"""Stream logs incrementally to the server.
|
|
The global logfile will always be streamed.
|
|
The logfiles associated with any builds
|
|
will also be streamed."""
|
|
global logfile
|
|
logs = {logfile: (os.path.basename(logfile), None)}
|
|
while handler.active:
|
|
for build in builds:
|
|
for relpath in build.logs:
|
|
logpath = os.path.join(build.source_dir, relpath)
|
|
if logpath not in logs:
|
|
logs[logpath] = (relpath, None)
|
|
for log, (relpath, fd) in logs.items():
|
|
if not fd:
|
|
if os.path.isfile(log):
|
|
try:
|
|
fd = open(log, 'rb')
|
|
logs[log] = (relpath, fd)
|
|
except Exception:
|
|
log_local('Error opening %s' % log)
|
|
continue
|
|
else:
|
|
continue
|
|
offset = fd.tell()
|
|
contents = fd.read(65536)
|
|
if contents:
|
|
size = len(contents)
|
|
data = base64.b64encode(contents).decode()
|
|
digest = hashlib.sha256(contents).hexdigest()
|
|
del contents
|
|
try:
|
|
server.uploadDirect(relpath, offset, size, ('sha256', digest), data)
|
|
except Exception:
|
|
log_local('error uploading %s' % relpath)
|
|
time.sleep(1)
|
|
|
|
|
|
def fail(server, handler):
|
|
"""do the right thing when a build fails"""
|
|
global logfile, logfd
|
|
logging.getLogger('koji.vm').error('error running build', exc_info=True)
|
|
tb = ''.join(traceback.format_exception(*sys.exc_info()))
|
|
handler.active = False
|
|
if server is not None:
|
|
try:
|
|
logfd.flush()
|
|
upload_file(server, os.path.dirname(logfile),
|
|
os.path.basename(logfile))
|
|
except Exception:
|
|
log_local('error calling upload_file()')
|
|
while True:
|
|
try:
|
|
# this is the very last thing we do, keep trying as long as we can
|
|
server.failTask(tb)
|
|
break
|
|
except Exception:
|
|
log_local('error calling server.failTask()')
|
|
sys.exit(1)
|
|
|
|
|
|
logfile = '/tmp/build.log'
|
|
logfd = None
|
|
|
|
|
|
def main():
|
|
prog = os.path.basename(sys.argv[0])
|
|
opts = get_options()
|
|
if opts.install:
|
|
ret, output = run(['/bin/cygrunsrv', '--install', prog,
|
|
'--path', sys.executable, '--args', os.path.abspath(prog),
|
|
'--type', 'auto', '--dep', 'Dhcp',
|
|
'--disp', 'Koji Windows Daemon',
|
|
'--desc', 'Runs Koji tasks assigned to a VM'],
|
|
log=False)
|
|
if ret:
|
|
print('Error installing %s service, output was: %s' % (prog, output))
|
|
sys.exit(1)
|
|
else:
|
|
print('Successfully installed the %s service' % prog)
|
|
sys.exit(0)
|
|
elif opts.uninstall:
|
|
ret, output = run(['/bin/cygrunsrv', '--remove', prog], log=False)
|
|
if ret:
|
|
print('Error removing the %s service, output was: %s' % (prog, output))
|
|
sys.exit(1)
|
|
else:
|
|
print('Successfully removed the %s service' % prog)
|
|
sys.exit(0)
|
|
|
|
handler = setup_logging(opts)
|
|
handler.active = True
|
|
server = None
|
|
try:
|
|
server = get_mgmt_server()
|
|
|
|
builds = []
|
|
thread = threading.Thread(target=stream_logs,
|
|
args=(server, handler, builds))
|
|
thread.daemon = True
|
|
thread.start()
|
|
|
|
# xmlrpclib is not thread-safe, create a new ServerProxy
|
|
# instance so we're not sharing with the stream_logs thread
|
|
server = get_mgmt_server()
|
|
|
|
build = WindowsBuild(server)
|
|
builds.append(build)
|
|
results = build.run()
|
|
|
|
for filename in results['output'].keys():
|
|
upload_file(server, build.source_dir, filename)
|
|
|
|
handler.active = False
|
|
thread.join()
|
|
|
|
for filename in results['logs']:
|
|
# reupload the log files to make sure the thread
|
|
# didn't miss anything
|
|
upload_file(server, build.source_dir, filename)
|
|
|
|
upload_file(server, os.path.dirname(logfile),
|
|
os.path.basename(logfile))
|
|
results['logs'].append(os.path.basename(logfile))
|
|
|
|
server.closeTask(results)
|
|
except Exception:
|
|
fail(server, handler)
|
|
sys.exit(0)
|
|
|
|
|
|
if __name__ == '__main__':
|
|
main()
|