From 2fdc271887db8c0ef0641472d00e850e8b3caa19 Mon Sep 17 00:00:00 2001 From: Xiaofeng Yan Date: Mon, 26 Mar 2012 18:49:26 +0800 Subject: archiver.bbclass: 1 Archive sources in ${S} in the different stage (do_unpack,do_patch,do_configure). 2 Archive patches including series 3 Archive logs including scripts (.bb and .inc files) 4 dump environment resources which show all variable and functions used to xxx.showdata.dump when running a task 5 dump all content in 's' including patches to file xxx.diff.gz All archiving packages will be deployed to ${DEPLOY_DIR}/sources/ [YOCTO #1977] Signed-off-by: Xiaofeng Yan Signed-off-by: Richard Purdie --- meta/classes/archiver.bbclass | 460 ++++++++++++++++++++++++++++++++++++++++++ 1 file changed, 460 insertions(+) create mode 100644 meta/classes/archiver.bbclass (limited to 'meta/classes') diff --git a/meta/classes/archiver.bbclass b/meta/classes/archiver.bbclass new file mode 100644 index 0000000000..471430e24f --- /dev/null +++ b/meta/classes/archiver.bbclass @@ -0,0 +1,460 @@ +# This file is used for archiving sources ,patches,and logs to tarball. +# It also output building environment to xxx.dump.data and create xxx.diff.gz to record +# all content in ${S} to a diff file. + +EXCLUDE_FROM ?= ".pc autom4te.cache" +ARCHIVE_TYPE ?= "TAR SRPM" +DISTRO ?= "poky" +PATCHES_ARCHIVE_WITH_SERIES = 'TRUE' + +def parse_var(d,var): + ''' parse variable like ${PV} in "require xxx_${PV}.inc" to a real value. for example, change "require xxx_${PV}.inc" to "require xxx_1.2.inc" ''' + import re + pat = re.compile('.*\$({(.*)}).*') + if '$' not in var and '/' not in var: + return var + else: + if '/' in var: + return [i for i in var.split('/') if i.endswith('.inc')][0] + elif '$' in var: + m = pat.match(var) + patstr = '\$' + m.group(1) + var_str = m.group(2) + return re.sub(patstr,d.getVar(var_str,True),var) + else: + return var + +def get_bb_inc(d): + '''create a directory "script-logs" including .bb and .inc file in ${WORKDIR}''' + import re + import os + import shutil + + bbinc = [] + pat=re.compile('require\s*([^\s]*\.*)(.*)') + file_dir = d.getVar('FILE', True) + bbdir = os.path.dirname(file_dir) + work_dir = d.getVar('WORKDIR', True) + os.chdir(work_dir) + bb.mkdirhier("script-logs") + os.chdir(bbdir) + bbfile = os.path.basename(file_dir) + bbinc.append(bbfile) + + def get_inc (file): + f = open(file,'r') + for line in f.readlines(): + if 'require' not in line: + bbinc.append(file) + else: + try: + incfile = pat.match(line).group(1) + incfile = parse_var(d,incfile) + bbinc.append(incfile) + get_inc(incfile) + except (IOError,AttributeError): + pass + get_inc(bbfile) + os.chdir(work_dir) + for root, dirs, files in os.walk(bbdir): + for file in bbinc: + if file in files: + shutil.copy(root + '/' + file,'script-logs') + oe.path.copytree('temp', 'script-logs') + return work_dir + '/script-logs' + +def get_all_patches(d): + '''copy patches and series file to a pointed directory which will be archived to tarball in ${WORKDIR}''' + import shutil + + src_patches=[] + pf = d.getVar('PF', True) + work_dir = d.getVar('WORKDIR', True) + s = d.getVar('S',True) + dest = os.path.join(work_dir, pf + '-patches') + shutil.rmtree(dest, ignore_errors=True) + bb.mkdirhier(dest) + + src_uri = d.getVar('SRC_URI', 1).split() + fetch = bb.fetch2.Fetch(src_uri, d) + locals = (fetch.localpath(url) for url in fetch.urls) + for local in locals: + src_patches.append(local) + if not cmp(work_dir,s): + tmp_list = src_patches + else: + tmp_list = src_patches[1:] + + for patch in tmp_list: + try: + shutil.copy(patch,dest) + except IOError: + if os.path.isdir(patch): + oe.path.copytree(patch,dest) + return dest + +def get_applying_patches(d): + """only copy applying patches to a pointed directory which will be archived to tarball""" + import os + import shutil + + + pf = d.getVar('PF', True) + work_dir = d.getVar('WORKDIR', True) + dest = os.path.join(work_dir, pf + '-patches') + shutil.rmtree(dest, ignore_errors=True) + bb.mkdirhier(dest) + + + patches = src_patches(d) + for patch in patches: + _, _, local, _, _, parm = bb.decodeurl(patch) + if local: + shutil.copy(local,dest) + return dest + +def not_tarball(d): + '''packages including key words 'work-shared','native', 'task-' will be passed''' + import os + + workdir = d.getVar('WORKDIR',True) + s = d.getVar('S',True) + if 'work-shared' in s or 'task-' in workdir or 'native' in workdir: + return True + else: + return False + +def get_source_from_downloads(d,middle_name): + '''copy tarball of $P to $WORKDIR when this tarball exists in $DL_DIR''' + if middle_name in 'patched' 'configured': + return + pf = d.getVar('PF', True) + dl_dir = d.getVar('DL_DIR',True) + try: + source = os.path.basename(d.getVar('SRC_URI', 1).split()[0]) + os.chdir(dl_dir) + if os.path.exists(source) and not os.path.isdir(source): + return source + except (IndexError, OSError): + pass + +def archive_sources_from_directory(d,middle_name): + '''archive sources codes tree to tarball when tarball of $P doesn't exist in $DL_DIR''' + import tarfile + import shutil + + s = d.getVar('S',True) + workdir=d.getVar('WORKDIR', True) + PF = d.getVar('PF',True) + tarname = PF + '-' + middle_name + ".tar.gz" + + if os.path.exists(s) and s is not workdir: + sourcedir = os.path.basename(s) + tarbase = os.path.dirname(s) + if not sourcedir or os.path.dirname(tarbase) == workdir: + sourcedir = os.path.basename(os.path.dirname(s)) + tarbase = os.path.dirname(os.path.dirname(s)) + os.chdir(tarbase) + else: + sourcedir = os.path.basename(s) + if not os.path.exists(sourcedir): + os.mkdir(sourcedir) + try: + for file in os.listdir(s): + if file is not 'temp' and file is not sourcedir: + shutil.copy(file,sourcedir) + except (IOError,OSError): + pass + + if (len(os.listdir(sourcedir))) != 0: + tar = tarfile.open( tarname, "w:gz") + tar.add(sourcedir) + tar.close() + if cmp(workdir,os.path.dirname(s)) and not os.path.exists(workdir + '/' + tarname): + shutil.move(os.path.dirname(s) + '/' + tarname,workdir) + else: + return + return tarname + +def archive_sources(d,middle_name): + '''copy tarball from $DL_DIR to $WORKDIR if have tarball, archive source codes tree in $WORKDIR if $P is directory instead of tarball''' + import shutil + work_dir = d.getVar('WORKDIR',True) + file = get_source_from_downloads(d,middle_name) + if file: + shutil.copy(file,work_dir) + else: + file = archive_sources_from_directory(d,middle_name) + return file + + +def archive_patches(d,patchdir,series): + '''archive patches to tarball and also include series files if 'series' is True''' + import tarfile + import shutil + + s = d.getVar('S',True) + work_dir = d.getVar('WORKDIR', True) + os.chdir(work_dir) + patch_dir = os.path.basename(patchdir) + tarname = patch_dir + ".tar.gz" + if series == 'all' and os.path.exists(s + '/patches/series'): + shutil.copy(s + '/patches/series',patch_dir) + tar = tarfile.open(tarname, "w:gz") + tar.add(patch_dir) + tar.close() + shutil.rmtree(patch_dir, ignore_errors=True) + return tarname + +def select_archive_patches(d,option): + '''select to archive all patches including non-applying and series or applying patches ''' + if option == "all": + patchdir = get_all_patches(d) + elif option == "applying": + patchdir = get_applying_patches(d) + try: + os.rmdir(patchdir) + except OSError: + tarpatch = archive_patches(d,patchdir,option) + return tarpatch + return + +def archive_logs(d,logdir,bbinc=False): + '''archive logs in temp to tarball and .bb and .inc files if bbinc is True ''' + import tarfile + import shutil + + log_dir = os.path.basename(logdir) + pf = d.getVar('PF',True) + tarname = pf + '-' + log_dir + ".tar.gz" + tar = tarfile.open(tarname, "w:gz") + tar.add(log_dir) + tar.close() + if bbinc: + shutil.rmtree(log_dir, ignore_errors=True) + return tarname + +def get_licenses(d): + '''get licenses for running .bb file''' + licenses = d.getVar('LICENSE', 1).replace('&', '|') + licenses = licenses.replace('(', '').replace(')', '') + clean_licenses = "" + for x in licenses.split(): + if x.strip() == '' or x == 'CLOSED': + continue + if x != "|": + clean_licenses += x + if '|' in clean_licenses: + clean_licenses = clean_licenses.replace('|','') + return clean_licenses + + +def move_tarball_deploy(d,tarball_list): + '''move tarball in location to ${DEPLOY_DIR}/sources''' + import shutil + + if tarball_list is []: + return + target_sys = d.getVar('TARGET_SYS', True) + pf = d.getVar('PF', True) + licenses = get_licenses(d) + tar_sources = d.getVar('DEPLOY_DIR', True) + '/sources/' + target_sys + '/' + licenses + '/' + pf + if not os.path.exists(tar_sources): + bb.mkdirhier(tar_sources) + for source in tarball_list: + if source: + if os.path.exists(tar_sources + '/' + source): + os.remove(tar_sources + '/' + source) + shutil.move(source,tar_sources) + +def verify_var(d): + '''check the type for archiving package('tar' or 'srpm')''' + try: + if d.getVar('SOURCE_ARCHIVE_PACKAGE_TYPE', True).upper() not in d.getVar('ARCHIVE_TYPE', True).split(): + raise AttributeError + except AttributeError: + bb.fatal("\"SOURCE_ARCHIVE_PACKAGE_TYPE\" is \'tar\' or \'srpm\', no other types") + +def store_package(d,package_name): + '''store tarbablls name to file "tar-package"''' + try: + f = open(d.getVar('WORKDIR',True )+ '/tar-package','a') + f.write(package_name + ' ') + f.close() + except IOError: + pass + +def get_package(d): + '''get tarballs name from "tar-package"''' + try: + os.chdir(d.getVar('WORKDIR', True)) + f = open('tar-package','r') + line = list(set(f.readline().replace('\n','').split())) + f.close() + return line + except IOError: + pass + + +def archive_sources_patches(d,middle_name): + '''archive sources and patches to tarball. middle_name will append strings ${middle_name} to ${PR} as middle name. for example, zlib-1.4.6-prepatch(middle_name).tar.gz ''' + import shutil + verify_var(d) + if not_tarball(d): + return + + source_tar_name = archive_sources(d,middle_name) + if middle_name == "prepatch": + if d.getVar('PATCHES_ARCHIVE_WITH_SERIES',True).upper() == 'TRUE': + patch_tar_name = select_archive_patches(d,"all") + elif d.getVar('PATCHES_ARCHIVE_WITH_SERIES',True).upper() == 'FALSE': + patch_tar_name = select_archive_patches(d,"applying") + else: + bb.fatal("Please define 'PATCHES_ARCHIVE_WITH_SERIES' is strings 'True' or 'False' ") + else: + patch_tar_name = '' + + if d.getVar('SOURCE_ARCHIVE_PACKAGE_TYPE', True).upper() not in 'SRPM': + move_tarball_deploy(d,[source_tar_name,patch_tar_name]) + else: + tarpackage = d.getVar('WORKDIR', True) + '/tar-package' + if os.path.exists(tarpackage): + os.remove(tarpackage) + for package in source_tar_name, patch_tar_name: + if package: + store_package(d,str(package) + ' ') + +def archive_scripts_logs(d): + '''archive scripts and logs. scripts include .bb and .inc files and logs include stuff in "temp".''' + + work_dir = d.getVar('WORKDIR', True) + os.chdir(work_dir) + source_archive_log_with_scripts = d.getVar('SOURCE_ARCHIVE_LOG_WITH_SCRIPTS', True) + if source_archive_log_with_scripts == 'logs_with_scripts': + logdir = get_bb_inc(d) + tarlog = archive_logs(d,logdir,True) + elif source_archive_log_with_scripts == 'logs': + if os.path.exists('temp'): + tarlog = archive_logs(d,'temp',False) + else: + return + + if d.getVar('SOURCE_ARCHIVE_PACKAGE_TYPE', True).upper() not in 'SRPM': + move_tarball_deploy(d,[tarlog]) + + else: + store_package(d,tarlog) + +def dumpdata(d): + '''dump environment to "${P}-${PR}.showdata.dump" including all kinds of variables and functions when running a task''' + workdir = bb.data.getVar('WORKDIR', d, 1) + distro = bb.data.getVar('DISTRO', d, 1) + s = d.getVar('S', True) + pf = d.getVar('PF', True) + target_sys = d.getVar('TARGET_SYS', True) + licenses = get_licenses(d) + dumpdir = d.getVar('DEPLOY_DIR', True) + '/sources/' + target_sys + '/' + licenses + '/' + pf + if not os.path.exists(dumpdir): + bb.mkdirhier(dumpdir) + + dumpfile = os.path.join(dumpdir, bb.data.expand("${P}-${PR}.showdata.dump",d)) + + bb.note("Dumping metadata into '%s'" % dumpfile) + f = open(dumpfile, "w") + # emit variables and shell functions + bb.data.emit_env(f, d, True) + # emit the metadata which isnt valid shell + for e in d.keys(): + if bb.data.getVarFlag(e, 'python', d): + f.write("\npython %s () {\n%s}\n" % (e, bb.data.getVar(e, d, 1))) + f.close() + +def create_diff_gz(d): + '''creating .diff.gz in ${DEPLOY_DIR_SRC}/${P}-${PR}.diff.g gz for mapping all content in 's' including patches to xxx.diff.gz''' + import shutil + + work_dir = d.getVar('WORKDIR', True) + exclude_from = d.getVar('EXCLUDE_FROM', True).split() + pf = d.getVar('PF', True) + licenses = get_licenses(d) + target_sys = d.getVar('TARGET_SYS', True) + diff_dir = d.getVar('DEPLOY_DIR', True) + '/sources/' + target_sys + '/' + licenses + '/' + pf + diff_file = os.path.join(diff_dir, bb.data.expand("${P}-${PR}.diff.gz",d)) + os.chdir(work_dir) + f = open('temp/exclude-from-file', 'a') + for i in exclude_from: + f.write(i) + f.write("\n") + f.close() + + s=d.getVar('S', True) + distro = d.getVar('DISTRO',True) + dest = s + '/' + distro + '/files' + if not os.path.exists(dest): + bb.mkdirhier(dest) + for i in os.listdir(os.getcwd()): + if os.path.isfile(i): + shutil.copy(i, dest) + + bb.note("Creating .diff.gz in ${DEPLOY_DIR_SRC}/${P}-${PR}.diff.gz") + cmd = "LC_ALL=C TZ=UTC0 diff --exclude-from=" + work_dir + "/temp/exclude-from-file -Naur " + s + '.org' + ' ' + s + " | gzip -c > " + diff_file + d.setVar('DIFF', cmd + "\n") + d.setVarFlag('DIFF', 'func', '1') + bb.build.exec_func('DIFF', d) + shutil.rmtree(s + '.org', ignore_errors=True) + +# This function will run when user want to get tarball for sources and patches after do_unpack +python do_archive_original_sources_patches(){ + archive_sources_patches(d,'prepatch') +} + +# This function will run when user want to get tarball for patched sources after do_patch +python do_archive_patched_sources(){ + archive_sources_patches(d,'patched') +} + +# This function will run when user want to get tarball for configured sources after do_configure +python do_archive_configured_sources(){ + archive_sources_patches(d,'configured') +} + +# This function will run when user want to get tarball for logs or both logs and scripts(.bb and .inc files) +python do_archive_scripts_logs(){ + archive_scripts_logs(d) +} + +# This function will run when user want to know what variable and functions in a running task are and also can get a diff file including +# all content a package should include. +python do_dumpdata_create_diff_gz(){ + dumpdata(d) + create_diff_gz(d) +} + +# This functions prepare for archiving "linux-yocto" because this package create directory 's' before do_patch instead of after do_unpack. +# This is special control for archiving linux-yocto only. +python do_archive_linux_yocto(){ + s = d.getVar('S', True) + if 'linux-yocto' in s: + source_tar_name = archive_sources(d,'') + if d.getVar('SOURCE_ARCHIVE_PACKAGE_TYPE', True).upper() not in 'SRPM': + move_tarball_deploy(d,[source_tar_name,'']) +} +do_kernel_checkout[postfuncs] += "do_archive_linux_yocto " + +# remove tarball for sources, patches and logs after creating srpm. +python do_remove_tarball(){ + if d.getVar('SOURCE_ARCHIVE_PACKAGE_TYPE', True).upper() == 'SRPM': + work_dir = d.getVar('WORKDIR', True) + os.chdir(work_dir) + try: + for file in os.listdir(os.getcwd()): + if file in get_package(d): + os.remove(file) + os.remove('tar-package') + except (TypeError,OSError): + pass +} +do_remove_taball[deptask] = "do_archive_scripts_logs" +do_package_write_rpm[postfuncs] += "do_remove_tarball " +export get_licenses +export get_package -- cgit v1.2.3