diff options
author | Robert Yang <liezhi.yang@windriver.com> | 2012-08-23 23:03:10 +0800 |
---|---|---|
committer | Saul Wold <sgw@linux.intel.com> | 2012-08-28 08:03:40 -0700 |
commit | 5eacbcdf306d9e743164a1563559cd24eb5fffe0 (patch) | |
tree | f5f55d1a4c69f948d8fff64421bbab92ce7d46a2 /meta/classes/archiver.bbclass | |
parent | 9a8b4fcac639404caa8ac87717118b3380239838 (diff) | |
download | openembedded-core-5eacbcdf306d9e743164a1563559cd24eb5fffe0.tar.gz openembedded-core-5eacbcdf306d9e743164a1563559cd24eb5fffe0.tar.bz2 openembedded-core-5eacbcdf306d9e743164a1563559cd24eb5fffe0.zip |
archiver.bbclass: indent fixes
Several fixes:
* It uses mixed tab and whitespace as the indent in one function,
Fix them to use "4 spaces" as the indent.
* Remove the unwanted blank, for example, more than one blank lines
appeared together.
* Remove the tail whitespace.
[YOCTO #2619]
Signed-off-by: Robert Yang <liezhi.yang@windriver.com>
Signed-off-by: Saul Wold <sgw@linux.intel.com>
Diffstat (limited to 'meta/classes/archiver.bbclass')
-rw-r--r-- | meta/classes/archiver.bbclass | 757 |
1 files changed, 379 insertions, 378 deletions
diff --git a/meta/classes/archiver.bbclass b/meta/classes/archiver.bbclass index 8e97e447c9..117ad02da4 100644 --- a/meta/classes/archiver.bbclass +++ b/meta/classes/archiver.bbclass @@ -1,6 +1,10 @@ +# ex:ts=4:sw=4:sts=4:et +# -*- tab-width: 4; c-basic-offset: 4; indent-tabs-mode: nil -*- +# # This file is used for archiving sources ,patches,and logs to tarball. -# It also output building environment to xxx.dump.data and create xxx.diff.gz to record -# all content in ${S} to a diff file. +# It also output building environment to xxx.dump.data and create xxx.diff.gz to record +# all content in ${S} to a diff file. +# ARCHIVE_EXCLUDE_FROM ?= ".pc autom4te.cache" ARCHIVE_TYPE ?= "TAR SRPM" @@ -67,454 +71,451 @@ def tar_filter(d): if d.getVar('FILTER', True).upper() == "YES": included, reason = copyleft_should_include(d) if not included: - return False + return False else: - return True + return True else: return False def get_bb_inc(d): - '''create a directory "script-logs" including .bb and .inc file in ${WORKDIR}''' - import re - import shutil - - bbinc = [] - pat=re.compile('require\s*([^\s]*\.*)(.*)') - work_dir = d.getVar('WORKDIR', True) - bbfile = d.getVar('FILE', True) - bbdir = os.path.dirname(bbfile) - script_logs = os.path.join(work_dir,'script-logs') - bb_inc = os.path.join(script_logs,'bb_inc') - bb.mkdirhier(script_logs) - bb.mkdirhier(bb_inc) - - def find_file(dir,file): - for root, dirs, files in os.walk(dir): - if file in files: - return os.path.join(root,file) - - def get_inc (file): - f = open(file,'r') - for line in f.readlines(): - if 'require' not in line: - bbinc.append(file) - else: - try: - incfile = pat.match(line).group(1) - incfile = bb.data.expand(os.path.basename(incfile),d) - abs_incfile = find_file(bbdir,incfile) - if abs_incfile: - bbinc.append(abs_incfile) - get_inc(abs_incfile) - except AttributeError: - pass - get_inc(bbfile) - bbinc = list(set(bbinc)) - for bbincfile in bbinc: - shutil.copy(bbincfile,bb_inc) - - try: - bb.mkdirhier(os.path.join(script_logs,'temp')) - oe.path.copytree(os.path.join(work_dir,'temp'), os.path.join(script_logs,'temp')) - except (IOError,AttributeError): - pass - return script_logs - -def get_series(d): - '''copy patches and series file to a pointed directory which will be archived to tarball in ${WORKDIR}''' - import shutil - - src_patches=[] - pf = d.getVar('PF', True) - work_dir = d.getVar('WORKDIR', True) - s = d.getVar('S',True) - dest = os.path.join(work_dir, pf + '-series') - shutil.rmtree(dest, ignore_errors=True) - bb.mkdirhier(dest) - - src_uri = d.getVar('SRC_URI', True).split() - fetch = bb.fetch2.Fetch(src_uri, d) - locals = (fetch.localpath(url) for url in fetch.urls) - for local in locals: - src_patches.append(local) - if not cmp(work_dir,s): - tmp_list = src_patches - else: - tmp_list = src_patches[1:] - - for patch in tmp_list: - try: - shutil.copy(patch,dest) - except IOError: - if os.path.isdir(patch): - bb.mkdirhier(os.path.join(dest,patch)) - oe.path.copytree(patch, os.path.join(dest,patch)) - return dest - -def get_applying_patches(d): - """only copy applying patches to a pointed directory which will be archived to tarball""" - import shutil + '''create a directory "script-logs" including .bb and .inc file in ${WORKDIR}''' + import re + import shutil + + bbinc = [] + pat=re.compile('require\s*([^\s]*\.*)(.*)') + work_dir = d.getVar('WORKDIR', True) + bbfile = d.getVar('FILE', True) + bbdir = os.path.dirname(bbfile) + script_logs = os.path.join(work_dir,'script-logs') + bb_inc = os.path.join(script_logs,'bb_inc') + bb.mkdirhier(script_logs) + bb.mkdirhier(bb_inc) + + def find_file(dir,file): + for root, dirs, files in os.walk(dir): + if file in files: + return os.path.join(root,file) + + def get_inc (file): + f = open(file,'r') + for line in f.readlines(): + if 'require' not in line: + bbinc.append(file) + else: + try: + incfile = pat.match(line).group(1) + incfile = bb.data.expand(os.path.basename(incfile),d) + abs_incfile = find_file(bbdir,incfile) + if abs_incfile: + bbinc.append(abs_incfile) + get_inc(abs_incfile) + except AttributeError: + pass + get_inc(bbfile) + bbinc = list(set(bbinc)) + for bbincfile in bbinc: + shutil.copy(bbincfile,bb_inc) + try: + bb.mkdirhier(os.path.join(script_logs,'temp')) + oe.path.copytree(os.path.join(work_dir,'temp'), os.path.join(script_logs,'temp')) + except (IOError,AttributeError): + pass + return script_logs - pf = d.getVar('PF', True) - work_dir = d.getVar('WORKDIR', True) - dest = os.path.join(work_dir, pf + '-patches') - shutil.rmtree(dest, ignore_errors=True) - bb.mkdirhier(dest) +def get_series(d): + '''copy patches and series file to a pointed directory which will be archived to tarball in ${WORKDIR}''' + import shutil + + src_patches=[] + pf = d.getVar('PF', True) + work_dir = d.getVar('WORKDIR', True) + s = d.getVar('S',True) + dest = os.path.join(work_dir, pf + '-series') + shutil.rmtree(dest, ignore_errors=True) + bb.mkdirhier(dest) + + src_uri = d.getVar('SRC_URI', True).split() + fetch = bb.fetch2.Fetch(src_uri, d) + locals = (fetch.localpath(url) for url in fetch.urls) + for local in locals: + src_patches.append(local) + if not cmp(work_dir,s): + tmp_list = src_patches + else: + tmp_list = src_patches[1:] + for patch in tmp_list: + try: + shutil.copy(patch,dest) + except IOError: + if os.path.isdir(patch): + bb.mkdirhier(os.path.join(dest,patch)) + oe.path.copytree(patch, os.path.join(dest,patch)) + return dest - patches = src_patches(d) - for patch in patches: - _, _, local, _, _, parm = bb.decodeurl(patch) - if local: - shutil.copy(local,dest) - return dest +def get_applying_patches(d): + """only copy applying patches to a pointed directory which will be archived to tarball""" + import shutil + + pf = d.getVar('PF', True) + work_dir = d.getVar('WORKDIR', True) + dest = os.path.join(work_dir, pf + '-patches') + shutil.rmtree(dest, ignore_errors=True) + bb.mkdirhier(dest) + + patches = src_patches(d) + for patch in patches: + _, _, local, _, _, parm = bb.decodeurl(patch) + if local: + shutil.copy(local,dest) + return dest def not_tarball(d): - '''packages including key words 'work-shared','native', 'task-' will be passed''' + '''packages including key words 'work-shared','native', 'task-' will be passed''' - workdir = d.getVar('WORKDIR',True) - s = d.getVar('S',True) - if 'work-shared' in s or 'task-' in workdir or 'native' in workdir: - return True - else: - return False + workdir = d.getVar('WORKDIR',True) + s = d.getVar('S',True) + if 'work-shared' in s or 'task-' in workdir or 'native' in workdir: + return True + else: + return False def get_source_from_downloads(d,stage_name): - '''copy tarball of $P to $WORKDIR when this tarball exists in $DL_DIR''' - if stage_name in 'patched' 'configured': - return - pf = d.getVar('PF', True) - dl_dir = d.getVar('DL_DIR',True) - try: - source = os.path.join(dl_dir,os.path.basename(d.getVar('SRC_URI', True).split()[0])) - if os.path.exists(source) and not os.path.isdir(source): - return source - except (IndexError, OSError): - pass - return '' + '''copy tarball of $P to $WORKDIR when this tarball exists in $DL_DIR''' + if stage_name in 'patched' 'configured': + return + pf = d.getVar('PF', True) + dl_dir = d.getVar('DL_DIR',True) + try: + source = os.path.join(dl_dir,os.path.basename(d.getVar('SRC_URI', True).split()[0])) + if os.path.exists(source) and not os.path.isdir(source): + return source + except (IndexError, OSError): + pass + return '' def do_tarball(workdir,srcdir,tarname): - '''tar "srcdir" under "workdir" to "tarname"''' - import tarfile - - sav_dir = os.getcwd() - os.chdir(workdir) - if (len(os.listdir(srcdir))) != 0: - tar = tarfile.open(tarname, "w:gz") - tar.add(srcdir) - tar.close() - else: - tarname = '' - os.chdir(sav_dir) - return tarname - -def archive_sources_from_directory(d,stage_name): - '''archive sources codes tree to tarball when tarball of $P doesn't exist in $DL_DIR''' - import shutil - - s = d.getVar('S',True) - work_dir=d.getVar('WORKDIR', True) - PF = d.getVar('PF',True) - tarname = PF + '-' + stage_name + ".tar.gz" - - if os.path.exists(s) and work_dir in s: - try: - source_dir = os.path.join(work_dir,[ i for i in s.replace(work_dir,'').split('/') if i][0]) - except IndexError: - if not cmp(s,work_dir): - return '' - else: - return '' - source = os.path.basename(source_dir) - return do_tarball(work_dir,source,tarname) + '''tar "srcdir" under "workdir" to "tarname"''' + import tarfile + + sav_dir = os.getcwd() + os.chdir(workdir) + if (len(os.listdir(srcdir))) != 0: + tar = tarfile.open(tarname, "w:gz") + tar.add(srcdir) + tar.close() + else: + tarname = '' + os.chdir(sav_dir) + return tarname + +def archive_sources_from_directory(d,stage_name): + '''archive sources codes tree to tarball when tarball of $P doesn't exist in $DL_DIR''' + import shutil + + s = d.getVar('S',True) + work_dir=d.getVar('WORKDIR', True) + PF = d.getVar('PF',True) + tarname = PF + '-' + stage_name + ".tar.gz" + + if os.path.exists(s) and work_dir in s: + try: + source_dir = os.path.join(work_dir,[ i for i in s.replace(work_dir,'').split('/') if i][0]) + except IndexError: + if not cmp(s,work_dir): + return '' + else: + return '' + source = os.path.basename(source_dir) + return do_tarball(work_dir,source,tarname) def archive_sources(d,stage_name): - '''copy tarball from $DL_DIR to $WORKDIR if have tarball, archive source codes tree in $WORKDIR if $P is directory instead of tarball''' - import shutil - work_dir = d.getVar('WORKDIR',True) - file = get_source_from_downloads(d,stage_name) - if file: - shutil.copy(file,work_dir) - file = os.path.basename(file) - else: - file = archive_sources_from_directory(d,stage_name) - return file - + '''copy tarball from $DL_DIR to $WORKDIR if have tarball, archive source codes tree in $WORKDIR if $P is directory instead of tarball''' + import shutil + work_dir = d.getVar('WORKDIR',True) + file = get_source_from_downloads(d,stage_name) + if file: + shutil.copy(file,work_dir) + file = os.path.basename(file) + else: + file = archive_sources_from_directory(d,stage_name) + return file def archive_patches(d,patchdir,series): - '''archive patches to tarball and also include series files if 'series' is True''' - import shutil - - s = d.getVar('S',True) - work_dir = d.getVar('WORKDIR', True) - patch_dir = os.path.basename(patchdir) - tarname = patch_dir + ".tar.gz" - if series == 'all' and os.path.exists(os.path.join(s,'patches/series')): - shutil.copy(os.path.join(s,'patches/series'),patchdir) - tarname = do_tarball(work_dir,patch_dir,tarname) - shutil.rmtree(patchdir, ignore_errors=True) - return tarname + '''archive patches to tarball and also include series files if 'series' is True''' + import shutil + + s = d.getVar('S',True) + work_dir = d.getVar('WORKDIR', True) + patch_dir = os.path.basename(patchdir) + tarname = patch_dir + ".tar.gz" + if series == 'all' and os.path.exists(os.path.join(s,'patches/series')): + shutil.copy(os.path.join(s,'patches/series'),patchdir) + tarname = do_tarball(work_dir,patch_dir,tarname) + shutil.rmtree(patchdir, ignore_errors=True) + return tarname def select_archive_patches(d,option): - '''select to archive all patches including non-applying and series or applying patches ''' - if option == "all": - patchdir = get_series(d) - elif option == "applying": - patchdir = get_applying_patches(d) - try: - os.rmdir(patchdir) - except OSError: - tarpatch = archive_patches(d,patchdir,option) - return tarpatch - return + '''select to archive all patches including non-applying and series or applying patches ''' + if option == "all": + patchdir = get_series(d) + elif option == "applying": + patchdir = get_applying_patches(d) + try: + os.rmdir(patchdir) + except OSError: + tarpatch = archive_patches(d,patchdir,option) + return tarpatch + return def archive_logs(d,logdir,bbinc=False): - '''archive logs in temp to tarball and .bb and .inc files if bbinc is True ''' - import shutil - - pf = d.getVar('PF',True) - work_dir = d.getVar('WORKDIR',True) - log_dir = os.path.basename(logdir) - tarname = pf + '-' + log_dir + ".tar.gz" - tarname = do_tarball(work_dir,log_dir,tarname) - if bbinc: - shutil.rmtree(logdir, ignore_errors=True) - return tarname + '''archive logs in temp to tarball and .bb and .inc files if bbinc is True ''' + import shutil + + pf = d.getVar('PF',True) + work_dir = d.getVar('WORKDIR',True) + log_dir = os.path.basename(logdir) + tarname = pf + '-' + log_dir + ".tar.gz" + tarname = do_tarball(work_dir,log_dir,tarname) + if bbinc: + shutil.rmtree(logdir, ignore_errors=True) + return tarname def get_licenses(d): - '''get licenses for running .bb file''' - import oe.license - - licenses_type = d.getVar('LICENSE', True) or "" - lics = oe.license.is_included(licenses_type)[1:][0] - lice = '' - for lic in lics: - licens = d.getVarFlag('SPDXLICENSEMAP', lic) - if licens != None: - lice += licens - else: - lice += lic - return lice - + '''get licenses for running .bb file''' + import oe.license + + licenses_type = d.getVar('LICENSE', True) or "" + lics = oe.license.is_included(licenses_type)[1:][0] + lice = '' + for lic in lics: + licens = d.getVarFlag('SPDXLICENSEMAP', lic) + if licens != None: + lice += licens + else: + lice += lic + return lice + def move_tarball_deploy(d,tarball_list): - '''move tarball in location to ${DEPLOY_DIR}/sources''' - import shutil - - if tarball_list is []: - return - target_sys = d.getVar('TARGET_SYS', True) - pf = d.getVar('PF', True) - licenses = get_licenses(d) - work_dir = d.getVar('WORKDIR',True) - tar_sources = d.getVar('DEPLOY_DIR', True) + '/sources/' + target_sys + '/' + licenses + '/' + pf - if not os.path.exists(tar_sources): - bb.mkdirhier(tar_sources) - for source in tarball_list: - if source: - if os.path.exists(os.path.join(tar_sources, source)): - os.remove(os.path.join(tar_sources,source)) - shutil.move(os.path.join(work_dir,source),tar_sources) + '''move tarball in location to ${DEPLOY_DIR}/sources''' + import shutil + + if tarball_list is []: + return + target_sys = d.getVar('TARGET_SYS', True) + pf = d.getVar('PF', True) + licenses = get_licenses(d) + work_dir = d.getVar('WORKDIR',True) + tar_sources = d.getVar('DEPLOY_DIR', True) + '/sources/' + target_sys + '/' + licenses + '/' + pf + if not os.path.exists(tar_sources): + bb.mkdirhier(tar_sources) + for source in tarball_list: + if source: + if os.path.exists(os.path.join(tar_sources, source)): + os.remove(os.path.join(tar_sources,source)) + shutil.move(os.path.join(work_dir,source),tar_sources) def check_archiving_type(d): - '''check the type for archiving package('tar' or 'srpm')''' - try: - if d.getVar('SOURCE_ARCHIVE_PACKAGE_TYPE', True).upper() not in d.getVar('ARCHIVE_TYPE', True).split(): - raise AttributeError - except AttributeError: - bb.fatal("\"SOURCE_ARCHIVE_PACKAGE_TYPE\" is \'tar\' or \'srpm\', no other types") + '''check the type for archiving package('tar' or 'srpm')''' + try: + if d.getVar('SOURCE_ARCHIVE_PACKAGE_TYPE', True).upper() not in d.getVar('ARCHIVE_TYPE', True).split(): + raise AttributeError + except AttributeError: + bb.fatal("\"SOURCE_ARCHIVE_PACKAGE_TYPE\" is \'tar\' or \'srpm\', no other types") def store_package(d,package_name): - '''store tarbablls name to file "tar-package"''' - try: - f = open(os.path.join(d.getVar('WORKDIR',True),'tar-package'),'a') - f.write(package_name + ' ') - f.close() - except IOError: - pass + '''store tarbablls name to file "tar-package"''' + try: + f = open(os.path.join(d.getVar('WORKDIR',True),'tar-package'),'a') + f.write(package_name + ' ') + f.close() + except IOError: + pass def get_package(d): - '''get tarballs name from "tar-package"''' - work_dir = (d.getVar('WORKDIR', True)) - tarpackage = os.path.join(work_dir,'tar-package') - try: - f = open(tarpackage,'r') - line = list(set(f.readline().replace('\n','').split())) - except UnboundLocalError,IOError: - pass - f.close() - return line + '''get tarballs name from "tar-package"''' + work_dir = (d.getVar('WORKDIR', True)) + tarpackage = os.path.join(work_dir,'tar-package') + try: + f = open(tarpackage,'r') + line = list(set(f.readline().replace('\n','').split())) + except UnboundLocalError,IOError: + pass + f.close() + return line def archive_sources_patches(d,stage_name): - '''archive sources and patches to tarball. stage_name will append strings ${stage_name} to ${PR} as middle name. for example, zlib-1.4.6-prepatch(stage_name).tar.gz ''' - import shutil - - check_archiving_type(d) - if not_tarball(d) or tar_filter(d): - return - - source_tar_name = archive_sources(d,stage_name) - if stage_name == "prepatch": - if d.getVar('PATCHES_ARCHIVE_WITH_SERIES',True).upper() == 'TRUE': - patch_tar_name = select_archive_patches(d,"all") - elif d.getVar('PATCHES_ARCHIVE_WITH_SERIES',True).upper() == 'FALSE': - patch_tar_name = select_archive_patches(d,"applying") - else: - bb.fatal("Please define 'PATCHES_ARCHIVE_WITH_SERIES' is strings 'True' or 'False' ") - else: - patch_tar_name = '' - - if d.getVar('SOURCE_ARCHIVE_PACKAGE_TYPE', True).upper() not in 'SRPM': - move_tarball_deploy(d,[source_tar_name,patch_tar_name]) - else: - tarpackage = os.path.join(d.getVar('WORKDIR', True),'tar-package') - if os.path.exists(tarpackage): - os.remove(tarpackage) - for package in os.path.basename(source_tar_name), patch_tar_name: - if package: - store_package(d,str(package) + ' ') + '''archive sources and patches to tarball. stage_name will append strings ${stage_name} to ${PR} as middle name. for example, zlib-1.4.6-prepatch(stage_name).tar.gz ''' + import shutil + + check_archiving_type(d) + if not_tarball(d) or tar_filter(d): + return + + source_tar_name = archive_sources(d,stage_name) + if stage_name == "prepatch": + if d.getVar('PATCHES_ARCHIVE_WITH_SERIES',True).upper() == 'TRUE': + patch_tar_name = select_archive_patches(d,"all") + elif d.getVar('PATCHES_ARCHIVE_WITH_SERIES',True).upper() == 'FALSE': + patch_tar_name = select_archive_patches(d,"applying") + else: + bb.fatal("Please define 'PATCHES_ARCHIVE_WITH_SERIES' is strings 'True' or 'False' ") + else: + patch_tar_name = '' + + if d.getVar('SOURCE_ARCHIVE_PACKAGE_TYPE', True).upper() not in 'SRPM': + move_tarball_deploy(d,[source_tar_name,patch_tar_name]) + else: + tarpackage = os.path.join(d.getVar('WORKDIR', True),'tar-package') + if os.path.exists(tarpackage): + os.remove(tarpackage) + for package in os.path.basename(source_tar_name), patch_tar_name: + if package: + store_package(d,str(package) + ' ') def archive_scripts_logs(d): - '''archive scripts and logs. scripts include .bb and .inc files and logs include stuff in "temp".''' - - if tar_filter(d): - return - work_dir = d.getVar('WORKDIR', True) - temp_dir = os.path.join(work_dir,'temp') - source_archive_log_with_scripts = d.getVar('SOURCE_ARCHIVE_LOG_WITH_SCRIPTS', True) - if source_archive_log_with_scripts == 'logs_with_scripts': - logdir = get_bb_inc(d) - tarlog = archive_logs(d,logdir,True) - elif source_archive_log_with_scripts == 'logs': - if os.path.exists(temp_dir): - tarlog = archive_logs(d,temp_dir,False) - else: - return - - if d.getVar('SOURCE_ARCHIVE_PACKAGE_TYPE', True).upper() not in 'SRPM': - move_tarball_deploy(d,[tarlog]) - - else: - store_package(d,tarlog) + '''archive scripts and logs. scripts include .bb and .inc files and logs include stuff in "temp".''' + + if tar_filter(d): + return + work_dir = d.getVar('WORKDIR', True) + temp_dir = os.path.join(work_dir,'temp') + source_archive_log_with_scripts = d.getVar('SOURCE_ARCHIVE_LOG_WITH_SCRIPTS', True) + if source_archive_log_with_scripts == 'logs_with_scripts': + logdir = get_bb_inc(d) + tarlog = archive_logs(d,logdir,True) + elif source_archive_log_with_scripts == 'logs': + if os.path.exists(temp_dir): + tarlog = archive_logs(d,temp_dir,False) + else: + return + + if d.getVar('SOURCE_ARCHIVE_PACKAGE_TYPE', True).upper() not in 'SRPM': + move_tarball_deploy(d,[tarlog]) + + else: + store_package(d,tarlog) def dumpdata(d): - '''dump environment to "${P}-${PR}.showdata.dump" including all kinds of variables and functions when running a task''' - - if tar_filter(d): - return - workdir = bb.data.getVar('WORKDIR', d, 1) - distro = bb.data.getVar('DISTRO', d, 1) - s = d.getVar('S', True) - pf = d.getVar('PF', True) - target_sys = d.getVar('TARGET_SYS', True) - licenses = get_licenses(d) - dumpdir = d.getVar('DEPLOY_DIR', True) + '/sources/' + target_sys + '/' + licenses + '/' + pf - if not os.path.exists(dumpdir): - bb.mkdirhier(dumpdir) - - dumpfile = os.path.join(dumpdir, bb.data.expand("${P}-${PR}.showdata.dump",d)) - - bb.note("Dumping metadata into '%s'" % dumpfile) - f = open(dumpfile, "w") - # emit variables and shell functions - bb.data.emit_env(f, d, True) - # emit the metadata which isnt valid shell - for e in d.keys(): - if bb.data.getVarFlag(e, 'python', d): - f.write("\npython %s () {\n%s}\n" % (e, bb.data.getVar(e, d, 1))) - f.close() + '''dump environment to "${P}-${PR}.showdata.dump" including all kinds of variables and functions when running a task''' + + if tar_filter(d): + return + workdir = bb.data.getVar('WORKDIR', d, 1) + distro = bb.data.getVar('DISTRO', d, 1) + s = d.getVar('S', True) + pf = d.getVar('PF', True) + target_sys = d.getVar('TARGET_SYS', True) + licenses = get_licenses(d) + dumpdir = d.getVar('DEPLOY_DIR', True) + '/sources/' + target_sys + '/' + licenses + '/' + pf + if not os.path.exists(dumpdir): + bb.mkdirhier(dumpdir) + + dumpfile = os.path.join(dumpdir, bb.data.expand("${P}-${PR}.showdata.dump",d)) + + bb.note("Dumping metadata into '%s'" % dumpfile) + f = open(dumpfile, "w") + # emit variables and shell functions + bb.data.emit_env(f, d, True) + # emit the metadata which isnt valid shell + for e in d.keys(): + if bb.data.getVarFlag(e, 'python', d): + f.write("\npython %s () {\n%s}\n" % (e, bb.data.getVar(e, d, 1))) + f.close() def create_diff_gz(d): - '''creating .diff.gz in ${DEPLOY_DIR_SRC}/${P}-${PR}.diff.g gz for mapping all content in 's' including patches to xxx.diff.gz''' - import shutil - import subprocess - - if tar_filter(d): - return - work_dir = d.getVar('WORKDIR', True) - exclude_from = d.getVar('ARCHIVE_EXCLUDE_FROM', True).split() - pf = d.getVar('PF', True) - licenses = get_licenses(d) - target_sys = d.getVar('TARGET_SYS', True) - diff_dir = d.getVar('DEPLOY_DIR', True) + '/sources/' + target_sys + '/' + licenses + '/' + pf - diff_file = os.path.join(diff_dir, bb.data.expand("${P}-${PR}.diff.gz",d)) - - f = open(os.path.join(work_dir,'temp/exclude-from-file'), 'a') - for i in exclude_from: - f.write(i) - f.write("\n") - f.close() - - s=d.getVar('S', True) - distro = d.getVar('DISTRO',True) - dest = s + '/' + distro + '/files' - if not os.path.exists(dest): - bb.mkdirhier(dest) - for i in os.listdir(os.getcwd()): - if os.path.isfile(i): - try: - shutil.copy(i, dest) - except IOError: - subprocess.call('fakeroot cp -rf ' + i + " " + dest, shell=True) - - bb.note("Creating .diff.gz in ${DEPLOY_DIR_SRC}/${P}-${PR}.diff.gz") - cmd = "LC_ALL=C TZ=UTC0 diff --exclude-from=" + work_dir + "/temp/exclude-from-file -Naur " + s + '.org' + ' ' + s + " | gzip -c > " + diff_file - d.setVar('DIFF', cmd + "\n") - d.setVarFlag('DIFF', 'func', '1') - bb.build.exec_func('DIFF', d) - shutil.rmtree(s + '.org', ignore_errors=True) + '''creating .diff.gz in ${DEPLOY_DIR_SRC}/${P}-${PR}.diff.g gz for mapping all content in 's' including patches to xxx.diff.gz''' + import shutil + import subprocess + + if tar_filter(d): + return + work_dir = d.getVar('WORKDIR', True) + exclude_from = d.getVar('ARCHIVE_EXCLUDE_FROM', True).split() + pf = d.getVar('PF', True) + licenses = get_licenses(d) + target_sys = d.getVar('TARGET_SYS', True) + diff_dir = d.getVar('DEPLOY_DIR', True) + '/sources/' + target_sys + '/' + licenses + '/' + pf + diff_file = os.path.join(diff_dir, bb.data.expand("${P}-${PR}.diff.gz",d)) + + f = open(os.path.join(work_dir,'temp/exclude-from-file'), 'a') + for i in exclude_from: + f.write(i) + f.write("\n") + f.close() + + s=d.getVar('S', True) + distro = d.getVar('DISTRO',True) + dest = s + '/' + distro + '/files' + if not os.path.exists(dest): + bb.mkdirhier(dest) + for i in os.listdir(os.getcwd()): + if os.path.isfile(i): + try: + shutil.copy(i, dest) + except IOError: + subprocess.call('fakeroot cp -rf ' + i + " " + dest, shell=True) + + bb.note("Creating .diff.gz in ${DEPLOY_DIR_SRC}/${P}-${PR}.diff.gz") + cmd = "LC_ALL=C TZ=UTC0 diff --exclude-from=" + work_dir + "/temp/exclude-from-file -Naur " + s + '.org' + ' ' + s + " | gzip -c > " + diff_file + d.setVar('DIFF', cmd + "\n") + d.setVarFlag('DIFF', 'func', '1') + bb.build.exec_func('DIFF', d) + shutil.rmtree(s + '.org', ignore_errors=True) # This function will run when user want to get tarball for sources and patches after do_unpack python do_archive_original_sources_patches(){ - archive_sources_patches(d,'prepatch') + archive_sources_patches(d,'prepatch') } # This function will run when user want to get tarball for patched sources after do_patch python do_archive_patched_sources(){ - archive_sources_patches(d,'patched') + archive_sources_patches(d,'patched') } # This function will run when user want to get tarball for configured sources after do_configure python do_archive_configured_sources(){ - archive_sources_patches(d,'configured') + archive_sources_patches(d,'configured') } # This function will run when user want to get tarball for logs or both logs and scripts(.bb and .inc files) python do_archive_scripts_logs(){ - archive_scripts_logs(d) + archive_scripts_logs(d) } -# This function will run when user want to know what variable and functions in a running task are and also can get a diff file including +# This function will run when user want to know what variable and functions in a running task are and also can get a diff file including # all content a package should include. python do_dumpdata_create_diff_gz(){ - dumpdata(d) - create_diff_gz(d) + dumpdata(d) + create_diff_gz(d) } # This functions prepare for archiving "linux-yocto" because this package create directory 's' before do_patch instead of after do_unpack. # This is special control for archiving linux-yocto only. python do_archive_linux_yocto(){ - s = d.getVar('S', True) - if 'linux-yocto' in s: - source_tar_name = archive_sources(d,'') - if d.getVar('SOURCE_ARCHIVE_PACKAGE_TYPE', True).upper() not in 'SRPM': - move_tarball_deploy(d,[source_tar_name,'']) + s = d.getVar('S', True) + if 'linux-yocto' in s: + source_tar_name = archive_sources(d,'') + if d.getVar('SOURCE_ARCHIVE_PACKAGE_TYPE', True).upper() not in 'SRPM': + move_tarball_deploy(d,[source_tar_name,'']) } do_kernel_checkout[postfuncs] += "do_archive_linux_yocto " # remove tarball for sources, patches and logs after creating srpm. python do_remove_tarball(){ - if d.getVar('SOURCE_ARCHIVE_PACKAGE_TYPE', True).upper() == 'SRPM': - work_dir = d.getVar('WORKDIR', True) - try: - for file in os.listdir(os.getcwd()): - if file in get_package(d): - os.remove(file) - os.remove(os.path.join(work_dir,'tar-package')) - except (TypeError,OSError): - pass + if d.getVar('SOURCE_ARCHIVE_PACKAGE_TYPE', True).upper() == 'SRPM': + work_dir = d.getVar('WORKDIR', True) + try: + for file in os.listdir(os.getcwd()): + if file in get_package(d): + os.remove(file) + os.remove(os.path.join(work_dir,'tar-package')) + except (TypeError,OSError): + pass } do_remove_taball[deptask] = "do_archive_scripts_logs" do_package_write_rpm[postfuncs] += "do_remove_tarball " |