summaryrefslogtreecommitdiff
path: root/meta/classes/distrodata.bbclass
diff options
context:
space:
mode:
authorMei Lei <lei.mei@intel.com>2011-05-16 19:06:21 +0800
committerRichard Purdie <richard.purdie@linuxfoundation.org>2011-05-17 14:38:52 +0100
commitb41148cda9f0cc292b662a8473f26bc1ee0148f3 (patch)
treedd004068a668e82e9eb6d4b88c61707e94122082 /meta/classes/distrodata.bbclass
parentb4f1845f7cf42059984112e3f41a323b4c9d6dfd (diff)
downloadopenembedded-core-b41148cda9f0cc292b662a8473f26bc1ee0148f3.tar.gz
openembedded-core-b41148cda9f0cc292b662a8473f26bc1ee0148f3.tar.bz2
openembedded-core-b41148cda9f0cc292b662a8473f26bc1ee0148f3.zip
Add a new task checklicense and fix some bugs in distro_check.py
distro_check.py: Create a new function called create_log_file to reduce a lot of repeat code in distrodata.bbclass. We needn't to create log file in function save_distro_check_result, because the log file has been generated in check_eventhandler. Another bug is that we maybe access the /tmp/Meego-1.0 before we create this file. Add a judge statement to decide whether we need to create this file firstly. distrodata.bbclass: Add a new task checklicense to collect missing text license information. This can help package-report system to know how many recipes are missing license text. Signed-off-by: Mei Lei <lei.mei@intel.com>
Diffstat (limited to 'meta/classes/distrodata.bbclass')
-rw-r--r--meta/classes/distrodata.bbclass177
1 files changed, 101 insertions, 76 deletions
diff --git a/meta/classes/distrodata.bbclass b/meta/classes/distrodata.bbclass
index 905dad7b9c..f24cff8dc7 100644
--- a/meta/classes/distrodata.bbclass
+++ b/meta/classes/distrodata.bbclass
@@ -4,19 +4,9 @@ addhandler distro_eventhandler
python distro_eventhandler() {
if bb.event.getName(e) == "BuildStarted":
- """initialize log files."""
- logpath = bb.data.getVar('LOG_DIR', e.data, 1)
- bb.utils.mkdirhier(logpath)
- logfile = os.path.join(logpath, "distrodata.%s.csv" % bb.data.getVar('DATETIME', e.data, 1))
- if not os.path.exists(logfile):
- slogfile = os.path.join(logpath, "distrodata.csv")
- if os.path.exists(slogfile):
- os.remove(slogfile)
- os.system("touch %s" % logfile)
- os.symlink(logfile, slogfile)
- bb.data.setVar('LOG_FILE', logfile, e.data)
-
- lf = bb.utils.lockfile(logfile + ".lock")
+ import oe.distro_check as dc
+ logfile = dc.create_log_file(e.data, "distrodata.csv")
+ lf = bb.utils.lockfile("%s.lock" % logfile)
f = open(logfile, "a")
f.write("Package,Description,Owner,License,ChkSum,Status,VerMatch,Version,Upsteam,Non-Update,Reason,Recipe Status\n")
f.close()
@@ -33,9 +23,9 @@ python do_distrodata_np() {
bb.note("Package Name: %s" % pn)
import oe.distro_check as dist_check
- tmpdir = bb.data.getVar('TMPDIR', d, 1)
+ tmpdir = bb.data.getVar('TMPDIR', d, True)
distro_check_dir = os.path.join(tmpdir, "distro_check")
- datetime = bb.data.getVar('DATETIME', localdata, 1)
+ datetime = bb.data.getVar('DATETIME', localdata, True)
dist_check.update_distro_data(distro_check_dir, datetime)
if pn.find("-native") != -1:
@@ -111,15 +101,15 @@ python do_distrodata_np() {
addtask distrodata
do_distrodata[nostamp] = "1"
python do_distrodata() {
- logpath = bb.data.getVar('LOG_DIR', d, 1)
+ logpath = bb.data.getVar('LOG_DIR', d, True)
bb.utils.mkdirhier(logpath)
logfile = os.path.join(logpath, "distrodata.csv")
import oe.distro_check as dist_check
localdata = bb.data.createCopy(d)
- tmpdir = bb.data.getVar('TMPDIR', d, 1)
+ tmpdir = bb.data.getVar('TMPDIR', d, True)
distro_check_dir = os.path.join(tmpdir, "distro_check")
- datetime = bb.data.getVar('DATETIME', localdata, 1)
+ datetime = bb.data.getVar('DATETIME', localdata, True)
dist_check.update_distro_data(distro_check_dir, datetime)
pn = bb.data.getVar("PN", d, True)
@@ -189,7 +179,7 @@ python do_distrodata() {
# do the comparison
result = dist_check.compare_in_distro_packages_list(distro_check_dir, localdata)
- lf = bb.utils.lockfile(logfile + ".lock")
+ lf = bb.utils.lockfile("%s.lock" % logfile)
f = open(logfile, "a")
f.write("%s,%s,%s,%s,%s,%s,%s,%s,%s,%s,%s,%s," % \
(pname, pdesc, maintainer, plicense, pchksum, hasrstatus, vermatch, pcurver, pupver, noupdate, noupdate_reason, rstatus))
@@ -211,19 +201,10 @@ do_distrodataall() {
addhandler checkpkg_eventhandler
python checkpkg_eventhandler() {
if bb.event.getName(e) == "BuildStarted":
- """initialize log files."""
- logpath = bb.data.getVar('LOG_DIR', e.data, 1)
- bb.utils.mkdirhier(logpath)
- logfile = os.path.join(logpath, "checkpkg.%s.csv" % bb.data.getVar('DATETIME', e.data, 1))
- if not os.path.exists(logfile):
- slogfile = os.path.join(logpath, "checkpkg.csv")
- if os.path.exists(slogfile):
- os.remove(slogfile)
- os.system("touch %s" % logfile)
- os.symlink(logfile, slogfile)
- bb.data.setVar('LOG_FILE', logfile, e.data)
-
- lf = bb.utils.lockfile(logfile + ".lock")
+ import oe.distro_check as dc
+ logfile = dc.create_log_file(e.data, "checkpkg.csv")
+
+ lf = bb.utils.lockfile("%s.lock" % logfile)
f = open(logfile, "a")
f.write("Package\tVersion\tUpver\tLicense\tSection\tHome\tRelease\tPriority\tDepends\tBugTracker\tPE\tDescription\tStatus\tTracking\tURI\tMAINTAINER\n")
f.close()
@@ -304,7 +285,7 @@ python do_checkpkg() {
Clear internal url cache as it's a temporary check. Not doing so will have
bitbake check url multiple times when looping through a single url
"""
- fn = bb.data.getVar('FILE', d, 1)
+ fn = bb.data.getVar('FILE', d, True)
bb.fetch2.urldata_cache[fn] = {}
"""
@@ -335,7 +316,7 @@ python do_checkpkg() {
Return new version if success, or else error in "Errxxxx" style
"""
def check_new_dir(url, curver, d):
- pn = bb.data.getVar('PN', d, 1)
+ pn = bb.data.getVar('PN', d, True)
f = tempfile.NamedTemporaryFile(delete=False, prefix="%s-1-" % pn)
status = internal_fetch_wget(url, d, f)
fhtml = f.read()
@@ -394,7 +375,7 @@ python do_checkpkg() {
"""possible to have no version in pkg name, such as spectrum-fw"""
if not re.search("\d+", curname):
return pcurver
- pn = bb.data.getVar('PN', d, 1)
+ pn = bb.data.getVar('PN', d, True)
f = tempfile.NamedTemporaryFile(delete=False, prefix="%s-2-" % pn)
status = internal_fetch_wget(url, d, f)
fhtml = f.read()
@@ -437,35 +418,35 @@ python do_checkpkg() {
f.close()
"""if host hasn't directory information, no need to save tmp file"""
if status != "ErrHostNoDir" and re.match("Err", status):
- logpath = bb.data.getVar('LOG_DIR', d, 1)
+ logpath = bb.data.getVar('LOG_DIR', d, True)
os.system("cp %s %s/" % (f.name, logpath))
os.unlink(f.name)
return status
"""first check whether a uri is provided"""
- src_uri = bb.data.getVar('SRC_URI', d, 1)
+ src_uri = bb.data.getVar('SRC_URI', d, True)
if not src_uri:
return
"""initialize log files."""
- logpath = bb.data.getVar('LOG_DIR', d, 1)
+ logpath = bb.data.getVar('LOG_DIR', d, True)
bb.utils.mkdirhier(logpath)
logfile = os.path.join(logpath, "checkpkg.csv")
"""generate package information from .bb file"""
- pname = bb.data.getVar('PN', d, 1)
- pdesc = bb.data.getVar('DESCRIPTION', d, 1)
- pgrp = bb.data.getVar('SECTION', d, 1)
- pversion = bb.data.getVar('PV', d, 1)
- plicense = bb.data.getVar('LICENSE',d,1)
- psection = bb.data.getVar('SECTION',d,1)
- phome = bb.data.getVar('HOMEPAGE', d, 1)
- prelease = bb.data.getVar('PR',d,1)
- ppriority = bb.data.getVar('PRIORITY',d,1)
- pdepends = bb.data.getVar('DEPENDS',d,1)
- pbugtracker = bb.data.getVar('BUGTRACKER',d,1)
- ppe = bb.data.getVar('PE',d,1)
- psrcuri = bb.data.getVar('SRC_URI',d,1)
+ pname = bb.data.getVar('PN', d, True)
+ pdesc = bb.data.getVar('DESCRIPTION', d, True)
+ pgrp = bb.data.getVar('SECTION', d, True)
+ pversion = bb.data.getVar('PV', d, True)
+ plicense = bb.data.getVar('LICENSE', d, True)
+ psection = bb.data.getVar('SECTION', d, True)
+ phome = bb.data.getVar('HOMEPAGE', d, True)
+ prelease = bb.data.getVar('PR', d, True)
+ ppriority = bb.data.getVar('PRIORITY', d, True)
+ pdepends = bb.data.getVar('DEPENDS', d, True)
+ pbugtracker = bb.data.getVar('BUGTRACKER', d, True)
+ ppe = bb.data.getVar('PE', d, True)
+ psrcuri = bb.data.getVar('SRC_URI', d, True)
found = 0
for uri in src_uri.split():
@@ -483,9 +464,9 @@ python do_checkpkg() {
(type, host, path, user, pswd, parm) = bb.decodeurl(uri)
if type in ['http', 'https', 'ftp']:
- pcurver = bb.data.getVar('PV', d, 1)
+ pcurver = bb.data.getVar('PV', d, True)
else:
- pcurver = bb.data.getVar("SRCREV", d, 1)
+ pcurver = bb.data.getVar("SRCREV", d, True)
if type in ['http', 'https', 'ftp']:
newver = pcurver
@@ -509,7 +490,7 @@ python do_checkpkg() {
newver = check_new_dir(alturi, dirver, d)
altpath = path
if not re.match("Err", newver) and dirver != newver:
- altpath = altpath.replace(dirver, newver, 1)
+ altpath = altpath.replace(dirver, newver, True)
"""Now try to acquire all remote files in current directory"""
if not re.match("Err", newver):
@@ -625,7 +606,7 @@ python do_checkpkg() {
pstatus += ":%s%s" % (host, path)
"""Read from manual distro tracking fields as alternative"""
- pmver = bb.data.getVar("RECIPE_LATEST_VERSION", d, 1)
+ pmver = bb.data.getVar("RECIPE_LATEST_VERSION", d, True)
if not pmver:
pmver = "N/A"
pmstatus = "ErrNoRecipeData"
@@ -639,7 +620,7 @@ python do_checkpkg() {
psrcuri = psrcuri.split()[0]
pdepends = "".join(pdepends.split("\t"))
pdesc = "".join(pdesc.split("\t"))
- lf = bb.utils.lockfile(logfile + ".lock")
+ lf = bb.utils.lockfile("%s.lock" % logfile)
f = open(logfile, "a")
f.write("%s\t%s\t%s\t%s\t%s\t%s\t%s\t%s\t%s\t%s\t%s\t%s\t%s\t%s\t%s\t%s\n" % \
(pname,pversion,pupver,plicense,psection, phome,prelease, ppriority,pdepends,pbugtracker,ppe,pdesc,pstatus,pmver,psrcuri,maintainer))
@@ -654,25 +635,12 @@ do_checkpkgall() {
:
}
-#addhandler check_eventhandler
-python check_eventhandler() {
+addhandler distro_check_eventhandler
+python distro_check_eventhandler() {
if bb.event.getName(e) == "BuildStarted":
- import oe.distro_check as dc
- tmpdir = bb.data.getVar('TMPDIR', e.data, 1)
- distro_check_dir = os.path.join(tmpdir, "distro_check")
- datetime = bb.data.getVar('DATETIME', e.data, 1)
"""initialize log files."""
- logpath = bb.data.getVar('LOG_DIR', e.data, 1)
- bb.utils.mkdirhier(logpath)
- logfile = os.path.join(logpath, "distrocheck.%s.csv" % bb.data.getVar('DATETIME', e.data, 1))
- if not os.path.exists(logfile):
- slogfile = os.path.join(logpath, "distrocheck.csv")
- if os.path.exists(slogfile):
- os.remove(slogfile)
- os.system("touch %s" % logfile)
- os.symlink(logfile, slogfile)
- bb.data.setVar('LOG_FILE', logfile, e.data)
-
+ import oe.distro_check as dc
+ result_file = dc.create_log_file(e.data, "distrocheck.csv")
return
}
@@ -681,18 +649,23 @@ do_distro_check[nostamp] = "1"
python do_distro_check() {
"""checks if the package is present in other public Linux distros"""
import oe.distro_check as dc
+ import bb
+ import shutil
localdata = bb.data.createCopy(d)
bb.data.update_data(localdata)
- tmpdir = bb.data.getVar('TMPDIR', d, 1)
+ tmpdir = bb.data.getVar('TMPDIR', d, True)
distro_check_dir = os.path.join(tmpdir, "distro_check")
- datetime = bb.data.getVar('DATETIME', localdata, 1)
+ logpath = bb.data.getVar('LOG_DIR', d, True)
+ bb.utils.mkdirhier(logpath)
+ result_file = os.path.join(logpath, "distrocheck.csv")
+ datetime = bb.data.getVar('DATETIME', localdata, True)
dc.update_distro_data(distro_check_dir, datetime)
# do the comparison
result = dc.compare_in_distro_packages_list(distro_check_dir, d)
# save the results
- dc.save_distro_check_result(result, datetime, d)
+ dc.save_distro_check_result(result, datetime, result_file, d)
}
addtask distro_checkall after do_distro_check
@@ -701,3 +674,55 @@ do_distro_checkall[nostamp] = "1"
do_distro_checkall() {
:
}
+#
+#Check Missing License Text.
+#Use this task to generate the missing license text data for pkg-report system,
+#then we can search those recipes which license text isn't exsit in common-licenses directory
+#
+addhandler checklicense_eventhandler
+python checklicense_eventhandler() {
+ if bb.event.getName(e) == "BuildStarted":
+ """initialize log files."""
+ import oe.distro_check as dc
+ logfile = dc.create_log_file(e.data, "missinglicense.csv")
+ lf = bb.utils.lockfile("%s.lock" % logfile)
+ f = open(logfile, "a")
+ f.write("Package\tLicense\tMissingLicense\n")
+ f.close()
+ bb.utils.unlockfile(lf)
+ return
+}
+
+addtask checklicense
+do_checklicense[nostamp] = "1"
+python do_checklicense() {
+ import os
+ import bb
+ import shutil
+ logpath = bb.data.getVar('LOG_DIR', d, True)
+ bb.utils.mkdirhier(logpath)
+ pn = bb.data.getVar('PN', d, True)
+ logfile = os.path.join(logpath, "missinglicense.csv")
+ generic_directory = bb.data.getVar('COMMON_LICENSE_DIR', d, True)
+ license_types = bb.data.getVar('LICENSE', d, True)
+ for license_type in ((license_types.replace('+', '').replace('|', '&')
+ .replace('(', '').replace(')', '').replace(';', '')
+ .replace(',', '').replace(" ", "").split("&"))):
+ if not os.path.isfile(os.path.join(generic_directory, license_type)):
+ lf = bb.utils.lockfile("%s.lock" % logfile)
+ f = open(logfile, "a")
+ f.write("%s\t%s\t%s\n" % \
+ (pn,license_types,license_type))
+ f.close()
+ bb.utils.unlockfile(lf)
+ return
+}
+
+addtask checklicenseall after do_checklicense
+do_checklicenseall[recrdeptask] = "do_checklicense"
+do_checklicenseall[nostamp] = "1"
+do_checklicenseall() {
+ :
+}
+
+