diff options
Diffstat (limited to 'meta/classes/distrodata.bbclass')
-rw-r--r-- | meta/classes/distrodata.bbclass | 1090 |
1 files changed, 545 insertions, 545 deletions
diff --git a/meta/classes/distrodata.bbclass b/meta/classes/distrodata.bbclass index 0da10a1dd15..945ff5344c8 100644 --- a/meta/classes/distrodata.bbclass +++ b/meta/classes/distrodata.bbclass @@ -3,11 +3,11 @@ addhandler distro_eventhandler python distro_eventhandler() { if bb.event.getName(e) == "BuildStarted": - import oe.distro_check as dc - logfile = dc.create_log_file(e.data, "distrodata.csv") - lf = bb.utils.lockfile("%s.lock" % logfile) - f = open(logfile, "a") - f.write("Package,Description,Owner,License,VerMatch,Version,Upsteam,Reason,Recipe Status,Distro 1,Distro 2,Distro 3\n") + import oe.distro_check as dc + logfile = dc.create_log_file(e.data, "distrodata.csv") + lf = bb.utils.lockfile("%s.lock" % logfile) + f = open(logfile, "a") + f.write("Package,Description,Owner,License,VerMatch,Version,Upsteam,Reason,Recipe Status,Distro 1,Distro 2,Distro 3\n") f.close() bb.utils.unlockfile(lf) @@ -17,7 +17,7 @@ python distro_eventhandler() { addtask distrodata_np do_distrodata_np[nostamp] = "1" python do_distrodata_np() { - localdata = bb.data.createCopy(d) + localdata = bb.data.createCopy(d) pn = d.getVar("PN", True) bb.note("Package Name: %s" % pn) @@ -27,69 +27,69 @@ python do_distrodata_np() { datetime = localdata.getVar('DATETIME', True) dist_check.update_distro_data(distro_check_dir, datetime) - if pn.find("-native") != -1: - pnstripped = pn.split("-native") - bb.note("Native Split: %s" % pnstripped) - localdata.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + d.getVar('OVERRIDES', True)) - bb.data.update_data(localdata) - - if pn.find("nativesdk-") != -1: - pnstripped = pn.replace("nativesdk-", "") - bb.note("Native Split: %s" % pnstripped) - localdata.setVar('OVERRIDES', "pn-" + pnstripped + ":" + d.getVar('OVERRIDES', True)) - bb.data.update_data(localdata) - - if pn.find("-cross") != -1: - pnstripped = pn.split("-cross") - bb.note("cross Split: %s" % pnstripped) - localdata.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + d.getVar('OVERRIDES', True)) - bb.data.update_data(localdata) - - if pn.find("-crosssdk") != -1: - pnstripped = pn.split("-crosssdk") - bb.note("cross Split: %s" % pnstripped) - localdata.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + d.getVar('OVERRIDES', True)) - bb.data.update_data(localdata) - - if pn.find("-initial") != -1: - pnstripped = pn.split("-initial") - bb.note("initial Split: %s" % pnstripped) - localdata.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + d.getVar('OVERRIDES', True)) - bb.data.update_data(localdata) - - """generate package information from .bb file""" - pname = localdata.getVar('PN', True) - pcurver = localdata.getVar('PV', True) - pdesc = localdata.getVar('DESCRIPTION', True) + if pn.find("-native") != -1: + pnstripped = pn.split("-native") + bb.note("Native Split: %s" % pnstripped) + localdata.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + d.getVar('OVERRIDES', True)) + bb.data.update_data(localdata) + + if pn.find("nativesdk-") != -1: + pnstripped = pn.replace("nativesdk-", "") + bb.note("Native Split: %s" % pnstripped) + localdata.setVar('OVERRIDES', "pn-" + pnstripped + ":" + d.getVar('OVERRIDES', True)) + bb.data.update_data(localdata) + + if pn.find("-cross") != -1: + pnstripped = pn.split("-cross") + bb.note("cross Split: %s" % pnstripped) + localdata.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + d.getVar('OVERRIDES', True)) + bb.data.update_data(localdata) + + if pn.find("-crosssdk") != -1: + pnstripped = pn.split("-crosssdk") + bb.note("cross Split: %s" % pnstripped) + localdata.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + d.getVar('OVERRIDES', True)) + bb.data.update_data(localdata) + + if pn.find("-initial") != -1: + pnstripped = pn.split("-initial") + bb.note("initial Split: %s" % pnstripped) + localdata.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + d.getVar('OVERRIDES', True)) + bb.data.update_data(localdata) + + """generate package information from .bb file""" + pname = localdata.getVar('PN', True) + pcurver = localdata.getVar('PV', True) + pdesc = localdata.getVar('DESCRIPTION', True) if pdesc is not None: pdesc = pdesc.replace(',','') pdesc = pdesc.replace('\n','') - pgrp = localdata.getVar('SECTION', True) - plicense = localdata.getVar('LICENSE', True).replace(',','_') + pgrp = localdata.getVar('SECTION', True) + plicense = localdata.getVar('LICENSE', True).replace(',','_') - rstatus = localdata.getVar('RECIPE_COLOR', True) + rstatus = localdata.getVar('RECIPE_COLOR', True) if rstatus is not None: rstatus = rstatus.replace(',','') - - pupver = localdata.getVar('RECIPE_UPSTREAM_VERSION', True) - if pcurver == pupver: - vermatch="1" - else: - vermatch="0" - noupdate_reason = localdata.getVar('RECIPE_NO_UPDATE_REASON', True) - if noupdate_reason is None: - noupdate="0" - else: - noupdate="1" + + pupver = localdata.getVar('RECIPE_UPSTREAM_VERSION', True) + if pcurver == pupver: + vermatch="1" + else: + vermatch="0" + noupdate_reason = localdata.getVar('RECIPE_NO_UPDATE_REASON', True) + if noupdate_reason is None: + noupdate="0" + else: + noupdate="1" noupdate_reason = noupdate_reason.replace(',','') - maintainer = localdata.getVar('RECIPE_MAINTAINER', True) - rlrd = localdata.getVar('RECIPE_UPSTREAM_DATE', True) + maintainer = localdata.getVar('RECIPE_MAINTAINER', True) + rlrd = localdata.getVar('RECIPE_UPSTREAM_DATE', True) result = dist_check.compare_in_distro_packages_list(distro_check_dir, localdata) - bb.note("DISTRO: %s,%s,%s,%s,%s,%s,%s,%s,%s\n" % \ - (pname, pdesc, maintainer, plicense, vermatch, pcurver, pupver, noupdate_reason, rstatus)) + bb.note("DISTRO: %s,%s,%s,%s,%s,%s,%s,%s,%s\n" % \ + (pname, pdesc, maintainer, plicense, vermatch, pcurver, pupver, noupdate_reason, rstatus)) line = pn for i in result: line = line + "," + i @@ -99,12 +99,12 @@ python do_distrodata_np() { addtask distrodata do_distrodata[nostamp] = "1" python do_distrodata() { - logpath = d.getVar('LOG_DIR', True) - bb.utils.mkdirhier(logpath) - logfile = os.path.join(logpath, "distrodata.csv") + logpath = d.getVar('LOG_DIR', True) + bb.utils.mkdirhier(logpath) + logfile = os.path.join(logpath, "distrodata.csv") import oe.distro_check as dist_check - localdata = bb.data.createCopy(d) + localdata = bb.data.createCopy(d) tmpdir = d.getVar('TMPDIR', True) distro_check_dir = os.path.join(tmpdir, "distro_check") datetime = localdata.getVar('DATETIME', True) @@ -113,61 +113,61 @@ python do_distrodata() { pn = d.getVar("PN", True) bb.note("Package Name: %s" % pn) - if pn.find("-native") != -1: - pnstripped = pn.split("-native") - bb.note("Native Split: %s" % pnstripped) - localdata.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + d.getVar('OVERRIDES', True)) - bb.data.update_data(localdata) - - if pn.find("-cross") != -1: - pnstripped = pn.split("-cross") - bb.note("cross Split: %s" % pnstripped) - localdata.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + d.getVar('OVERRIDES', True)) - bb.data.update_data(localdata) - - if pn.find("-initial") != -1: - pnstripped = pn.split("-initial") - bb.note("initial Split: %s" % pnstripped) - localdata.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + d.getVar('OVERRIDES', True)) - bb.data.update_data(localdata) - - """generate package information from .bb file""" - pname = localdata.getVar('PN', True) - pcurver = localdata.getVar('PV', True) - pdesc = localdata.getVar('DESCRIPTION', True) + if pn.find("-native") != -1: + pnstripped = pn.split("-native") + bb.note("Native Split: %s" % pnstripped) + localdata.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + d.getVar('OVERRIDES', True)) + bb.data.update_data(localdata) + + if pn.find("-cross") != -1: + pnstripped = pn.split("-cross") + bb.note("cross Split: %s" % pnstripped) + localdata.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + d.getVar('OVERRIDES', True)) + bb.data.update_data(localdata) + + if pn.find("-initial") != -1: + pnstripped = pn.split("-initial") + bb.note("initial Split: %s" % pnstripped) + localdata.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + d.getVar('OVERRIDES', True)) + bb.data.update_data(localdata) + + """generate package information from .bb file""" + pname = localdata.getVar('PN', True) + pcurver = localdata.getVar('PV', True) + pdesc = localdata.getVar('DESCRIPTION', True) if pdesc is not None: pdesc = pdesc.replace(',','') pdesc = pdesc.replace('\n','') - pgrp = localdata.getVar('SECTION', True) - plicense = localdata.getVar('LICENSE', True).replace(',','_') + pgrp = localdata.getVar('SECTION', True) + plicense = localdata.getVar('LICENSE', True).replace(',','_') - rstatus = localdata.getVar('RECIPE_COLOR', True) + rstatus = localdata.getVar('RECIPE_COLOR', True) if rstatus is not None: rstatus = rstatus.replace(',','') - - pupver = localdata.getVar('RECIPE_UPSTREAM_VERSION', True) - if pcurver == pupver: - vermatch="1" - else: - vermatch="0" - - noupdate_reason = localdata.getVar('RECIPE_NO_UPDATE_REASON', True) - if noupdate_reason is None: - noupdate="0" - else: - noupdate="1" + + pupver = localdata.getVar('RECIPE_UPSTREAM_VERSION', True) + if pcurver == pupver: + vermatch="1" + else: + vermatch="0" + + noupdate_reason = localdata.getVar('RECIPE_NO_UPDATE_REASON', True) + if noupdate_reason is None: + noupdate="0" + else: + noupdate="1" noupdate_reason = noupdate_reason.replace(',','') - maintainer = localdata.getVar('RECIPE_MAINTAINER', True) - rlrd = localdata.getVar('RECIPE_UPSTREAM_DATE', True) + maintainer = localdata.getVar('RECIPE_MAINTAINER', True) + rlrd = localdata.getVar('RECIPE_UPSTREAM_DATE', True) # do the comparison result = dist_check.compare_in_distro_packages_list(distro_check_dir, localdata) - lf = bb.utils.lockfile("%s.lock" % logfile) - f = open(logfile, "a") - f.write("%s,%s,%s,%s,%s,%s,%s,%s,%s" % \ - (pname, pdesc, maintainer, plicense, vermatch, pcurver, pupver, noupdate_reason, rstatus)) + lf = bb.utils.lockfile("%s.lock" % logfile) + f = open(logfile, "a") + f.write("%s,%s,%s,%s,%s,%s,%s,%s,%s" % \ + (pname, pdesc, maintainer, plicense, vermatch, pcurver, pupver, noupdate_reason, rstatus)) line = "" for i in result: line = line + "," + i @@ -180,18 +180,18 @@ addtask distrodataall after do_distrodata do_distrodataall[recrdeptask] = "do_distrodataall do_distrodata" do_distrodataall[nostamp] = "1" do_distrodataall() { - : + : } addhandler checkpkg_eventhandler python checkpkg_eventhandler() { if bb.event.getName(e) == "BuildStarted": - import oe.distro_check as dc - logfile = dc.create_log_file(e.data, "checkpkg.csv") + import oe.distro_check as dc + logfile = dc.create_log_file(e.data, "checkpkg.csv") - lf = bb.utils.lockfile("%s.lock" % logfile) - f = open(logfile, "a") - f.write("Package\tVersion\tUpver\tLicense\tSection\tHome\tRelease\tDepends\tBugTracker\tPE\tDescription\tStatus\tTracking\tURI\tMAINTAINER\n") + lf = bb.utils.lockfile("%s.lock" % logfile) + f = open(logfile, "a") + f.write("Package\tVersion\tUpver\tLicense\tSection\tHome\tRelease\tDepends\tBugTracker\tPE\tDescription\tStatus\tTracking\tURI\tMAINTAINER\n") f.close() bb.utils.unlockfile(lf) return @@ -200,444 +200,444 @@ python checkpkg_eventhandler() { addtask checkpkg do_checkpkg[nostamp] = "1" python do_checkpkg() { - localdata = bb.data.createCopy(d) - import sys - import re - import tempfile - import subprocess - - """ - sanity check to ensure same name and type. Match as many patterns as possible - such as: - gnome-common-2.20.0.tar.gz (most common format) - gtk+-2.90.1.tar.gz - xf86-input-synaptics-12.6.9.tar.gz - dri2proto-2.3.tar.gz - blktool_4.orig.tar.gz - libid3tag-0.15.1b.tar.gz - unzip552.tar.gz - icu4c-3_6-src.tgz - genext2fs_1.3.orig.tar.gz - gst-fluendo-mp3 - """ - prefix1 = "[a-zA-Z][a-zA-Z0-9]*([\-_][a-zA-Z]\w+)*[\-_]" # match most patterns which uses "-" as separator to version digits - prefix2 = "[a-zA-Z]+" # a loose pattern such as for unzip552.tar.gz - prefix3 = "[0-9a-zA-Z]+" # a loose pattern such as for 80325-quicky-0.4.tar.gz - prefix = "(%s|%s|%s)" % (prefix1, prefix2, prefix3) - suffix = "(tar\.gz|tgz|tar\.bz2|zip|xz|rpm)" - suffixtuple = ("tar.gz", "tgz", "zip", "tar.bz2", "tar.xz", "src.rpm") - - sinterstr = "(?P<name>%s?)(?P<ver>.*)" % prefix - sdirstr = "(?P<name>%s)(?P<ver>.*)\.(?P<type>%s$)" % (prefix, suffix) - - def parse_inter(s): - m = re.search(sinterstr, s) - if not m: - return None - else: - return (m.group('name'), m.group('ver'), "") - - def parse_dir(s): - m = re.search(sdirstr, s) - if not m: - return None - else: - return (m.group('name'), m.group('ver'), m.group('type')) - - """ - Check whether 'new' is newer than 'old' version. We use existing vercmp() for the - purpose. PE is cleared in comparison as it's not for build, and PV is cleared too - for simplicity as it's somehow difficult to get from various upstream format - """ - def __vercmp(old, new): - (on, ov, ot) = old - (en, ev, et) = new - if on != en or (et and et not in suffixtuple): - return 0 - ov = re.search("[\d|\.]+[^a-zA-Z]+", ov).group() - ev = re.search("[\d|\.]+[^a-zA-Z]+", ev).group() - return bb.utils.vercmp(("0", ov, ""), ("0", ev, "")) - - """ - wrapper for fetch upstream directory info - 'url' - upstream link customized by regular expression - 'd' - database - 'tmpf' - tmpfile for fetcher output - We don't want to exit whole build due to one recipe error. So handle all exceptions - gracefully w/o leaking to outer. - """ - def internal_fetch_wget(url, d, tmpf): - status = "ErrFetchUnknown" - """ - Clear internal url cache as it's a temporary check. Not doing so will have - bitbake check url multiple times when looping through a single url - """ - fn = d.getVar('FILE', True) - bb.fetch2.urldata_cache[fn] = {} - - """ - To avoid impacting bitbake build engine, this trick is required for reusing bitbake - interfaces. bb.fetch.go() is not appliable as it checks downloaded content in ${DL_DIR} - while we don't want to pollute that place. So bb.fetch2.checkstatus() is borrowed here - which is designed for check purpose but we override check command for our own purpose - """ - ld = bb.data.createCopy(d) - d.setVar('CHECKCOMMAND_wget', "/usr/bin/env wget -t 1 --passive-ftp -O %s --user-agent=\"Mozilla/5.0 (X11; U; Linux i686; en-US; rv:1.9.2.12) Gecko/20101027 Ubuntu/9.10 (karmic) Firefox/3.6.12\" '${URI}'" \ - % tmpf.name) - bb.data.update_data(ld) - - try: - fetcher = bb.fetch2.Fetch([url], ld) - fetcher.checkstatus() - status = "SUCC" - except bb.fetch2.BBFetchException, e: - status = "ErrFetch" - - return status - - """ - Check on middle version directory such as "2.4/" in "http://xxx/2.4/pkg-2.4.1.tar.gz", - 'url' - upstream link customized by regular expression - 'd' - database - 'curver' - current version - Return new version if success, or else error in "Errxxxx" style - """ - def check_new_dir(url, curver, d): - pn = d.getVar('PN', True) - f = tempfile.NamedTemporaryFile(delete=False, prefix="%s-1-" % pn) - status = internal_fetch_wget(url, d, f) - fhtml = f.read() - if status == "SUCC" and len(fhtml): - newver = parse_inter(curver) - - """ - match "*4.1/">*4.1/ where '*' matches chars - N.B. add package name, only match for digits - """ - m = re.search("^%s" % prefix, curver) - if m: - s = "%s[^\d\"]*?(\d+[\.\-_])+\d+/?" % m.group() - else: - s = "(\d+[\.\-_])+\d+/?" - - searchstr = "[hH][rR][eE][fF]=\"%s\">" % s - reg = re.compile(searchstr) - - valid = 0 - for line in fhtml.split("\n"): - if line.find(curver) >= 0: - valid = 1 - m = reg.search(line) - if m: - ver = m.group().split("\"")[1] - ver = ver.strip("/") - ver = parse_inter(ver) - if ver and __vercmp(newver, ver) < 0: - newver = ver - - """Expect a match for curver in directory list, or else it indicates unknown format""" - if not valid: - status = "ErrParseInterDir" - else: - """rejoin the path name""" - status = newver[0] + newver[1] - elif not len(fhtml): - status = "ErrHostNoDir" - - f.close() - if status != "ErrHostNoDir" and re.match("Err", status): - logpath = d.getVar('LOG_DIR', True) - subprocess.call("cp %s %s/" % (f.name, logpath), shell=True) - os.unlink(f.name) - return status - - """ - Check on the last directory to search '2.4.1' in "http://xxx/2.4/pkg-2.4.1.tar.gz", - 'url' - upstream link customized by regular expression - 'd' - database - 'curname' - current package name - Return new version if success, or else error in "Errxxxx" style - """ - def check_new_version(url, curname, d): - """possible to have no version in pkg name, such as spectrum-fw""" - if not re.search("\d+", curname): - return pcurver - pn = d.getVar('PN', True) - f = tempfile.NamedTemporaryFile(delete=False, prefix="%s-2-" % pn) - status = internal_fetch_wget(url, d, f) - fhtml = f.read() - - if status == "SUCC" and len(fhtml): - newver = parse_dir(curname) - - """match "{PN}-5.21.1.tar.gz">{PN}-5.21.1.tar.gz """ - pn1 = re.search("^%s" % prefix, curname).group() - - s = "[^\"]*%s[^\d\"]*?(\d+[\.\-_])+[^\"]*" % pn1 - searchstr = "[hH][rR][eE][fF]=\"%s\".*[>\"]" % s - reg = re.compile(searchstr) - - valid = 0 - for line in fhtml.split("\n"): - m = reg.search(line) - if m: - valid = 1 - ver = m.group().split("\"")[1].split("/")[-1] - if ver == "download": - ver = m.group().split("\"")[1].split("/")[-2] - ver = parse_dir(ver) - if ver and __vercmp(newver, ver) < 0: - newver = ver - - """Expect a match for curver in directory list, or else it indicates unknown format""" - if not valid: - status = "ErrParseDir" - else: - """newver still contains a full package name string""" - status = re.search("(\d+[\.\-_])*(\d+[0-9a-zA-Z]*)", newver[1]).group() - if "_" in status: - status = re.sub("_",".",status) - elif "-" in status: - status = re.sub("-",".",status) - elif not len(fhtml): - status = "ErrHostNoDir" - - f.close() - """if host hasn't directory information, no need to save tmp file""" - if status != "ErrHostNoDir" and re.match("Err", status): - logpath = d.getVar('LOG_DIR', True) - subprocess.call("cp %s %s/" % (f.name, logpath), shell=True) - os.unlink(f.name) - return status - - """first check whether a uri is provided""" - src_uri = d.getVar('SRC_URI', True) - if not src_uri: - return - - """initialize log files.""" - logpath = d.getVar('LOG_DIR', True) - bb.utils.mkdirhier(logpath) - logfile = os.path.join(logpath, "checkpkg.csv") - - """generate package information from .bb file""" - pname = d.getVar('PN', True) - - if pname.find("-native") != -1: - pnstripped = pname.split("-native") - bb.note("Native Split: %s" % pnstripped) - localdata.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + d.getVar('OVERRIDES', True)) - bb.data.update_data(localdata) - - if pname.find("-cross") != -1: - pnstripped = pname.split("-cross") - bb.note("cross Split: %s" % pnstripped) - localdata.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + d.getVar('OVERRIDES', True)) - bb.data.update_data(localdata) - - if pname.find("-initial") != -1: - pnstripped = pname.split("-initial") - bb.note("initial Split: %s" % pnstripped) - localdata.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + d.getVar('OVERRIDES', True)) - bb.data.update_data(localdata) - - pdesc = localdata.getVar('DESCRIPTION', True) - pgrp = localdata.getVar('SECTION', True) - pversion = localdata.getVar('PV', True) - plicense = localdata.getVar('LICENSE', True) - psection = localdata.getVar('SECTION', True) - phome = localdata.getVar('HOMEPAGE', True) - prelease = localdata.getVar('PR', True) - pdepends = localdata.getVar('DEPENDS', True) - pbugtracker = localdata.getVar('BUGTRACKER', True) - ppe = localdata.getVar('PE', True) - psrcuri = localdata.getVar('SRC_URI', True) - maintainer = localdata.getVar('RECIPE_MAINTAINER', True) - - found = 0 - for uri in src_uri.split(): - m = re.compile('(?P<type>[^:]*)').match(uri) - if not m: - raise MalformedUrl(uri) - elif m.group('type') in ('http', 'https', 'ftp', 'cvs', 'svn', 'git'): - found = 1 - pproto = m.group('type') - break - if not found: - pproto = "file" - pupver = "N/A" - pstatus = "ErrUnknown" - - (type, host, path, user, pswd, parm) = bb.decodeurl(uri) - if type in ['http', 'https', 'ftp']: - pcurver = d.getVar('PV', True) - else: - pcurver = d.getVar("SRCREV", True) - - if type in ['http', 'https', 'ftp']: - newver = pcurver - altpath = path - dirver = "-" - curname = "-" - - """ - match version number amid the path, such as "5.7" in: - http://download.gnome.org/sources/${PN}/5.7/${PN}-${PV}.tar.gz - N.B. how about sth. like "../5.7/5.8/..."? Not find such example so far :-P - """ - m = re.search(r"[^/]*(\d+\.)+\d+([\-_]r\d+)*/", path) - if m: - altpath = path.split(m.group())[0] - dirver = m.group().strip("/") - - """use new path and remove param. for wget only param is md5sum""" - alturi = bb.encodeurl([type, host, altpath, user, pswd, {}]) - - newver = check_new_dir(alturi, dirver, d) - altpath = path - if not re.match("Err", newver) and dirver != newver: - altpath = altpath.replace(dirver, newver, True) - - """Now try to acquire all remote files in current directory""" - if not re.match("Err", newver): - curname = altpath.split("/")[-1] - - """get remote name by skipping pacakge name""" - m = re.search(r"/.*/", altpath) - if not m: - altpath = "/" - else: - altpath = m.group() - - alturi = bb.encodeurl([type, host, altpath, user, pswd, {}]) - newver = check_new_version(alturi, curname, d) - while(newver == "ErrHostNoDir"): - if alturi == "/download": - break - else: - alturi = "/".join(alturi.split("/")[0:-2]) + "/download" - newver = check_new_version(alturi, curname, d) - if not re.match("Err", newver): - pupver = newver - if pupver != pcurver: - pstatus = "UPDATE" - else: - pstatus = "MATCH" - - if re.match("Err", newver): - pstatus = newver + ":" + altpath + ":" + dirver + ":" + curname - elif type == 'git': - if user: - gituser = user + '@' - else: - gituser = "" - - if 'protocol' in parm: - gitproto = parm['protocol'] - else: - gitproto = "git" - gitcmd = "git ls-remote %s://%s%s%s *tag* 2>&1" % (gitproto, gituser, host, path) - gitcmd2 = "git ls-remote %s://%s%s%s HEAD 2>&1" % (gitproto, gituser, host, path) - tmp = os.popen(gitcmd).read() - tmp2 = os.popen(gitcmd2).read() - #This is for those repo have tag like: refs/tags/1.2.2 - if tmp: - tmpline = tmp.split("\n") - verflag = 0 - for line in tmpline: - if len(line)==0: - break; - puptag = line.split("/")[-1] - puptag = re.search("[0-9][0-9|\.|_]+[0-9]", puptag) - if puptag == None: - continue; - puptag = puptag.group() - puptag = re.sub("_",".",puptag) - plocaltag = pversion.split("+")[0] - if "git" in plocaltag: - plocaltag = plocaltag.split("-")[0] - result = bb.utils.vercmp(("0", puptag, ""), ("0", plocaltag, "")) - if result > 0: - verflag = 1 - pstatus = "UPDATE" - pupver = puptag - elif verflag == 0 : - pupver = plocaltag - pstatus = "MATCH" - #This is for those no tag repo - elif tmp2: - pupver = tmp2.split("\t")[0] - if pupver in pversion: - pstatus = "MATCH" - else: - pstatus = "UPDATE" - else: - pstatus = "ErrGitAccess" - elif type == 'svn': - options = [] - if user: - options.append("--username %s" % user) - if pswd: - options.append("--password %s" % pswd) - svnproto = 'svn' - if 'proto' in parm: - svnproto = parm['proto'] - if 'rev' in parm: - pcurver = parm['rev'] - - svncmd = "svn info %s %s://%s%s/%s/ 2>&1" % (" ".join(options), svnproto, host, path, parm["module"]) - print svncmd - svninfo = os.popen(svncmd).read() - for line in svninfo.split("\n"): - if re.search("^Last Changed Rev:", line): - pupver = line.split(" ")[-1] - if pupver in pversion: - pstatus = "MATCH" - else: - pstatus = "UPDATE" - - if re.match("Err", pstatus): - pstatus = "ErrSvnAccess" - elif type == 'cvs': - pupver = "HEAD" - pstatus = "UPDATE" - elif type == 'file': - """local file is always up-to-date""" - pupver = pcurver - pstatus = "MATCH" - else: - pstatus = "ErrUnsupportedProto" - - if re.match("Err", pstatus): - pstatus += ":%s%s" % (host, path) - - """Read from manual distro tracking fields as alternative""" - pmver = d.getVar("RECIPE_UPSTREAM_VERSION", True) - if not pmver: - pmver = "N/A" - pmstatus = "ErrNoRecipeData" - else: - if pmver == pcurver: - pmstatus = "MATCH" - else: - pmstatus = "UPDATE" - - psrcuri = psrcuri.split()[0] - pdepends = "".join(pdepends.split("\t")) - pdesc = "".join(pdesc.split("\t")) - lf = bb.utils.lockfile("%s.lock" % logfile) - f = open(logfile, "a") - f.write("%s\t%s\t%s\t%s\t%s\t%s\t%s\t%s\t%s\t%s\t%s\t%s\t%s\t%s\t%s\n" % \ - (pname,pversion,pupver,plicense,psection, phome,prelease, pdepends,pbugtracker,ppe,pdesc,pstatus,pmver,psrcuri,maintainer)) - f.close() - bb.utils.unlockfile(lf) + localdata = bb.data.createCopy(d) + import sys + import re + import tempfile + import subprocess + + """ + sanity check to ensure same name and type. Match as many patterns as possible + such as: + gnome-common-2.20.0.tar.gz (most common format) + gtk+-2.90.1.tar.gz + xf86-input-synaptics-12.6.9.tar.gz + dri2proto-2.3.tar.gz + blktool_4.orig.tar.gz + libid3tag-0.15.1b.tar.gz + unzip552.tar.gz + icu4c-3_6-src.tgz + genext2fs_1.3.orig.tar.gz + gst-fluendo-mp3 + """ + prefix1 = "[a-zA-Z][a-zA-Z0-9]*([\-_][a-zA-Z]\w+)*[\-_]" # match most patterns which uses "-" as separator to version digits + prefix2 = "[a-zA-Z]+" # a loose pattern such as for unzip552.tar.gz + prefix3 = "[0-9a-zA-Z]+" # a loose pattern such as for 80325-quicky-0.4.tar.gz + prefix = "(%s|%s|%s)" % (prefix1, prefix2, prefix3) + suffix = "(tar\.gz|tgz|tar\.bz2|zip|xz|rpm)" + suffixtuple = ("tar.gz", "tgz", "zip", "tar.bz2", "tar.xz", "src.rpm") + + sinterstr = "(?P<name>%s?)(?P<ver>.*)" % prefix + sdirstr = "(?P<name>%s)(?P<ver>.*)\.(?P<type>%s$)" % (prefix, suffix) + + def parse_inter(s): + m = re.search(sinterstr, s) + if not m: + return None + else: + return (m.group('name'), m.group('ver'), "") + + def parse_dir(s): + m = re.search(sdirstr, s) + if not m: + return None + else: + return (m.group('name'), m.group('ver'), m.group('type')) + + """ + Check whether 'new' is newer than 'old' version. We use existing vercmp() for the + purpose. PE is cleared in comparison as it's not for build, and PV is cleared too + for simplicity as it's somehow difficult to get from various upstream format + """ + def __vercmp(old, new): + (on, ov, ot) = old + (en, ev, et) = new + if on != en or (et and et not in suffixtuple): + return 0 + ov = re.search("[\d|\.]+[^a-zA-Z]+", ov).group() + ev = re.search("[\d|\.]+[^a-zA-Z]+", ev).group() + return bb.utils.vercmp(("0", ov, ""), ("0", ev, "")) + + """ + wrapper for fetch upstream directory info + 'url' - upstream link customized by regular expression + 'd' - database + 'tmpf' - tmpfile for fetcher output + We don't want to exit whole build due to one recipe error. So handle all exceptions + gracefully w/o leaking to outer. + """ + def internal_fetch_wget(url, d, tmpf): + status = "ErrFetchUnknown" + """ + Clear internal url cache as it's a temporary check. Not doing so will have + bitbake check url multiple times when looping through a single url + """ + fn = d.getVar('FILE', True) + bb.fetch2.urldata_cache[fn] = {} + + """ + To avoid impacting bitbake build engine, this trick is required for reusing bitbake + interfaces. bb.fetch.go() is not appliable as it checks downloaded content in ${DL_DIR} + while we don't want to pollute that place. So bb.fetch2.checkstatus() is borrowed here + which is designed for check purpose but we override check command for our own purpose + """ + ld = bb.data.createCopy(d) + d.setVar('CHECKCOMMAND_wget', "/usr/bin/env wget -t 1 --passive-ftp -O %s --user-agent=\"Mozilla/5.0 (X11; U; Linux i686; en-US; rv:1.9.2.12) Gecko/20101027 Ubuntu/9.10 (karmic) Firefox/3.6.12\" '${URI}'" \ + % tmpf.name) + bb.data.update_data(ld) + + try: + fetcher = bb.fetch2.Fetch([url], ld) + fetcher.checkstatus() + status = "SUCC" + except bb.fetch2.BBFetchException, e: + status = "ErrFetch" + + return status + + """ + Check on middle version directory such as "2.4/" in "http://xxx/2.4/pkg-2.4.1.tar.gz", + 'url' - upstream link customized by regular expression + 'd' - database + 'curver' - current version + Return new version if success, or else error in "Errxxxx" style + """ + def check_new_dir(url, curver, d): + pn = d.getVar('PN', True) + f = tempfile.NamedTemporaryFile(delete=False, prefix="%s-1-" % pn) + status = internal_fetch_wget(url, d, f) + fhtml = f.read() + if status == "SUCC" and len(fhtml): + newver = parse_inter(curver) + + """ + match "*4.1/">*4.1/ where '*' matches chars + N.B. add package name, only match for digits + """ + m = re.search("^%s" % prefix, curver) + if m: + s = "%s[^\d\"]*?(\d+[\.\-_])+\d+/?" % m.group() + else: + s = "(\d+[\.\-_])+\d+/?" + + searchstr = "[hH][rR][eE][fF]=\"%s\">" % s + reg = re.compile(searchstr) + + valid = 0 + for line in fhtml.split("\n"): + if line.find(curver) >= 0: + valid = 1 + m = reg.search(line) + if m: + ver = m.group().split("\"")[1] + ver = ver.strip("/") + ver = parse_inter(ver) + if ver and __vercmp(newver, ver) < 0: + newver = ver + + """Expect a match for curver in directory list, or else it indicates unknown format""" + if not valid: + status = "ErrParseInterDir" + else: + """rejoin the path name""" + status = newver[0] + newver[1] + elif not len(fhtml): + status = "ErrHostNoDir" + + f.close() + if status != "ErrHostNoDir" and re.match("Err", status): + logpath = d.getVar('LOG_DIR', True) + subprocess.call("cp %s %s/" % (f.name, logpath), shell=True) + os.unlink(f.name) + return status + + """ + Check on the last directory to search '2.4.1' in "http://xxx/2.4/pkg-2.4.1.tar.gz", + 'url' - upstream link customized by regular expression + 'd' - database + 'curname' - current package name + Return new version if success, or else error in "Errxxxx" style + """ + def check_new_version(url, curname, d): + """possible to have no version in pkg name, such as spectrum-fw""" + if not re.search("\d+", curname): + return pcurver + pn = d.getVar('PN', True) + f = tempfile.NamedTemporaryFile(delete=False, prefix="%s-2-" % pn) + status = internal_fetch_wget(url, d, f) + fhtml = f.read() + + if status == "SUCC" and len(fhtml): + newver = parse_dir(curname) + + """match "{PN}-5.21.1.tar.gz">{PN}-5.21.1.tar.gz """ + pn1 = re.search("^%s" % prefix, curname).group() + + s = "[^\"]*%s[^\d\"]*?(\d+[\.\-_])+[^\"]*" % pn1 + searchstr = "[hH][rR][eE][fF]=\"%s\".*[>\"]" % s + reg = re.compile(searchstr) + + valid = 0 + for line in fhtml.split("\n"): + m = reg.search(line) + if m: + valid = 1 + ver = m.group().split("\"")[1].split("/")[-1] + if ver == "download": + ver = m.group().split("\"")[1].split("/")[-2] + ver = parse_dir(ver) + if ver and __vercmp(newver, ver) < 0: + newver = ver + + """Expect a match for curver in directory list, or else it indicates unknown format""" + if not valid: + status = "ErrParseDir" + else: + """newver still contains a full package name string""" + status = re.search("(\d+[\.\-_])*(\d+[0-9a-zA-Z]*)", newver[1]).group() + if "_" in status: + status = re.sub("_",".",status) + elif "-" in status: + status = re.sub("-",".",status) + elif not len(fhtml): + status = "ErrHostNoDir" + + f.close() + """if host hasn't directory information, no need to save tmp file""" + if status != "ErrHostNoDir" and re.match("Err", status): + logpath = d.getVar('LOG_DIR', True) + subprocess.call("cp %s %s/" % (f.name, logpath), shell=True) + os.unlink(f.name) + return status + + """first check whether a uri is provided""" + src_uri = d.getVar('SRC_URI', True) + if not src_uri: + return + + """initialize log files.""" + logpath = d.getVar('LOG_DIR', True) + bb.utils.mkdirhier(logpath) + logfile = os.path.join(logpath, "checkpkg.csv") + + """generate package information from .bb file""" + pname = d.getVar('PN', True) + + if pname.find("-native") != -1: + pnstripped = pname.split("-native") + bb.note("Native Split: %s" % pnstripped) + localdata.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + d.getVar('OVERRIDES', True)) + bb.data.update_data(localdata) + + if pname.find("-cross") != -1: + pnstripped = pname.split("-cross") + bb.note("cross Split: %s" % pnstripped) + localdata.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + d.getVar('OVERRIDES', True)) + bb.data.update_data(localdata) + + if pname.find("-initial") != -1: + pnstripped = pname.split("-initial") + bb.note("initial Split: %s" % pnstripped) + localdata.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + d.getVar('OVERRIDES', True)) + bb.data.update_data(localdata) + + pdesc = localdata.getVar('DESCRIPTION', True) + pgrp = localdata.getVar('SECTION', True) + pversion = localdata.getVar('PV', True) + plicense = localdata.getVar('LICENSE', True) + psection = localdata.getVar('SECTION', True) + phome = localdata.getVar('HOMEPAGE', True) + prelease = localdata.getVar('PR', True) + pdepends = localdata.getVar('DEPENDS', True) + pbugtracker = localdata.getVar('BUGTRACKER', True) + ppe = localdata.getVar('PE', True) + psrcuri = localdata.getVar('SRC_URI', True) + maintainer = localdata.getVar('RECIPE_MAINTAINER', True) + + found = 0 + for uri in src_uri.split(): + m = re.compile('(?P<type>[^:]*)').match(uri) + if not m: + raise MalformedUrl(uri) + elif m.group('type') in ('http', 'https', 'ftp', 'cvs', 'svn', 'git'): + found = 1 + pproto = m.group('type') + break + if not found: + pproto = "file" + pupver = "N/A" + pstatus = "ErrUnknown" + + (type, host, path, user, pswd, parm) = bb.decodeurl(uri) + if type in ['http', 'https', 'ftp']: + pcurver = d.getVar('PV', True) + else: + pcurver = d.getVar("SRCREV", True) + + if type in ['http', 'https', 'ftp']: + newver = pcurver + altpath = path + dirver = "-" + curname = "-" + + """ + match version number amid the path, such as "5.7" in: + http://download.gnome.org/sources/${PN}/5.7/${PN}-${PV}.tar.gz + N.B. how about sth. like "../5.7/5.8/..."? Not find such example so far :-P + """ + m = re.search(r"[^/]*(\d+\.)+\d+([\-_]r\d+)*/", path) + if m: + altpath = path.split(m.group())[0] + dirver = m.group().strip("/") + + """use new path and remove param. for wget only param is md5sum""" + alturi = bb.encodeurl([type, host, altpath, user, pswd, {}]) + + newver = check_new_dir(alturi, dirver, d) + altpath = path + if not re.match("Err", newver) and dirver != newver: + altpath = altpath.replace(dirver, newver, True) + + """Now try to acquire all remote files in current directory""" + if not re.match("Err", newver): + curname = altpath.split("/")[-1] + + """get remote name by skipping pacakge name""" + m = re.search(r"/.*/", altpath) + if not m: + altpath = "/" + else: + altpath = m.group() + + alturi = bb.encodeurl([type, host, altpath, user, pswd, {}]) + newver = check_new_version(alturi, curname, d) + while(newver == "ErrHostNoDir"): + if alturi == "/download": + break + else: + alturi = "/".join(alturi.split("/")[0:-2]) + "/download" + newver = check_new_version(alturi, curname, d) + if not re.match("Err", newver): + pupver = newver + if pupver != pcurver: + pstatus = "UPDATE" + else: + pstatus = "MATCH" + + if re.match("Err", newver): + pstatus = newver + ":" + altpath + ":" + dirver + ":" + curname + elif type == 'git': + if user: + gituser = user + '@' + else: + gituser = "" + + if 'protocol' in parm: + gitproto = parm['protocol'] + else: + gitproto = "git" + gitcmd = "git ls-remote %s://%s%s%s *tag* 2>&1" % (gitproto, gituser, host, path) + gitcmd2 = "git ls-remote %s://%s%s%s HEAD 2>&1" % (gitproto, gituser, host, path) + tmp = os.popen(gitcmd).read() + tmp2 = os.popen(gitcmd2).read() + #This is for those repo have tag like: refs/tags/1.2.2 + if tmp: + tmpline = tmp.split("\n") + verflag = 0 + for line in tmpline: + if len(line)==0: + break; + puptag = line.split("/")[-1] + puptag = re.search("[0-9][0-9|\.|_]+[0-9]", puptag) + if puptag == None: + continue; + puptag = puptag.group() + puptag = re.sub("_",".",puptag) + plocaltag = pversion.split("+")[0] + if "git" in plocaltag: + plocaltag = plocaltag.split("-")[0] + result = bb.utils.vercmp(("0", puptag, ""), ("0", plocaltag, "")) + if result > 0: + verflag = 1 + pstatus = "UPDATE" + pupver = puptag + elif verflag == 0 : + pupver = plocaltag + pstatus = "MATCH" + #This is for those no tag repo + elif tmp2: + pupver = tmp2.split("\t")[0] + if pupver in pversion: + pstatus = "MATCH" + else: + pstatus = "UPDATE" + else: + pstatus = "ErrGitAccess" + elif type == 'svn': + options = [] + if user: + options.append("--username %s" % user) + if pswd: + options.append("--password %s" % pswd) + svnproto = 'svn' + if 'proto' in parm: + svnproto = parm['proto'] + if 'rev' in parm: + pcurver = parm['rev'] + + svncmd = "svn info %s %s://%s%s/%s/ 2>&1" % (" ".join(options), svnproto, host, path, parm["module"]) + print svncmd + svninfo = os.popen(svncmd).read() + for line in svninfo.split("\n"): + if re.search("^Last Changed Rev:", line): + pupver = line.split(" ")[-1] + if pupver in pversion: + pstatus = "MATCH" + else: + pstatus = "UPDATE" + + if re.match("Err", pstatus): + pstatus = "ErrSvnAccess" + elif type == 'cvs': + pupver = "HEAD" + pstatus = "UPDATE" + elif type == 'file': + """local file is always up-to-date""" + pupver = pcurver + pstatus = "MATCH" + else: + pstatus = "ErrUnsupportedProto" + + if re.match("Err", pstatus): + pstatus += ":%s%s" % (host, path) + + """Read from manual distro tracking fields as alternative""" + pmver = d.getVar("RECIPE_UPSTREAM_VERSION", True) + if not pmver: + pmver = "N/A" + pmstatus = "ErrNoRecipeData" + else: + if pmver == pcurver: + pmstatus = "MATCH" + else: + pmstatus = "UPDATE" + + psrcuri = psrcuri.split()[0] + pdepends = "".join(pdepends.split("\t")) + pdesc = "".join(pdesc.split("\t")) + lf = bb.utils.lockfile("%s.lock" % logfile) + f = open(logfile, "a") + f.write("%s\t%s\t%s\t%s\t%s\t%s\t%s\t%s\t%s\t%s\t%s\t%s\t%s\t%s\t%s\n" % \ + (pname,pversion,pupver,plicense,psection, phome,prelease, pdepends,pbugtracker,ppe,pdesc,pstatus,pmver,psrcuri,maintainer)) + f.close() + bb.utils.unlockfile(lf) } addtask checkpkgall after do_checkpkg do_checkpkgall[recrdeptask] = "do_checkpkgall do_checkpkg" do_checkpkgall[nostamp] = "1" do_checkpkgall() { - : + : } addhandler distro_check_eventhandler @@ -679,7 +679,7 @@ addtask distro_checkall after do_distro_check do_distro_checkall[recrdeptask] = "do_distro_checkall do_distro_check" do_distro_checkall[nostamp] = "1" do_distro_checkall() { - : + : } # #Check Missing License Text. @@ -727,7 +727,7 @@ addtask checklicenseall after do_checklicense do_checklicenseall[recrdeptask] = "do_checklicenseall do_checklicense" do_checklicenseall[nostamp] = "1" do_checklicenseall() { - : + : } |