summaryrefslogtreecommitdiffstats
path: root/lib/bb/fetch2/__init__.py
diff options
context:
space:
mode:
Diffstat (limited to 'lib/bb/fetch2/__init__.py')
-rw-r--r--lib/bb/fetch2/__init__.py78
1 files changed, 39 insertions, 39 deletions
diff --git a/lib/bb/fetch2/__init__.py b/lib/bb/fetch2/__init__.py
index d3c761d00d..19a79fdbfe 100644
--- a/lib/bb/fetch2/__init__.py
+++ b/lib/bb/fetch2/__init__.py
@@ -28,7 +28,7 @@ from __future__ import absolute_import
from __future__ import print_function
import os, re
import logging
-import bb.data, bb.persist_data, bb.utils
+import bb.persist_data, bb.utils
from bb import data
__version__ = "2"
@@ -211,7 +211,7 @@ def fetcher_init(d):
Calls before this must not hit the cache.
"""
# When to drop SCM head revisions controlled by user policy
- srcrev_policy = bb.data.getVar('BB_SRCREV_POLICY', d, True) or "clear"
+ srcrev_policy = d.getVar('BB_SRCREV_POLICY', True) or "clear"
if srcrev_policy == "cache":
logger.debug(1, "Keeping SRCREV cache due to cache policy of: %s", srcrev_policy)
elif srcrev_policy == "clear":
@@ -271,7 +271,7 @@ def verify_checksum(u, ud, d):
sha256data = bb.utils.sha256_file(ud.localpath)
# If strict checking enabled and neither sum defined, raise error
- strict = bb.data.getVar("BB_STRICT_CHECKSUM", d, True) or None
+ strict = d.getVar("BB_STRICT_CHECKSUM", True) or None
if (strict and ud.md5_expected == None and ud.sha256_expected == None):
raise FetchError('No checksum specified for %s, please add at least one to the recipe:\n'
'SRC_URI[%s] = "%s"\nSRC_URI[%s] = "%s"', u,
@@ -336,8 +336,8 @@ def subprocess_setup():
def get_autorev(d):
# only not cache src rev in autorev case
- if bb.data.getVar('BB_SRCREV_POLICY', d, True) != "cache":
- bb.data.setVar('__BB_DONT_CACHE', '1', d)
+ if d.getVar('BB_SRCREV_POLICY', True) != "cache":
+ d.setVar('__BB_DONT_CACHE', '1')
return "AUTOINC"
def get_srcrev(d):
@@ -350,7 +350,7 @@ def get_srcrev(d):
"""
scms = []
- fetcher = Fetch(bb.data.getVar('SRC_URI', d, True).split(), d)
+ fetcher = Fetch(d.getVar('SRC_URI', True).split(), d)
urldata = fetcher.ud
for u in urldata:
if urldata[u].method.supports_srcrev():
@@ -365,7 +365,7 @@ def get_srcrev(d):
#
# Mutiple SCMs are in SRC_URI so we resort to SRCREV_FORMAT
#
- format = bb.data.getVar('SRCREV_FORMAT', d, True)
+ format = d.getVar('SRCREV_FORMAT', True)
if not format:
raise FetchError("The SRCREV_FORMAT variable must be set when multiple SCMs are used.")
@@ -400,7 +400,7 @@ def runfetchcmd(cmd, d, quiet = False, cleanup = []):
'GIT_PROXY_IGNORE', 'SOCKS5_USER', 'SOCKS5_PASSWD']
for var in exportvars:
- val = bb.data.getVar(var, d, True)
+ val = d.getVar(var, True)
if val:
cmd = 'export ' + var + '=\"%s\"; %s' % (val, cmd)
@@ -440,7 +440,7 @@ def check_network_access(d, info = "", url = None):
"""
log remote network access, and error if BB_NO_NETWORK is set
"""
- if bb.data.getVar("BB_NO_NETWORK", d, True) == "1":
+ if d.getVar("BB_NO_NETWORK", True) == "1":
raise NetworkAccess(url, info)
else:
logger.debug(1, "Fetcher accessed the network with the command %s" % info)
@@ -526,15 +526,15 @@ def srcrev_internal_helper(ud, d, name):
return ud.parm['tag']
rev = None
- pn = bb.data.getVar("PN", d, True)
+ pn = d.getVar("PN", True)
if name != '':
- rev = bb.data.getVar("SRCREV_%s_pn-%s" % (name, pn), d, True)
+ rev = d.getVar("SRCREV_%s_pn-%s" % (name, pn), True)
if not rev:
- rev = bb.data.getVar("SRCREV_%s" % name, d, True)
+ rev = d.getVar("SRCREV_%s" % name, True)
if not rev:
- rev = bb.data.getVar("SRCREV_pn-%s" % pn, d, True)
+ rev = d.getVar("SRCREV_pn-%s" % pn, True)
if not rev:
- rev = bb.data.getVar("SRCREV", d, True)
+ rev = d.getVar("SRCREV", True)
if rev == "INVALID":
raise FetchError("Please set SRCREV to a valid value", ud.url)
if rev == "AUTOINC":
@@ -572,11 +572,11 @@ class FetchData(object):
if self.md5_name in self.parm:
self.md5_expected = self.parm[self.md5_name]
else:
- self.md5_expected = bb.data.getVarFlag("SRC_URI", self.md5_name, d)
+ self.md5_expected = d.getVarFlag("SRC_URI", self.md5_name)
if self.sha256_name in self.parm:
self.sha256_expected = self.parm[self.sha256_name]
else:
- self.sha256_expected = bb.data.getVarFlag("SRC_URI", self.sha256_name, d)
+ self.sha256_expected = d.getVarFlag("SRC_URI", self.sha256_name)
self.names = self.parm.get("name",'default').split(',')
@@ -600,7 +600,7 @@ class FetchData(object):
self.localpath = self.method.localpath(self.url, self, d)
# Note: These files should always be in DL_DIR whereas localpath may not be.
- basepath = bb.data.expand("${DL_DIR}/%s" % os.path.basename(self.localpath or self.basename), d)
+ basepath = d.expand("${DL_DIR}/%s" % os.path.basename(self.localpath or self.basename))
self.donestamp = basepath + '.done'
self.lockfile = basepath + '.lock'
@@ -626,12 +626,12 @@ class FetchData(object):
if "srcdate" in self.parm:
return self.parm['srcdate']
- pn = bb.data.getVar("PN", d, True)
+ pn = d.getVar("PN", True)
if pn:
- return bb.data.getVar("SRCDATE_%s" % pn, d, True) or bb.data.getVar("SRCDATE", d, True) or bb.data.getVar("DATE", d, True)
+ return d.getVar("SRCDATE_%s" % pn, True) or d.getVar("SRCDATE", True) or d.getVar("DATE", True)
- return bb.data.getVar("SRCDATE", d, True) or bb.data.getVar("DATE", d, True)
+ return d.getVar("SRCDATE", True) or d.getVar("DATE", True)
class FetchMethod(object):
"""Base class for 'fetch'ing data"""
@@ -703,7 +703,7 @@ class FetchMethod(object):
dots = file.split(".")
if dots[-1] in ['gz', 'bz2', 'Z']:
- efile = os.path.join(bb.data.getVar('WORKDIR', data, True),os.path.basename('.'.join(dots[0:-1])))
+ efile = os.path.join(data.getVar('WORKDIR', True),os.path.basename('.'.join(dots[0:-1])))
else:
efile = file
cmd = None
@@ -747,7 +747,7 @@ class FetchMethod(object):
dest = os.path.join(rootdir, os.path.basename(file))
if (file != dest) and not (os.path.exists(dest) and os.path.samefile(file, dest)):
if os.path.isdir(file):
- filesdir = os.path.realpath(bb.data.getVar("FILESDIR", data, True))
+ filesdir = os.path.realpath(data.getVar("FILESDIR", True))
destdir = "."
if file[0:len(filesdir)] == filesdir:
destdir = file[len(filesdir):file.rfind('/')]
@@ -779,7 +779,7 @@ class FetchMethod(object):
bb.utils.mkdirhier(newdir)
os.chdir(newdir)
- cmd = "PATH=\"%s\" %s" % (bb.data.getVar('PATH', data, True), cmd)
+ cmd = "PATH=\"%s\" %s" % (data.getVar('PATH', True), cmd)
bb.note("Unpacking %s to %s/" % (file, os.getcwd()))
ret = subprocess.call(cmd, preexec_fn=subprocess_setup, shell=True)
@@ -824,10 +824,10 @@ class FetchMethod(object):
localcount = None
if name != '':
- pn = bb.data.getVar("PN", d, True)
- localcount = bb.data.getVar("LOCALCOUNT_" + name, d, True)
+ pn = d.getVar("PN", True)
+ localcount = d.getVar("LOCALCOUNT_" + name, True)
if not localcount:
- localcount = bb.data.getVar("LOCALCOUNT", d, True)
+ localcount = d.getVar("LOCALCOUNT", True)
return localcount
localcount_internal_helper = staticmethod(localcount_internal_helper)
@@ -859,7 +859,7 @@ class FetchMethod(object):
latest_rev = self._build_revision(url, ud, d, name)
last_rev = localcounts.get(key + '_rev')
- uselocalcount = bb.data.getVar("BB_LOCALCOUNT_OVERRIDE", d, True) or False
+ uselocalcount = d.getVar("BB_LOCALCOUNT_OVERRIDE", True) or False
count = None
if uselocalcount:
count = FetchMethod.localcount_internal_helper(ud, d, name)
@@ -887,7 +887,7 @@ class FetchMethod(object):
def generate_revision_key(self, url, ud, d, name):
key = self._revision_key(url, ud, d, name)
- return "%s-%s" % (key, bb.data.getVar("PN", d, True) or "")
+ return "%s-%s" % (key, d.getVar("PN", True) or "")
class Fetch(object):
def __init__(self, urls, d, cache = True):
@@ -897,7 +897,7 @@ class Fetch(object):
self.d = d
self.ud = {}
- fn = bb.data.getVar('FILE', d, True)
+ fn = d.getVar('FILE', True)
if cache and fn in urldata_cache:
self.ud = urldata_cache[fn]
@@ -913,7 +913,7 @@ class Fetch(object):
self.ud[url] = FetchData(url, self.d)
self.ud[url].setup_localpath(self.d)
- return bb.data.expand(self.ud[url].localpath, self.d)
+ return self.d.expand(self.ud[url].localpath)
def localpaths(self):
"""
@@ -935,8 +935,8 @@ class Fetch(object):
if len(urls) == 0:
urls = self.urls
- network = bb.data.getVar("BB_NO_NETWORK", self.d, True)
- premirroronly = (bb.data.getVar("BB_FETCH_PREMIRRORONLY", self.d, True) == "1")
+ network = self.d.getVar("BB_NO_NETWORK", True)
+ premirroronly = (self.d.getVar("BB_FETCH_PREMIRRORONLY", True) == "1")
for u in urls:
ud = self.ud[u]
@@ -947,17 +947,17 @@ class Fetch(object):
lf = bb.utils.lockfile(ud.lockfile)
try:
- bb.data.setVar("BB_NO_NETWORK", network, self.d)
+ self.d.setVar("BB_NO_NETWORK", network)
if not m.need_update(u, ud, self.d):
localpath = ud.localpath
elif m.try_premirror(u, ud, self.d):
logger.debug(1, "Trying PREMIRRORS")
- mirrors = mirror_from_string(bb.data.getVar('PREMIRRORS', self.d, True))
+ mirrors = mirror_from_string(self.d.getVar('PREMIRRORS', True))
localpath = try_mirrors(self.d, ud, mirrors, False)
if premirroronly:
- bb.data.setVar("BB_NO_NETWORK", "1", self.d)
+ self.d.setVar("BB_NO_NETWORK", "1")
if not localpath and m.need_update(u, ud, self.d):
try:
@@ -979,7 +979,7 @@ class Fetch(object):
if os.path.isfile(ud.localpath):
bb.utils.remove(ud.localpath)
logger.debug(1, "Trying MIRRORS")
- mirrors = mirror_from_string(bb.data.getVar('MIRRORS', self.d, True))
+ mirrors = mirror_from_string(self.d.getVar('MIRRORS', True))
localpath = try_mirrors (self.d, ud, mirrors)
if not localpath or ((not os.path.exists(localpath)) and localpath.find("*") == -1):
@@ -1004,7 +1004,7 @@ class Fetch(object):
m = ud.method
logger.debug(1, "Testing URL %s", u)
# First try checking uri, u, from PREMIRRORS
- mirrors = mirror_from_string(bb.data.getVar('PREMIRRORS', self.d, True))
+ mirrors = mirror_from_string(self.d.getVar('PREMIRRORS', True))
ret = try_mirrors(self.d, ud, mirrors, True)
if not ret:
# Next try checking from the original uri, u
@@ -1012,7 +1012,7 @@ class Fetch(object):
ret = m.checkstatus(u, ud, self.d)
except:
# Finally, try checking uri, u, from MIRRORS
- mirrors = mirror_from_string(bb.data.getVar('MIRRORS', self.d, True))
+ mirrors = mirror_from_string(self.d.getVar('MIRRORS', True))
ret = try_mirrors (self.d, ud, mirrors, True)
if not ret:
@@ -1030,7 +1030,7 @@ class Fetch(object):
ud = self.ud[u]
ud.setup_localpath(self.d)
- if bb.data.expand(self.localpath, self.d) is None:
+ if self.d.expand(self.localpath) is None:
continue
if ud.lockfile: