licenses = get_licenses(d)
script_logs = os.path.join(work_dir, 'script-logs/'+ target_sys + '/' + licenses + '/' + pf + '/script-logs')
bb_inc = os.path.join(script_logs, 'bb_inc')
- bb.mkdirhier(bb_inc)
+ bb.utils.mkdirhier(bb_inc)
def find_file(dir, file):
for root, dirs, files in os.walk(dir):
script_logs = os.path.join(work_dir, 'script-logs/'+ target_sys + '/' + licenses + '/' + pf + '/script-logs')
try:
- bb.mkdirhier(os.path.join(script_logs, 'temp'))
+ bb.utils.mkdirhier(os.path.join(script_logs, 'temp'))
oe.path.copytree(os.path.join(work_dir, 'temp'), os.path.join(script_logs, 'temp'))
except (IOError, AttributeError):
pass
s = d.getVar('S', True)
dest = os.path.join(work_dir, pf + '-series')
shutil.rmtree(dest, ignore_errors=True)
- bb.mkdirhier(dest)
+ bb.utils.mkdirhier(dest)
src_uri = d.getVar('SRC_URI', True).split()
fetch = bb.fetch2.Fetch(src_uri, d)
shutil.copy(patch, dest)
except IOError:
if os.path.isdir(patch):
- bb.mkdirhier(os.path.join(dest, patch))
+ bb.utils.mkdirhier(os.path.join(dest, patch))
oe.path.copytree(patch, os.path.join(dest, patch))
return dest
work_dir = d.getVar('WORKDIR', True)
dest = os.path.join(work_dir, pf + '-patches')
shutil.rmtree(dest, ignore_errors=True)
- bb.mkdirhier(dest)
+ bb.utils.mkdirhier(dest)
patches = src_patches(d)
for patch in patches:
- _, _, local, _, _, parm = bb.decodeurl(patch)
+ _, _, local, _, _, parm = bb.fetch.decodeurl(patch)
if local:
shutil.copy(local, dest)
return dest
work_dir = d.getVar('WORKDIR', True)
tar_sources = d.getVar('DEPLOY_DIR', True) + '/sources/' + target_sys + '/' + licenses + '/' + pf
if not os.path.exists(tar_sources):
- bb.mkdirhier(tar_sources)
+ bb.utils.mkdirhier(tar_sources)
for source in tarball_list:
if source:
if os.path.exists(os.path.join(tar_sources, source)):
licenses = get_licenses(d)
dumpdir = os.path.join(workdir, 'diffgz-envdata/'+ target_sys + '/' + licenses + '/' + pf )
if not os.path.exists(dumpdir):
- bb.mkdirhier(dumpdir)
+ bb.utils.mkdirhier(dumpdir)
dumpfile = os.path.join(dumpdir, bb.data.expand("${P}-${PR}.showdata.dump", d))
distro = d.getVar('DISTRO',True) or ""
dest = s + '/' + distro + '/files'
if not os.path.exists(dest):
- bb.mkdirhier(dest)
+ bb.utils.mkdirhier(dest)
for i in os.listdir(os.getcwd()):
if os.path.isfile(i):
try:
d.appendVarFlag('do_fetch', 'depends', ' osc-native:do_populate_sysroot')
# *.xz should depends on xz-native for unpacking
- # Not endswith because of "*.patch.xz;patch=1". Need bb.decodeurl in future
+ # Not endswith because of "*.patch.xz;patch=1". Need bb.fetch.decodeurl in future
if '.xz' in srcuri:
d.appendVarFlag('do_unpack', 'depends', ' xz-native:do_populate_sysroot')
# set the buildname
########################################################################
try:
- bb.mkdirhier(e.data.getVar('BUILDSTATS_BASE', True))
+ bb.utils.mkdirhier(e.data.getVar('BUILDSTATS_BASE', True))
except:
pass
set_bn(e)
bsdir = os.path.join(e.data.getVar('BUILDSTATS_BASE', True), bn)
try:
- bb.mkdirhier(bsdir)
+ bb.utils.mkdirhier(bsdir)
except:
pass
if device != "NoLogicalDevice":
set_diskdata("__diskdata_task", device, e.data)
set_timedata("__timedata_task", e.data)
try:
- bb.mkdirhier(taskdir)
+ bb.utils.mkdirhier(taskdir)
except:
pass
# write into the task event file the name and start time
-CCACHE = "${@bb.which(d.getVar('PATH', True), 'ccache') and 'ccache '}"
+CCACHE = "${@bb.utils.which(d.getVar('PATH', True), 'ccache') and 'ccache '}"
export CCACHE_DIR ?= "${TMPDIR}/ccache/${MULTIMACH_HOST_SYS}/${PN}"
CCACHE_DISABLE[unexport] = "1"
pf = d.getVar('PF', True)
dest = os.path.join(sources_dir, pf)
shutil.rmtree(dest, ignore_errors=True)
- bb.mkdirhier(dest)
+ bb.utils.mkdirhier(dest)
for u in ud.values():
local = os.path.normpath(fetch.localpath(u.url))
patches = src_patches(d)
for patch in patches:
- _, _, local, _, _, parm = bb.decodeurl(patch)
+ _, _, local, _, _, parm = bb.fetch.decodeurl(patch)
patchdir = parm.get('patchdir')
if patchdir:
series = os.path.join(dest, 'series.subdir.%s' % patchdir.replace('/', '_'))
pupver = "N/A"
pstatus = "ErrUnknown"
- (type, host, path, user, pswd, parm) = bb.decodeurl(uri)
+ (type, host, path, user, pswd, parm) = bb.fetch.decodeurl(uri)
if type in ['http', 'https', 'ftp']:
if d.getVar('PRSPV', True):
pcurver = d.getVar('PRSPV', True)
dirver = m.group().strip("/")
"""use new path and remove param. for wget only param is md5sum"""
- alturi = bb.encodeurl([type, host, altpath, user, pswd, {}])
+ alturi = bb.fetch.encodeurl([type, host, altpath, user, pswd, {}])
my_uri = d.getVar('REGEX_URI', True)
if my_uri:
if d.getVar('PRSPV', True):
chk_uri = d.getVar('REGEX_URI', True)
if not chk_uri:
- alturi = bb.encodeurl([type, host, altpath, user, pswd, {}])
+ alturi = bb.fetch.encodeurl([type, host, altpath, user, pswd, {}])
else:
alturi = chk_uri
newver = check_new_version(alturi, curname, d)
if devtables == None:
devtables = 'files/device_table-minimal.txt'
for devtable in devtables.split():
- str += " %s" % bb.which(d.getVar('BBPATH', True), devtable)
+ str += " %s" % bb.utils.which(d.getVar('BBPATH', True), devtable)
return str
IMAGE_CLASSES ?= "image_types"
srcdir = d.getVar('S', True)
for url in lic_files.split():
- (type, host, path, user, pswd, parm) = bb.decodeurl(url)
+ (type, host, path, user, pswd, parm) = bb.fetch.decodeurl(url)
srclicfile = os.path.join(srcdir, path)
if not os.path.isfile(srclicfile):
raise bb.build.FuncFailed( pn + ": LIC_FILES_CHKSUM points to an invalid file: " + srclicfile)
patches = src_patches(d)
patch_list=[]
for p in patches:
- _, _, local, _, _, _ = bb.decodeurl(p)
+ _, _, local, _, _, _ = bb.fetch.decodeurl(p)
patch_list.append(local)
return patch_list
d.setVar('RRECOMMENDS_' + pn, "%s" % (pn_lic))
def copy_license_files(lic_files_paths, destdir):
- bb.mkdirhier(destdir)
+ bb.utils.mkdirhier(destdir)
for (basename, path) in lic_files_paths:
- ret = bb.copyfile(path, os.path.join(destdir, basename))
+ ret = bb.utils.copyfile(path, os.path.join(destdir, basename))
# If the copy didn't occur, something horrible went wrong and we fail out
if not ret:
bb.warn("%s could not be copied for some reason. It may not exist. WARN for now." % path)
def find_license(license_type):
try:
- bb.mkdirhier(gen_lic_dest)
+ bb.utils.mkdirhier(gen_lic_dest)
except:
pass
spdx_generic = None
return lic_files_paths
for url in lic_files.split():
- (type, host, path, user, pswd, parm) = bb.decodeurl(url)
+ (type, host, path, user, pswd, parm) = bb.fetch.decodeurl(url)
# We want the license filename and path
srclicfile = os.path.join(srcdir, path)
lic_files_paths.append((os.path.basename(path), srclicfile))
basedir = os.path.join(os.path.dirname(root))
pkgoutdir = os.path.join(outdir, localdata.getVar('PACKAGE_ARCH', True))
- bb.mkdirhier(pkgoutdir)
+ bb.utils.mkdirhier(pkgoutdir)
os.chdir(root)
from glob import glob
continue
controldir = os.path.join(root, 'DEBIAN')
- bb.mkdirhier(controldir)
+ bb.utils.mkdirhier(controldir)
os.chmod(controldir, 0755)
try:
ctrlfile = open(os.path.join(controldir, 'control'), 'w')
basedir = os.path.join(os.path.dirname(root))
arch = localdata.getVar('PACKAGE_ARCH', True)
pkgoutdir = "%s/%s" % (outdir, arch)
- bb.mkdirhier(pkgoutdir)
+ bb.utils.mkdirhier(pkgoutdir)
os.chdir(root)
from glob import glob
g = glob('*')
continue
controldir = os.path.join(root, 'CONTROL')
- bb.mkdirhier(controldir)
+ bb.utils.mkdirhier(controldir)
try:
ctrlfile = open(os.path.join(controldir, 'control'), 'w')
except OSError:
clean_licenses = get_licenses(d)
pkgwritesrpmdir = bb.data.expand('${PKGWRITEDIRSRPM}/${PACKAGE_ARCH_EXTEND}', d)
pkgwritesrpmdir = pkgwritesrpmdir + '/' + clean_licenses
- bb.mkdirhier(pkgwritesrpmdir)
+ bb.utils.mkdirhier(pkgwritesrpmdir)
os.chmod(pkgwritesrpmdir, 0755)
return pkgwritesrpmdir
pkgwritedir = d.expand('${PKGWRITEDIRRPM}/${PACKAGE_ARCH_EXTEND}')
pkgarch = d.expand('${PACKAGE_ARCH_EXTEND}${TARGET_VENDOR}-${TARGET_OS}')
magicfile = d.expand('${STAGING_DIR_NATIVE}${datadir_native}/misc/magic.mgc')
- bb.mkdirhier(pkgwritedir)
+ bb.utils.mkdirhier(pkgwritedir)
os.chmod(pkgwritedir, 0755)
cmd = rpmbuild
if patchdir:
patchparm['patchdir'] = patchdir
- localurl = bb.encodeurl(('file', '', local, '', '', patchparm))
+ localurl = bb.fetch.encodeurl(('file', '', local, '', '', patchparm))
patches.append(localurl)
if all:
os.environ['TMPDIR'] = process_tmpdir
for patch in src_patches(d):
- _, _, local, _, _, parm = bb.decodeurl(patch)
+ _, _, local, _, _, parm = bb.fetch.decodeurl(patch)
if "patchdir" in parm:
patchdir = parm["patchdir"]
def can_delete_FILESDIR(cfgdata, d):
expected = cfgdata.get("FILESDIR")
- #expected = "${@bb.which(d.getVar('FILESPATH', True), '.')}"
+ #expected = "${@bb.utils.which(d.getVar('FILESPATH', True), '.')}"
unexpanded = d.getVar("FILESDIR", 0)
if unexpanded is None:
return False
cur_ver_code = get_ver_code( info['sourcedir'] )
cache_cur = False
if not os.path.exists( spdx_sstate_dir ):
- bb.mkdirhier( spdx_sstate_dir )
+ bb.utils.mkdirhier( spdx_sstate_dir )
if not os.path.exists( info['spdx_temp_dir'] ):
- bb.mkdirhier( info['spdx_temp_dir'] )
+ bb.utils.mkdirhier( info['spdx_temp_dir'] )
if os.path.exists( sstatefile ):
## cache for this package exists. read it in
cached_spdx = get_cached_spdx( sstatefile )
sharedfiles = []
shareddirs = []
- bb.mkdirhier(d.expand("${SSTATE_MANIFESTS}"))
+ bb.utils.mkdirhier(d.expand("${SSTATE_MANIFESTS}"))
d2 = d.createCopy()
extrainf = d.getVarFlag("do_" + ss['task'], 'stamp-extra-info', True)
# remove dir if it exists, ensure any parent directories do exist
if os.path.exists(dir):
oe.path.remove(dir)
- bb.mkdirhier(dir)
+ bb.utils.mkdirhier(dir)
oe.path.remove(dir)
sstateinst = d.expand("${WORKDIR}/sstate-install-%s/" % ss['name'])
workdir = d.getVar('WORKDIR', True)
src = sstateinst + "/" + plain.replace(workdir, '')
dest = plain
- bb.mkdirhier(src)
+ bb.utils.mkdirhier(src)
prepdir(dest)
os.rename(src, dest)
sstatebuild = d.expand("${WORKDIR}/sstate-build-%s/" % ss['name'])
sstatepkg = d.getVar('SSTATE_PKG', True) + '_'+ ss['name'] + ".tgz"
bb.utils.remove(sstatebuild, recurse=True)
- bb.mkdirhier(sstatebuild)
- bb.mkdirhier(os.path.dirname(sstatepkg))
+ bb.utils.mkdirhier(sstatebuild)
+ bb.utils.mkdirhier(os.path.dirname(sstatepkg))
for state in ss['dirs']:
if not os.path.exists(state[1]):
continue
workdir = d.getVar('WORKDIR', True)
for plain in ss['plaindirs']:
pdir = plain.replace(workdir, sstatebuild)
- bb.mkdirhier(plain)
- bb.mkdirhier(pdir)
+ bb.utils.mkdirhier(plain)
+ bb.utils.mkdirhier(pdir)
oe.path.copyhardlinktree(plain, pdir)
d.setVar('SSTATE_BUILDDIR', sstatebuild)
bb.data.update_data(localdata)
dldir = localdata.expand("${SSTATE_DIR}")
- bb.mkdirhier(dldir)
+ bb.utils.mkdirhier(dldir)
localdata.delVar('MIRRORS')
localdata.delVar('FILESPATH')
runfmt = d.getVar('BB_RUNFMT', True) or "run.{func}.{pid}"
runfile = runfmt.format(func=cmd_func, task=cmd_func, taskfunc=cmd_func, pid=os.getpid())
runfile = os.path.join(d.getVar('T', True), runfile)
- bb.mkdirhier(os.path.dirname(runfile))
+ bb.utils.mkdirhier(os.path.dirname(runfile))
with open(runfile, 'w') as script:
script.write('#!/bin/sh -e\n')
bb.fatal("Unable to populate toolchain binary symlinks in %s" % pattern)
bindir = d.getVar('STAGING_BINDIR_TOOLCHAIN', True)
- bb.mkdirhier(bindir)
+ bb.utils.mkdirhier(bindir)
for f in files:
base = os.path.basename(f)
newpath = os.path.join(bindir, base)
patch = self.patches[kwargs["patch"]]
if not patch:
raise PatchError("No patch found at index %s in patchset." % kwargs["patch"])
- (type, host, path, user, pswd, parm) = bb.decodeurl(patch["remote"])
+ (type, host, path, user, pswd, parm) = bb.fetch.decodeurl(patch["remote"])
if type == "file":
import shutil
if not patch.get("file") and patch.get("remote"):
oe_runconf
}
-export AUTOMAKE = "${@bb.which('automake', d.getVar('PATH', True))}"
+export AUTOMAKE = "${@bb.utils.which('automake', d.getVar('PATH', True))}"
FILES_${PN} += "${datadir}/automake* ${datadir}/aclocal*"
rel_path = os.path.relpath(fn, src_conf_dir)
parent_dir = os.path.dirname(rel_path)
bb.utils.mkdirhier('%s/%s' % (build_conf_dir, parent_dir))
- bb.copyfile(fn, '%s/%s' % (build_conf_dir, rel_path))
+ bb.utils.copyfile(fn, '%s/%s' % (build_conf_dir, rel_path))
multilibs = (d.getVar('MULTILIB_VARIANTS', True) or '').split()
if not multilibs:
S = "${WORKDIR}/pcre-${PV}"
-FILESPATH .= ":${@base_set_filespath([bb.which(BBPATH, 'recipes-support/libpcre/files', direction=True)], d)}"
+FILESPATH .= ":${@base_set_filespath([bb.utils.which(BBPATH, 'recipes-support/libpcre/files', direction=True)], d)}"
PROVIDES += "pcre"
DEPENDS += "bzip2 zlib"