2024-09-09 08:52:07 +00:00
|
|
|
inherit package
|
|
|
|
|
|
|
|
IMAGE_PKGTYPE ?= "ipk"
|
|
|
|
|
|
|
|
IPKGCONF_TARGET = "${WORKDIR}/opkg.conf"
|
|
|
|
IPKGCONF_SDK = "${WORKDIR}/opkg-sdk.conf"
|
|
|
|
|
|
|
|
PKGWRITEDIRIPK = "${WORKDIR}/deploy-ipks"
|
|
|
|
|
|
|
|
# Program to be used to build opkg packages
|
|
|
|
OPKGBUILDCMD ??= "opkg-build"
|
|
|
|
|
2024-09-09 08:57:42 +00:00
|
|
|
OPKG_ARGS = "--force_postinstall --prefer-arch-to-version"
|
|
|
|
OPKG_ARGS += "${@['', '--no-install-recommends'][d.getVar("NO_RECOMMENDATIONS", True) == "1"]}"
|
|
|
|
OPKG_ARGS += "${@['', '--add-exclude ' + ' --add-exclude '.join((d.getVar('PACKAGE_EXCLUDE', True) or "").split())][(d.getVar("PACKAGE_EXCLUDE", True) or "") != ""]}"
|
2024-09-09 08:52:07 +00:00
|
|
|
|
2024-09-09 08:57:42 +00:00
|
|
|
OPKGLIBDIR = "${localstatedir}/lib"
|
2024-09-09 08:52:07 +00:00
|
|
|
|
|
|
|
python do_package_ipk () {
|
2024-09-09 08:57:42 +00:00
|
|
|
import re, copy
|
|
|
|
import textwrap
|
|
|
|
import subprocess
|
|
|
|
|
|
|
|
workdir = d.getVar('WORKDIR', True)
|
|
|
|
outdir = d.getVar('PKGWRITEDIRIPK', True)
|
|
|
|
tmpdir = d.getVar('TMPDIR', True)
|
|
|
|
pkgdest = d.getVar('PKGDEST', True)
|
|
|
|
if not workdir or not outdir or not tmpdir:
|
|
|
|
bb.error("Variables incorrectly set, unable to package")
|
|
|
|
return
|
|
|
|
|
|
|
|
packages = d.getVar('PACKAGES', True)
|
|
|
|
if not packages or packages == '':
|
|
|
|
bb.debug(1, "No packages; nothing to do")
|
|
|
|
return
|
|
|
|
|
|
|
|
# We're about to add new packages so the index needs to be checked
|
|
|
|
# so remove the appropriate stamp file.
|
|
|
|
if os.access(os.path.join(tmpdir, "stamps", "IPK_PACKAGE_INDEX_CLEAN"), os.R_OK):
|
|
|
|
os.unlink(os.path.join(tmpdir, "stamps", "IPK_PACKAGE_INDEX_CLEAN"))
|
|
|
|
|
|
|
|
def cleanupcontrol(root):
|
|
|
|
for p in ['CONTROL', 'DEBIAN']:
|
|
|
|
p = os.path.join(root, p)
|
|
|
|
if os.path.exists(p):
|
|
|
|
bb.utils.prunedir(p)
|
|
|
|
|
|
|
|
for pkg in packages.split():
|
|
|
|
localdata = bb.data.createCopy(d)
|
|
|
|
root = "%s/%s" % (pkgdest, pkg)
|
|
|
|
|
|
|
|
lf = bb.utils.lockfile(root + ".lock")
|
|
|
|
|
|
|
|
localdata.setVar('ROOT', '')
|
|
|
|
localdata.setVar('ROOT_%s' % pkg, root)
|
|
|
|
pkgname = localdata.getVar('PKG_%s' % pkg, True)
|
|
|
|
if not pkgname:
|
|
|
|
pkgname = pkg
|
|
|
|
localdata.setVar('PKG', pkgname)
|
|
|
|
|
|
|
|
localdata.setVar('OVERRIDES', pkg)
|
|
|
|
|
|
|
|
bb.data.update_data(localdata)
|
|
|
|
basedir = os.path.join(os.path.dirname(root))
|
|
|
|
arch = localdata.getVar('PACKAGE_ARCH', True)
|
|
|
|
|
|
|
|
if localdata.getVar('IPK_HIERARCHICAL_FEED') == "1":
|
|
|
|
# Spread packages across subdirectories so each isn't too crowded
|
|
|
|
if pkgname.startswith('lib'):
|
|
|
|
pkg_prefix = 'lib' + pkgname[3]
|
|
|
|
else:
|
|
|
|
pkg_prefix = pkgname[0]
|
|
|
|
|
|
|
|
# Keep -dbg, -dev, -doc, -staticdev, -locale and -locale-* packages
|
|
|
|
# together. These package suffixes are taken from the definitions of
|
|
|
|
# PACKAGES and PACKAGES_DYNAMIC in meta/conf/bitbake.conf
|
|
|
|
if pkgname[-4:] in ('-dbg', '-dev', '-doc'):
|
|
|
|
pkg_subdir = pkgname[:-4]
|
|
|
|
elif pkgname.endswith('-staticdev'):
|
|
|
|
pkg_subdir = pkgname[:-10]
|
|
|
|
elif pkgname.endswith('-locale'):
|
|
|
|
pkg_subdir = pkgname[:-7]
|
|
|
|
elif '-locale-' in pkgname:
|
|
|
|
pkg_subdir = pkgname[:pkgname.find('-locale-')]
|
|
|
|
else:
|
|
|
|
pkg_subdir = pkgname
|
|
|
|
|
|
|
|
pkgoutdir = "%s/%s/%s/%s" % (outdir, arch, pkg_prefix, pkg_subdir)
|
|
|
|
else:
|
|
|
|
pkgoutdir = "%s/%s" % (outdir, arch)
|
|
|
|
|
|
|
|
bb.utils.mkdirhier(pkgoutdir)
|
|
|
|
os.chdir(root)
|
|
|
|
cleanupcontrol(root)
|
|
|
|
from glob import glob
|
|
|
|
g = glob('*')
|
|
|
|
if not g and localdata.getVar('ALLOW_EMPTY') != "1":
|
|
|
|
bb.note("Not creating empty archive for %s-%s-%s" % (pkg, localdata.getVar('PKGV', True), localdata.getVar('PKGR', True)))
|
|
|
|
bb.utils.unlockfile(lf)
|
|
|
|
continue
|
|
|
|
|
|
|
|
controldir = os.path.join(root, 'CONTROL')
|
|
|
|
bb.utils.mkdirhier(controldir)
|
|
|
|
try:
|
|
|
|
ctrlfile = open(os.path.join(controldir, 'control'), 'w')
|
|
|
|
except OSError:
|
|
|
|
bb.utils.unlockfile(lf)
|
|
|
|
raise bb.build.FuncFailed("unable to open control file for writing.")
|
|
|
|
|
|
|
|
fields = []
|
|
|
|
pe = d.getVar('PKGE', True)
|
|
|
|
if pe and int(pe) > 0:
|
|
|
|
fields.append(["Version: %s:%s-%s\n", ['PKGE', 'PKGV', 'PKGR']])
|
|
|
|
else:
|
|
|
|
fields.append(["Version: %s-%s\n", ['PKGV', 'PKGR']])
|
|
|
|
fields.append(["Description: %s\n", ['DESCRIPTION']])
|
|
|
|
fields.append(["Section: %s\n", ['SECTION']])
|
|
|
|
fields.append(["Priority: %s\n", ['PRIORITY']])
|
|
|
|
fields.append(["Maintainer: %s\n", ['MAINTAINER']])
|
|
|
|
fields.append(["License: %s\n", ['LICENSE']])
|
|
|
|
fields.append(["Architecture: %s\n", ['PACKAGE_ARCH']])
|
|
|
|
fields.append(["OE: %s\n", ['PN']])
|
|
|
|
if d.getVar('HOMEPAGE', True):
|
|
|
|
fields.append(["Homepage: %s\n", ['HOMEPAGE']])
|
|
|
|
|
|
|
|
def pullData(l, d):
|
|
|
|
l2 = []
|
|
|
|
for i in l:
|
|
|
|
l2.append(d.getVar(i, True))
|
|
|
|
return l2
|
|
|
|
|
|
|
|
ctrlfile.write("Package: %s\n" % pkgname)
|
|
|
|
# check for required fields
|
|
|
|
try:
|
|
|
|
for (c, fs) in fields:
|
|
|
|
for f in fs:
|
|
|
|
if localdata.getVar(f) is None:
|
|
|
|
raise KeyError(f)
|
|
|
|
# Special behavior for description...
|
|
|
|
if 'DESCRIPTION' in fs:
|
|
|
|
summary = localdata.getVar('SUMMARY', True) or localdata.getVar('DESCRIPTION', True) or "."
|
|
|
|
ctrlfile.write('Description: %s\n' % summary)
|
|
|
|
description = localdata.getVar('DESCRIPTION', True) or "."
|
|
|
|
description = textwrap.dedent(description).strip()
|
|
|
|
if '\\n' in description:
|
|
|
|
# Manually indent
|
|
|
|
for t in description.split('\\n'):
|
|
|
|
# We don't limit the width when manually indent, but we do
|
|
|
|
# need the textwrap.fill() to set the initial_indent and
|
|
|
|
# subsequent_indent, so set a large width
|
|
|
|
ctrlfile.write('%s\n' % textwrap.fill(t.strip(), width=100000, initial_indent=' ', subsequent_indent=' '))
|
|
|
|
else:
|
|
|
|
# Auto indent
|
|
|
|
ctrlfile.write('%s\n' % textwrap.fill(description, width=74, initial_indent=' ', subsequent_indent=' '))
|
|
|
|
else:
|
|
|
|
ctrlfile.write(c % tuple(pullData(fs, localdata)))
|
|
|
|
except KeyError:
|
|
|
|
import sys
|
|
|
|
(type, value, traceback) = sys.exc_info()
|
|
|
|
ctrlfile.close()
|
|
|
|
bb.utils.unlockfile(lf)
|
|
|
|
raise bb.build.FuncFailed("Missing field for ipk generation: %s" % value)
|
|
|
|
# more fields
|
|
|
|
|
|
|
|
custom_fields_chunk = get_package_additional_metadata("ipk", localdata)
|
|
|
|
if custom_fields_chunk is not None:
|
|
|
|
ctrlfile.write(custom_fields_chunk)
|
|
|
|
ctrlfile.write("\n")
|
|
|
|
|
|
|
|
mapping_rename_hook(localdata)
|
|
|
|
|
|
|
|
def debian_cmp_remap(var):
|
|
|
|
# In debian '>' and '<' do not mean what it appears they mean
|
|
|
|
# '<' = less or equal
|
|
|
|
# '>' = greater or equal
|
|
|
|
# adjust these to the '<<' and '>>' equivalents
|
|
|
|
#
|
|
|
|
for dep in var:
|
|
|
|
for i, v in enumerate(var[dep]):
|
|
|
|
if (v or "").startswith("< "):
|
|
|
|
var[dep][i] = var[dep][i].replace("< ", "<< ")
|
|
|
|
elif (v or "").startswith("> "):
|
|
|
|
var[dep][i] = var[dep][i].replace("> ", ">> ")
|
|
|
|
|
|
|
|
rdepends = bb.utils.explode_dep_versions2(localdata.getVar("RDEPENDS", True) or "")
|
|
|
|
debian_cmp_remap(rdepends)
|
|
|
|
rrecommends = bb.utils.explode_dep_versions2(localdata.getVar("RRECOMMENDS", True) or "")
|
|
|
|
debian_cmp_remap(rrecommends)
|
|
|
|
rsuggests = bb.utils.explode_dep_versions2(localdata.getVar("RSUGGESTS", True) or "")
|
|
|
|
debian_cmp_remap(rsuggests)
|
|
|
|
rprovides = bb.utils.explode_dep_versions2(localdata.getVar("RPROVIDES", True) or "")
|
|
|
|
debian_cmp_remap(rprovides)
|
|
|
|
rreplaces = bb.utils.explode_dep_versions2(localdata.getVar("RREPLACES", True) or "")
|
|
|
|
debian_cmp_remap(rreplaces)
|
|
|
|
rconflicts = bb.utils.explode_dep_versions2(localdata.getVar("RCONFLICTS", True) or "")
|
|
|
|
debian_cmp_remap(rconflicts)
|
|
|
|
|
|
|
|
if rdepends:
|
|
|
|
ctrlfile.write("Depends: %s\n" % bb.utils.join_deps(rdepends))
|
|
|
|
if rsuggests:
|
|
|
|
ctrlfile.write("Suggests: %s\n" % bb.utils.join_deps(rsuggests))
|
|
|
|
if rrecommends:
|
|
|
|
ctrlfile.write("Recommends: %s\n" % bb.utils.join_deps(rrecommends))
|
|
|
|
if rprovides:
|
|
|
|
ctrlfile.write("Provides: %s\n" % bb.utils.join_deps(rprovides))
|
|
|
|
if rreplaces:
|
|
|
|
ctrlfile.write("Replaces: %s\n" % bb.utils.join_deps(rreplaces))
|
|
|
|
if rconflicts:
|
|
|
|
ctrlfile.write("Conflicts: %s\n" % bb.utils.join_deps(rconflicts))
|
|
|
|
src_uri = localdata.getVar("SRC_URI", True).strip() or "None"
|
|
|
|
if src_uri:
|
|
|
|
src_uri = re.sub("\s+", " ", src_uri)
|
|
|
|
ctrlfile.write("Source: %s\n" % " ".join(src_uri.split()))
|
|
|
|
ctrlfile.close()
|
|
|
|
|
|
|
|
for script in ["preinst", "postinst", "prerm", "postrm"]:
|
|
|
|
scriptvar = localdata.getVar('pkg_%s' % script, True)
|
|
|
|
if not scriptvar:
|
|
|
|
continue
|
|
|
|
try:
|
|
|
|
scriptfile = open(os.path.join(controldir, script), 'w')
|
|
|
|
except OSError:
|
|
|
|
bb.utils.unlockfile(lf)
|
|
|
|
raise bb.build.FuncFailed("unable to open %s script file for writing." % script)
|
|
|
|
scriptfile.write(scriptvar)
|
|
|
|
scriptfile.close()
|
|
|
|
os.chmod(os.path.join(controldir, script), 0755)
|
|
|
|
|
|
|
|
conffiles_str = ' '.join(get_conffiles(pkg, d))
|
|
|
|
if conffiles_str:
|
|
|
|
try:
|
|
|
|
conffiles = open(os.path.join(controldir, 'conffiles'), 'w')
|
|
|
|
except OSError:
|
|
|
|
bb.utils.unlockfile(lf)
|
|
|
|
raise bb.build.FuncFailed("unable to open conffiles for writing.")
|
|
|
|
for f in conffiles_str.split():
|
|
|
|
if os.path.exists(oe.path.join(root, f)):
|
|
|
|
conffiles.write('%s\n' % f)
|
|
|
|
conffiles.close()
|
|
|
|
|
|
|
|
os.chdir(basedir)
|
|
|
|
ret = subprocess.call("PATH=\"%s\" %s %s %s" % (localdata.getVar("PATH", True),
|
|
|
|
d.getVar("OPKGBUILDCMD",1), pkg, pkgoutdir), shell=True)
|
|
|
|
if ret != 0:
|
|
|
|
bb.utils.unlockfile(lf)
|
|
|
|
raise bb.build.FuncFailed("opkg-build execution failed")
|
|
|
|
|
|
|
|
cleanupcontrol(root)
|
|
|
|
bb.utils.unlockfile(lf)
|
2024-09-09 08:52:07 +00:00
|
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
SSTATETASKS += "do_package_write_ipk"
|
|
|
|
do_package_write_ipk[sstate-inputdirs] = "${PKGWRITEDIRIPK}"
|
|
|
|
do_package_write_ipk[sstate-outputdirs] = "${DEPLOY_DIR_IPK}"
|
|
|
|
|
|
|
|
python do_package_write_ipk_setscene () {
|
2024-09-09 08:57:42 +00:00
|
|
|
tmpdir = d.getVar('TMPDIR', True)
|
|
|
|
|
|
|
|
if os.access(os.path.join(tmpdir, "stamps", "IPK_PACKAGE_INDEX_CLEAN"), os.R_OK):
|
|
|
|
os.unlink(os.path.join(tmpdir, "stamps", "IPK_PACKAGE_INDEX_CLEAN"))
|
|
|
|
|
|
|
|
sstate_setscene(d)
|
2024-09-09 08:52:07 +00:00
|
|
|
}
|
|
|
|
addtask do_package_write_ipk_setscene
|
|
|
|
|
|
|
|
python () {
|
|
|
|
if d.getVar('PACKAGES', True) != '':
|
2024-09-09 08:57:42 +00:00
|
|
|
deps = ' opkg-utils-native:do_populate_sysroot virtual/fakeroot-native:do_populate_sysroot'
|
|
|
|
d.appendVarFlag('do_package_write_ipk', 'depends', deps)
|
2024-09-09 08:52:07 +00:00
|
|
|
d.setVarFlag('do_package_write_ipk', 'fakeroot', "1")
|
|
|
|
}
|
|
|
|
|
|
|
|
python do_package_write_ipk () {
|
2024-09-09 08:57:42 +00:00
|
|
|
bb.build.exec_func("read_subpackage_metadata", d)
|
|
|
|
bb.build.exec_func("do_package_ipk", d)
|
2024-09-09 08:52:07 +00:00
|
|
|
}
|
|
|
|
do_package_write_ipk[dirs] = "${PKGWRITEDIRIPK}"
|
2024-09-09 08:57:42 +00:00
|
|
|
do_package_write_ipk[cleandirs] = "${PKGWRITEDIRIPK}"
|
|
|
|
do_package_write_ipk[umask] = "022"
|
|
|
|
addtask package_write_ipk after do_packagedata do_package
|
2024-09-09 08:52:07 +00:00
|
|
|
|
|
|
|
PACKAGEINDEXDEPS += "opkg-utils-native:do_populate_sysroot"
|
|
|
|
PACKAGEINDEXDEPS += "opkg-native:do_populate_sysroot"
|
2024-09-09 08:57:42 +00:00
|
|
|
|
|
|
|
do_build[recrdeptask] += "do_package_write_ipk"
|