diff options
-rw-r--r-- | classes/base.bbclass | 96 | ||||
-rw-r--r-- | classes/patch.bbclass | 476 | ||||
-rw-r--r-- | conf/bitbake.conf | 12 | ||||
-rw-r--r-- | conf/machine/dht-walnut.conf | 20 | ||||
-rw-r--r-- | conf/machine/rb500.conf | 8 | ||||
-rw-r--r-- | packages/gawk/gawk-native_3.1.4.bb | 4 | ||||
-rw-r--r-- | packages/gobby/gobby_0.4.1.bb | 8 | ||||
-rw-r--r-- | packages/gobby/gtksourceview_1.7.2.bb | 10 | ||||
-rw-r--r-- | packages/gobby/libxml++_2.14.0.bb | 10 | ||||
-rw-r--r-- | packages/gobby/net6_1.2.2.bb | 8 | ||||
-rw-r--r-- | packages/gobby/net6_1.3.1.bb | 15 | ||||
-rw-r--r-- | packages/gobby/obby_0.4.1.bb | 19 | ||||
-rw-r--r-- | packages/patcher/patcher-native_20040913.bb | 4 | ||||
-rw-r--r-- | packages/quilt/quilt-native.inc | 3 |
14 files changed, 593 insertions, 100 deletions
diff --git a/classes/base.bbclass b/classes/base.bbclass index 670b97a72d..51b6d2111f 100644 --- a/classes/base.bbclass +++ b/classes/base.bbclass @@ -1,5 +1,4 @@ BB_DEFAULT_TASK = "build" -PATCHES_DIR="${S}" def base_dep_prepend(d): import bb; @@ -13,9 +12,11 @@ def base_dep_prepend(d): # INHIBIT_DEFAULT_DEPS doesn't apply to the patch command. Whether or not # we need that built is the responsibility of the patch function / class, not # the application. - patchdeps = bb.data.getVar("PATCH_DEPENDS", d, 1) - if patchdeps and not patchdeps in bb.data.getVar("PROVIDES", d, 1): - deps = patchdeps + patchdeps = bb.data.getVar("PATCHTOOL", d, 1) + if patchdeps: + patchdeps = "%s-native" % patchdeps + if not patchdeps in bb.data.getVar("PROVIDES", d, 1): + deps = patchdeps if not bb.data.getVar('INHIBIT_DEFAULT_DEPS', d): if (bb.data.getVar('HOST_SYS', d, 1) != @@ -451,86 +452,6 @@ python base_do_unpack() { raise bb.build.FuncFailed() } -addtask patch after do_unpack -do_patch[dirs] = "${WORKDIR}" -python base_do_patch() { - import re - import bb.fetch - - src_uri = (bb.data.getVar('SRC_URI', d, 1) or '').split() - if not src_uri: - return - - patchcleancmd = bb.data.getVar('PATCHCLEANCMD', d, 1) - if patchcleancmd: - bb.data.setVar("do_patchcleancmd", patchcleancmd, d) - bb.data.setVarFlag("do_patchcleancmd", "func", 1, d) - bb.build.exec_func("do_patchcleancmd", d) - - workdir = bb.data.getVar('WORKDIR', d, 1) - for url in src_uri: - - (type, host, path, user, pswd, parm) = bb.decodeurl(url) - if not "patch" in parm: - continue - - bb.fetch.init([url],d) - url = bb.encodeurl((type, host, path, user, pswd, [])) - local = os.path.join('/', bb.fetch.localpath(url, d)) - - # did it need to be unpacked? - dots = os.path.basename(local).split(".") - if dots[-1] in ['gz', 'bz2', 'Z']: - unpacked = os.path.join(bb.data.getVar('WORKDIR', d),'.'.join(dots[0:-1])) - else: - unpacked = local - unpacked = bb.data.expand(unpacked, d) - - if "pnum" in parm: - pnum = parm["pnum"] - else: - pnum = "1" - - if "pname" in parm: - pname = parm["pname"] - else: - pname = os.path.basename(unpacked) - - if "mindate" in parm: - mindate = parm["mindate"] - else: - mindate = 0 - - if "maxdate" in parm: - maxdate = parm["maxdate"] - else: - maxdate = "20711226" - - pn = bb.data.getVar('PN', d, 1) - srcdate = bb.data.getVar('SRCDATE_%s' % pn, d, 1) - - if not srcdate: - srcdate = bb.data.getVar('SRCDATE', d, 1) - - if srcdate == "now": - srcdate = bb.data.getVar('DATE', d, 1) - - if (maxdate < srcdate) or (mindate > srcdate): - if (maxdate < srcdate): - bb.note("Patch '%s' is outdated" % pname) - - if (mindate > srcdate): - bb.note("Patch '%s' is predated" % pname) - - continue - - bb.note("Applying patch '%s'" % pname) - bb.data.setVar("do_patchcmd", bb.data.getVar("PATCHCMD", d, 1) % (pnum, pname, unpacked), d) - bb.data.setVarFlag("do_patchcmd", "func", 1, d) - bb.data.setVarFlag("do_patchcmd", "dirs", "${WORKDIR} ${S}", d) - bb.build.exec_func("do_patchcmd", d) -} - addhandler base_eventhandler python base_eventhandler() { @@ -615,6 +536,8 @@ base_do_compile() { fi } + +addtask stage after do_compile base_do_stage () { : } @@ -759,7 +682,10 @@ python () { return } -EXPORT_FUNCTIONS do_clean do_mrproper do_fetch do_unpack do_configure do_compile do_install do_package do_patch do_populate_pkgs do_stage +# Patch handling +inherit patch + +EXPORT_FUNCTIONS do_clean do_mrproper do_fetch do_unpack do_configure do_compile do_install do_package do_populate_pkgs do_stage MIRRORS[func] = "0" MIRRORS () { diff --git a/classes/patch.bbclass b/classes/patch.bbclass new file mode 100644 index 0000000000..ea0182484e --- /dev/null +++ b/classes/patch.bbclass @@ -0,0 +1,476 @@ +# Copyright (C) 2006 OpenedHand LTD + +def patch_init(d): + import os, sys + + def md5sum(fname): + import md5, sys + + f = file(fname, 'rb') + m = md5.new() + while True: + d = f.read(8096) + if not d: + break + m.update(d) + f.close() + return m.hexdigest() + + class CmdError(Exception): + def __init__(self, exitstatus, output): + self.status = exitstatus + self.output = output + + def __str__(self): + return "Command Error: exit status: %d Output:\n%s" % (self.status, self.output) + + class NotFoundError(Exception): + def __init__(self, path): + self.path = path + def __str__(self): + return "Error: %s not found." % self.path + + def runcmd(args, dir = None): + import commands + + if dir: + olddir = os.path.abspath(os.curdir) + if not os.path.exists(dir): + raise NotFoundError(dir) + os.chdir(dir) + # print("cwd: %s -> %s" % (olddir, self.dir)) + + try: + args = [ commands.mkarg(str(arg)) for arg in args ] + cmd = " ".join(args) + # print("cmd: %s" % cmd) + (exitstatus, output) = commands.getstatusoutput(cmd) + if exitstatus != 0: + raise CmdError(exitstatus >> 8, output) + return output + + finally: + if dir: + os.chdir(olddir) + + class PatchError(Exception): + def __init__(self, msg): + self.msg = msg + + def __str__(self): + return "Patch Error: %s" % self.msg + + import bb, bb.data, bb.fetch + + class PatchSet(object): + defaults = { + "strippath": 1 + } + + def __init__(self, dir, d): + self.dir = dir + self.d = d + self.patches = [] + self._current = None + + def current(self): + return self._current + + def Clean(self): + """ + Clean out the patch set. Generally includes unapplying all + patches and wiping out all associated metadata. + """ + raise NotImplementedError() + + def Import(self, patch, force): + if not patch.get("file"): + if not patch.get("remote"): + raise PatchError("Patch file must be specified in patch import.") + else: + patch["file"] = bb.fetch.localpath(patch["remote"], self.d) + + for param in PatchSet.defaults: + if not patch.get(param): + patch[param] = PatchSet.defaults[param] + + if patch.get("remote"): + patch["file"] = bb.data.expand(bb.fetch.localpath(patch["remote"], self.d), self.d) + + patch["filemd5"] = md5sum(patch["file"]) + + def Push(self, force): + raise NotImplementedError() + + def Pop(self, force): + raise NotImplementedError() + + def Refresh(self, remote = None, all = None): + raise NotImplementedError() + + + class PatchTree(PatchSet): + def __init__(self, dir, d): + PatchSet.__init__(self, dir, d) + + def Import(self, patch, force = None): + """""" + PatchSet.Import(self, patch, force) + + self.patches.insert(self._current or 0, patch) + + def _applypatch(self, patch, force = None, reverse = None): + shellcmd = ["patch", "<", patch['file'], "-p", patch['strippath']] + if reverse: + shellcmd.append('-R') + shellcmd.append('--dry-run') + + try: + output = runcmd(["sh", "-c", " ".join(shellcmd)], self.dir) + except CmdError: + if force: + shellcmd.pop(len(shellcmd) - 1) + output = runcmd(["sh", "-c", " ".join(shellcmd)], self.dir) + else: + import sys + raise sys.exc_value + + return output + + def Push(self, force = None, all = None): + if all: + for i in self.patches: + if self._current is not None: + self._current = self._current + 1 + else: + self._current = 0 + self._applypatch(i, force) + else: + if self._current is not None: + self._current = self._current + 1 + else: + self._current = 0 + self._applypatch(self.patches[self._current], force) + + + def Pop(self, force = None, all = None): + if all: + for i in self.patches: + self._applypatch(i, force, True) + else: + self._applypatch(self.patches[self._current], force, True) + + def Clean(self): + """""" + + class QuiltTree(PatchSet): + def _runcmd(self, args): + runcmd(["quilt"] + args, self.dir) + + def _quiltpatchpath(self, file): + return os.path.join(self.dir, "patches", os.path.basename(file)) + + + def __init__(self, dir, d): + PatchSet.__init__(self, dir, d) + self.initialized = False + + def Clean(self): + try: + self._runcmd(["pop", "-a", "-f"]) + except CmdError: + pass + except NotFoundError: + pass + # runcmd(["rm", "-rf", os.path.join(self.dir, "patches"), os.path.join(self.dir, ".pc")]) + self.initialized = True + + def InitFromDir(self): + # read series -> self.patches + seriespath = os.path.join(self.dir, 'patches', 'series') + if not os.path.exists(self.dir): + raise Exception("Error: %s does not exist." % self.dir) + if os.path.exists(seriespath): + series = file(seriespath, 'r') + for line in series.readlines(): + patch = {} + parts = line.strip().split() + patch["quiltfile"] = self._quiltpatchpath(parts[0]) + patch["quiltfilemd5"] = md5sum(patch["quiltfile"]) + if len(parts) > 1: + patch["strippath"] = parts[1][2:] + self.patches.append(patch) + series.close() + + # determine which patches are applied -> self._current + try: + output = runcmd(["quilt", "applied"], self.dir) + except CmdError: + if sys.exc_value.output.strip() == "No patches applied": + return + else: + raise sys.exc_value + output = [val for val in output.split('\n') if not val.startswith('#')] + for patch in self.patches: + if os.path.basename(patch["quiltfile"]) == output[-1]: + self._current = self.patches.index(patch) + self.initialized = True + + def Import(self, patch, force = None): + if not self.initialized: + self.InitFromDir() + PatchSet.Import(self, patch, force) + + args = ["import", "-p", patch["strippath"]] + if force: + args.append("-f") + args.append(patch["file"]) + + self._runcmd(args) + + patch["quiltfile"] = self._quiltpatchpath(patch["file"]) + patch["quiltfilemd5"] = md5sum(patch["quiltfile"]) + + # TODO: determine if the file being imported: + # 1) is already imported, and is the same + # 2) is already imported, but differs + + self.patches.insert(self._current or 0, patch) + + + def Push(self, force = None, all = None): + # quilt push [-f] + + args = ["push"] + if force: + args.append("-f") + if all: + args.append("-a") + + self._runcmd(args) + + if self._current is not None: + self._current = self._current + 1 + else: + self._current = 0 + + def Pop(self, force = None, all = None): + # quilt pop [-f] + args = ["pop"] + if force: + args.append("-f") + if all: + args.append("-a") + + self._runcmd(args) + + if self._current == 0: + self._current = None + + if self._current is not None: + self._current = self._current - 1 + + def Refresh(self, **kwargs): + if kwargs.get("remote"): + patch = self.patches[kwargs["patch"]] + if not patch: + raise PatchError("No patch found at index %s in patchset." % kwargs["patch"]) + (type, host, path, user, pswd, parm) = bb.decodeurl(patch["remote"]) + if type == "file": + import shutil + if not patch.get("file") and patch.get("remote"): + patch["file"] = bb.fetch.localpath(patch["remote"], self.d) + + shutil.copyfile(patch["quiltfile"], patch["file"]) + else: + raise PatchError("Unable to do a remote refresh of %s, unsupported remote url scheme %s." % (os.path.basename(patch["quiltfile"]), type)) + else: + # quilt refresh + args = ["refresh"] + if kwargs.get("quiltfile"): + args.append(os.path.basename(kwargs["quiltfile"])) + elif kwargs.get("patch"): + args.append(os.path.basename(self.patches[kwargs["patch"]]["quiltfile"])) + self._runcmd(args) + + class Resolver(object): + def __init__(self, patchset): + raise NotImplementedError() + + def Resolve(self): + raise NotImplementedError() + + def Revert(self): + raise NotImplementedError() + + def Finalize(self): + raise NotImplementedError() + + # Patch resolver which relies on the user doing all the work involved in the + # resolution, with the exception of refreshing the remote copy of the patch + # files (the urls). + class UserResolver(Resolver): + def __init__(self, patchset): + self.patchset = patchset + + # Force a push in the patchset, then drop to a shell for the user to + # resolve any rejected hunks + def Resolve(self): + + olddir = os.path.abspath(os.curdir) + os.chdir(self.patchset.dir) + try: + self.patchset.Push(True) + except CmdError, v: + # Patch application failed + if sys.exc_value.output.strip() == "No patches applied": + return + print(sys.exc_value) + print('NOTE: dropping user into a shell, so that patch rejects can be fixed manually.') + + os.system('/bin/sh') + + # Construct a new PatchSet after the user's changes, compare the + # sets, checking patches for modifications, and doing a remote + # refresh on each. + oldpatchset = self.patchset + self.patchset = oldpatchset.__class__(self.patchset.dir, self.patchset.d) + + for patch in self.patchset.patches: + oldpatch = None + for opatch in oldpatchset.patches: + if opatch["quiltfile"] == patch["quiltfile"]: + oldpatch = opatch + + if oldpatch: + patch["remote"] = oldpatch["remote"] + if patch["quiltfile"] == oldpatch["quiltfile"]: + if patch["quiltfilemd5"] != oldpatch["quiltfilemd5"]: + bb.note("Patch %s has changed, updating remote url %s" % (os.path.basename(patch["quiltfile"]), patch["remote"])) + # user change? remote refresh + self.patchset.Refresh(remote=True, patch=self.patchset.patches.index(patch)) + else: + # User did not fix the problem. Abort. + raise PatchError("Patch application failed, and user did not fix and refresh the patch.") + except Exception: + os.chdir(olddir) + raise + os.chdir(olddir) + + # Throw away the changes to the patches in the patchset made by resolve() + def Revert(self): + raise NotImplementedError() + + # Apply the changes to the patches in the patchset made by resolve() + def Finalize(self): + raise NotImplementedError() + + g = globals() + g["PatchSet"] = PatchSet + g["PatchTree"] = PatchTree + g["QuiltTree"] = QuiltTree + g["Resolver"] = Resolver + g["UserResolver"] = UserResolver + g["NotFoundError"] = NotFoundError + g["CmdError"] = CmdError + +addtask patch after do_unpack +do_patch[dirs] = "${WORKDIR}" +python patch_do_patch() { + import re + import bb.fetch + + patch_init(d) + + src_uri = (bb.data.getVar('SRC_URI', d, 1) or '').split() + if not src_uri: + return + + patchsetmap = { + "patch": PatchTree, + "quilt": QuiltTree, + } + + cls = patchsetmap[bb.data.getVar('PATCHTOOL', d, 1) or 'quilt'] + + resolvermap = { + "user": UserResolver, + } + + rcls = resolvermap[bb.data.getVar('PATCHRESOLVE', d, 1) or 'user'] + + s = bb.data.getVar('S', d, 1) + + path = os.getenv('PATH') + os.putenv('PATH', bb.data.getVar('PATH', d, 1)) + patchset = cls(s, d) + patchset.Clean() + + resolver = rcls(patchset) + + workdir = bb.data.getVar('WORKDIR', d, 1) + for url in src_uri: + (type, host, path, user, pswd, parm) = bb.decodeurl(url) + if not "patch" in parm: + continue + + bb.fetch.init([url],d) + url = bb.encodeurl((type, host, path, user, pswd, [])) + local = os.path.join('/', bb.fetch.localpath(url, d)) + + # did it need to be unpacked? + dots = os.path.basename(local).split(".") + if dots[-1] in ['gz', 'bz2', 'Z']: + unpacked = os.path.join(bb.data.getVar('WORKDIR', d),'.'.join(dots[0:-1])) + else: + unpacked = local + unpacked = bb.data.expand(unpacked, d) + + if "pnum" in parm: + pnum = parm["pnum"] + else: + pnum = "1" + + if "pname" in parm: + pname = parm["pname"] + else: + pname = os.path.basename(unpacked) + + if "mindate" in parm: + mindate = parm["mindate"] + else: + mindate = 0 + + if "maxdate" in parm: + maxdate = parm["maxdate"] + else: + maxdate = "20711226" + + pn = bb.data.getVar('PN', d, 1) + srcdate = bb.data.getVar('SRCDATE_%s' % pn, d, 1) + + if not srcdate: + srcdate = bb.data.getVar('SRCDATE', d, 1) + + if srcdate == "now": + srcdate = bb.data.getVar('DATE', d, 1) + + if (maxdate < srcdate) or (mindate > srcdate): + if (maxdate < srcdate): + bb.note("Patch '%s' is outdated" % pname) + + if (mindate > srcdate): + bb.note("Patch '%s' is predated" % pname) + + continue + + bb.note("Applying patch '%s'" % pname) + try: + patchset.Import({"file":unpacked, "remote":url, "strippath": pnum}, True) + except NotFoundError: + import sys + raise bb.build.FuncFailed(str(sys.exc_value)) + resolver.Resolve() +} + +EXPORT_FUNCTIONS do_patch diff --git a/conf/bitbake.conf b/conf/bitbake.conf index 9f94bae3f5..2e3cb0f980 100644 --- a/conf/bitbake.conf +++ b/conf/bitbake.conf @@ -239,6 +239,12 @@ export MAKE = "make" EXTRA_OEMAKE = "-e MAKEFLAGS=" ################################################################## +# Patch handling. +################################################################## +PATCHTOOL = 'quilt' +PATCHRESOLVE = 'user' + +################################################################## # Build flags and options. ################################################################## @@ -336,12 +342,6 @@ SRC_URI = "file://${FILE}" MKTEMPDIRCMD = "mktemp -d -q ${TMPBASE}" MKTEMPCMD = "mktemp -q ${TMPBASE}" -# Program to be used to patch sources, use 'inherit patcher' to overwrite this: - -PATCHCLEANCMD = 'if [ -n "`quilt applied`" ]; then quilt pop -a -R -f || exit 1; fi' -PATCHCMD = "pnum='%s'; name='%s'; patch='%s'; mkdir -p patches ; quilt upgrade >/dev/null 2>&1; quilt import -f -p $pnum -P $name $patch; chmod u+w patches/$name; quilt push" -PATCH_DEPENDS = "quilt-native" - # GNU patch tries to be intellgent about checking out read-only files from # a RCS, which freaks out those special folks with active Perforce clients # the following makes patch ignore RCS: diff --git a/conf/machine/dht-walnut.conf b/conf/machine/dht-walnut.conf new file mode 100644 index 0000000000..68b754f2d3 --- /dev/null +++ b/conf/machine/dht-walnut.conf @@ -0,0 +1,20 @@ +TARGET_ARCH = "powerpc" +IPKG_ARCHS = "all powerpc ${MACHINE}" + +PREFERRED_PROVIDER_virtual/kernel = "linux-dht-walnut" + +#TARGET_FPU = "soft" +TARGET_CPU = "405" +OLDEST_KERNEL = "2.6.9" + +# TARGET_VENDOR = "-oe" + +BOOTSTRAP_EXTRA_DEPENDS = "virtual/kernel pciutils udev" +BOOTSTRAP_EXTRA_RDEPENDS = "kernel pciutils udev" + +udevdir = "/dev" +OLDEST_KERNEL = "2.6.5" +# GLIBC_ADDONS = "nptl" +# GLIBC_EXTRA_OECONF = "--with-tls" + + diff --git a/conf/machine/rb500.conf b/conf/machine/rb500.conf new file mode 100644 index 0000000000..aec2a8f726 --- /dev/null +++ b/conf/machine/rb500.conf @@ -0,0 +1,8 @@ +#@TYPE: Machine +#@NAME: Mikrotik RB500 +#@DESCRIPTION: Machine configuration for the MIPS based Routerboard + +TARGET_ARCH = "mipsel" +TARGET_CC_ARCH = "-Os -mips2" +IPKG_ARCHS = "all mipsel ${MACHINE}" +PREFERRED_PROVIDER_virtual/kernel = "linux-rb500" diff --git a/packages/gawk/gawk-native_3.1.4.bb b/packages/gawk/gawk-native_3.1.4.bb index ba8efd1af7..0a33f71684 100644 --- a/packages/gawk/gawk-native_3.1.4.bb +++ b/packages/gawk/gawk-native_3.1.4.bb @@ -4,9 +4,7 @@ require gawk_${PV}.bb inherit native DEPENDS = "" -PATCH_DEPENDS = "" -PATCHCLEANCMD = "" -PATCHCMD = "num='%s'; name='%s'; file='%s'; patch -p "$num" -i "$file"" +PATCHTOOL = "patch" S = "${WORKDIR}/gawk-${PV}" diff --git a/packages/gobby/gobby_0.4.1.bb b/packages/gobby/gobby_0.4.1.bb new file mode 100644 index 0000000000..8b7ccad199 --- /dev/null +++ b/packages/gobby/gobby_0.4.1.bb @@ -0,0 +1,8 @@ +LICENSE = "LGPL" +HOMEPAGE = "http://darcs.0x539.de/trac/obby/cgi-bin/trac.cgi/wiki/" +MAINTAINER = "Patrick Steiner <patrick.steiner@a1.net>" + +DEPENDS = "net6 gtkmm obby gtksourceview libxml++" +inherit autotools pkgconfig + +SRC_URI = "http://releases.0x539.de/gobby/gobby-${PV}.tar.gz" diff --git a/packages/gobby/gtksourceview_1.7.2.bb b/packages/gobby/gtksourceview_1.7.2.bb new file mode 100644 index 0000000000..084af175f7 --- /dev/null +++ b/packages/gobby/gtksourceview_1.7.2.bb @@ -0,0 +1,10 @@ +LICENSE = "GPL" +MAINTAINER = "Patrick Steiner <patrick.steiner@a1.net>" + +DEPENDS = "gtk+ libgnomeprint" + +inherit gnome pkgconfig + +do_stage() { +autotools_stage_all +} diff --git a/packages/gobby/libxml++_2.14.0.bb b/packages/gobby/libxml++_2.14.0.bb new file mode 100644 index 0000000000..6eb296c8c2 --- /dev/null +++ b/packages/gobby/libxml++_2.14.0.bb @@ -0,0 +1,10 @@ +LICENSE = "GPL" +MAINTAINER = "Patrick Steiner <patrick.steiner@a1.net>" + +DEPENDS = "gtk+" + +inherit gnome pkgconfig + +do_stage() { +autotools_stage_all +} diff --git a/packages/gobby/net6_1.2.2.bb b/packages/gobby/net6_1.2.2.bb index daa7e9b5b9..f9154fcca9 100644 --- a/packages/gobby/net6_1.2.2.bb +++ b/packages/gobby/net6_1.2.2.bb @@ -7,4 +7,10 @@ inherit autotools pkgconfig SRC_URI = "http://releases.0x539.de/${PN}/${P}.tar.gz" - +do_stage() { + autotools_stage_all + install -d ${STAGING_LIBDIR} + install -d ${STAGING_INCDIR}/net6 + install -m 644 inc/*.hpp ${STAGING_INCDIR}/net6 + install -m 755 .libs/*so* ${STAGING_LIBDIR}/ +} diff --git a/packages/gobby/net6_1.3.1.bb b/packages/gobby/net6_1.3.1.bb new file mode 100644 index 0000000000..3bcf8811d9 --- /dev/null +++ b/packages/gobby/net6_1.3.1.bb @@ -0,0 +1,15 @@ +LICENSE = "LGPL" +HOMEPAGE = "http://darcs.0x539.de/trac/obby/cgi-bin/trac.cgi/wiki/" +MAINTAINER = "Koen Kooi <koen@handhelds.org>" + +DEPENDS = "libsigc++-2.0 gnutls" +inherit autotools pkgconfig + +SRC_URI = "http://releases.0x539.de/${PN}/${P}.tar.gz" + +do_stage() { + install -d ${STAGING_LIBDIR} + install -d ${STAGING_INCDIR}/net6 + install -m 644 inc/*.hpp ${STAGING_INCDIR}/net6 + install -m 755 .libs/*so* ${STAGING_LIBDIR}/ +} diff --git a/packages/gobby/obby_0.4.1.bb b/packages/gobby/obby_0.4.1.bb new file mode 100644 index 0000000000..adf3a72571 --- /dev/null +++ b/packages/gobby/obby_0.4.1.bb @@ -0,0 +1,19 @@ +LICENSE = "LGPL" +HOMEPAGE = "http://darcs.0x539.de/trac/obby/cgi-bin/trac.cgi/wiki/" +MAINTAINER = "Patrick Steiner <patrick.steiner@a1.net>" + +DEPENDS = "net6 gtkmm gmp" +inherit autotools pkgconfig + +SRC_URI = "http://releases.0x539.de/obby/obby-${PV}.tar.gz" + +EXTRA_OECONF += " --with-libgmp-prefix=${STAGING_LIBDIR}" + +do_stage() { + install -d ${STAGING_LIBDIR} + install -d ${STAGING_INCDIR}/obby + install -d ${STAGING_INCDIR}/obby/serialise + install -m 644 inc/*.hpp ${STAGING_INCDIR}/obby + install -m 644 inc/serialise/*.hpp ${STAGING_INCDIR}/obby/serialise + install -m 755 src/.libs/*so* ${STAGING_LIBDIR}/ +} diff --git a/packages/patcher/patcher-native_20040913.bb b/packages/patcher/patcher-native_20040913.bb index 833ca21a53..7da61e91c2 100644 --- a/packages/patcher/patcher-native_20040913.bb +++ b/packages/patcher/patcher-native_20040913.bb @@ -14,9 +14,7 @@ SRC_URI = "http://www.holgerschurig.de/files/linux/patcher-${PV}.tar.bz2" SRC_URI_append_build-freebsd = " file://freebsd_gpatch.patch;patch=1 " S = "${WORKDIR}/patcher" -PATCHCLEANCMD = "" -PATCHCMD = "num='%s'; name='%s'; file='%s'; patch -p "$num" -i "$file"" -PATCHCMD_build-freebsd = "num='%s'; name='%s'; file='%s'; gpatch -p "$num" -i "$file"" +PATCHTOOL = "patch" do_stage() { install -m 0755 patcher.py ${STAGING_BINDIR}/patcher diff --git a/packages/quilt/quilt-native.inc b/packages/quilt/quilt-native.inc index d19d864904..c8bbdc8e07 100644 --- a/packages/quilt/quilt-native.inc +++ b/packages/quilt/quilt-native.inc @@ -8,8 +8,7 @@ INHIBIT_AUTOTOOLS_DEPS = "1" inherit autotools native -PATCHCLEANCMD = "" -PATCHCMD = "num='%s'; name='%s'; file='%s'; patch -p "$num" -i "$file"" +PATCHTOOL = "patch" EXTRA_OECONF = "--disable-nls" do_configure () { |