Browse Source

Convert to use direct access to the data store (instead of bb.data.*Var*())

This is the result of running the following over the metadata:

sed \
-e 's:bb.data.\(setVar([^,()]*,[^,()]*\), *\([^ )]*\) *):\2.\1):g' \
-e 's:bb.data.\(setVarFlag([^,()]*,[^,()]*,[^,()]*\), *\([^) ]*\) *):\2.\1):g' \
-e 's:bb.data.\(getVar([^,()]*\), *\([^(), ]*\) *,\([^)]*\)):\2.\1,\3):g' \
-e 's:bb.data.\(getVarFlag([^,()]*,[^,()]*\), *\([^(), ]*\) *,\([^)]*\)):\2.\1,\3):g' \
-e 's:bb.data.\(getVarFlag([^,()]*,[^,()]*\), *\([^() ]*\) *):\2.\1):g' \
-e 's:bb.data.\(getVar([^,()]*\), *\([^) ]*\) *):\2.\1):g' \
-i `grep -ril bb.data *`

Signed-off-by: Richard Purdie <richard.purdie@linuxfoundation.org>
Richard Purdie 12 years ago
parent
commit
b22831fd63
100 changed files with 976 additions and 976 deletions
  1. 1 1
      meta-demoapps/recipes-gnome/abiword/abiword-2.5.inc
  2. 1 1
      meta-demoapps/recipes-gnome/abiword/abiword.inc
  3. 1 1
      meta-demoapps/recipes-support/poppler/poppler-fpu.inc
  4. 3 3
      meta/classes/autotools.bbclass
  5. 57 57
      meta/classes/base.bbclass
  6. 15 15
      meta/classes/bugzilla.bbclass
  7. 18 18
      meta/classes/buildstats.bbclass
  8. 1 1
      meta/classes/cpan-base.bbclass
  9. 2 2
      meta/classes/cpan_build.bbclass
  10. 1 1
      meta/classes/cross-canadian.bbclass
  11. 9 9
      meta/classes/debian.bbclass
  12. 81 81
      meta/classes/distrodata.bbclass
  13. 1 1
      meta/classes/distutils-base.bbclass
  14. 1 1
      meta/classes/distutils-native-base.bbclass
  15. 8 8
      meta/classes/gconf.bbclass
  16. 12 12
      meta/classes/gtk-icon-cache.bbclass
  17. 8 8
      meta/classes/icecc.bbclass
  18. 2 2
      meta/classes/image-swab.bbclass
  19. 10 10
      meta/classes/image.bbclass
  20. 2 2
      meta/classes/image_types.bbclass
  21. 28 28
      meta/classes/imagetest-qemu.bbclass
  22. 39 39
      meta/classes/insane.bbclass
  23. 3 3
      meta/classes/kernel-arch.bbclass
  24. 1 1
      meta/classes/kernel-yocto.bbclass
  25. 34 34
      meta/classes/kernel.bbclass
  26. 5 5
      meta/classes/libc-common.bbclass
  27. 33 33
      meta/classes/libc-package.bbclass
  28. 9 9
      meta/classes/license.bbclass
  29. 1 1
      meta/classes/metadata_scm.bbclass
  30. 6 6
      meta/classes/native.bbclass
  31. 7 7
      meta/classes/nativesdk.bbclass
  32. 116 116
      meta/classes/package.bbclass
  33. 41 41
      meta/classes/package_deb.bbclass
  34. 41 41
      meta/classes/package_ipk.bbclass
  35. 68 68
      meta/classes/package_rpm.bbclass
  36. 19 19
      meta/classes/package_tar.bbclass
  37. 4 4
      meta/classes/packagedata.bbclass
  38. 4 4
      meta/classes/packagehistory.bbclass
  39. 14 14
      meta/classes/patch.bbclass
  40. 1 1
      meta/classes/pkg_distribute.bbclass
  41. 6 6
      meta/classes/pkg_metainfo.bbclass
  42. 2 2
      meta/classes/populate_sdk_deb.bbclass
  43. 1 1
      meta/classes/populate_sdk_rpm.bbclass
  44. 1 1
      meta/classes/qemu.bbclass
  45. 1 1
      meta/classes/qt4e.bbclass
  46. 1 1
      meta/classes/qt4x11.bbclass
  47. 1 1
      meta/classes/relocatable.bbclass
  48. 5 5
      meta/classes/rootfs_ipk.bbclass
  49. 7 7
      meta/classes/rootfs_rpm.bbclass
  50. 5 5
      meta/classes/sanity.bbclass
  51. 1 1
      meta/classes/siteconfig.bbclass
  52. 1 1
      meta/classes/siteinfo.bbclass
  53. 6 6
      meta/classes/sourcepkg.bbclass
  54. 8 8
      meta/classes/src_distribute.bbclass
  55. 47 47
      meta/classes/sstate.bbclass
  56. 3 3
      meta/classes/staging.bbclass
  57. 17 17
      meta/classes/syslinux.bbclass
  58. 1 1
      meta/classes/task.bbclass
  59. 3 3
      meta/classes/toolchain-scripts.bbclass
  60. 19 19
      meta/classes/update-alternatives.bbclass
  61. 18 18
      meta/classes/update-rc.d.bbclass
  62. 4 4
      meta/classes/useradd.bbclass
  63. 4 4
      meta/classes/utility-tasks.bbclass
  64. 2 2
      meta/classes/utils.bbclass
  65. 18 18
      meta/conf/bitbake.conf
  66. 1 1
      meta/conf/distro/defaultsetup.conf
  67. 1 1
      meta/conf/machine/include/arm/feature-arm-thumb.inc
  68. 4 4
      meta/conf/machine/include/tune-thumb.inc
  69. 11 11
      meta/lib/oe/distro_check.py
  70. 2 2
      meta/lib/oe/patch.py
  71. 4 4
      meta/lib/oe/utils.py
  72. 1 1
      meta/recipes-bsp/grub/grub_0.97.bb
  73. 1 1
      meta/recipes-bsp/grub/grub_1.99.bb
  74. 1 1
      meta/recipes-bsp/uboot/u-boot_2011.03.bb
  75. 1 1
      meta/recipes-bsp/uboot/u-boot_2011.06.bb
  76. 1 1
      meta/recipes-bsp/x-load/x-load_git.bb
  77. 1 1
      meta/recipes-connectivity/connman/connman.inc
  78. 3 3
      meta/recipes-core/busybox/busybox.inc
  79. 3 3
      meta/recipes-core/eglibc/eglibc-options.inc
  80. 2 2
      meta/recipes-core/eglibc/eglibc-package.inc
  81. 2 2
      meta/recipes-core/eglibc/eglibc_2.13.bb
  82. 2 2
      meta/recipes-core/eglibc/eglibc_2.14.bb
  83. 1 1
      meta/recipes-core/glib-2.0/glib-2.0_2.30.0.bb
  84. 2 2
      meta/recipes-core/libxml/libxml2.inc
  85. 5 5
      meta/recipes-core/tasks/task-base.bb
  86. 3 3
      meta/recipes-core/tasks/task-core-sdk.bb
  87. 8 8
      meta/recipes-core/uclibc/uclibc-config.inc
  88. 5 5
      meta/recipes-core/uclibc/uclibc.inc
  89. 2 2
      meta/recipes-devtools/apt/apt-native.inc
  90. 4 4
      meta/recipes-devtools/apt/apt-package.inc
  91. 1 1
      meta/recipes-devtools/automake/automake.inc
  92. 1 1
      meta/recipes-devtools/cmake/cmake.inc
  93. 3 3
      meta/recipes-devtools/cmake/cmake_2.8.5.bb
  94. 4 4
      meta/recipes-devtools/gcc/gcc-common.inc
  95. 1 1
      meta/recipes-devtools/gcc/gcc-configure-common.inc
  96. 1 1
      meta/recipes-devtools/gnu-config/gnu-config_20080123.bb
  97. 1 1
      meta/recipes-devtools/intltool/intltool.inc
  98. 1 1
      meta/recipes-devtools/opkg/opkg.inc
  99. 1 1
      meta/recipes-devtools/perl/perl_5.14.2.bb
  100. 1 1
      meta/recipes-devtools/python/python-pygobject_2.27.91.bb

+ 1 - 1
meta-demoapps/recipes-gnome/abiword/abiword-2.5.inc

@@ -9,7 +9,7 @@ RDEPENDS_${PN}    = "glibc-gconv-ibm850 glibc-gconv-cp1252 \
 SRC_URI = "http://www.abiword.org/downloads/abiword/${PV}/source/abiword-${PV}.tar.gz"
 
 #want 2.x from 2.x.y for the installation directory
-SHRT_VER = "${@bb.data.getVar('PV',d,1).split('.')[0]}.${@bb.data.getVar('PV',d,1).split('.')[1]}"
+SHRT_VER = "${@d.getVar('PV',1).split('.')[0]}.${@d.getVar('PV',1).split('.')[1]}"
 
 FILES_${PN} += " \
 		${datadir}/icons/* \

+ 1 - 1
meta-demoapps/recipes-gnome/abiword/abiword.inc

@@ -17,7 +17,7 @@ SVNURI = "svn://svn.abisource.com/abiword/trunk;module=abiword;proto=http"
 SVNSRC = "${WORKDIR}/abi"
 
 #want 2.x from 2.x.y for the installation directory
-SHRT_VER = "${@bb.data.getVar('PV',d,1).split('.')[0]}.${@bb.data.getVar('PV',d,1).split('.')[1]}"
+SHRT_VER = "${@d.getVar('PV',1).split('.')[0]}.${@d.getVar('PV',1).split('.')[1]}"
 
 FILES_${PN} += " \
 		${datadir}/icons/* \

+ 1 - 1
meta-demoapps/recipes-support/poppler/poppler-fpu.inc

@@ -1,6 +1,6 @@
 
 def get_poppler_fpu_setting(bb, d):
-	if bb.data.getVar('TARGET_FPU', d, 1) in [ 'soft' ]:
+	if d.getVar('TARGET_FPU', 1) in [ 'soft' ]:
 		return "--enable-fixedpoint"
 	return ""
 

+ 3 - 3
meta/classes/autotools.bbclass

@@ -1,8 +1,8 @@
 def autotools_dep_prepend(d):
-	if bb.data.getVar('INHIBIT_AUTOTOOLS_DEPS', d, 1):
+	if d.getVar('INHIBIT_AUTOTOOLS_DEPS', 1):
 		return ''
 
-	pn = bb.data.getVar('PN', d, 1)
+	pn = d.getVar('PN', 1)
 	deps = ''
 
 	if pn in ['autoconf-native', 'automake-native', 'help2man-native']:
@@ -13,7 +13,7 @@ def autotools_dep_prepend(d):
 		deps += 'libtool-native '
 		if not bb.data.inherits_class('native', d) \
                         and not bb.data.inherits_class('cross', d) \
-                        and not bb.data.getVar('INHIBIT_DEFAULT_DEPS', d, 1):
+                        and not d.getVar('INHIBIT_DEFAULT_DEPS', 1):
                     deps += 'libtool-cross '
 
 	return deps + 'gnu-config-native '

+ 57 - 57
meta/classes/base.bbclass

@@ -60,9 +60,9 @@ def base_dep_prepend(d):
 	# INHIBIT_DEFAULT_DEPS doesn't apply to the patch command.  Whether or  not
 	# we need that built is the responsibility of the patch function / class, not
 	# the application.
-	if not bb.data.getVar('INHIBIT_DEFAULT_DEPS', d):
-		if (bb.data.getVar('HOST_SYS', d, 1) !=
-	     	    bb.data.getVar('BUILD_SYS', d, 1)):
+	if not d.getVar('INHIBIT_DEFAULT_DEPS'):
+		if (d.getVar('HOST_SYS', 1) !=
+	     	    d.getVar('BUILD_SYS', 1)):
 			deps += " virtual/${TARGET_PREFIX}gcc virtual/${TARGET_PREFIX}compilerlibs virtual/libc "
 	return deps
 
@@ -73,13 +73,13 @@ DEPENDS_prepend="${BASEDEPENDS} "
 FILESPATH = "${@base_set_filespath([ "${FILE_DIRNAME}/${PF}", "${FILE_DIRNAME}/${P}", "${FILE_DIRNAME}/${PN}", "${FILE_DIRNAME}/${BP}", "${FILE_DIRNAME}/${BPN}", "${FILE_DIRNAME}/files", "${FILE_DIRNAME}" ], d)}"
 # THISDIR only works properly with imediate expansion as it has to run
 # in the context of the location its used (:=)
-THISDIR = "${@os.path.dirname(bb.data.getVar('FILE', d, True))}"
+THISDIR = "${@os.path.dirname(d.getVar('FILE', True))}"
 
 addtask fetch
 do_fetch[dirs] = "${DL_DIR}"
 python base_do_fetch() {
 
-	src_uri = (bb.data.getVar('SRC_URI', d, True) or "").split()
+	src_uri = (d.getVar('SRC_URI', True) or "").split()
 	if len(src_uri) == 0:
 		return
 
@@ -96,14 +96,14 @@ python base_do_fetch() {
 addtask unpack after do_fetch
 do_unpack[dirs] = "${WORKDIR}"
 python base_do_unpack() {
-	src_uri = (bb.data.getVar('SRC_URI', d, True) or "").split()
+	src_uri = (d.getVar('SRC_URI', True) or "").split()
 	if len(src_uri) == 0:
 		return
 
 	localdata = bb.data.createCopy(d)
 	bb.data.update_data(localdata)
 
-	rootdir = bb.data.getVar('WORKDIR', localdata, True)
+	rootdir = localdata.getVar('WORKDIR', True)
 
         try:
             fetcher = bb.fetch2.Fetch(src_uri, localdata)
@@ -118,7 +118,7 @@ def generate_git_config(e):
         from bb import data
 
         if data.getVar('GIT_CORE_CONFIG', e.data, True):
-                gitconfig_path = bb.data.getVar('GIT_CONFIG', e.data, True)
+                gitconfig_path = e.data.getVar('GIT_CONFIG', True)
                 proxy_command = "    gitproxy = %s\n" % data.getVar('GIT_PROXY_COMMAND', e.data, True)
 
                 bb.mkdirhier(bb.data.expand("${STAGING_DIR_NATIVE}/usr/etc/", e.data))
@@ -207,11 +207,11 @@ python base_eventhandler() {
 	name = getName(e)
 
 	if name.startswith("BuildStarted"):
-		bb.data.setVar( 'BB_VERSION', bb.__version__, e.data )
+		e.data.setVar( 'BB_VERSION', bb.__version__)
 		statusvars = ['BB_VERSION', 'TARGET_ARCH', 'TARGET_OS', 'MACHINE', 'DISTRO', 'DISTRO_VERSION','TUNE_FEATURES', 'TARGET_FPU']
-		statuslines = ["%-17s = \"%s\"" % (i, bb.data.getVar(i, e.data, 1) or '') for i in statusvars]
+		statuslines = ["%-17s = \"%s\"" % (i, e.data.getVar(i, 1) or '') for i in statusvars]
 
-		layers = (bb.data.getVar("BBLAYERS", e.data, 1) or "").split()
+		layers = (e.data.getVar("BBLAYERS", 1) or "").split()
 		layers_branch_rev = ["%-17s = \"%s:%s\"" % (os.path.basename(i), \
 			base_get_metadata_git_branch(i, None).strip(), \
 			base_get_metadata_git_revision(i, None)) \
@@ -237,7 +237,7 @@ python base_eventhandler() {
 		needed_vars = [ "TARGET_ARCH", "TARGET_OS" ]
 		pesteruser = []
 		for v in needed_vars:
-			val = bb.data.getVar(v, e.data, 1)
+			val = e.data.getVar(v, 1)
 			if not val or val == 'INVALID':
 				pesteruser.append(v)
 		if pesteruser:
@@ -330,23 +330,23 @@ python () {
         appendVar('EXTRA_OECONF', extraconf)
 
     # If PRINC is set, try and increase the PR value by the amount specified
-    princ = bb.data.getVar('PRINC', d, True)
+    princ = d.getVar('PRINC', True)
     if princ:
-        pr = bb.data.getVar('PR', d, True)
+        pr = d.getVar('PR', True)
         pr_prefix = re.search("\D+",pr)
         prval = re.search("\d+",pr)
         if pr_prefix is None or prval is None:
             bb.error("Unable to analyse format of PR variable: %s" % pr)
         nval = int(prval.group(0)) + int(princ)
         pr = pr_prefix.group(0) + str(nval) + pr[prval.end():]
-        bb.data.setVar('PR', pr, d)
+        d.setVar('PR', pr)
 
-    pn = bb.data.getVar('PN', d, 1)
-    license = bb.data.getVar('LICENSE', d, True)
+    pn = d.getVar('PN', 1)
+    license = d.getVar('LICENSE', True)
     if license == "INVALID":
         bb.fatal('This recipe does not have the LICENSE field set (%s)' % pn)
 
-    commercial_license = " %s " % bb.data.getVar('COMMERCIAL_LICENSE', d, 1)
+    commercial_license = " %s " % d.getVar('COMMERCIAL_LICENSE', 1)
     import re
     pnr = "[ \t]%s[ \t]" % pn.replace('+', "\+")
     if commercial_license and re.search(pnr, commercial_license):
@@ -356,86 +356,86 @@ python () {
     # If we're building a target package we need to use fakeroot (pseudo)
     # in order to capture permissions, owners, groups and special files
     if not bb.data.inherits_class('native', d) and not bb.data.inherits_class('cross', d):
-        bb.data.setVarFlag('do_configure', 'umask', 022, d)
-        bb.data.setVarFlag('do_compile', 'umask', 022, d)
-        deps = (bb.data.getVarFlag('do_install', 'depends', d) or "").split()
+        d.setVarFlag('do_configure', 'umask', 022)
+        d.setVarFlag('do_compile', 'umask', 022)
+        deps = (d.getVarFlag('do_install', 'depends') or "").split()
         deps.append('virtual/fakeroot-native:do_populate_sysroot')
         bb.data.setVarFlag('do_install', 'depends', " ".join(deps),d)
-        bb.data.setVarFlag('do_install', 'fakeroot', 1, d)
-        bb.data.setVarFlag('do_install', 'umask', 022, d)
-        deps = (bb.data.getVarFlag('do_package', 'depends', d) or "").split()
+        d.setVarFlag('do_install', 'fakeroot', 1)
+        d.setVarFlag('do_install', 'umask', 022)
+        deps = (d.getVarFlag('do_package', 'depends') or "").split()
         deps.append('virtual/fakeroot-native:do_populate_sysroot')
         bb.data.setVarFlag('do_package', 'depends', " ".join(deps),d)
-        bb.data.setVarFlag('do_package', 'fakeroot', 1, d)
-        bb.data.setVarFlag('do_package', 'umask', 022, d)
-        bb.data.setVarFlag('do_package_setscene', 'fakeroot', 1, d)
-    source_mirror_fetch = bb.data.getVar('SOURCE_MIRROR_FETCH', d, 0)
+        d.setVarFlag('do_package', 'fakeroot', 1)
+        d.setVarFlag('do_package', 'umask', 022)
+        d.setVarFlag('do_package_setscene', 'fakeroot', 1)
+    source_mirror_fetch = d.getVar('SOURCE_MIRROR_FETCH', 0)
     if not source_mirror_fetch:
-        need_host = bb.data.getVar('COMPATIBLE_HOST', d, 1)
+        need_host = d.getVar('COMPATIBLE_HOST', 1)
         if need_host:
             import re
-            this_host = bb.data.getVar('HOST_SYS', d, 1)
+            this_host = d.getVar('HOST_SYS', 1)
             if not re.match(need_host, this_host):
                 raise bb.parse.SkipPackage("incompatible with host %s" % this_host)
 
-        need_machine = bb.data.getVar('COMPATIBLE_MACHINE', d, 1)
+        need_machine = d.getVar('COMPATIBLE_MACHINE', 1)
         if need_machine:
             import re
-            this_machine = bb.data.getVar('MACHINE', d, 1)
+            this_machine = d.getVar('MACHINE', 1)
             if this_machine and not re.match(need_machine, this_machine):
-                this_soc_family = bb.data.getVar('SOC_FAMILY', d, 1)
+                this_soc_family = d.getVar('SOC_FAMILY', 1)
                 if (this_soc_family and not re.match(need_machine, this_soc_family)) or not this_soc_family:
                     raise bb.parse.SkipPackage("incompatible with machine %s" % this_machine)
 
 
-        dont_want_license = bb.data.getVar('INCOMPATIBLE_LICENSE', d, 1)
+        dont_want_license = d.getVar('INCOMPATIBLE_LICENSE', 1)
         if dont_want_license and not pn.endswith("-native") and not pn.endswith("-cross") and not pn.endswith("-cross-initial") and not pn.endswith("-cross-intermediate") and not pn.endswith("-crosssdk-intermediate") and not pn.endswith("-crosssdk") and not pn.endswith("-crosssdk-initial"):
-            hosttools_whitelist = (bb.data.getVar('HOSTTOOLS_WHITELIST_%s' % dont_want_license, d, 1) or "").split()
-            lgplv2_whitelist = (bb.data.getVar('LGPLv2_WHITELIST_%s' % dont_want_license, d, 1) or "").split()
-            dont_want_whitelist = (bb.data.getVar('WHITELIST_%s' % dont_want_license, d, 1) or "").split()
+            hosttools_whitelist = (d.getVar('HOSTTOOLS_WHITELIST_%s' % dont_want_license, 1) or "").split()
+            lgplv2_whitelist = (d.getVar('LGPLv2_WHITELIST_%s' % dont_want_license, 1) or "").split()
+            dont_want_whitelist = (d.getVar('WHITELIST_%s' % dont_want_license, 1) or "").split()
             if pn not in hosttools_whitelist and pn not in lgplv2_whitelist and pn not in dont_want_whitelist:
 
                 import re
-                this_license = bb.data.getVar('LICENSE', d, 1)
+                this_license = d.getVar('LICENSE', 1)
                 if this_license and re.search(dont_want_license, this_license):
                     bb.note("SKIPPING %s because it's %s" % (pn, this_license))
                     raise bb.parse.SkipPackage("incompatible with license %s" % this_license)
 
     # Git packages should DEPEND on git-native
-    srcuri = bb.data.getVar('SRC_URI', d, 1)
+    srcuri = d.getVar('SRC_URI', 1)
     if "git://" in srcuri:
-        depends = bb.data.getVarFlag('do_fetch', 'depends', d) or ""
+        depends = d.getVarFlag('do_fetch', 'depends') or ""
         depends = depends + " git-native:do_populate_sysroot"
-        bb.data.setVarFlag('do_fetch', 'depends', depends, d)
+        d.setVarFlag('do_fetch', 'depends', depends)
 
     # Mercurial packages should DEPEND on mercurial-native
     elif "hg://" in srcuri:
-        depends = bb.data.getVarFlag('do_fetch', 'depends', d) or ""
+        depends = d.getVarFlag('do_fetch', 'depends') or ""
         depends = depends + " mercurial-native:do_populate_sysroot"
-        bb.data.setVarFlag('do_fetch', 'depends', depends, d)
+        d.setVarFlag('do_fetch', 'depends', depends)
 
     # OSC packages should DEPEND on osc-native
     elif "osc://" in srcuri:
-        depends = bb.data.getVarFlag('do_fetch', 'depends', d) or ""
+        depends = d.getVarFlag('do_fetch', 'depends') or ""
         depends = depends + " osc-native:do_populate_sysroot"
-        bb.data.setVarFlag('do_fetch', 'depends', depends, d)
+        d.setVarFlag('do_fetch', 'depends', depends)
 
     # *.xz should depends on xz-native for unpacking
     # Not endswith because of "*.patch.xz;patch=1". Need bb.decodeurl in future
     if '.xz' in srcuri:
-        depends = bb.data.getVarFlag('do_unpack', 'depends', d) or ""
+        depends = d.getVarFlag('do_unpack', 'depends') or ""
         depends = depends + " xz-native:do_populate_sysroot"
-        bb.data.setVarFlag('do_unpack', 'depends', depends, d)
+        d.setVarFlag('do_unpack', 'depends', depends)
 
     # unzip-native should already be staged before unpacking ZIP recipes
     if ".zip" in srcuri:
-        depends = bb.data.getVarFlag('do_unpack', 'depends', d) or ""
+        depends = d.getVarFlag('do_unpack', 'depends') or ""
         depends = depends + " unzip-native:do_populate_sysroot"
-        bb.data.setVarFlag('do_unpack', 'depends', depends, d)
+        d.setVarFlag('do_unpack', 'depends', depends)
 
     # 'multimachine' handling
-    mach_arch = bb.data.getVar('MACHINE_ARCH', d, 1)
-    pkg_arch = bb.data.getVar('PACKAGE_ARCH', d, 1)
+    mach_arch = d.getVar('MACHINE_ARCH', 1)
+    pkg_arch = d.getVar('PACKAGE_ARCH', 1)
 
     if (pkg_arch == mach_arch):
         # Already machine specific - nothing further to do
@@ -445,7 +445,7 @@ python () {
     # We always try to scan SRC_URI for urls with machine overrides
     # unless the package sets SRC_URI_OVERRIDES_PACKAGE_ARCH=0
     #
-    override = bb.data.getVar('SRC_URI_OVERRIDES_PACKAGE_ARCH', d, 1)
+    override = d.getVar('SRC_URI_OVERRIDES_PACKAGE_ARCH', 1)
     if override != '0':
         paths = []
         for p in [ "${PF}", "${P}", "${PN}", "files", "" ]:
@@ -461,18 +461,18 @@ python () {
                 for mp in paths:
                     if local.startswith(mp):
                         #bb.note("overriding PACKAGE_ARCH from %s to %s" % (pkg_arch, mach_arch))
-                        bb.data.setVar('PACKAGE_ARCH', "${MACHINE_ARCH}", d)
+                        d.setVar('PACKAGE_ARCH', "${MACHINE_ARCH}")
                         return
 
-    packages = bb.data.getVar('PACKAGES', d, 1).split()
+    packages = d.getVar('PACKAGES', 1).split()
     for pkg in packages:
-        pkgarch = bb.data.getVar("PACKAGE_ARCH_%s" % pkg, d, 1)
+        pkgarch = d.getVar("PACKAGE_ARCH_%s" % pkg, 1)
 
         # We could look for != PACKAGE_ARCH here but how to choose
         # if multiple differences are present?
         # Look through PACKAGE_ARCHS for the priority order?
         if pkgarch and pkgarch == mach_arch:
-            bb.data.setVar('PACKAGE_ARCH', "${MACHINE_ARCH}", d)
+            d.setVar('PACKAGE_ARCH', "${MACHINE_ARCH}")
             bb.warn("Recipe %s is marked as only being architecture specific but seems to have machine specific packages?! The recipe may as well mark itself as machine specific directly." % d.getVar("PN", True))
 }
 
@@ -483,7 +483,7 @@ python do_cleansstate() {
 
 addtask cleanall after do_cleansstate
 python do_cleanall() {
-        src_uri = (bb.data.getVar('SRC_URI', d, True) or "").split()
+        src_uri = (d.getVar('SRC_URI', True) or "").split()
         if len(src_uri) == 0:
             return
 

+ 15 - 15
meta/classes/bugzilla.bbclass

@@ -109,14 +109,14 @@ python bugzilla_eventhandler() {
        return
 
     if name == "TaskFailed":
-        xmlrpc  = bb.data.getVar("BUGZILLA_XMLRPC", data, True)
-        user    = bb.data.getVar("BUGZILLA_USER",  data, True)
-        passw   = bb.data.getVar("BUGZILLA_PASS",  data, True)
-        product = bb.data.getVar("BUGZILLA_PRODUCT", data, True)
-        compon  = bb.data.getVar("BUGZILLA_COMPONENT", data, True)
-        version = bb.data.getVar("BUGZILLA_VERSION", data, True)
-
-        proxy   = bb.data.getVar('http_proxy', data, True )
+        xmlrpc  = data.getVar("BUGZILLA_XMLRPC", True)
+        user    = data.getVar("BUGZILLA_USER", True)
+        passw   = data.getVar("BUGZILLA_PASS", True)
+        product = data.getVar("BUGZILLA_PRODUCT", True)
+        compon  = data.getVar("BUGZILLA_COMPONENT", True)
+        version = data.getVar("BUGZILLA_VERSION", True)
+
+        proxy   = data.getVar('http_proxy', True )
         if (proxy):
             import urllib2
             s, u, p, hostport = urllib2._parse_proxy(proxy)
@@ -132,14 +132,14 @@ python bugzilla_eventhandler() {
             'component': compon}
 
         # evil hack to figure out what is going on
-        debug_file = open(os.path.join(bb.data.getVar("TMPDIR", data, True),"..","bugzilla-log"),"a")
+        debug_file = open(os.path.join(data.getVar("TMPDIR", True),"..","bugzilla-log"),"a")
 
         file = None
-        bugname = "%(package)s-%(pv)s-autobuild" % { "package" : bb.data.getVar("PN", data, True),
-                                                           "pv"      : bb.data.getVar("PV", data, True),
+        bugname = "%(package)s-%(pv)s-autobuild" % { "package" : data.getVar("PN", True),
+                                                           "pv"      : data.getVar("PV", True),
                                                            }
-        log_file = glob.glob("%s/log.%s.*" % (bb.data.getVar('T', event.data, True), event.task))
-        text     = "The %s step in %s failed at %s for machine %s" % (e.task, bb.data.getVar("PN", data, True), bb.data.getVar('DATETIME', data, True), bb.data.getVar( 'MACHINE', data, True ) )
+        log_file = glob.glob("%s/log.%s.*" % (event.data.getVar('T', True), event.task))
+        text     = "The %s step in %s failed at %s for machine %s" % (e.task, data.getVar("PN", True), data.getVar('DATETIME', True), data.getVar( 'MACHINE', True ) )
         if len(log_file) != 0:
             print >> debug_file, "Adding log file %s" % log_file[0]
             file = open(log_file[0], 'r')
@@ -167,7 +167,7 @@ python bugzilla_eventhandler() {
 
         if bug_number and log:
             print >> debug_file, "The bug is known as '%s'" % bug_number
-            desc = "Build log for machine %s" % (bb.data.getVar('MACHINE', data, True))
+            desc = "Build log for machine %s" % (data.getVar('MACHINE', True))
             if not bugzilla_create_attachment(debug_file, server, args.copy(), bug_number, text, log_file[0], log, desc):
                  print >> debug_file, "Failed to attach the build log for bug #%s" % bug_number
             else:
@@ -181,6 +181,6 @@ python bugzilla_eventhandler() {
 
         # store bug number for oestats-client
         if bug_number:
-            bb.data.setVar('OESTATS_BUG_NUMBER', bug_number, data)
+            data.setVar('OESTATS_BUG_NUMBER', bug_number)
 }
 

+ 18 - 18
meta/classes/buildstats.bbclass

@@ -21,25 +21,25 @@ def get_cputime():
     return sum(int(field) for field in fields)
 
 def set_bn(e):
-    bn = e.getPkgs()[0] + "-" + bb.data.getVar('MACHINE',e.data, True)
+    bn = e.getPkgs()[0] + "-" + e.data.getVar('MACHINE', True)
     try:
-        os.remove(bb.data.getVar('BNFILE', e.data, True))
+        os.remove(e.data.getVar('BNFILE', True))
     except:
         pass
-    file = open(bb.data.getVar('BNFILE', e.data, True), "w")
-    file.write(os.path.join(bn, bb.data.getVar('BUILDNAME', e.data, True)))
+    file = open(e.data.getVar('BNFILE', True), "w")
+    file.write(os.path.join(bn, e.data.getVar('BUILDNAME', True)))
     file.close()
 
 def get_bn(e):
-    file = open(bb.data.getVar('BNFILE', e.data, True))
+    file = open(e.data.getVar('BNFILE', True))
     bn = file.readline()
     file.close()
     return bn
 
 def set_device(e):
-    tmpdir = bb.data.getVar('TMPDIR', e.data, True)
+    tmpdir = e.data.getVar('TMPDIR', True)
     try:
-        os.remove(bb.data.getVar('DEVFILE', e.data, True))
+        os.remove(e.data.getVar('DEVFILE', True))
     except:
         pass
     ############################################################################
@@ -66,12 +66,12 @@ def set_device(e):
            rdev=line.split()[2]
         else:
            rdev="NoLogicalDevice"
-    file = open(bb.data.getVar('DEVFILE', e.data, True), "w")
+    file = open(e.data.getVar('DEVFILE', True), "w")
     file.write(rdev)
     file.close()
     
 def get_device(e):
-    file = open(bb.data.getVar('DEVFILE', e.data, True))
+    file = open(e.data.getVar('DEVFILE', True))
     device = file.readline()
     file.close()
     return device
@@ -126,7 +126,7 @@ def get_timedata(var, data):
     
 def write_task_data(status, logfile, dev, e):
     bn = get_bn(e)
-    bsdir = os.path.join(bb.data.getVar('BUILDSTATS_BASE', e.data, True), bn)
+    bsdir = os.path.join(e.data.getVar('BUILDSTATS_BASE', True), bn)
     taskdir = os.path.join(bsdir, bb.data.expand("${PF}", e.data))
     file = open(os.path.join(logfile), "a")
     timedata = get_timedata("__timedata_task", e.data)
@@ -168,7 +168,7 @@ python run_buildstats () {
         # set the buildname
         ########################################################################
         try:
-            bb.mkdirhier(bb.data.getVar('BUILDSTATS_BASE', e.data, True))
+            bb.mkdirhier(e.data.getVar('BUILDSTATS_BASE', True))
         except:
             pass
         set_bn(e)
@@ -176,7 +176,7 @@ python run_buildstats () {
         set_device(e)
         device = get_device(e)
         
-        bsdir = os.path.join(bb.data.getVar('BUILDSTATS_BASE', e.data, True), bn)
+        bsdir = os.path.join(e.data.getVar('BUILDSTATS_BASE', True), bn)
         try:
             bb.mkdirhier(bsdir)
         except:
@@ -199,7 +199,7 @@ python run_buildstats () {
     elif isinstance(e, bb.event.BuildCompleted):
         bn = get_bn(e)
         device = get_device(e)
-        bsdir = os.path.join(bb.data.getVar('BUILDSTATS_BASE', e.data, True), bn)
+        bsdir = os.path.join(e.data.getVar('BUILDSTATS_BASE', True), bn)
         taskdir = os.path.join(bsdir, bb.data.expand("${PF}", e.data))
         build_time = os.path.join(bsdir, "build_stats")
         file = open(build_time, "a")
@@ -224,7 +224,7 @@ python run_buildstats () {
     if isinstance(e, bb.build.TaskStarted):
         bn = get_bn(e)
         device = get_device(e)
-        bsdir = os.path.join(bb.data.getVar('BUILDSTATS_BASE', e.data, True), bn)
+        bsdir = os.path.join(e.data.getVar('BUILDSTATS_BASE', True), bn)
         taskdir = os.path.join(bsdir, bb.data.expand("${PF}", e.data))
         if device != "NoLogicalDevice":
             set_diskdata("__diskdata_task", device, e.data)
@@ -242,14 +242,14 @@ python run_buildstats () {
     elif isinstance(e, bb.build.TaskSucceeded):
         bn = get_bn(e)
         device = get_device(e)
-        bsdir = os.path.join(bb.data.getVar('BUILDSTATS_BASE', e.data, True), bn)
+        bsdir = os.path.join(e.data.getVar('BUILDSTATS_BASE', True), bn)
         taskdir = os.path.join(bsdir, bb.data.expand("${PF}", e.data))
         write_task_data("passed", os.path.join(taskdir, e.task), device, e)
         if e.task == "do_rootfs":
-            bsdir = os.path.join(bb.data.getVar('BUILDSTATS_BASE', e.data, True), bn)
+            bsdir = os.path.join(e.data.getVar('BUILDSTATS_BASE', True), bn)
             bs=os.path.join(bsdir, "build_stats")
             file = open(bs,"a") 
-            rootfs = bb.data.getVar('IMAGE_ROOTFS', e.data, True)
+            rootfs = e.data.getVar('IMAGE_ROOTFS', True)
             rootfs_size = subprocess.Popen(["du", "-sh", rootfs], stdout=subprocess.PIPE).stdout.read() 
             file.write("Uncompressed Rootfs size: %s" % rootfs_size)
             file.close()
@@ -257,7 +257,7 @@ python run_buildstats () {
     elif isinstance(e, bb.build.TaskFailed):
         bn = get_bn(e)
         device = get_device(e)
-        bsdir = os.path.join(bb.data.getVar('BUILDSTATS_BASE', e.data, True), bn)
+        bsdir = os.path.join(e.data.getVar('BUILDSTATS_BASE', True), bn)
         taskdir = os.path.join(bsdir, bb.data.expand("${PF}", e.data))
         write_task_data("failed", os.path.join(taskdir, e.task), device, e)
         ########################################################################

+ 1 - 1
meta/classes/cpan-base.bbclass

@@ -28,7 +28,7 @@ def get_perl_version(d):
 
 # Determine where the library directories are
 def perl_get_libdirs(d):
-	libdir = bb.data.getVar('libdir', d, 1)
+	libdir = d.getVar('libdir', 1)
 	if is_target(d) == "no":
 		libdir += '/perl-native'
 	libdir += '/perl'

+ 2 - 2
meta/classes/cpan_build.bbclass

@@ -10,9 +10,9 @@ inherit cpan-base
 # libmodule-build-perl)
 #
 def cpan_build_dep_prepend(d):
-	if bb.data.getVar('CPAN_BUILD_DEPS', d, 1):
+	if d.getVar('CPAN_BUILD_DEPS', 1):
 		return ''
-	pn = bb.data.getVar('PN', d, 1)
+	pn = d.getVar('PN', 1)
 	if pn in ['libmodule-build-perl', 'libmodule-build-perl-native']:
 		return ''
 	return 'libmodule-build-perl-native '

+ 1 - 1
meta/classes/cross-canadian.bbclass

@@ -16,7 +16,7 @@ STAGING_BINDIR_TOOLCHAIN = "${STAGING_DIR_NATIVE}${bindir_native}/${SDK_ARCH}${S
 #
 PACKAGE_ARCH = "${SDK_ARCH}-nativesdk"
 python () {
-    archs = bb.data.getVar('PACKAGE_ARCHS', d, True).split()
+    archs = d.getVar('PACKAGE_ARCHS', True).split()
     sdkarchs = []
     for arch in archs:
         sdkarchs.append(arch + '-nativesdk')

+ 9 - 9
meta/classes/debian.bbclass

@@ -22,8 +22,8 @@ python () {
 python debian_package_name_hook () {
 	import glob, copy, stat, errno, re
 
-	pkgdest = bb.data.getVar('PKGDEST', d, 1)
-	packages = bb.data.getVar('PACKAGES', d, 1)
+	pkgdest = d.getVar('PKGDEST', 1)
+	packages = d.getVar('PACKAGES', 1)
 	bin_re = re.compile(".*/s?" + os.path.basename(d.getVar("bindir", True)) + "$")
 	lib_re = re.compile(".*/" + os.path.basename(d.getVar("libdir", True)) + "$")
 	so_re = re.compile("lib.*\.so")
@@ -60,7 +60,7 @@ python debian_package_name_hook () {
 				for f in files:
 					if so_re.match(f):
 						fp = os.path.join(root, f)
-						cmd = (bb.data.getVar('BUILD_PREFIX', d, 1) or "") + "objdump -p " + fp + " 2>/dev/null"
+						cmd = (d.getVar('BUILD_PREFIX', 1) or "") + "objdump -p " + fp + " 2>/dev/null"
 						fd = os.popen(cmd)
 						lines = fd.readlines()
 						fd.close()
@@ -74,7 +74,7 @@ python debian_package_name_hook () {
 		if len(sonames) == 1:
 			soname = sonames[0]
 		elif len(sonames) > 1:
-			lead = bb.data.getVar('LEAD_SONAME', d, 1)
+			lead = d.getVar('LEAD_SONAME', 1)
 			if lead:
 				r = re.compile(lead)
 				filtered = []
@@ -95,21 +95,21 @@ python debian_package_name_hook () {
 			if soname_result:
 				(pkgname, devname) = soname_result
 				for pkg in packages.split():
-					if (bb.data.getVar('PKG_' + pkg, d) or bb.data.getVar('DEBIAN_NOAUTONAME_' + pkg, d)):
+					if (d.getVar('PKG_' + pkg) or d.getVar('DEBIAN_NOAUTONAME_' + pkg)):
 						continue
-					debian_pn = bb.data.getVar('DEBIANNAME_' + pkg, d)
+					debian_pn = d.getVar('DEBIANNAME_' + pkg)
 					if debian_pn:
 						newpkg = debian_pn
 					elif pkg == orig_pkg:
 						newpkg = pkgname
 					else:
 						newpkg = pkg.replace(orig_pkg, devname, 1)
-					mlpre=bb.data.getVar('MLPREFIX', d, True)
+					mlpre=d.getVar('MLPREFIX', True)
 					if mlpre:
 						if not newpkg.find(mlpre) == 0:
 							newpkg = mlpre + newpkg
 					if newpkg != pkg:
-						bb.data.setVar('PKG_' + pkg, newpkg, d)
+						d.setVar('PKG_' + pkg, newpkg)
 
 	# reversed sort is needed when some package is substring of another
 	# ie in ncurses we get without reverse sort: 
@@ -117,7 +117,7 @@ python debian_package_name_hook () {
 	# and later
 	# DEBUG: LIBNAMES: pkgname libtic5 devname libtic pkg ncurses-libticw orig_pkg ncurses-libtic debian_pn None newpkg libticw
 	# so we need to handle ncurses-libticw->libticw5 before ncurses-libtic->libtic5
-	for pkg in sorted((bb.data.getVar('AUTO_LIBNAME_PKGS', d, 1) or "").split(), reverse=True):
+	for pkg in sorted((d.getVar('AUTO_LIBNAME_PKGS', 1) or "").split(), reverse=True):
 		auto_libname(packages, pkg)
 }
 

+ 81 - 81
meta/classes/distrodata.bbclass

@@ -19,87 +19,87 @@ addtask distrodata_np
 do_distrodata_np[nostamp] = "1"
 python do_distrodata_np() {
 	localdata = bb.data.createCopy(d)
-        pn = bb.data.getVar("PN", d, True)
+        pn = d.getVar("PN", True)
         bb.note("Package Name: %s" % pn)
 
         import oe.distro_check as dist_check
-        tmpdir = bb.data.getVar('TMPDIR', d, True)
+        tmpdir = d.getVar('TMPDIR', True)
         distro_check_dir = os.path.join(tmpdir, "distro_check")
-        datetime = bb.data.getVar('DATETIME', localdata, True)
+        datetime = localdata.getVar('DATETIME', True)
         dist_check.update_distro_data(distro_check_dir, datetime)
 
 	if pn.find("-native") != -1:
 	    pnstripped = pn.split("-native")
 	    bb.note("Native Split: %s" % pnstripped)
-	    bb.data.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + bb.data.getVar('OVERRIDES', d, True), localdata)
+	    bb.data.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + d.getVar('OVERRIDES', True), localdata)
 	    bb.data.update_data(localdata)
 
 	if pn.find("-nativesdk") != -1:
 	    pnstripped = pn.split("-nativesdk")
 	    bb.note("Native Split: %s" % pnstripped)
-	    bb.data.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + bb.data.getVar('OVERRIDES', d, True), localdata)
+	    bb.data.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + d.getVar('OVERRIDES', True), localdata)
 	    bb.data.update_data(localdata)
 
 	if pn.find("-cross") != -1:
 	    pnstripped = pn.split("-cross")
 	    bb.note("cross Split: %s" % pnstripped)
-	    bb.data.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + bb.data.getVar('OVERRIDES', d, True), localdata)
+	    bb.data.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + d.getVar('OVERRIDES', True), localdata)
 	    bb.data.update_data(localdata)
 
 	if pn.find("-crosssdk") != -1:
 	    pnstripped = pn.split("-crosssdk")
 	    bb.note("cross Split: %s" % pnstripped)
-	    bb.data.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + bb.data.getVar('OVERRIDES', d, True), localdata)
+	    bb.data.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + d.getVar('OVERRIDES', True), localdata)
 	    bb.data.update_data(localdata)
 
 	if pn.find("-initial") != -1:
 	    pnstripped = pn.split("-initial")
 	    bb.note("initial Split: %s" % pnstripped)
-	    bb.data.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + bb.data.getVar('OVERRIDES', d, True), localdata)
+	    bb.data.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + d.getVar('OVERRIDES', True), localdata)
 	    bb.data.update_data(localdata)
 
 	"""generate package information from .bb file"""
-	pname = bb.data.getVar('PN', localdata, True)
-	pcurver = bb.data.getVar('PV', localdata, True)
-	pdesc = bb.data.getVar('DESCRIPTION', localdata, True)
+	pname = localdata.getVar('PN', True)
+	pcurver = localdata.getVar('PV', True)
+	pdesc = localdata.getVar('DESCRIPTION', True)
         if pdesc is not None:
                 pdesc = pdesc.replace(',','')
                 pdesc = pdesc.replace('\n','')
 
-	pgrp = bb.data.getVar('SECTION', localdata, True)
-	plicense = bb.data.getVar('LICENSE', localdata, True).replace(',','_')
-	if bb.data.getVar('LIC_FILES_CHKSUM', localdata, True):
+	pgrp = localdata.getVar('SECTION', True)
+	plicense = localdata.getVar('LICENSE', True).replace(',','_')
+	if localdata.getVar('LIC_FILES_CHKSUM', True):
 		pchksum="1"
 	else:
 		pchksum="0"
 
-	if bb.data.getVar('RECIPE_STATUS', localdata, True):
+	if localdata.getVar('RECIPE_STATUS', True):
 		hasrstatus="1"
 	else:
 		hasrstatus="0"
 
-	rstatus = bb.data.getVar('RECIPE_STATUS', localdata, True)
+	rstatus = localdata.getVar('RECIPE_STATUS', True)
         if rstatus is not None:
                 rstatus = rstatus.replace(',','')
 		
-	pupver = bb.data.getVar('RECIPE_LATEST_VERSION', localdata, True)
+	pupver = localdata.getVar('RECIPE_LATEST_VERSION', True)
 	if pcurver == pupver:
 		vermatch="1"
 	else:
 		vermatch="0"
-	noupdate_reason = bb.data.getVar('RECIPE_NO_UPDATE_REASON', localdata, True)
+	noupdate_reason = localdata.getVar('RECIPE_NO_UPDATE_REASON', True)
 	if noupdate_reason is None:
 		noupdate="0"
 	else:
 		noupdate="1"
                 noupdate_reason = noupdate_reason.replace(',','')
 
-	ris = bb.data.getVar('RECIPE_INTEL_SECTION', localdata, True)
-	maintainer = bb.data.getVar('RECIPE_MAINTAINER', localdata, True)
-	rttr = bb.data.getVar('RECIPE_TIME_BETWEEN_LAST_TWO_RELEASES', localdata, True)
-	rlrd = bb.data.getVar('RECIPE_LATEST_RELEASE_DATE', localdata, True)
-	dc = bb.data.getVar('DEPENDENCY_CHECK', localdata, True)
-	rc = bb.data.getVar('RECIPE_COMMENTS', localdata, True)
+	ris = localdata.getVar('RECIPE_INTEL_SECTION', True)
+	maintainer = localdata.getVar('RECIPE_MAINTAINER', True)
+	rttr = localdata.getVar('RECIPE_TIME_BETWEEN_LAST_TWO_RELEASES', True)
+	rlrd = localdata.getVar('RECIPE_LATEST_RELEASE_DATE', True)
+	dc = localdata.getVar('DEPENDENCY_CHECK', True)
+	rc = localdata.getVar('RECIPE_COMMENTS', True)
         result = dist_check.compare_in_distro_packages_list(distro_check_dir, localdata)
 
 	bb.note("DISTRO: %s,%s,%s,%s,%s,%s,%s,%s,%s, %s, %s, %s\n" % \
@@ -113,81 +113,81 @@ python do_distrodata_np() {
 addtask distrodata
 do_distrodata[nostamp] = "1"
 python do_distrodata() {
-	logpath = bb.data.getVar('LOG_DIR', d, True)
+	logpath = d.getVar('LOG_DIR', True)
 	bb.utils.mkdirhier(logpath)
 	logfile = os.path.join(logpath, "distrodata.csv")
 
         import oe.distro_check as dist_check
 	localdata = bb.data.createCopy(d)
-        tmpdir = bb.data.getVar('TMPDIR', d, True)
+        tmpdir = d.getVar('TMPDIR', True)
         distro_check_dir = os.path.join(tmpdir, "distro_check")
-        datetime = bb.data.getVar('DATETIME', localdata, True)
+        datetime = localdata.getVar('DATETIME', True)
         dist_check.update_distro_data(distro_check_dir, datetime)
 
-        pn = bb.data.getVar("PN", d, True)
+        pn = d.getVar("PN", True)
         bb.note("Package Name: %s" % pn)
 
 	if pn.find("-native") != -1:
 	    pnstripped = pn.split("-native")
 	    bb.note("Native Split: %s" % pnstripped)
-	    bb.data.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + bb.data.getVar('OVERRIDES', d, True), localdata)
+	    bb.data.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + d.getVar('OVERRIDES', True), localdata)
 	    bb.data.update_data(localdata)
 
 	if pn.find("-cross") != -1:
 	    pnstripped = pn.split("-cross")
 	    bb.note("cross Split: %s" % pnstripped)
-	    bb.data.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + bb.data.getVar('OVERRIDES', d, True), localdata)
+	    bb.data.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + d.getVar('OVERRIDES', True), localdata)
 	    bb.data.update_data(localdata)
 
 	if pn.find("-initial") != -1:
 	    pnstripped = pn.split("-initial")
 	    bb.note("initial Split: %s" % pnstripped)
-	    bb.data.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + bb.data.getVar('OVERRIDES', d, True), localdata)
+	    bb.data.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + d.getVar('OVERRIDES', True), localdata)
 	    bb.data.update_data(localdata)
 
 	"""generate package information from .bb file"""
-	pname = bb.data.getVar('PN', localdata, True)
-	pcurver = bb.data.getVar('PV', localdata, True)
-	pdesc = bb.data.getVar('DESCRIPTION', localdata, True)
+	pname = localdata.getVar('PN', True)
+	pcurver = localdata.getVar('PV', True)
+	pdesc = localdata.getVar('DESCRIPTION', True)
         if pdesc is not None:
                 pdesc = pdesc.replace(',','')
                 pdesc = pdesc.replace('\n','')
 
-	pgrp = bb.data.getVar('SECTION', localdata, True)
-	plicense = bb.data.getVar('LICENSE', localdata, True).replace(',','_')
-	if bb.data.getVar('LIC_FILES_CHKSUM', localdata, True):
+	pgrp = localdata.getVar('SECTION', True)
+	plicense = localdata.getVar('LICENSE', True).replace(',','_')
+	if localdata.getVar('LIC_FILES_CHKSUM', True):
 		pchksum="1"
 	else:
 		pchksum="0"
 
-	if bb.data.getVar('RECIPE_STATUS', localdata, True):
+	if localdata.getVar('RECIPE_STATUS', True):
 		hasrstatus="1"
 	else:
 		hasrstatus="0"
 
-	rstatus = bb.data.getVar('RECIPE_STATUS', localdata, True)
+	rstatus = localdata.getVar('RECIPE_STATUS', True)
         if rstatus is not None:
                 rstatus = rstatus.replace(',','')
 		
-	pupver = bb.data.getVar('RECIPE_LATEST_VERSION', localdata, True)
+	pupver = localdata.getVar('RECIPE_LATEST_VERSION', True)
 	if pcurver == pupver:
 		vermatch="1"
 	else:
 		vermatch="0"
 
-	noupdate_reason = bb.data.getVar('RECIPE_NO_UPDATE_REASON', localdata, True)
+	noupdate_reason = localdata.getVar('RECIPE_NO_UPDATE_REASON', True)
 	if noupdate_reason is None:
 		noupdate="0"
 	else:
 		noupdate="1"
                 noupdate_reason = noupdate_reason.replace(',','')
 
-	ris = bb.data.getVar('RECIPE_INTEL_SECTION', localdata, True)
-	maintainer = bb.data.getVar('RECIPE_MAINTAINER', localdata, True)
-	rttr = bb.data.getVar('RECIPE_TIME_BETWEEN_LAST_TWO_RELEASES', localdata, True)
-	rlrd = bb.data.getVar('RECIPE_LATEST_RELEASE_DATE', localdata, True)
-	dc = bb.data.getVar('DEPENDENCY_CHECK', localdata, True)
-	rc = bb.data.getVar('RECIPE_COMMENTS', localdata, True)
+	ris = localdata.getVar('RECIPE_INTEL_SECTION', True)
+	maintainer = localdata.getVar('RECIPE_MAINTAINER', True)
+	rttr = localdata.getVar('RECIPE_TIME_BETWEEN_LAST_TWO_RELEASES', True)
+	rlrd = localdata.getVar('RECIPE_LATEST_RELEASE_DATE', True)
+	dc = localdata.getVar('DEPENDENCY_CHECK', True)
+	rc = localdata.getVar('RECIPE_COMMENTS', True)
         # do the comparison
         result = dist_check.compare_in_distro_packages_list(distro_check_dir, localdata)
 
@@ -298,7 +298,7 @@ python do_checkpkg() {
 		Clear internal url cache as it's a temporary check. Not doing so will have 
 		bitbake check url multiple times when looping through a single url
 		"""
-		fn = bb.data.getVar('FILE', d, True)
+		fn = d.getVar('FILE', True)
 		bb.fetch2.urldata_cache[fn] = {}
 
 		"""
@@ -329,7 +329,7 @@ python do_checkpkg() {
 	Return new version if success, or else error in "Errxxxx" style
 	"""
 	def check_new_dir(url, curver, d):
-		pn = bb.data.getVar('PN', d, True)
+		pn = d.getVar('PN', True)
 		f = tempfile.NamedTemporaryFile(delete=False, prefix="%s-1-" % pn)
 		status = internal_fetch_wget(url, d, f)
 		fhtml = f.read()
@@ -372,7 +372,7 @@ python do_checkpkg() {
 
 		f.close()
 		if status != "ErrHostNoDir" and re.match("Err", status):
-			logpath = bb.data.getVar('LOG_DIR', d, 1)
+			logpath = d.getVar('LOG_DIR', 1)
 			os.system("cp %s %s/" % (f.name, logpath))
 		os.unlink(f.name)
 		return status
@@ -388,7 +388,7 @@ python do_checkpkg() {
 		"""possible to have no version in pkg name, such as spectrum-fw"""
 		if not re.search("\d+", curname):
 			return pcurver
-		pn = bb.data.getVar('PN', d, True)
+		pn = d.getVar('PN', True)
 		f = tempfile.NamedTemporaryFile(delete=False, prefix="%s-2-" % pn)
 		status = internal_fetch_wget(url, d, f)
 		fhtml = f.read()
@@ -431,55 +431,55 @@ python do_checkpkg() {
 		f.close()
 		"""if host hasn't directory information, no need to save tmp file"""
 		if status != "ErrHostNoDir" and re.match("Err", status):
-			logpath = bb.data.getVar('LOG_DIR', d, True)
+			logpath = d.getVar('LOG_DIR', True)
 			os.system("cp %s %s/" % (f.name, logpath))
 		os.unlink(f.name)
 		return status
 
 	"""first check whether a uri is provided"""
-	src_uri = bb.data.getVar('SRC_URI', d, True)
+	src_uri = d.getVar('SRC_URI', True)
 	if not src_uri:
 		return
 
 	"""initialize log files."""
-	logpath = bb.data.getVar('LOG_DIR', d, True)
+	logpath = d.getVar('LOG_DIR', True)
 	bb.utils.mkdirhier(logpath)
 	logfile = os.path.join(logpath, "checkpkg.csv")
 
 	"""generate package information from .bb file"""
-	pname = bb.data.getVar('PN', d, True)
+	pname = d.getVar('PN', True)
 
 	if pname.find("-native") != -1:
 	    pnstripped = pname.split("-native")
 	    bb.note("Native Split: %s" % pnstripped)
-	    bb.data.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + bb.data.getVar('OVERRIDES', d, True), localdata)
+	    bb.data.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + d.getVar('OVERRIDES', True), localdata)
 	    bb.data.update_data(localdata)
 
 	if pname.find("-cross") != -1:
 	    pnstripped = pname.split("-cross")
 	    bb.note("cross Split: %s" % pnstripped)
-	    bb.data.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + bb.data.getVar('OVERRIDES', d, True), localdata)
+	    bb.data.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + d.getVar('OVERRIDES', True), localdata)
 	    bb.data.update_data(localdata)
 
 	if pname.find("-initial") != -1:
 	    pnstripped = pname.split("-initial")
 	    bb.note("initial Split: %s" % pnstripped)
-	    bb.data.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + bb.data.getVar('OVERRIDES', d, True), localdata)
+	    bb.data.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + d.getVar('OVERRIDES', True), localdata)
 	    bb.data.update_data(localdata)
 
-	pdesc = bb.data.getVar('DESCRIPTION', localdata, True)
-	pgrp = bb.data.getVar('SECTION', localdata, True)
-	pversion = bb.data.getVar('PV', localdata, True)
-	plicense = bb.data.getVar('LICENSE', localdata, True)
-	psection = bb.data.getVar('SECTION', localdata, True)
-	phome = bb.data.getVar('HOMEPAGE', localdata, True)
-	prelease = bb.data.getVar('PR', localdata, True)
-	ppriority = bb.data.getVar('PRIORITY', localdata, True)
-	pdepends = bb.data.getVar('DEPENDS', localdata, True)
-	pbugtracker = bb.data.getVar('BUGTRACKER', localdata, True)
-	ppe = bb.data.getVar('PE', localdata, True)
-	psrcuri = bb.data.getVar('SRC_URI', localdata, True)
-	maintainer = bb.data.getVar('RECIPE_MAINTAINER', localdata, True)
+	pdesc = localdata.getVar('DESCRIPTION', True)
+	pgrp = localdata.getVar('SECTION', True)
+	pversion = localdata.getVar('PV', True)
+	plicense = localdata.getVar('LICENSE', True)
+	psection = localdata.getVar('SECTION', True)
+	phome = localdata.getVar('HOMEPAGE', True)
+	prelease = localdata.getVar('PR', True)
+	ppriority = localdata.getVar('PRIORITY', True)
+	pdepends = localdata.getVar('DEPENDS', True)
+	pbugtracker = localdata.getVar('BUGTRACKER', True)
+	ppe = localdata.getVar('PE', True)
+	psrcuri = localdata.getVar('SRC_URI', True)
+	maintainer = localdata.getVar('RECIPE_MAINTAINER', True)
 
 	found = 0
 	for uri in src_uri.split():
@@ -497,9 +497,9 @@ python do_checkpkg() {
 
 	(type, host, path, user, pswd, parm) = bb.decodeurl(uri)
 	if type in ['http', 'https', 'ftp']:
-		pcurver = bb.data.getVar('PV', d, True)
+		pcurver = d.getVar('PV', True)
 	else:
-		pcurver = bb.data.getVar("SRCREV", d, True)
+		pcurver = d.getVar("SRCREV", True)
 
 	if type in ['http', 'https', 'ftp']:
 		newver = pcurver
@@ -639,7 +639,7 @@ python do_checkpkg() {
 		pstatus += ":%s%s" % (host, path)
 
 	"""Read from manual distro tracking fields as alternative"""
-	pmver = bb.data.getVar("RECIPE_LATEST_VERSION", d, True)
+	pmver = d.getVar("RECIPE_LATEST_VERSION", True)
 	if not pmver:
 		pmver = "N/A"
 		pmstatus = "ErrNoRecipeData"
@@ -688,12 +688,12 @@ python do_distro_check() {
 
     localdata = bb.data.createCopy(d)
     bb.data.update_data(localdata)
-    tmpdir = bb.data.getVar('TMPDIR', d, True)
+    tmpdir = d.getVar('TMPDIR', True)
     distro_check_dir = os.path.join(tmpdir, "distro_check")
-    logpath = bb.data.getVar('LOG_DIR', d, True)
+    logpath = d.getVar('LOG_DIR', True)
     bb.utils.mkdirhier(logpath)
     result_file = os.path.join(logpath, "distrocheck.csv")
-    datetime = bb.data.getVar('DATETIME', localdata, True)
+    datetime = localdata.getVar('DATETIME', True)
     dc.update_distro_data(distro_check_dir, datetime)
 
     # do the comparison
@@ -734,12 +734,12 @@ python do_checklicense() {
     import os
     import bb
     import shutil
-    logpath = bb.data.getVar('LOG_DIR', d, True)
+    logpath = d.getVar('LOG_DIR', True)
     bb.utils.mkdirhier(logpath)
-    pn = bb.data.getVar('PN', d, True)
+    pn = d.getVar('PN', True)
     logfile = os.path.join(logpath, "missinglicense.csv")
-    generic_directory = bb.data.getVar('COMMON_LICENSE_DIR', d, True)
-    license_types = bb.data.getVar('LICENSE', d, True)
+    generic_directory = d.getVar('COMMON_LICENSE_DIR', True)
+    license_types = d.getVar('LICENSE', True)
     for license_type in ((license_types.replace('+', '').replace('|', '&')
                           .replace('(', '').replace(')', '').replace(';', '')
                           .replace(',', '').replace(" ", "").split("&"))):

+ 1 - 1
meta/classes/distutils-base.bbclass

@@ -1,4 +1,4 @@
-DEPENDS  += "${@["python-native python", ""][(bb.data.getVar('PACKAGES', d, 1) == '')]}"
+DEPENDS  += "${@["python-native python", ""][(d.getVar('PACKAGES', 1) == '')]}"
 RDEPENDS_${PN} += "${@['', 'python-core']['${PN}' == '${BPN}']}"
 
 inherit distutils-common-base

+ 1 - 1
meta/classes/distutils-native-base.bbclass

@@ -1,3 +1,3 @@
-DEPENDS  += "${@["python-native", ""][(bb.data.getVar('PACKAGES', d, 1) == '')]}"
+DEPENDS  += "${@["python-native", ""][(d.getVar('PACKAGES', 1) == '')]}"
 
 inherit distutils-common-base

+ 8 - 8
meta/classes/gconf.bbclass

@@ -27,8 +27,8 @@ done
 
 python populate_packages_append () {
 	import re
-	packages = bb.data.getVar('PACKAGES', d, 1).split()
-	pkgdest =  bb.data.getVar('PKGDEST', d, 1)
+	packages = d.getVar('PACKAGES', 1).split()
+	pkgdest =  d.getVar('PKGDEST', 1)
 	
 	for pkg in packages:
 		schema_dir = '%s/%s/etc/gconf/schemas' % (pkgdest, pkg)
@@ -41,15 +41,15 @@ python populate_packages_append () {
 		if schemas != []:
 			bb.note("adding gconf postinst and prerm scripts to %s" % pkg)
 			bb.data.setVar('SCHEMA_FILES', " ".join(schemas), d)
-			postinst = bb.data.getVar('pkg_postinst_%s' % pkg, d, 1) or bb.data.getVar('pkg_postinst', d, 1)
+			postinst = d.getVar('pkg_postinst_%s' % pkg, 1) or d.getVar('pkg_postinst', 1)
 			if not postinst:
 				postinst = '#!/bin/sh\n'
-			postinst += bb.data.getVar('gconf_postinst', d, 1)
-			bb.data.setVar('pkg_postinst_%s' % pkg, postinst, d)
-			prerm = bb.data.getVar('pkg_prerm_%s' % pkg, d, 1) or bb.data.getVar('pkg_prerm', d, 1)
+			postinst += d.getVar('gconf_postinst', 1)
+			d.setVar('pkg_postinst_%s' % pkg, postinst)
+			prerm = d.getVar('pkg_prerm_%s' % pkg, 1) or d.getVar('pkg_prerm', 1)
 			if not prerm:
 				prerm = '#!/bin/sh\n'
-			prerm += bb.data.getVar('gconf_prerm', d, 1)
-			bb.data.setVar('pkg_prerm_%s' % pkg, prerm, d)
+			prerm += d.getVar('gconf_prerm', 1)
+			d.setVar('pkg_prerm_%s' % pkg, prerm)
 
 }

+ 12 - 12
meta/classes/gtk-icon-cache.bbclass

@@ -28,31 +28,31 @@ done
 }
 
 python populate_packages_append () {
-	packages = bb.data.getVar('PACKAGES', d, 1).split()
-	pkgdest =  bb.data.getVar('PKGDEST', d, 1)
+	packages = d.getVar('PACKAGES', 1).split()
+	pkgdest =  d.getVar('PKGDEST', 1)
 	
 	for pkg in packages:
-		icon_dir = '%s/%s/%s/icons' % (pkgdest, pkg, bb.data.getVar('datadir', d, 1))
+		icon_dir = '%s/%s/%s/icons' % (pkgdest, pkg, d.getVar('datadir', 1))
 		if not os.path.exists(icon_dir):
 			continue
 
 		bb.note("adding hicolor-icon-theme dependency to %s" % pkg)	
-		rdepends = bb.data.getVar('RDEPENDS_%s' % pkg, d, 1)
-		rdepends = rdepends + ' ' + bb.data.getVar('MLPREFIX', d) + "hicolor-icon-theme"
-		bb.data.setVar('RDEPENDS_%s' % pkg, rdepends, d)
+		rdepends = d.getVar('RDEPENDS_%s' % pkg, 1)
+		rdepends = rdepends + ' ' + d.getVar('MLPREFIX') + "hicolor-icon-theme"
+		d.setVar('RDEPENDS_%s' % pkg, rdepends)
 	
 		bb.note("adding gtk-icon-cache postinst and postrm scripts to %s" % pkg)
 		
-		postinst = bb.data.getVar('pkg_postinst_%s' % pkg, d, 1) or bb.data.getVar('pkg_postinst', d, 1)
+		postinst = d.getVar('pkg_postinst_%s' % pkg, 1) or d.getVar('pkg_postinst', 1)
 		if not postinst:
 			postinst = '#!/bin/sh\n'
-		postinst += bb.data.getVar('gtk_icon_cache_postinst', d, 1)
-		bb.data.setVar('pkg_postinst_%s' % pkg, postinst, d)
+		postinst += d.getVar('gtk_icon_cache_postinst', 1)
+		d.setVar('pkg_postinst_%s' % pkg, postinst)
 
-		postrm = bb.data.getVar('pkg_postrm_%s' % pkg, d, 1) or bb.data.getVar('pkg_postrm', d, 1)
+		postrm = d.getVar('pkg_postrm_%s' % pkg, 1) or d.getVar('pkg_postrm', 1)
 		if not postrm:
 			postrm = '#!/bin/sh\n'
-		postrm += bb.data.getVar('gtk_icon_cache_postrm', d, 1)
-		bb.data.setVar('pkg_postrm_%s' % pkg, postrm, d)
+		postrm += d.getVar('gtk_icon_cache_postrm', 1)
+		d.setVar('pkg_postrm_%s' % pkg, postrm)
 }
 

+ 8 - 8
meta/classes/icecc.bbclass

@@ -32,7 +32,7 @@ def icecc_dep_prepend(d):
     # INHIBIT_DEFAULT_DEPS doesn't apply to the patch command.  Whether or  not
     # we need that built is the responsibility of the patch function / class, not
     # the application.
-    if not bb.data.getVar('INHIBIT_DEFAULT_DEPS', d):
+    if not d.getVar('INHIBIT_DEFAULT_DEPS'):
         return "icecc-create-env-native"
     return ""
 
@@ -54,7 +54,7 @@ def create_path(compilers, bb, d):
         staging += "-kernel"
 
     #check if the icecc path is set by the user
-    icecc   = bb.data.getVar('ICECC_PATH', d) or os.popen("which icecc").read()[:-1]
+    icecc   = d.getVar('ICECC_PATH') or os.popen("which icecc").read()[:-1]
 
     # Create the dir if necessary
     try:
@@ -81,7 +81,7 @@ def use_icc(bb,d):
     package_tmp = bb.data.expand('${PN}', d)
 
     system_class_blacklist = [ "none" ] 
-    user_class_blacklist = (bb.data.getVar('ICECC_USER_CLASS_BL', d) or "none").split()
+    user_class_blacklist = (d.getVar('ICECC_USER_CLASS_BL') or "none").split()
     package_class_blacklist = system_class_blacklist + user_class_blacklist
 
     for black in package_class_blacklist:
@@ -92,7 +92,7 @@ def use_icc(bb,d):
     #"system" package blacklist contains a list of packages that can not distribute compile tasks
     #for one reason or the other
     system_package_blacklist = [ "uclibc", "glibc", "gcc", "bind", "u-boot", "dhcp-forwarder", "enchant", "connman", "orbit2" ]
-    user_package_blacklist = (bb.data.getVar('ICECC_USER_PACKAGE_BL', d) or "").split()
+    user_package_blacklist = (d.getVar('ICECC_USER_PACKAGE_BL') or "").split()
     package_blacklist = system_package_blacklist + user_package_blacklist
 
     for black in package_blacklist:
@@ -100,7 +100,7 @@ def use_icc(bb,d):
             #bb.note(package_tmp, ' found in blacklist, disable icecc')
             return "no"
 
-    if bb.data.getVar('PARALLEL_MAKE', d) == "":
+    if d.getVar('PARALLEL_MAKE') == "":
         bb.note(package_tmp, " ", bb.data.expand('${PV}', d), " has empty PARALLEL_MAKE, disable icecc")
         return "no"
 
@@ -119,8 +119,8 @@ def icc_version(bb, d):
     if use_icc(bb, d) == "no":
         return ""
 
-    parallel = bb.data.getVar('ICECC_PARALLEL_MAKE', d) or ""
-    bb.data.setVar("PARALLEL_MAKE", parallel, d)
+    parallel = d.getVar('ICECC_PARALLEL_MAKE') or ""
+    d.setVar("PARALLEL_MAKE", parallel)
 
     if icc_is_native(bb, d):
         archive_name = "local-host-env"
@@ -130,7 +130,7 @@ def icc_version(bb, d):
         prefix = bb.data.expand('${HOST_PREFIX}' , d)
         distro = bb.data.expand('${DISTRO}', d)
         target_sys = bb.data.expand('${TARGET_SYS}', d)
-        float = bb.data.getVar('TARGET_FPU', d) or "hard"
+        float = d.getVar('TARGET_FPU') or "hard"
         archive_name = prefix + distro + "-"        + target_sys + "-" + float
         if icc_is_kernel(bb, d):
             archive_name += "-kernel"

+ 2 - 2
meta/classes/image-swab.bbclass

@@ -51,13 +51,13 @@ python() {
     # and cross packages which aren't swabber-native or one of its dependencies
     # I have ignored them for now...
     if not bb.data.inherits_class('native', d) and not bb.data.inherits_class('nativesdk', d) and not bb.data.inherits_class('cross', d):
-       deps = (bb.data.getVarFlag('do_setscene', 'depends', d) or "").split()
+       deps = (d.getVarFlag('do_setscene', 'depends') or "").split()
        deps.append('strace-native:do_populate_sysroot')
        bb.data.setVarFlag('do_setscene', 'depends', " ".join(deps), d)
        logdir = bb.data.expand("${TRACE_LOGDIR}", d)
        bb.utils.mkdirhier(logdir)
     else:
-       bb.data.setVar('STRACEFUNC', '', d)
+       d.setVar('STRACEFUNC', '')
 }
 
 STRACEPID = "${@os.getpid()}"

+ 10 - 10
meta/classes/image.bbclass

@@ -74,17 +74,17 @@ IMAGE_TYPE = ${@base_contains("IMAGE_FSTYPES", "live", "live", "empty", d)}
 inherit image-${IMAGE_TYPE}
 
 python () {
-    deps = bb.data.getVarFlag('do_rootfs', 'depends', d) or ""
-    for type in (bb.data.getVar('IMAGE_FSTYPES', d, True) or "").split():
-        for dep in ((bb.data.getVar('IMAGE_DEPENDS_%s' % type, d) or "").split() or []):
+    deps = d.getVarFlag('do_rootfs', 'depends') or ""
+    for type in (d.getVar('IMAGE_FSTYPES', True) or "").split():
+        for dep in ((d.getVar('IMAGE_DEPENDS_%s' % type) or "").split() or []):
             deps += " %s:do_populate_sysroot" % dep
-    for dep in (bb.data.getVar('EXTRA_IMAGEDEPENDS', d, True) or "").split():
+    for dep in (d.getVar('EXTRA_IMAGEDEPENDS', True) or "").split():
         deps += " %s:do_populate_sysroot" % dep
-    bb.data.setVarFlag('do_rootfs', 'depends', deps, d)
+    d.setVarFlag('do_rootfs', 'depends', deps)
 
     # If we don't do this we try and run the mapping hooks while parsing which is slow
     # bitbake should really provide something to let us know this...
-    if bb.data.getVar('BB_WORKERCONTEXT', d, True) is not None:
+    if d.getVar('BB_WORKERCONTEXT', True) is not None:
         runtime_mapping_rename("PACKAGE_INSTALL", d)
         runtime_mapping_rename("PACKAGE_INSTALL_ATTEMPTONLY", d)
 }
@@ -98,15 +98,15 @@ python () {
 # is searched for in the BBPATH (same as the old version.)
 #
 def get_devtable_list(d):
-    devtable = bb.data.getVar('IMAGE_DEVICE_TABLE', d, 1)
+    devtable = d.getVar('IMAGE_DEVICE_TABLE', 1)
     if devtable != None:
         return devtable
     str = ""
-    devtables = bb.data.getVar('IMAGE_DEVICE_TABLES', d, 1)
+    devtables = d.getVar('IMAGE_DEVICE_TABLES', 1)
     if devtables == None:
         devtables = 'files/device_table-minimal.txt'
     for devtable in devtables.split():
-        str += " %s" % bb.which(bb.data.getVar('BBPATH', d, 1), devtable)
+        str += " %s" % bb.which(d.getVar('BBPATH', 1), devtable)
     return str
 
 IMAGE_CLASSES ?= "image_types"
@@ -119,7 +119,7 @@ ROOTFS_POSTPROCESS_COMMAND ?= ""
 # some default locales
 IMAGE_LINGUAS ?= "de-de fr-fr en-gb"
 
-LINGUAS_INSTALL = "${@" ".join(map(lambda s: "locale-base-%s" % s, bb.data.getVar('IMAGE_LINGUAS', d, 1).split()))}"
+LINGUAS_INSTALL = "${@" ".join(map(lambda s: "locale-base-%s" % s, d.getVar('IMAGE_LINGUAS', 1).split()))}"
 
 do_rootfs[nostamp] = "1"
 do_rootfs[dirs] = "${TOPDIR}"

+ 2 - 2
meta/classes/image_types.bbclass

@@ -1,8 +1,8 @@
 def get_imagecmds(d):
     cmds = "\n"
-    old_overrides = bb.data.getVar('OVERRIDES', d, 0)
+    old_overrides = d.getVar('OVERRIDES', 0)
 
-    types = bb.data.getVar('IMAGE_FSTYPES', d, True).split()
+    types = d.getVar('IMAGE_FSTYPES', True).split()
     # Live images will be processed via inheriting bbclass and 
     # does not get processed here.
     # live images also depend on ext3 so ensure its present

+ 28 - 28
meta/classes/imagetest-qemu.bbclass

@@ -35,12 +35,12 @@ def qemuimagetest_main(d):
     
     casestr = re.compile(r'(?P<scen>\w+\b):(?P<case>\S+$)')
     resultstr = re.compile(r'\s*(?P<case>\w+)\s*(?P<pass>\d+)\s*(?P<fail>\d+)\s*(?P<noresult>\d+)')
-    machine = bb.data.getVar('MACHINE', d, 1)
-    pname = bb.data.getVar('PN', d, 1)
+    machine = d.getVar('MACHINE', 1)
+    pname = d.getVar('PN', 1)
     
     """function to save test cases running status"""
     def teststatus(test, status, index, length):
-        test_status = bb.data.getVar('TEST_STATUS', d, 1)
+        test_status = d.getVar('TEST_STATUS', 1)
         if not os.path.exists(test_status):
             raise bb.build.FuncFailed("No test status file existing under TEST_TMP")
 
@@ -51,30 +51,30 @@ def qemuimagetest_main(d):
 
     """funtion to run each case under scenario"""
     def runtest(scen, case, fulltestpath):
-        resultpath = bb.data.getVar('TEST_RESULT', d, 1)
-        tmppath = bb.data.getVar('TEST_TMP', d, 1)
+        resultpath = d.getVar('TEST_RESULT', 1)
+        tmppath = d.getVar('TEST_TMP', 1)
 
         """initialize log file for testcase"""
-        logpath = bb.data.getVar('TEST_LOG', d, 1)
+        logpath = d.getVar('TEST_LOG', 1)
         bb.utils.mkdirhier("%s/%s" % (logpath, scen))
-        caselog = os.path.join(logpath, "%s/log_%s.%s" % (scen, case, bb.data.getVar('DATETIME', d, 1)))
+        caselog = os.path.join(logpath, "%s/log_%s.%s" % (scen, case, d.getVar('DATETIME', 1)))
         os.system("touch %s" % caselog)
         
         """export TEST_TMP, TEST_RESULT, DEPLOY_DIR and QEMUARCH"""
-        os.environ["PATH"] = bb.data.getVar("PATH", d, True)
+        os.environ["PATH"] = d.getVar("PATH", True)
         os.environ["TEST_TMP"] = tmppath
         os.environ["TEST_RESULT"] = resultpath
-        os.environ["DEPLOY_DIR"] = bb.data.getVar("DEPLOY_DIR", d, True)
+        os.environ["DEPLOY_DIR"] = d.getVar("DEPLOY_DIR", True)
         os.environ["QEMUARCH"] = machine
         os.environ["QEMUTARGET"] = pname
-        os.environ["DISPLAY"] = bb.data.getVar("DISPLAY", d, True)
-        os.environ["COREBASE"] = bb.data.getVar("COREBASE", d, True)
-        os.environ["TOPDIR"] = bb.data.getVar("TOPDIR", d, True)
-        os.environ["OE_TMPDIR"] = bb.data.getVar("TMPDIR", d, True)
-        os.environ["TEST_STATUS"] = bb.data.getVar("TEST_STATUS", d, True)
-        os.environ["TARGET_IPSAVE"] = bb.data.getVar("TARGET_IPSAVE", d, True)
-        os.environ["TEST_SERIALIZE"] = bb.data.getVar("TEST_SERIALIZE", d, True)
-        os.environ["SDK_NAME"] = bb.data.getVar("SDK_NAME", d, True)
+        os.environ["DISPLAY"] = d.getVar("DISPLAY", True)
+        os.environ["COREBASE"] = d.getVar("COREBASE", True)
+        os.environ["TOPDIR"] = d.getVar("TOPDIR", True)
+        os.environ["OE_TMPDIR"] = d.getVar("TMPDIR", True)
+        os.environ["TEST_STATUS"] = d.getVar("TEST_STATUS", True)
+        os.environ["TARGET_IPSAVE"] = d.getVar("TARGET_IPSAVE", True)
+        os.environ["TEST_SERIALIZE"] = d.getVar("TEST_SERIALIZE", True)
+        os.environ["SDK_NAME"] = d.getVar("SDK_NAME", True)
 
         """run Test Case"""
         bb.note("Run %s test in scenario %s" % (case, scen))
@@ -92,13 +92,13 @@ def qemuimagetest_main(d):
             if n:
                 item = n.group('scen')
                 casefile = n.group('case')
-                for dir in bb.data.getVar("QEMUIMAGETESTS", d, True).split():
+                for dir in d.getVar("QEMUIMAGETESTS", True).split():
                     fulltestcase = os.path.join(dir, item, casefile)
                     if not os.path.isfile(fulltestcase):
                         raise bb.build.FuncFailed("Testcase %s not found" % fulltestcase)
                     list.append((item, casefile, fulltestcase))
             else:
-                for dir in bb.data.getVar("QEMUIMAGETESTS", d, True).split():
+                for dir in d.getVar("QEMUIMAGETESTS", True).split():
                     scenlist = os.path.join(dir, "scenario", machine, pname)
                     if not os.path.isfile(scenlist):
                         raise bb.build.FuncFailed("No scenario list file named %s found" % scenlist)
@@ -118,7 +118,7 @@ def qemuimagetest_main(d):
 
     """Clean tmp folder for testing"""
     def clean_tmp():
-        tmppath = bb.data.getVar('TEST_TMP', d, 1)
+        tmppath = d.getVar('TEST_TMP', 1)
 
         if os.path.isdir(tmppath):
             for f in os.listdir(tmppath):
@@ -132,28 +132,28 @@ def qemuimagetest_main(d):
     clean_tmp()
 
     """check testcase folder and create test log folder"""
-    testpath = bb.data.getVar('TEST_DIR', d, 1)
+    testpath = d.getVar('TEST_DIR', 1)
     bb.utils.mkdirhier(testpath)
     
-    logpath = bb.data.getVar('TEST_LOG', d, 1)
+    logpath = d.getVar('TEST_LOG', 1)
     bb.utils.mkdirhier(logpath)
 
-    tmppath = bb.data.getVar('TEST_TMP', d, 1)
+    tmppath = d.getVar('TEST_TMP', 1)
     bb.utils.mkdirhier(tmppath)
 
     """initialize test status file"""
-    test_status = bb.data.getVar('TEST_STATUS', d, 1)
+    test_status = d.getVar('TEST_STATUS', 1)
     if os.path.exists(test_status):
         os.remove(test_status)
     os.system("touch %s" % test_status)
 
     """initialize result file"""
-    resultpath = bb.data.getVar('TEST_RESULT', d, 1)
+    resultpath = d.getVar('TEST_RESULT', 1)
     bb.utils.mkdirhier(resultpath)
-    resultfile = os.path.join(resultpath, "testresult.%s" % bb.data.getVar('DATETIME', d, 1))
+    resultfile = os.path.join(resultpath, "testresult.%s" % d.getVar('DATETIME', 1))
     sresultfile = os.path.join(resultpath, "testresult.log")
 
-    machine = bb.data.getVar('MACHINE', d, 1)
+    machine = d.getVar('MACHINE', 1)
 
     if os.path.exists(sresultfile):
         os.remove(sresultfile)
@@ -165,7 +165,7 @@ def qemuimagetest_main(d):
     f.close()
     
     """generate pre-defined testcase list"""
-    testlist = bb.data.getVar('TEST_SCEN', d, 1)
+    testlist = d.getVar('TEST_SCEN', 1)
     fulllist = generate_list(testlist)
 
     """Begin testing"""

+ 39 - 39
meta/classes/insane.bbclass

@@ -105,7 +105,7 @@ ERROR_QA ?= "dev-so debug-deps dev-deps debug-files arch la2 pkgconfig la perms"
 
 def package_qa_clean_path(path,d):
     """ Remove the common prefix from the path. In this case it is the TMPDIR"""
-    return path.replace(bb.data.getVar('TMPDIR',d,True),"")
+    return path.replace(d.getVar('TMPDIR',True),"")
 
 def package_qa_write_error(error, d):
     logfile = d.getVar('QA_LOGFILE', True)
@@ -132,13 +132,13 @@ def package_qa_check_rpath(file,name, d, elf, messages):
     if not elf:
         return
 
-    scanelf = os.path.join(bb.data.getVar('STAGING_BINDIR_NATIVE',d,True),'scanelf')
-    bad_dirs = [bb.data.getVar('TMPDIR', d, True) + "/work", bb.data.getVar('STAGING_DIR_TARGET', d, True)]
-    bad_dir_test = bb.data.getVar('TMPDIR', d, True)
+    scanelf = os.path.join(d.getVar('STAGING_BINDIR_NATIVE',True),'scanelf')
+    bad_dirs = [d.getVar('TMPDIR', True) + "/work", d.getVar('STAGING_DIR_TARGET', True)]
+    bad_dir_test = d.getVar('TMPDIR', True)
     if not os.path.exists(scanelf):
         bb.fatal("Can not check RPATH, scanelf (part of pax-utils-native) not found")
 
-    if not bad_dirs[0] in bb.data.getVar('WORKDIR', d, True):
+    if not bad_dirs[0] in d.getVar('WORKDIR', True):
         bb.fatal("This class assumed that WORKDIR is ${TMPDIR}/work... Not doing any check")
 
     output = os.popen("%s -B -F%%r#F '%s'" % (scanelf,file))
@@ -156,11 +156,11 @@ def package_qa_check_useless_rpaths(file, name, d, elf, messages):
     if not elf:
         return
 
-    objdump = bb.data.getVar('OBJDUMP', d, True)
-    env_path = bb.data.getVar('PATH', d, True)
+    objdump = d.getVar('OBJDUMP', True)
+    env_path = d.getVar('PATH', True)
 
-    libdir = bb.data.getVar("libdir", d, True)
-    base_libdir = bb.data.getVar("base_libdir", d, True)
+    libdir = d.getVar("libdir", True)
+    base_libdir = d.getVar("base_libdir", True)
 
     import re
     rpath_re = re.compile("\s+RPATH\s+(.*)")
@@ -209,8 +209,8 @@ def package_qa_check_arch(path,name,d, elf, messages):
     if not elf:
         return
 
-    target_os   = bb.data.getVar('TARGET_OS',   d, True)
-    target_arch = bb.data.getVar('TARGET_ARCH', d, True)
+    target_os   = d.getVar('TARGET_OS', True)
+    target_arch = d.getVar('TARGET_ARCH', True)
 
     # FIXME: Cross package confuse this check, so just skip them
     for s in ['cross', 'nativesdk', 'cross-canadian']:
@@ -243,7 +243,7 @@ def package_qa_check_desktop(path, name, d, elf, messages):
     Run all desktop files through desktop-file-validate.
     """
     if path.endswith(".desktop"):
-        desktop_file_validate = os.path.join(bb.data.getVar('STAGING_BINDIR_NATIVE',d,True),'desktop-file-validate')
+        desktop_file_validate = os.path.join(d.getVar('STAGING_BINDIR_NATIVE',True),'desktop-file-validate')
         output = os.popen("%s %s" % (desktop_file_validate, path))
         # This only produces output on errors
         for l in output:
@@ -261,14 +261,14 @@ def package_qa_hash_style(path, name, d, elf, messages):
     if os.path.islink(path):
         return
 
-    gnu_hash = "--hash-style=gnu" in bb.data.getVar('LDFLAGS', d, True)
+    gnu_hash = "--hash-style=gnu" in d.getVar('LDFLAGS', True)
     if not gnu_hash:
-        gnu_hash = "--hash-style=both" in bb.data.getVar('LDFLAGS', d, True)
+        gnu_hash = "--hash-style=both" in d.getVar('LDFLAGS', True)
     if not gnu_hash:
         return
 
-    objdump = bb.data.getVar('OBJDUMP', d, True)
-    env_path = bb.data.getVar('PATH', d, True)
+    objdump = d.getVar('OBJDUMP', True)
+    env_path = d.getVar('PATH', True)
 
     sane = False
     has_syms = False
@@ -299,7 +299,7 @@ def package_qa_check_buildpaths(path, name, d, elf, messages):
     if os.path.islink(path):
         return
 
-    tmpdir = bb.data.getVar('TMPDIR', d, True)
+    tmpdir = d.getVar('TMPDIR', True)
     file_content = open(path).read()
     if tmpdir in file_content:
         messages.append("File %s in package contained reference to tmpdir" % package_qa_clean_path(path,d))
@@ -311,9 +311,9 @@ def package_qa_check_license(workdir, d):
     import tempfile
     sane = True
 
-    lic_files = bb.data.getVar('LIC_FILES_CHKSUM', d, True)
-    lic = bb.data.getVar('LICENSE', d, True)
-    pn = bb.data.getVar('PN', d, True)
+    lic_files = d.getVar('LIC_FILES_CHKSUM', True)
+    lic = d.getVar('LICENSE', True)
+    pn = d.getVar('PN', True)
 
     if lic == "CLOSED":
         return True
@@ -324,7 +324,7 @@ def package_qa_check_license(workdir, d):
         bb.error(pn + ": Recipe file does not have license file information (LIC_FILES_CHKSUM)")
         return False
 
-    srcdir = bb.data.getVar('S', d, True)
+    srcdir = d.getVar('S', True)
 
     for url in lic_files.split():
         (type, host, path, user, pswd, parm) = bb.decodeurl(url)
@@ -384,7 +384,7 @@ def package_qa_check_staged(path,d):
     """
 
     sane = True
-    tmpdir = bb.data.getVar('TMPDIR', d, True)
+    tmpdir = d.getVar('TMPDIR', True)
     workdir = os.path.join(tmpdir, "work")
 
     installed = "installed=yes"
@@ -417,8 +417,8 @@ def package_qa_walk(path, warnfuncs, errorfuncs, skip, package, d):
     import oe.qa
 
     #if this will throw an exception, then fix the dict above
-    target_os   = bb.data.getVar('TARGET_OS',   d, True)
-    target_arch = bb.data.getVar('TARGET_ARCH', d, True)
+    target_os   = d.getVar('TARGET_OS', True)
+    target_arch = d.getVar('TARGET_ARCH', True)
 
     warnings = []
     errors = []
@@ -457,19 +457,19 @@ def package_qa_check_rdepends(pkg, pkgdest, skip, d):
         localdata = bb.data.createCopy(d)
         root = "%s/%s" % (pkgdest, pkg)
 
-        bb.data.setVar('ROOT', '', localdata) 
-        bb.data.setVar('ROOT_%s' % pkg, root, localdata)
-        pkgname = bb.data.getVar('PKG_%s' % pkg, localdata, True)
+        localdata.setVar('ROOT', '') 
+        localdata.setVar('ROOT_%s' % pkg, root)
+        pkgname = localdata.getVar('PKG_%s' % pkg, True)
         if not pkgname:
             pkgname = pkg
-        bb.data.setVar('PKG', pkgname, localdata)
+        localdata.setVar('PKG', pkgname)
 
-        bb.data.setVar('OVERRIDES', pkg, localdata)
+        localdata.setVar('OVERRIDES', pkg)
 
         bb.data.update_data(localdata)
 
         # Now check the RDEPENDS
-        rdepends = bb.utils.explode_deps(bb.data.getVar('RDEPENDS', localdata, True) or "")
+        rdepends = bb.utils.explode_deps(localdata.getVar('RDEPENDS', True) or "")
 
 
         # Now do the sanity check!!!
@@ -487,8 +487,8 @@ def package_qa_check_rdepends(pkg, pkgdest, skip, d):
 python do_package_qa () {
     bb.note("DO PACKAGE QA")
 
-    logdir = bb.data.getVar('T', d, True)
-    pkg = bb.data.getVar('PN', d, True)
+    logdir = d.getVar('T', True)
+    pkg = d.getVar('PN', True)
 
     # Check the compile log for host contamination
     compilelog = os.path.join(logdir,"log.do_compile")
@@ -508,8 +508,8 @@ python do_package_qa () {
                 (pkg, installlog))
 
     # Scan the packages...
-    pkgdest = bb.data.getVar('PKGDEST', d, True)
-    packages = bb.data.getVar('PACKAGES',d, True)
+    pkgdest = d.getVar('PKGDEST', True)
+    packages = d.getVar('PACKAGES', True)
 
     # no packages should be scanned
     if not packages:
@@ -521,7 +521,7 @@ python do_package_qa () {
     walk_sane = True
     rdepends_sane = True
     for package in packages.split():
-        skip = (bb.data.getVar('INSANE_SKIP_' + package, d, True) or "").split()
+        skip = (d.getVar('INSANE_SKIP_' + package, True) or "").split()
         if skip:
             bb.note("Package %s skipping QA tests: %s" % (package, str(skip)))
         warnchecks = []
@@ -560,7 +560,7 @@ python do_qa_staging() {
 
 python do_qa_configure() {
     configs = []
-    workdir = bb.data.getVar('WORKDIR', d, True)
+    workdir = d.getVar('WORKDIR', True)
     bb.note("Checking autotools environment for common misconfiguration")
     for root, dirs, files in os.walk(workdir):
         statement = "grep -e 'CROSS COMPILE Badness:' -e 'is unsafe for cross-compilation' %s > /dev/null" % \
@@ -575,8 +575,8 @@ Rerun configure task after fixing this. The path was '%s'""" % root)
         if "configure.in" in files:
             configs.append(os.path.join(root, "configure.in"))
 
-    cnf = bb.data.getVar('EXTRA_OECONF', d, True) or ""
-    if "gettext" not in bb.data.getVar('P', d, True) and "gcc-runtime" not in bb.data.getVar('P', d, True) and "--disable-nls" not in cnf:
+    cnf = d.getVar('EXTRA_OECONF', True) or ""
+    if "gettext" not in d.getVar('P', True) and "gcc-runtime" not in d.getVar('P', True) and "--disable-nls" not in cnf:
        ml = d.getVar("MLPREFIX", True) or ""
        if bb.data.inherits_class('native', d) or bb.data.inherits_class('cross', d) or bb.data.inherits_class('crosssdk', d) or bb.data.inherits_class('nativesdk', d):
           gt = "gettext-native"
@@ -584,7 +584,7 @@ Rerun configure task after fixing this. The path was '%s'""" % root)
           gt = "gettext-nativesdk"
        else:
           gt = "virtual/" + ml + "gettext"
-       deps = bb.utils.explode_deps(bb.data.getVar('DEPENDS', d, True) or "")
+       deps = bb.utils.explode_deps(d.getVar('DEPENDS', True) or "")
        if gt not in deps:
           for config in configs:
               gnu = "grep \"^[[:space:]]*AM_GNU_GETTEXT\" %s >/dev/null" % config

+ 3 - 3
meta/classes/kernel-arch.bbclass

@@ -18,7 +18,7 @@ valid_archs = "alpha cris ia64 \
 def map_kernel_arch(a, d):
 	import re
 
-	valid_archs = bb.data.getVar('valid_archs', d, 1).split()
+	valid_archs = d.getVar('valid_archs', 1).split()
 
 	if   re.match('(i.86|athlon|x86.64)$', a):	return 'x86'
 	elif re.match('arm26$', a):		        return 'arm26'
@@ -32,7 +32,7 @@ def map_kernel_arch(a, d):
 	else:
 		bb.error("cannot map '%s' to a linux kernel architecture" % a)
 
-export ARCH = "${@map_kernel_arch(bb.data.getVar('TARGET_ARCH', d, 1), d)}"
+export ARCH = "${@map_kernel_arch(d.getVar('TARGET_ARCH', 1), d)}"
 
 def map_uboot_arch(a, d):
 	import re
@@ -41,5 +41,5 @@ def map_uboot_arch(a, d):
 	elif re.match('i.86$', a): return 'x86'
 	return a
 
-export UBOOT_ARCH = "${@map_uboot_arch(bb.data.getVar('ARCH', d, 1), d)}"
+export UBOOT_ARCH = "${@map_uboot_arch(d.getVar('ARCH', 1), d)}"
 

+ 1 - 1
meta/classes/kernel-yocto.bbclass

@@ -104,7 +104,7 @@ python do_kernel_configcheck() {
 
     bb.plain("NOTE: validating kernel configuration")
 
-    pathprefix = "export PATH=%s; " % bb.data.getVar('PATH', d, True)
+    pathprefix = "export PATH=%s; " % d.getVar('PATH', True)
     cmd = bb.data.expand("cd ${B}/..; kconf_check -config- ${B} ${S} ${B} ${KBRANCH}",d )
     ret, result = commands.getstatusoutput("%s%s" % (pathprefix, cmd))
 

+ 34 - 34
meta/classes/kernel.bbclass

@@ -11,15 +11,15 @@ INITRAMFS_IMAGE ?= ""
 INITRAMFS_TASK ?= ""
 
 python __anonymous () {
-    kerneltype = bb.data.getVar('KERNEL_IMAGETYPE', d, 1) or ''
+    kerneltype = d.getVar('KERNEL_IMAGETYPE', 1) or ''
     if kerneltype == 'uImage':
-    	depends = bb.data.getVar("DEPENDS", d, 1)
+    	depends = d.getVar("DEPENDS", 1)
     	depends = "%s u-boot-mkimage-native" % depends
-    	bb.data.setVar("DEPENDS", depends, d)
+    	d.setVar("DEPENDS", depends)
 
-    image = bb.data.getVar('INITRAMFS_IMAGE', d, True)
+    image = d.getVar('INITRAMFS_IMAGE', True)
     if image:
-        bb.data.setVar('INITRAMFS_TASK', '${INITRAMFS_IMAGE}:do_rootfs', d)
+        d.setVar('INITRAMFS_TASK', '${INITRAMFS_IMAGE}:do_rootfs')
 }
 
 inherit kernel-arch deploy
@@ -31,7 +31,7 @@ PACKAGES_DYNAMIC += "kernel-firmware-*"
 export OS = "${TARGET_OS}"
 export CROSS_COMPILE = "${TARGET_PREFIX}"
 
-KERNEL_PRIORITY = "${@bb.data.getVar('PV',d,1).split('-')[0].split('.')[-1]}"
+KERNEL_PRIORITY = "${@d.getVar('PV',1).split('-')[0].split('.')[-1]}"
 
 KERNEL_RELEASE ?= "${KERNEL_VERSION}"
 
@@ -56,7 +56,7 @@ KERNEL_IMAGEDEST = "boot"
 #
 # configuration
 #
-export CMDLINE_CONSOLE = "console=${@bb.data.getVar("KERNEL_CONSOLE",d,1) or "ttyS0"}"
+export CMDLINE_CONSOLE = "console=${@d.getVar("KERNEL_CONSOLE",1) or "ttyS0"}"
 
 KERNEL_VERSION = "${@get_kernelversion('${B}')}"
 
@@ -75,7 +75,7 @@ EXTRA_OEMAKE = ""
 
 KERNEL_ALT_IMAGETYPE ??= ""
 
-KERNEL_IMAGETYPE_FOR_MAKE = "${@(lambda s: s[:-3] if s[-3:] == ".gz" else s)(bb.data.getVar('KERNEL_IMAGETYPE', d, 1))}"
+KERNEL_IMAGETYPE_FOR_MAKE = "${@(lambda s: s[:-3] if s[-3:] == ".gz" else s)(d.getVar('KERNEL_IMAGETYPE', 1))}"
 
 kernel_do_compile() {
 	unset CFLAGS CPPFLAGS CXXFLAGS LDFLAGS MACHINE
@@ -306,10 +306,10 @@ module_conf_rfcomm = "alias bt-proto-3 rfcomm"
 python populate_packages_prepend () {
 	def extract_modinfo(file):
 		import tempfile, re
-		tempfile.tempdir = bb.data.getVar("WORKDIR", d, 1)
+		tempfile.tempdir = d.getVar("WORKDIR", 1)
 		tf = tempfile.mkstemp()
 		tmpfile = tf[1]
-		cmd = "PATH=\"%s\" %sobjcopy -j .modinfo -O binary %s %s" % (bb.data.getVar("PATH", d, 1), bb.data.getVar("HOST_PREFIX", d, 1) or "", file, tmpfile)
+		cmd = "PATH=\"%s\" %sobjcopy -j .modinfo -O binary %s %s" % (d.getVar("PATH", 1), d.getVar("HOST_PREFIX", 1) or "", file, tmpfile)
 		os.system(cmd)
 		f = open(tmpfile)
 		l = f.read().split("\000")
@@ -328,18 +328,18 @@ python populate_packages_prepend () {
 	def parse_depmod():
 		import re
 
-		dvar = bb.data.getVar('PKGD', d, 1)
+		dvar = d.getVar('PKGD', 1)
 		if not dvar:
 			bb.error("PKGD not defined")
 			return
 
-		kernelver = bb.data.getVar('KERNEL_VERSION', d, 1)
+		kernelver = d.getVar('KERNEL_VERSION', 1)
 		kernelver_stripped = kernelver
 		m = re.match('^(.*-hh.*)[\.\+].*$', kernelver)
 		if m:
 			kernelver_stripped = m.group(1)
-		path = bb.data.getVar("PATH", d, 1)
-		host_prefix = bb.data.getVar("HOST_PREFIX", d, 1) or ""
+		path = d.getVar("PATH", 1)
+		host_prefix = d.getVar("HOST_PREFIX", 1) or ""
 
 		cmd = "PATH=\"%s\" %sdepmod -n -a -r -b %s -F %s/boot/System.map-%s %s" % (path, host_prefix, dvar, dvar, kernelver, kernelver_stripped)
 		f = os.popen(cmd, 'r')
@@ -377,9 +377,9 @@ python populate_packages_prepend () {
 	
 	def get_dependencies(file, pattern, format):
                 # file no longer includes PKGD
-		file = file.replace(bb.data.getVar('PKGD', d, 1) or '', '', 1)
+		file = file.replace(d.getVar('PKGD', 1) or '', '', 1)
                 # instead is prefixed with /lib/modules/${KERNEL_VERSION}
-                file = file.replace("/lib/modules/%s/" % bb.data.getVar('KERNEL_VERSION', d, 1) or '', '', 1)
+                file = file.replace("/lib/modules/%s/" % d.getVar('KERNEL_VERSION', 1) or '', '', 1)
 
 		if module_deps.has_key(file):
 			import re
@@ -398,40 +398,40 @@ python populate_packages_prepend () {
 		import re
 		vals = extract_modinfo(file)
 
-		dvar = bb.data.getVar('PKGD', d, 1)
+		dvar = d.getVar('PKGD', 1)
 
 		# If autoloading is requested, output /etc/modutils/<name> and append
 		# appropriate modprobe commands to the postinst
-		autoload = bb.data.getVar('module_autoload_%s' % basename, d, 1)
+		autoload = d.getVar('module_autoload_%s' % basename, 1)
 		if autoload:
 			name = '%s/etc/modutils/%s' % (dvar, basename)
 			f = open(name, 'w')
 			for m in autoload.split():
 				f.write('%s\n' % m)
 			f.close()
-			postinst = bb.data.getVar('pkg_postinst_%s' % pkg, d, 1)
+			postinst = d.getVar('pkg_postinst_%s' % pkg, 1)
 			if not postinst:
 				bb.fatal("pkg_postinst_%s not defined" % pkg)
-			postinst += bb.data.getVar('autoload_postinst_fragment', d, 1) % autoload
-			bb.data.setVar('pkg_postinst_%s' % pkg, postinst, d)
+			postinst += d.getVar('autoload_postinst_fragment', 1) % autoload
+			d.setVar('pkg_postinst_%s' % pkg, postinst)
 
 		# Write out any modconf fragment
-		modconf = bb.data.getVar('module_conf_%s' % basename, d, 1)
+		modconf = d.getVar('module_conf_%s' % basename, 1)
 		if modconf:
 			name = '%s/etc/modprobe.d/%s.conf' % (dvar, basename)
 			f = open(name, 'w')
 			f.write("%s\n" % modconf)
 			f.close()
 
-		files = bb.data.getVar('FILES_%s' % pkg, d, 1)
+		files = d.getVar('FILES_%s' % pkg, 1)
 		files = "%s /etc/modutils/%s /etc/modutils/%s.conf /etc/modprobe.d/%s.conf" % (files, basename, basename, basename)
-		bb.data.setVar('FILES_%s' % pkg, files, d)
+		d.setVar('FILES_%s' % pkg, files)
 
 		if vals.has_key("description"):
-			old_desc = bb.data.getVar('DESCRIPTION_' + pkg, d, 1) or ""
-			bb.data.setVar('DESCRIPTION_' + pkg, old_desc + "; " + vals["description"], d)
+			old_desc = d.getVar('DESCRIPTION_' + pkg, 1) or ""
+			d.setVar('DESCRIPTION_' + pkg, old_desc + "; " + vals["description"])
 
-		rdepends_str = bb.data.getVar('RDEPENDS_' + pkg, d, 1)
+		rdepends_str = d.getVar('RDEPENDS_' + pkg, 1)
 		if rdepends_str:
 			rdepends = rdepends_str.split()
 		else:
@@ -443,29 +443,29 @@ python populate_packages_prepend () {
 	module_regex = '^(.*)\.k?o$'
 	module_pattern = 'kernel-module-%s'
 
-	postinst = bb.data.getVar('pkg_postinst_modules', d, 1)
-	postrm = bb.data.getVar('pkg_postrm_modules', d, 1)
+	postinst = d.getVar('pkg_postinst_modules', 1)
+	postrm = d.getVar('pkg_postrm_modules', 1)
 	do_split_packages(d, root='/lib/firmware', file_regex='^(.*)\.bin$', output_pattern='kernel-firmware-%s', description='Firmware for %s', recursive=True, extra_depends='')
 	do_split_packages(d, root='/lib/firmware', file_regex='^(.*)\.fw$', output_pattern='kernel-firmware-%s', description='Firmware for %s', recursive=True, extra_depends='')
 	do_split_packages(d, root='/lib/firmware', file_regex='^(.*)\.cis$', output_pattern='kernel-firmware-%s', description='Firmware for %s', recursive=True, extra_depends='')
-	do_split_packages(d, root='/lib/modules', file_regex=module_regex, output_pattern=module_pattern, description='%s kernel module', postinst=postinst, postrm=postrm, recursive=True, hook=frob_metadata, extra_depends='update-modules kernel-%s' % bb.data.getVar("KERNEL_VERSION", d, 1))
+	do_split_packages(d, root='/lib/modules', file_regex=module_regex, output_pattern=module_pattern, description='%s kernel module', postinst=postinst, postrm=postrm, recursive=True, hook=frob_metadata, extra_depends='update-modules kernel-%s' % d.getVar("KERNEL_VERSION", 1))
 
 	import re
 	metapkg = "kernel-modules"
-	bb.data.setVar('ALLOW_EMPTY_' + metapkg, "1", d)
-	bb.data.setVar('FILES_' + metapkg, "", d)
+	d.setVar('ALLOW_EMPTY_' + metapkg, "1")
+	d.setVar('FILES_' + metapkg, "")
 	blacklist = [ 'kernel-dev', 'kernel-image', 'kernel-base', 'kernel-vmlinux', 'perf', 'perf-dbg', 'kernel-misc' ]
 	for l in module_deps.values():
 		for i in l:
 			pkg = module_pattern % legitimize_package_name(re.match(module_regex, os.path.basename(i)).group(1))
 			blacklist.append(pkg)
 	metapkg_rdepends = []
-	packages = bb.data.getVar('PACKAGES', d, 1).split()
+	packages = d.getVar('PACKAGES', 1).split()
 	for pkg in packages[1:]:
 		if not pkg in blacklist and not pkg in metapkg_rdepends:
 			metapkg_rdepends.append(pkg)
 	bb.data.setVar('RDEPENDS_' + metapkg, ' '.join(metapkg_rdepends), d)
-	bb.data.setVar('DESCRIPTION_' + metapkg, 'Kernel modules meta package', d)
+	d.setVar('DESCRIPTION_' + metapkg, 'Kernel modules meta package')
 	packages.append(metapkg)
 	bb.data.setVar('PACKAGES', ' '.join(packages), d)
 }

+ 5 - 5
meta/classes/libc-common.bbclass

@@ -18,13 +18,13 @@ do_install() {
 }
 
 def get_libc_fpu_setting(bb, d):
-    if bb.data.getVar('TARGET_FPU', d, 1) in [ 'soft' ]:
+    if d.getVar('TARGET_FPU', 1) in [ 'soft' ]:
         return "--without-fp"
     return ""
 
 python populate_packages_prepend () {
-	if bb.data.getVar('DEBIAN_NAMES', d, 1):
-		bpn = bb.data.getVar('BPN', d, 1)
-		bb.data.setVar('PKG_'+bpn, 'libc6', d)
-		bb.data.setVar('PKG_'+bpn+'-dev', 'libc6-dev', d)
+	if d.getVar('DEBIAN_NAMES', 1):
+		bpn = d.getVar('BPN', 1)
+		d.setVar('PKG_'+bpn, 'libc6')
+		d.setVar('PKG_'+bpn+'-dev', 'libc6-dev')
 }

+ 33 - 33
meta/classes/libc-package.bbclass

@@ -10,7 +10,7 @@
 GLIBC_INTERNAL_USE_BINARY_LOCALE ?= "ondevice"
 
 python __anonymous () {
-    enabled = bb.data.getVar("ENABLE_BINARY_LOCALE_GENERATION", d, 1)
+    enabled = d.getVar("ENABLE_BINARY_LOCALE_GENERATION", 1)
 
     pn = d.getVar("PN", True)
     if pn.endswith("-initial"):
@@ -19,21 +19,21 @@ python __anonymous () {
     if enabled and int(enabled):
         import re
 
-        target_arch = bb.data.getVar("TARGET_ARCH", d, 1)
-        binary_arches = bb.data.getVar("BINARY_LOCALE_ARCHES", d, 1) or ""
-        use_cross_localedef = bb.data.getVar("LOCALE_GENERATION_WITH_CROSS-LOCALEDEF", d, 1) or ""
+        target_arch = d.getVar("TARGET_ARCH", 1)
+        binary_arches = d.getVar("BINARY_LOCALE_ARCHES", 1) or ""
+        use_cross_localedef = d.getVar("LOCALE_GENERATION_WITH_CROSS-LOCALEDEF", 1) or ""
 
         for regexp in binary_arches.split(" "):
             r = re.compile(regexp)
 
             if r.match(target_arch):
-                depends = bb.data.getVar("DEPENDS", d, 1)
+                depends = d.getVar("DEPENDS", 1)
 		if use_cross_localedef == "1" :
 	                depends = "%s cross-localedef-native" % depends
 		else:
 	                depends = "%s qemu-native" % depends
-                bb.data.setVar("DEPENDS", depends, d)
-                bb.data.setVar("GLIBC_INTERNAL_USE_BINARY_LOCALE", "compile", d)
+                d.setVar("DEPENDS", depends)
+                d.setVar("GLIBC_INTERNAL_USE_BINARY_LOCALE", "compile")
                 break
 }
 
@@ -109,19 +109,19 @@ inherit qemu
 
 python package_do_split_gconvs () {
 	import os, re
-	if (bb.data.getVar('PACKAGE_NO_GCONV', d, 1) == '1'):
+	if (d.getVar('PACKAGE_NO_GCONV', 1) == '1'):
 		bb.note("package requested not splitting gconvs")
 		return
 
-	if not bb.data.getVar('PACKAGES', d, 1):
+	if not d.getVar('PACKAGES', 1):
 		return
 
-	bpn = bb.data.getVar('BPN', d, 1)
-	libdir = bb.data.getVar('libdir', d, 1)
+	bpn = d.getVar('BPN', 1)
+	libdir = d.getVar('libdir', 1)
 	if not libdir:
 		bb.error("libdir not defined")
 		return
-	datadir = bb.data.getVar('datadir', d, 1)
+	datadir = d.getVar('datadir', 1)
 	if not datadir:
 		bb.error("datadir not defined")
 		return
@@ -191,17 +191,17 @@ python package_do_split_gconvs () {
 
 	do_split_packages(d, locales_dir, file_regex='(.*)', output_pattern=bpn+'-localedata-%s', \
 		description='locale definition for %s', hook=calc_locale_deps, extra_depends='')
-	bb.data.setVar('PACKAGES', bb.data.getVar('PACKAGES', d) + ' ' + bb.data.getVar('MLPREFIX', d) + bpn + '-gconv', d)
+	bb.data.setVar('PACKAGES', d.getVar('PACKAGES') + ' ' + d.getVar('MLPREFIX') + bpn + '-gconv', d)
 
-	use_bin = bb.data.getVar("GLIBC_INTERNAL_USE_BINARY_LOCALE", d, 1)
+	use_bin = d.getVar("GLIBC_INTERNAL_USE_BINARY_LOCALE", 1)
 
 	dot_re = re.compile("(.*)\.(.*)")
 
 #GLIBC_GENERATE_LOCALES var specifies which locales to be supported, empty or "all" means all locales 
 	if use_bin != "precompiled":
-		supported = bb.data.getVar('GLIBC_GENERATE_LOCALES', d, 1)
+		supported = d.getVar('GLIBC_GENERATE_LOCALES', 1)
 		if not supported or supported == "all":
-			f = open(base_path_join(bb.data.getVar('WORKDIR', d, 1), "SUPPORTED"), "r")
+			f = open(base_path_join(d.getVar('WORKDIR', 1), "SUPPORTED"), "r")
 			supported = f.readlines()
 			f.close()
 		else:
@@ -209,7 +209,7 @@ python package_do_split_gconvs () {
 			supported = map(lambda s:s.replace(".", " ") + "\n", supported)
 	else:
 		supported = []
-		full_bin_path = bb.data.getVar('PKGD', d, True) + binary_locales_dir
+		full_bin_path = d.getVar('PKGD', True) + binary_locales_dir
 		for dir in os.listdir(full_bin_path):
 			dbase = dir.split(".")
 			d2 = "  "
@@ -218,7 +218,7 @@ python package_do_split_gconvs () {
 			supported.append(dbase[0] + d2)
 
 	# Collate the locales by base and encoding
-	utf8_only = int(bb.data.getVar('LOCALE_UTF8_ONLY', d, 1) or 0)
+	utf8_only = int(d.getVar('LOCALE_UTF8_ONLY', 1) or 0)
 	encodings = {}
 	for l in supported:
 		l = l[:-1]
@@ -235,9 +235,9 @@ python package_do_split_gconvs () {
 	def output_locale_source(name, pkgname, locale, encoding):
 		bb.data.setVar('RDEPENDS_%s' % pkgname, 'localedef %s-localedata-%s %s-charmap-%s' % \
 		(bpn, legitimize_package_name(locale), bpn, legitimize_package_name(encoding)), d)
-		bb.data.setVar('pkg_postinst_%s' % pkgname, bb.data.getVar('locale_base_postinst', d, 1) \
+		bb.data.setVar('pkg_postinst_%s' % pkgname, d.getVar('locale_base_postinst', 1) \
 		% (locale, encoding, locale), d)
-		bb.data.setVar('pkg_postrm_%s' % pkgname, bb.data.getVar('locale_base_postrm', d, 1) % \
+		bb.data.setVar('pkg_postrm_%s' % pkgname, d.getVar('locale_base_postrm', 1) % \
 		(locale, encoding, locale), d)
 
 	def output_locale_binary_rdepends(name, pkgname, locale, encoding):
@@ -248,23 +248,23 @@ python package_do_split_gconvs () {
 			libc_name = name
 		bb.data.setVar('RDEPENDS_%s' % pkgname, legitimize_package_name('%s-binary-localedata-%s' \
 			% (bpn, libc_name)), d)
-		rprovides = (bb.data.getVar('RPROVIDES_%s' % pkgname, d, True) or "").split()
+		rprovides = (d.getVar('RPROVIDES_%s' % pkgname, True) or "").split()
 		rprovides.append(legitimize_package_name('%s-binary-localedata-%s' % (bpn, libc_name)))
 		bb.data.setVar('RPROVIDES_%s' % pkgname, " ".join(rprovides), d)
 
 	commands = {}
 
 	def output_locale_binary(name, pkgname, locale, encoding):
-		treedir = base_path_join(bb.data.getVar("WORKDIR", d, 1), "locale-tree")
-		ldlibdir = base_path_join(treedir, bb.data.getVar("base_libdir", d, 1))
-		path = bb.data.getVar("PATH", d, 1)
+		treedir = base_path_join(d.getVar("WORKDIR", 1), "locale-tree")
+		ldlibdir = base_path_join(treedir, d.getVar("base_libdir", 1))
+		path = d.getVar("PATH", 1)
 		i18npath = base_path_join(treedir, datadir, "i18n")
 		gconvpath = base_path_join(treedir, "iconvdata")
 		outputpath = base_path_join(treedir, libdir, "locale")
 
-		use_cross_localedef = bb.data.getVar("LOCALE_GENERATION_WITH_CROSS-LOCALEDEF", d, 1) or "0"
+		use_cross_localedef = d.getVar("LOCALE_GENERATION_WITH_CROSS-LOCALEDEF", 1) or "0"
 		if use_cross_localedef == "1":
-	    		target_arch = bb.data.getVar('TARGET_ARCH', d, True)
+	    		target_arch = d.getVar('TARGET_ARCH', True)
 			locale_arch_options = { \
 				"arm":     " --uint32-align=4 --little-endian ", \
 				"powerpc": " --uint32-align=4 --big-endian ",    \
@@ -292,9 +292,9 @@ python package_do_split_gconvs () {
 				--inputfile=%s/i18n/locales/%s --charmap=%s %s" \
 				% (treedir, datadir, locale, encoding, name)
 
-			qemu_options = bb.data.getVar("QEMU_OPTIONS_%s" % bb.data.getVar('PACKAGE_ARCH', d, 1), d, 1)
+			qemu_options = bb.data.getVar("QEMU_OPTIONS_%s" % d.getVar('PACKAGE_ARCH', 1), d, 1)
 			if not qemu_options:
-				qemu_options = bb.data.getVar('QEMU_OPTIONS', d, 1)
+				qemu_options = d.getVar('QEMU_OPTIONS', 1)
 
 			cmd = "PSEUDO_RELOADED=YES PATH=\"%s\" I18NPATH=\"%s\" %s -L %s \
 				-E LD_LIBRARY_PATH=%s %s %s/bin/localedef %s" % \
@@ -305,14 +305,14 @@ python package_do_split_gconvs () {
 		bb.note("generating locale %s (%s)" % (locale, encoding))
 
 	def output_locale(name, locale, encoding):
-		pkgname = bb.data.getVar('MLPREFIX', d) + 'locale-base-' + legitimize_package_name(name)
-		bb.data.setVar('ALLOW_EMPTY_%s' % pkgname, '1', d)
-		bb.data.setVar('PACKAGES', '%s %s' % (pkgname, bb.data.getVar('PACKAGES', d, 1)), d)
+		pkgname = d.getVar('MLPREFIX') + 'locale-base-' + legitimize_package_name(name)
+		d.setVar('ALLOW_EMPTY_%s' % pkgname, '1')
+		bb.data.setVar('PACKAGES', '%s %s' % (pkgname, d.getVar('PACKAGES', 1)), d)
 		rprovides = ' virtual-locale-%s' % legitimize_package_name(name)
 		m = re.match("(.*)_(.*)", name)
 		if m:
 			rprovides += ' virtual-locale-%s' % m.group(1)
-		bb.data.setVar('RPROVIDES_%s' % pkgname, rprovides, d)
+		d.setVar('RPROVIDES_%s' % pkgname, rprovides)
 
 		if use_bin == "compile":
 			output_locale_binary_rdepends(name, pkgname, locale, encoding)
@@ -347,7 +347,7 @@ python package_do_split_gconvs () {
 		bb.note("  " + " ".join(non_utf8))
 
 	if use_bin == "compile":
-		makefile = base_path_join(bb.data.getVar("WORKDIR", d, 1), "locale-tree", "Makefile")
+		makefile = base_path_join(d.getVar("WORKDIR", 1), "locale-tree", "Makefile")
 		m = open(makefile, "w")
 		m.write("all: %s\n\n" % " ".join(commands.keys()))
 		for cmd in commands:

+ 9 - 9
meta/classes/license.bbclass

@@ -104,10 +104,10 @@ python do_populate_lic() {
 
         # If the generic does not exist we need to check to see if there is an SPDX mapping to it
         if not os.path.isfile(os.path.join(generic_directory, license_type)):
-            if bb.data.getVarFlag('SPDXLICENSEMAP', license_type, d) != None:
+            if d.getVarFlag('SPDXLICENSEMAP', license_type) != None:
                 # Great, there is an SPDXLICENSEMAP. We can copy!
                 bb.note("We need to use a SPDXLICENSEMAP for %s" % (license_type))
-                spdx_generic = bb.data.getVarFlag('SPDXLICENSEMAP', license_type, d)
+                spdx_generic = d.getVarFlag('SPDXLICENSEMAP', license_type)
                 copy_license(generic_directory, gen_lic_dest, spdx_generic)            
                 link_license(gen_lic_dest, destdir, spdx_generic)            
             else:
@@ -120,16 +120,16 @@ python do_populate_lic() {
             link_license(gen_lic_dest, destdir, license_type)            
 
     # All the license types for the package
-    license_types = bb.data.getVar('LICENSE', d, True)
+    license_types = d.getVar('LICENSE', True)
     # All the license files for the package
-    lic_files = bb.data.getVar('LIC_FILES_CHKSUM', d, True)
-    pn = bb.data.getVar('PN', d, True)
+    lic_files = d.getVar('LIC_FILES_CHKSUM', True)
+    pn = d.getVar('PN', True)
     # The base directory we wrangle licenses to
-    destdir = os.path.join(bb.data.getVar('LICSSTATEDIR', d, True), pn)
+    destdir = os.path.join(d.getVar('LICSSTATEDIR', True), pn)
     # The license files are located in S/LIC_FILE_CHECKSUM.
-    srcdir = bb.data.getVar('S', d, True)
+    srcdir = d.getVar('S', True)
     # Directory we store the generic licenses as set in the distro configuration
-    generic_directory = bb.data.getVar('COMMON_LICENSE_DIR', d, True)
+    generic_directory = d.getVar('COMMON_LICENSE_DIR', True)
     
     try:
         bb.mkdirhier(destdir)
@@ -154,7 +154,7 @@ python do_populate_lic() {
         if ret is False or ret == 0:
             bb.warn("%s could not be copied for some reason. It may not exist. WARN for now." % srclicfile)
  
-    gen_lic_dest = os.path.join(bb.data.getVar('LICENSE_DIRECTORY', d, True), "common-licenses")
+    gen_lic_dest = os.path.join(d.getVar('LICENSE_DIRECTORY', True), "common-licenses")
     
     clean_licenses = ""
 

+ 1 - 1
meta/classes/metadata_scm.bbclass

@@ -27,7 +27,7 @@ def base_detect_branch(d):
 	return "<unknown>"	
 
 def base_get_scmbasepath(d):
-	return bb.data.getVar( 'COREBASE', d, 1 )
+	return d.getVar( 'COREBASE', 1 )
 
 def base_get_metadata_monotone_branch(path, d):
 	monotone_branch = "<unknown>"

+ 6 - 6
meta/classes/native.bbclass

@@ -98,18 +98,18 @@ python native_virtclass_handler () {
     if not isinstance(e, bb.event.RecipePreFinalise):
         return
 
-    classextend = bb.data.getVar('BBCLASSEXTEND', e.data, True) or ""
+    classextend = e.data.getVar('BBCLASSEXTEND', True) or ""
     if "native" not in classextend:
         return
 
-    pn = bb.data.getVar("PN", e.data, True)
+    pn = e.data.getVar("PN", True)
     if not pn.endswith("-native"):
         return
 
     def map_dependencies(varname, d, suffix = ""):
         if suffix:
             varname = varname + "_" + suffix
-        deps = bb.data.getVar(varname, d, True)
+        deps = d.getVar(varname, True)
         if not deps:
             return
         deps = bb.utils.explode_deps(deps)
@@ -131,15 +131,15 @@ python native_virtclass_handler () {
         map_dependencies("RPROVIDES", e.data, pkg)
         map_dependencies("RREPLACES", e.data, pkg)
 
-    provides = bb.data.getVar("PROVIDES", e.data, True)
+    provides = e.data.getVar("PROVIDES", True)
     for prov in provides.split():
         if prov.find(pn) != -1:
             continue
         if not prov.endswith("-native"):
             provides = provides.replace(prov, prov + "-native")
-    bb.data.setVar("PROVIDES", provides, e.data)
+    e.data.setVar("PROVIDES", provides)
 
-    bb.data.setVar("OVERRIDES", bb.data.getVar("OVERRIDES", e.data, False) + ":virtclass-native", e.data)
+    bb.data.setVar("OVERRIDES", e.data.getVar("OVERRIDES", False) + ":virtclass-native", e.data)
 }
 
 addhandler native_virtclass_handler

+ 7 - 7
meta/classes/nativesdk.bbclass

@@ -11,7 +11,7 @@ STAGING_BINDIR_TOOLCHAIN = "${STAGING_DIR_NATIVE}${bindir_native}/${SDK_ARCH}${S
 #
 PACKAGE_ARCH = "${SDK_ARCH}-nativesdk"
 python () {
-    archs = bb.data.getVar('PACKAGE_ARCHS', d, True).split()
+    archs = d.getVar('PACKAGE_ARCHS', True).split()
     sdkarchs = []
     for arch in archs:
         sdkarchs.append(arch + '-nativesdk')
@@ -62,22 +62,22 @@ python nativesdk_virtclass_handler () {
     if not isinstance(e, bb.event.RecipePreFinalise):
         return
 
-    pn = bb.data.getVar("PN", e.data, True)
+    pn = e.data.getVar("PN", True)
     if not pn.endswith("-nativesdk"):
         return
 
-    bb.data.setVar("OVERRIDES", bb.data.getVar("OVERRIDES", e.data, False) + ":virtclass-nativesdk", e.data)
+    bb.data.setVar("OVERRIDES", e.data.getVar("OVERRIDES", False) + ":virtclass-nativesdk", e.data)
 }
 
 python () {
-    pn = bb.data.getVar("PN", d, True)
+    pn = d.getVar("PN", True)
     if not pn.endswith("-nativesdk"):
         return
 
     def map_dependencies(varname, d, suffix = ""):
         if suffix:
             varname = varname + "_" + suffix
-        deps = bb.data.getVar(varname, d, True)
+        deps = d.getVar(varname, True)
         if not deps:
             return
         deps = bb.utils.explode_deps(deps)
@@ -101,13 +101,13 @@ python () {
     #    map_dependencies("RPROVIDES", d, pkg)
     #    map_dependencies("RREPLACES", d, pkg)
 
-    provides = bb.data.getVar("PROVIDES", d, True)
+    provides = d.getVar("PROVIDES", True)
     for prov in provides.split():
         if prov.find(pn) != -1:
             continue
         if not prov.endswith("-nativesdk"):
             provides = provides.replace(prov, prov + "-nativesdk")
-    bb.data.setVar("PROVIDES", provides, d)
+    d.setVar("PROVIDES", provides)
 }
 
 addhandler nativesdk_virtclass_handler

+ 116 - 116
meta/classes/package.bbclass

@@ -88,9 +88,9 @@ def do_split_packages(d, root, file_regex, output_pattern, description, postinst
 		if newdeps:
 			extra_depends = " ".join(newdeps)
 
-	dvar = bb.data.getVar('PKGD', d, True)
+	dvar = d.getVar('PKGD', True)
 
-	packages = bb.data.getVar('PACKAGES', d, True).split()
+	packages = d.getVar('PACKAGES', True).split()
 
 	if postinst:
 		postinst = '#!/bin/sh\n' + postinst + '\n'
@@ -136,7 +136,7 @@ def do_split_packages(d, root, file_regex, output_pattern, description, postinst
 				packages = [pkg] + packages
 			else:
 				packages.append(pkg)
-		oldfiles = bb.data.getVar('FILES_' + pkg, d, True)
+		oldfiles = d.getVar('FILES_' + pkg, True)
 		if not oldfiles:
 			the_files = [os.path.join(root, o)]
 			if aux_files_pattern:
@@ -153,17 +153,17 @@ def do_split_packages(d, root, file_regex, output_pattern, description, postinst
 					the_files.append(aux_files_pattern_verbatim % m.group(1))
 			bb.data.setVar('FILES_' + pkg, " ".join(the_files), d)
 			if extra_depends != '':
-				the_depends = bb.data.getVar('RDEPENDS_' + pkg, d, True)
+				the_depends = d.getVar('RDEPENDS_' + pkg, True)
 				if the_depends:
 					the_depends = '%s %s' % (the_depends, extra_depends)
 				else:
 					the_depends = extra_depends
-				bb.data.setVar('RDEPENDS_' + pkg, the_depends, d)
-			bb.data.setVar('DESCRIPTION_' + pkg, description % on, d)
+				d.setVar('RDEPENDS_' + pkg, the_depends)
+			d.setVar('DESCRIPTION_' + pkg, description % on)
 			if postinst:
-				bb.data.setVar('pkg_postinst_' + pkg, postinst, d)
+				d.setVar('pkg_postinst_' + pkg, postinst)
 			if postrm:
-				bb.data.setVar('pkg_postrm_' + pkg, postrm, d)
+				d.setVar('pkg_postrm_' + pkg, postrm)
 		else:
 			bb.data.setVar('FILES_' + pkg, oldfiles + " " + os.path.join(root, o), d)
 		if callable(hook):
@@ -174,13 +174,13 @@ def do_split_packages(d, root, file_regex, output_pattern, description, postinst
 PACKAGE_DEPENDS += "file-native"
 
 python () {
-    if bb.data.getVar('PACKAGES', d, True) != '':
-        deps = bb.data.getVarFlag('do_package', 'depends', d) or ""
-        for dep in (bb.data.getVar('PACKAGE_DEPENDS', d, True) or "").split():
+    if d.getVar('PACKAGES', True) != '':
+        deps = d.getVarFlag('do_package', 'depends') or ""
+        for dep in (d.getVar('PACKAGE_DEPENDS', True) or "").split():
             deps += " %s:do_populate_sysroot" % dep
-        bb.data.setVarFlag('do_package', 'depends', deps, d)
+        d.setVarFlag('do_package', 'depends', deps)
 
-        deps = (bb.data.getVarFlag('do_package', 'deptask', d) or "").split()
+        deps = (d.getVarFlag('do_package', 'deptask') or "").split()
         # shlibs requires any DEPENDS to have already packaged for the *.list files
         deps.append("do_package")
         bb.data.setVarFlag('do_package', 'deptask', " ".join(deps), d)
@@ -198,9 +198,9 @@ def splitfile(file, debugfile, debugsrcdir, d):
 
     import commands, stat
 
-    dvar = bb.data.getVar('PKGD', d, True)
-    pathprefix = "export PATH=%s; " % bb.data.getVar('PATH', d, True)
-    objcopy = bb.data.getVar("OBJCOPY", d, True)
+    dvar = d.getVar('PKGD', True)
+    pathprefix = "export PATH=%s; " % d.getVar('PATH', True)
+    objcopy = d.getVar("OBJCOPY", True)
     debugedit = bb.data.expand("${STAGING_LIBDIR_NATIVE}/rpm/bin/debugedit", d)
     workdir = bb.data.expand("${WORKDIR}", d)
     workparentdir = os.path.dirname(workdir)
@@ -240,10 +240,10 @@ def splitfile2(debugsrcdir, d):
 
     import commands, stat
 
-    dvar = bb.data.getVar('PKGD', d, True)
-    pathprefix = "export PATH=%s; " % bb.data.getVar('PATH', d, True)
-    strip = bb.data.getVar("STRIP", d, True)
-    objcopy = bb.data.getVar("OBJCOPY", d, True)
+    dvar = d.getVar('PKGD', True)
+    pathprefix = "export PATH=%s; " % d.getVar('PATH', True)
+    strip = d.getVar("STRIP", True)
+    objcopy = d.getVar("OBJCOPY", True)
     debugedit = bb.data.expand("${STAGING_LIBDIR_NATIVE}/rpm/bin/debugedit", d)
     workdir = bb.data.expand("${WORKDIR}", d)
     workparentdir = os.path.dirname(workdir)
@@ -279,8 +279,8 @@ def runstrip(file, elftype, d):
 
     import commands, stat
 
-    pathprefix = "export PATH=%s; " % bb.data.getVar('PATH', d, True)
-    strip = bb.data.getVar("STRIP", d, True)
+    pathprefix = "export PATH=%s; " % d.getVar('PATH', True)
+    strip = d.getVar("STRIP", True)
 
     # Handle kernel modules specifically - .debug directories here are pointless
     if file.find("/lib/modules/") != -1 and file.endswith(".ko"):
@@ -329,10 +329,10 @@ def get_package_mapping (pkg, d):
 	return pkg
 
 def runtime_mapping_rename (varname, d):
-	#bb.note("%s before: %s" % (varname, bb.data.getVar(varname, d, True)))	
+	#bb.note("%s before: %s" % (varname, d.getVar(varname, True)))	
 
 	new_depends = []
-	deps = bb.utils.explode_dep_versions(bb.data.getVar(varname, d, True) or "")
+	deps = bb.utils.explode_dep_versions(d.getVar(varname, True) or "")
 	for depend in deps:
 		# Have to be careful with any version component of the depend
 		new_depend = get_package_mapping(depend, d)
@@ -343,7 +343,7 @@ def runtime_mapping_rename (varname, d):
 
 	bb.data.setVar(varname, " ".join(new_depends) or None, d)
 
-	#bb.note("%s after: %s" % (varname, bb.data.getVar(varname, d, True)))
+	#bb.note("%s after: %s" % (varname, d.getVar(varname, True)))
 
 #
 # Package functions suitable for inclusion in PACKAGEFUNCS
@@ -359,19 +359,19 @@ python package_get_auto_pr() {
 }
 
 python package_do_split_locales() {
-	if (bb.data.getVar('PACKAGE_NO_LOCALE', d, True) == '1'):
+	if (d.getVar('PACKAGE_NO_LOCALE', True) == '1'):
 		bb.debug(1, "package requested not splitting locales")
 		return
 
-	packages = (bb.data.getVar('PACKAGES', d, True) or "").split()
+	packages = (d.getVar('PACKAGES', True) or "").split()
 
-	datadir = bb.data.getVar('datadir', d, True)
+	datadir = d.getVar('datadir', True)
 	if not datadir:
 		bb.note("datadir not defined")
 		return
 
-	dvar = bb.data.getVar('PKGD', d, True)
-	pn = bb.data.getVar('PN', d, True)
+	dvar = d.getVar('PKGD', True)
+	pn = d.getVar('PN', True)
 
 	if pn + '-locale' in packages:
 		packages.remove(pn + '-locale')
@@ -392,9 +392,9 @@ python package_do_split_locales() {
 	if mainpkg.find('-dev'):
 		mainpkg = mainpkg.replace('-dev', '')
 
-	summary = bb.data.getVar('SUMMARY', d, True) or pn
-	description = bb.data.getVar('DESCRIPTION', d, True) or "" 
-        locale_section = bb.data.getVar('LOCALE_SECTION', d, True)
+	summary = d.getVar('SUMMARY', True) or pn
+	description = d.getVar('DESCRIPTION', True) or "" 
+        locale_section = d.getVar('LOCALE_SECTION', True)
 	for l in locales:
 		ln = legitimize_package_name(l)
 		pkg = pn + '-locale-' + ln
@@ -405,7 +405,7 @@ python package_do_split_locales() {
 		bb.data.setVar('SUMMARY_' + pkg, '%s - %s translations' % (summary, l), d)
 		bb.data.setVar('DESCRIPTION_' + pkg, '%s  This package contains language translation files for the %s locale.' % (description, l), d)
 		if locale_section:
-			bb.data.setVar('SECTION_' + pkg, locale_section, d)
+			d.setVar('SECTION_' + pkg, locale_section)
 
 	bb.data.setVar('PACKAGES', ' '.join(packages), d)
 
@@ -415,14 +415,14 @@ python package_do_split_locales() {
 	# glibc-localedata-translit* won't install as a dependency
 	# for some other package which breaks meta-toolchain
 	# Probably breaks since virtual-locale- isn't provided anywhere
-	#rdep = (bb.data.getVar('RDEPENDS_%s' % mainpkg, d, True) or bb.data.getVar('RDEPENDS', d, True) or "").split()
+	#rdep = (d.getVar('RDEPENDS_%s' % mainpkg, True) or d.getVar('RDEPENDS', True) or "").split()
 	#rdep.append('%s-locale*' % pn)
 	#bb.data.setVar('RDEPENDS_%s' % mainpkg, ' '.join(rdep), d)
 }
 
 python perform_packagecopy () {
-	dest = bb.data.getVar('D', d, True)
-	dvar = bb.data.getVar('PKGD', d, True)
+	dest = d.getVar('D', True)
+	dvar = d.getVar('PKGD', True)
 
 	bb.mkdirhier(dvar)
 
@@ -549,16 +549,16 @@ python fixup_perms () {
 	# paths are resolved via BBPATH
 	def get_fs_perms_list(d):
 		str = ""
-		fs_perms_tables = bb.data.getVar('FILESYSTEM_PERMS_TABLES', d, True)
+		fs_perms_tables = d.getVar('FILESYSTEM_PERMS_TABLES', True)
 		if not fs_perms_tables:
 			fs_perms_tables = 'files/fs-perms.txt'
 		for conf_file in fs_perms_tables.split():
-			str += " %s" % bb.which(bb.data.getVar('BBPATH', d, True), conf_file)
+			str += " %s" % bb.which(d.getVar('BBPATH', True), conf_file)
 		return str
 
 
 
-	dvar = bb.data.getVar('PKGD', d, True)
+	dvar = d.getVar('PKGD', True)
 
 	fs_perms_table = {}
 
@@ -586,7 +586,7 @@ python fixup_perms () {
 				'oldincludedir' ]
 
 	for path in target_path_vars:
-		dir = bb.data.getVar(path, d, True) or ""
+		dir = d.getVar(path, True) or ""
 		if dir == "":
 			continue
 		fs_perms_table[dir] = fs_perms_entry(bb.data.expand("%s 0755 root root false - - -" % (dir), d))
@@ -664,11 +664,11 @@ python fixup_perms () {
 python split_and_strip_files () {
 	import commands, stat, errno
 
-	dvar = bb.data.getVar('PKGD', d, True)
-	pn = bb.data.getVar('PN', d, True)
+	dvar = d.getVar('PKGD', True)
+	pn = d.getVar('PN', True)
 
 	# We default to '.debug' style
-	if bb.data.getVar('PACKAGE_DEBUG_SPLIT_STYLE', d, True) == 'debug-file-directory':
+	if d.getVar('PACKAGE_DEBUG_SPLIT_STYLE', True) == 'debug-file-directory':
 		# Single debug-file-directory style debug info
 		debugappend = ".debug"
 		debugdir = ""
@@ -691,7 +691,7 @@ python split_and_strip_files () {
 	# 8 - shared library
 	def isELF(path):
 		type = 0
-		pathprefix = "export PATH=%s; " % bb.data.getVar('PATH', d, True)
+		pathprefix = "export PATH=%s; " % d.getVar('PATH', True)
 		ret, result = commands.getstatusoutput("%sfile '%s'" % (pathprefix, path))
 
 		if ret:
@@ -715,8 +715,8 @@ python split_and_strip_files () {
 	#
 	file_list = {}
 	file_links = {}
-	if (bb.data.getVar('INHIBIT_PACKAGE_DEBUG_SPLIT', d, True) != '1') and \
-	   (bb.data.getVar('INHIBIT_PACKAGE_STRIP', d, True) != '1'):
+	if (d.getVar('INHIBIT_PACKAGE_DEBUG_SPLIT', True) != '1') and \
+	   (d.getVar('INHIBIT_PACKAGE_STRIP', True) != '1'):
 		for root, dirs, files in os.walk(dvar):
 			for f in files:
 				file = os.path.join(root, f)
@@ -764,7 +764,7 @@ python split_and_strip_files () {
 	#
 	# First lets process debug splitting
 	#
-	if (bb.data.getVar('INHIBIT_PACKAGE_DEBUG_SPLIT', d, True) != '1'):
+	if (d.getVar('INHIBIT_PACKAGE_DEBUG_SPLIT', True) != '1'):
 		for file in file_list:
 			src = file[len(dvar):]
 			dest = debuglibdir + os.path.dirname(src) + debugdir + "/" + os.path.basename(src) + debugappend
@@ -842,7 +842,7 @@ python split_and_strip_files () {
 	#
 	# Now lets go back over things and strip them
 	#
-	if (bb.data.getVar('INHIBIT_PACKAGE_STRIP', d, True) != '1'):	
+	if (d.getVar('INHIBIT_PACKAGE_STRIP', True) != '1'):	
 		for file in file_list:
 			if file_list[file].startswith("ELF: "):
 				elf_file = int(file_list[file][5:])
@@ -856,11 +856,11 @@ python split_and_strip_files () {
 python populate_packages () {
 	import glob, stat, errno, re
 
-	workdir = bb.data.getVar('WORKDIR', d, True)
-	outdir = bb.data.getVar('DEPLOY_DIR', d, True)
-	dvar = bb.data.getVar('PKGD', d, True)
-	packages = bb.data.getVar('PACKAGES', d, True)
-	pn = bb.data.getVar('PN', d, True)
+	workdir = d.getVar('WORKDIR', True)
+	outdir = d.getVar('DEPLOY_DIR', True)
+	dvar = d.getVar('PKGD', True)
+	packages = d.getVar('PACKAGES', True)
+	pn = d.getVar('PN', True)
 
 	bb.mkdirhier(outdir)
 	os.chdir(dvar)
@@ -877,7 +877,7 @@ python populate_packages () {
 		else:
 			package_list.append(pkg)
 
-	pkgdest = bb.data.getVar('PKGDEST', d, True)
+	pkgdest = d.getVar('PKGDEST', True)
 	os.system('rm -rf %s' % pkgdest)
 
 	seen = []
@@ -887,14 +887,14 @@ python populate_packages () {
 		root = os.path.join(pkgdest, pkg)
 		bb.mkdirhier(root)
 
-		bb.data.setVar('PKG', pkg, localdata)
-		overrides = bb.data.getVar('OVERRIDES', localdata, True)
+		localdata.setVar('PKG', pkg)
+		overrides = localdata.getVar('OVERRIDES', True)
 		if not overrides:
 			raise bb.build.FuncFailed('OVERRIDES not defined')
-		bb.data.setVar('OVERRIDES', overrides + ':' + pkg, localdata)
+		localdata.setVar('OVERRIDES', overrides + ':' + pkg)
 		bb.data.update_data(localdata)
 
-		filesvar = bb.data.getVar('FILES', localdata, True) or ""
+		filesvar = localdata.getVar('FILES', True) or ""
 		files = filesvar.split()
 		file_links = {}
 		for file in files:
@@ -973,9 +973,9 @@ python populate_packages () {
 	bb.build.exec_func("package_name_hook", d)
 
 	for pkg in package_list:
-		pkgname = bb.data.getVar('PKG_%s' % pkg, d, True)
+		pkgname = d.getVar('PKG_%s' % pkg, True)
 		if pkgname is None:
-			bb.data.setVar('PKG_%s' % pkg, pkg, d)
+			d.setVar('PKG_%s' % pkg, pkg)
 
 	dangling_links = {}
 	pkg_files = {}
@@ -999,7 +999,7 @@ python populate_packages () {
 					dangling_links[pkg].append(os.path.normpath(target))
 
 	for pkg in package_list:
-		rdepends = bb.utils.explode_dep_versions(bb.data.getVar('RDEPENDS_' + pkg, d, True) or bb.data.getVar('RDEPENDS', d, True) or "")
+		rdepends = bb.utils.explode_dep_versions(d.getVar('RDEPENDS_' + pkg, True) or d.getVar('RDEPENDS', True) or "")
 
 		for l in dangling_links[pkg]:
 			found = False
@@ -1040,9 +1040,9 @@ python emit_pkgdata() {
 			f.write('%s: %s\n' % (var, encode(val)))
 		return
 
-	packages = bb.data.getVar('PACKAGES', d, True)
-	pkgdest = bb.data.getVar('PKGDEST', d, 1)
-	pkgdatadir = bb.data.getVar('PKGDESTWORK', d, True)
+	packages = d.getVar('PACKAGES', True)
+	pkgdest = d.getVar('PKGDEST', 1)
+	pkgdatadir = d.getVar('PKGDESTWORK', True)
 
 	# Take shared lock since we're only reading, not writing
 	lf = bb.utils.lockfile(bb.data.expand("${PACKAGELOCK}", d), True)
@@ -1052,7 +1052,7 @@ python emit_pkgdata() {
 	f.write("PACKAGES: %s\n" % packages)
 	f.close()
 
-	workdir = bb.data.getVar('WORKDIR', d, True)
+	workdir = d.getVar('WORKDIR', True)
 
 	for pkg in packages.split():
 		subdata_file = pkgdatadir + "/runtime/%s" % pkg
@@ -1080,19 +1080,19 @@ python emit_pkgdata() {
 		write_if_exists(sf, pkg, 'pkg_preinst')
 		write_if_exists(sf, pkg, 'pkg_prerm')
 		write_if_exists(sf, pkg, 'FILERPROVIDESFLIST')
-		for dfile in (bb.data.getVar('FILERPROVIDESFLIST_' + pkg, d, True) or "").split():
+		for dfile in (d.getVar('FILERPROVIDESFLIST_' + pkg, True) or "").split():
 			write_if_exists(sf, pkg, 'FILERPROVIDES_' + dfile)
 
 		write_if_exists(sf, pkg, 'FILERDEPENDSFLIST')
-		for dfile in (bb.data.getVar('FILERDEPENDSFLIST_' + pkg, d, True) or "").split():
+		for dfile in (d.getVar('FILERDEPENDSFLIST_' + pkg, True) or "").split():
 			write_if_exists(sf, pkg, 'FILERDEPENDS_' + dfile)
 
 		sf.close()
 
 
-		allow_empty = bb.data.getVar('ALLOW_EMPTY_%s' % pkg, d, True)
+		allow_empty = d.getVar('ALLOW_EMPTY_%s' % pkg, True)
 		if not allow_empty:
-			allow_empty = bb.data.getVar('ALLOW_EMPTY', d, True)
+			allow_empty = d.getVar('ALLOW_EMPTY', True)
 		root = "%s/%s" % (pkgdest, pkg)
 		os.chdir(root)
 		g = glob('*')
@@ -1123,8 +1123,8 @@ RPMDEPS = "${STAGING_LIBDIR_NATIVE}/rpm/bin/rpmdeps"
 python package_do_filedeps() {
 	import os, re
 
-	pkgdest = bb.data.getVar('PKGDEST', d, True)
-	packages = bb.data.getVar('PACKAGES', d, True)
+	pkgdest = d.getVar('PKGDEST', True)
+	packages = d.getVar('PACKAGES', True)
 
 	rpmdeps = bb.data.expand("${RPMDEPS}", d)
 	r = re.compile(r'[<>=]+ +[^ ]*')
@@ -1189,7 +1189,7 @@ SHLIBSWORKDIR = "${WORKDIR}/shlibs"
 python package_do_shlibs() {
 	import re
 
-	exclude_shlibs = bb.data.getVar('EXCLUDE_FROM_SHLIBS', d, 0)
+	exclude_shlibs = d.getVar('EXCLUDE_FROM_SHLIBS', 0)
 	if exclude_shlibs:
 		bb.note("not generating shlibs")
 		return
@@ -1197,27 +1197,27 @@ python package_do_shlibs() {
 	lib_re = re.compile("^.*\.so")
 	libdir_re = re.compile(".*/lib$")
 
-	packages = bb.data.getVar('PACKAGES', d, True)
-	targetos = bb.data.getVar('TARGET_OS', d, True)
+	packages = d.getVar('PACKAGES', True)
+	targetos = d.getVar('TARGET_OS', True)
 
-	workdir = bb.data.getVar('WORKDIR', d, True)
+	workdir = d.getVar('WORKDIR', True)
 
-	ver = bb.data.getVar('PKGV', d, True)
+	ver = d.getVar('PKGV', True)
 	if not ver:
 		bb.error("PKGV not defined")
 		return
 
-	pkgdest = bb.data.getVar('PKGDEST', d, True)
+	pkgdest = d.getVar('PKGDEST', True)
 
-	shlibs_dir = bb.data.getVar('SHLIBSDIR', d, True)
-	shlibswork_dir = bb.data.getVar('SHLIBSWORKDIR', d, True)
+	shlibs_dir = d.getVar('SHLIBSDIR', True)
+	shlibswork_dir = d.getVar('SHLIBSWORKDIR', True)
 
 	# Take shared lock since we're only reading, not writing
 	lf = bb.utils.lockfile(bb.data.expand("${PACKAGELOCK}", d))
 
 	def linux_so(root, path, file):
-		cmd = bb.data.getVar('OBJDUMP', d, True) + " -p " + os.path.join(root, file) + " 2>/dev/null"
-		cmd = "PATH=\"%s\" %s" % (bb.data.getVar('PATH', d, True), cmd)
+		cmd = d.getVar('OBJDUMP', True) + " -p " + os.path.join(root, file) + " 2>/dev/null"
+		cmd = "PATH=\"%s\" %s" % (d.getVar('PATH', True), cmd)
 		fd = os.popen(cmd)
 		lines = fd.readlines()
 		fd.close()
@@ -1264,7 +1264,7 @@ python package_do_shlibs() {
 				if not combo in sonames:
 					sonames.append(combo)
 		if file.endswith('.dylib') or file.endswith('.so'):
-			lafile = fullpath.replace(os.path.join(pkgdest, pkg), bb.data.getVar('PKGD', d, True))
+			lafile = fullpath.replace(os.path.join(pkgdest, pkg), d.getVar('PKGD', True))
 			# Drop suffix
 			lafile = lafile.rsplit(".",1)[0]
 			lapath = os.path.dirname(lafile)
@@ -1299,26 +1299,26 @@ python package_do_shlibs() {
 								needed[pkg].append(name)
 								#bb.note("Adding %s for %s" % (name, pkg))
 
-	if bb.data.getVar('PACKAGE_SNAP_LIB_SYMLINKS', d, True) == "1":
+	if d.getVar('PACKAGE_SNAP_LIB_SYMLINKS', True) == "1":
 		snap_symlinks = True
 	else:
 		snap_symlinks = False
 
-	if (bb.data.getVar('USE_LDCONFIG', d, True) or "1") == "1":
+	if (d.getVar('USE_LDCONFIG', True) or "1") == "1":
 		use_ldconfig = True
 	else:
 		use_ldconfig = False
 
 	needed = {}
 	shlib_provider = {}
-	private_libs = bb.data.getVar('PRIVATE_LIBS', d, True)
+	private_libs = d.getVar('PRIVATE_LIBS', True)
 	for pkg in packages.split():
 		needs_ldconfig = False
 		bb.debug(2, "calculating shlib provides for %s" % pkg)
 
-		pkgver = bb.data.getVar('PKGV_' + pkg, d, True)
+		pkgver = d.getVar('PKGV_' + pkg, True)
 		if not pkgver:
-			pkgver = bb.data.getVar('PV_' + pkg, d, True)
+			pkgver = d.getVar('PV_' + pkg, True)
 		if not pkgver:
 			pkgver = ver
 
@@ -1352,11 +1352,11 @@ python package_do_shlibs() {
 			fd.close()
 		if needs_ldconfig and use_ldconfig:
 			bb.debug(1, 'adding ldconfig call to postinst for %s' % pkg)
-			postinst = bb.data.getVar('pkg_postinst_%s' % pkg, d, True) or bb.data.getVar('pkg_postinst', d, True)
+			postinst = d.getVar('pkg_postinst_%s' % pkg, True) or d.getVar('pkg_postinst', True)
 			if not postinst:
 				postinst = '#!/bin/sh\n'
-			postinst += bb.data.getVar('ldconfig_postinst_fragment', d, True)
-			bb.data.setVar('pkg_postinst_%s' % pkg, postinst, d)
+			postinst += d.getVar('ldconfig_postinst_fragment', True)
+			d.setVar('pkg_postinst_%s' % pkg, postinst)
 
 	list_re = re.compile('^(.*)\.list$')
 	for dir in [shlibs_dir]: 
@@ -1380,7 +1380,7 @@ python package_do_shlibs() {
 
 	bb.utils.unlockfile(lf)
 
-	assumed_libs = bb.data.getVar('ASSUME_SHLIBS', d, True)
+	assumed_libs = d.getVar('ASSUME_SHLIBS', True)
 	if assumed_libs:
 	    for e in assumed_libs.split():
 		l, dep_pkg = e.split(":")
@@ -1424,12 +1424,12 @@ python package_do_shlibs() {
 python package_do_pkgconfig () {
 	import re
 
-	packages = bb.data.getVar('PACKAGES', d, True)
-	workdir = bb.data.getVar('WORKDIR', d, True)
-	pkgdest = bb.data.getVar('PKGDEST', d, True)
+	packages = d.getVar('PACKAGES', True)
+	workdir = d.getVar('WORKDIR', True)
+	pkgdest = d.getVar('PKGDEST', True)
 
-	shlibs_dir = bb.data.getVar('SHLIBSDIR', d, True)
-	shlibswork_dir = bb.data.getVar('SHLIBSWORKDIR', d, True)
+	shlibs_dir = d.getVar('SHLIBSDIR', True)
+	shlibswork_dir = d.getVar('SHLIBSWORKDIR', True)
 
 	pc_re = re.compile('(.*)\.pc$')
 	var_re = re.compile('(.*)=(.*)')
@@ -1515,9 +1515,9 @@ python package_do_pkgconfig () {
 }
 
 python read_shlibdeps () {
-	packages = bb.data.getVar('PACKAGES', d, True).split()
+	packages = d.getVar('PACKAGES', True).split()
 	for pkg in packages:
-		rdepends = bb.utils.explode_dep_versions(bb.data.getVar('RDEPENDS_' + pkg, d, 0) or bb.data.getVar('RDEPENDS', d, 0) or "")
+		rdepends = bb.utils.explode_dep_versions(d.getVar('RDEPENDS_' + pkg, 0) or d.getVar('RDEPENDS', 0) or "")
 
 		for extension in ".shlibdeps", ".pcdeps", ".clilibdeps":
 			depsfile = bb.data.expand("${PKGDEST}/" + pkg + extension, d)
@@ -1544,14 +1544,14 @@ python package_depchains() {
 	package.
 	"""
 
-	packages  = bb.data.getVar('PACKAGES', d, True)
-	postfixes = (bb.data.getVar('DEPCHAIN_POST', d, True) or '').split()
-	prefixes  = (bb.data.getVar('DEPCHAIN_PRE', d, True) or '').split()
+	packages  = d.getVar('PACKAGES', True)
+	postfixes = (d.getVar('DEPCHAIN_POST', True) or '').split()
+	prefixes  = (d.getVar('DEPCHAIN_PRE', True) or '').split()
 
 	def pkg_adddeprrecs(pkg, base, suffix, getname, depends, d):
 
 		#bb.note('depends for %s is %s' % (base, depends))
-		rreclist = bb.utils.explode_dep_versions(bb.data.getVar('RRECOMMENDS_' + pkg, d, True) or bb.data.getVar('RRECOMMENDS', d, True) or "")
+		rreclist = bb.utils.explode_dep_versions(d.getVar('RRECOMMENDS_' + pkg, True) or d.getVar('RRECOMMENDS', True) or "")
 
 		for depend in depends:
 			if depend.find('-native') != -1 or depend.find('-cross') != -1 or depend.startswith('virtual/'):
@@ -1572,7 +1572,7 @@ python package_depchains() {
 	def pkg_addrrecs(pkg, base, suffix, getname, rdepends, d):
 
 		#bb.note('rdepends for %s is %s' % (base, rdepends))
-		rreclist = bb.utils.explode_dep_versions(bb.data.getVar('RRECOMMENDS_' + pkg, d, True) or bb.data.getVar('RRECOMMENDS', d, True) or "")
+		rreclist = bb.utils.explode_dep_versions(d.getVar('RRECOMMENDS_' + pkg, True) or d.getVar('RRECOMMENDS', True) or "")
 
 		for depend in rdepends:
 			if depend.find('virtual-locale-') != -1:
@@ -1596,15 +1596,15 @@ python package_depchains() {
 			list.append(dep)
 
 	depends = []
-	for dep in bb.utils.explode_deps(bb.data.getVar('DEPENDS', d, True) or ""):
+	for dep in bb.utils.explode_deps(d.getVar('DEPENDS', True) or ""):
 		add_dep(depends, dep)
 
 	rdepends = []
-	for dep in bb.utils.explode_deps(bb.data.getVar('RDEPENDS', d, True) or ""):
+	for dep in bb.utils.explode_deps(d.getVar('RDEPENDS', True) or ""):
 		add_dep(rdepends, dep)
 
 	for pkg in packages.split():
-		for dep in bb.utils.explode_deps(bb.data.getVar('RDEPENDS_' + pkg, d, True) or ""):
+		for dep in bb.utils.explode_deps(d.getVar('RDEPENDS_' + pkg, True) or ""):
 			add_dep(rdepends, dep)
 
 	#bb.note('rdepends is %s' % rdepends)
@@ -1630,7 +1630,7 @@ python package_depchains() {
 
 	for suffix in pkgs:
 		for pkg in pkgs[suffix]:
-			if bb.data.getVarFlag('RRECOMMENDS_' + pkg, 'nodeprrecs', d):
+			if d.getVarFlag('RRECOMMENDS_' + pkg, 'nodeprrecs'):
 				continue
 			(base, func) = pkgs[suffix][pkg]
 			if suffix == "-dev":
@@ -1639,7 +1639,7 @@ python package_depchains() {
 				pkg_addrrecs(pkg, base, suffix, func, rdepends, d)
 			else:
 				rdeps = []
-				for dep in bb.utils.explode_deps(bb.data.getVar('RDEPENDS_' + base, d, True) or bb.data.getVar('RDEPENDS', d, True) or ""):
+				for dep in bb.utils.explode_deps(d.getVar('RDEPENDS_' + base, True) or d.getVar('RDEPENDS', True) or ""):
 					add_dep(rdeps, dep)
 				pkg_addrrecs(pkg, base, suffix, func, rdeps, d)
 }
@@ -1679,22 +1679,22 @@ python do_package () {
         # as any change to rpmdeps requires this to be rerun.
         # PACKAGE_BBCLASS_VERSION = "1"
 
-	packages = (bb.data.getVar('PACKAGES', d, True) or "").split()
+	packages = (d.getVar('PACKAGES', True) or "").split()
 	if len(packages) < 1:
 		bb.debug(1, "No packages to build, skipping do_package")
 		return
 
-	workdir = bb.data.getVar('WORKDIR', d, True)
-	outdir = bb.data.getVar('DEPLOY_DIR', d, True)
-	dest = bb.data.getVar('D', d, True)
-	dvar = bb.data.getVar('PKGD', d, True)
-	pn = bb.data.getVar('PN', d, True)
+	workdir = d.getVar('WORKDIR', True)
+	outdir = d.getVar('DEPLOY_DIR', True)
+	dest = d.getVar('D', True)
+	dvar = d.getVar('PKGD', True)
+	pn = d.getVar('PN', True)
 
 	if not workdir or not outdir or not dest or not dvar or not pn or not packages:
 		bb.error("WORKDIR, DEPLOY_DIR, D, PN and PKGD all must be defined, unable to package")
 		return
 
-	for f in (bb.data.getVar('PACKAGEFUNCS', d, True) or '').split():
+	for f in (d.getVar('PACKAGEFUNCS', True) or '').split():
 		bb.build.exec_func(f, d)
 }
 

+ 41 - 41
meta/classes/package_deb.bbclass

@@ -11,18 +11,18 @@ DPKG_ARCH ?= "${TARGET_ARCH}"
 PKGWRITEDIRDEB = "${WORKDIR}/deploy-debs"
 
 python package_deb_fn () {
-    bb.data.setVar('PKGFN', bb.data.getVar('PKG',d), d)
+    bb.data.setVar('PKGFN', d.getVar('PKG'), d)
 }
 
 addtask package_deb_install
 python do_package_deb_install () {
-    pkg = bb.data.getVar('PKG', d, True)
-    pkgfn = bb.data.getVar('PKGFN', d, True)
-    rootfs = bb.data.getVar('IMAGE_ROOTFS', d, True)
-    debdir = bb.data.getVar('DEPLOY_DIR_DEB', d, True)
+    pkg = d.getVar('PKG', True)
+    pkgfn = d.getVar('PKGFN', True)
+    rootfs = d.getVar('IMAGE_ROOTFS', True)
+    debdir = d.getVar('DEPLOY_DIR_DEB', True)
     apt_config = bb.data.expand('${STAGING_ETCDIR_NATIVE}/apt/apt.conf', d)
-    stagingbindir = bb.data.getVar('STAGING_BINDIR_NATIVE', d, True)
-    tmpdir = bb.data.getVar('TMPDIR', d, True)
+    stagingbindir = d.getVar('STAGING_BINDIR_NATIVE', True)
+    tmpdir = d.getVar('TMPDIR', True)
 
     if None in (pkg,pkgfn,rootfs):
         raise bb.build.FuncFailed("missing variables (one or more of PKG, PKGFN, IMAGE_ROOTFS)")
@@ -206,22 +206,22 @@ python do_package_deb () {
     import re, copy
     import textwrap
 
-    workdir = bb.data.getVar('WORKDIR', d, True)
+    workdir = d.getVar('WORKDIR', True)
     if not workdir:
         bb.error("WORKDIR not defined, unable to package")
         return
 
-    outdir = bb.data.getVar('PKGWRITEDIRDEB', d, True)
+    outdir = d.getVar('PKGWRITEDIRDEB', True)
     if not outdir:
         bb.error("PKGWRITEDIRDEB not defined, unable to package")
         return
 
-    packages = bb.data.getVar('PACKAGES', d, True)
+    packages = d.getVar('PACKAGES', True)
     if not packages:
         bb.debug(1, "PACKAGES not defined, nothing to package")
         return
 
-    tmpdir = bb.data.getVar('TMPDIR', d, True)
+    tmpdir = d.getVar('TMPDIR', True)
 
     if os.access(os.path.join(tmpdir, "stamps", "DEB_PACKAGE_INDEX_CLEAN"),os.R_OK):
         os.unlink(os.path.join(tmpdir, "stamps", "DEB_PACKAGE_INDEX_CLEAN"))
@@ -230,7 +230,7 @@ python do_package_deb () {
         bb.debug(1, "No packages; nothing to do")
         return
 
-    pkgdest = bb.data.getVar('PKGDEST', d, True)
+    pkgdest = d.getVar('PKGDEST', True)
 
     for pkg in packages.split():
         localdata = bb.data.createCopy(d)
@@ -238,19 +238,19 @@ python do_package_deb () {
 
         lf = bb.utils.lockfile(root + ".lock")
 
-        bb.data.setVar('ROOT', '', localdata)
-        bb.data.setVar('ROOT_%s' % pkg, root, localdata)
-        pkgname = bb.data.getVar('PKG_%s' % pkg, localdata, True)
+        localdata.setVar('ROOT', '')
+        localdata.setVar('ROOT_%s' % pkg, root)
+        pkgname = localdata.getVar('PKG_%s' % pkg, True)
         if not pkgname:
             pkgname = pkg
-        bb.data.setVar('PKG', pkgname, localdata)
+        localdata.setVar('PKG', pkgname)
 
-        bb.data.setVar('OVERRIDES', pkg, localdata)
+        localdata.setVar('OVERRIDES', pkg)
 
         bb.data.update_data(localdata)
         basedir = os.path.join(os.path.dirname(root))
 
-        pkgoutdir = os.path.join(outdir, bb.data.getVar('PACKAGE_ARCH', localdata, True))
+        pkgoutdir = os.path.join(outdir, localdata.getVar('PACKAGE_ARCH', True))
         bb.mkdirhier(pkgoutdir)
 
         os.chdir(root)
@@ -261,8 +261,8 @@ python do_package_deb () {
             del g[g.index('./DEBIAN')]
         except ValueError:
             pass
-        if not g and bb.data.getVar('ALLOW_EMPTY', localdata) != "1":
-            bb.note("Not creating empty archive for %s-%s-%s" % (pkg, bb.data.getVar('PKGV', localdata, True), bb.data.getVar('PKGR', localdata, True)))
+        if not g and localdata.getVar('ALLOW_EMPTY') != "1":
+            bb.note("Not creating empty archive for %s-%s-%s" % (pkg, localdata.getVar('PKGV', True), localdata.getVar('PKGR', True)))
             bb.utils.unlockfile(lf)
             continue
 
@@ -278,7 +278,7 @@ python do_package_deb () {
             raise bb.build.FuncFailed("unable to open control file for writing.")
 
         fields = []
-        pe = bb.data.getVar('PKGE', d, True)
+        pe = d.getVar('PKGE', True)
         if pe and int(pe) > 0:
             fields.append(["Version: %s:%s-%s\n", ['PKGE', 'PKGV', 'PKGR']])
         else:
@@ -298,10 +298,10 @@ python do_package_deb () {
         def pullData(l, d):
             l2 = []
             for i in l:
-                data = bb.data.getVar(i, d, True)
+                data = d.getVar(i, True)
                 if data is None:
                     raise KeyError(f)
-                if i == 'DPKG_ARCH' and bb.data.getVar('PACKAGE_ARCH', d, True) == 'all':
+                if i == 'DPKG_ARCH' and d.getVar('PACKAGE_ARCH', True) == 'all':
                     data = 'all'
                 l2.append(data)
             return l2
@@ -311,12 +311,12 @@ python do_package_deb () {
         try:
             for (c, fs) in fields:
                 for f in fs:
-                     if bb.data.getVar(f, localdata) is None:
+                     if localdata.getVar(f) is None:
                          raise KeyError(f)
                 # Special behavior for description...
                 if 'DESCRIPTION' in fs:
-                     summary = bb.data.getVar('SUMMARY', localdata, True) or bb.data.getVar('DESCRIPTION', localdata, True) or "."
-                     description = bb.data.getVar('DESCRIPTION', localdata, True) or "."
+                     summary = localdata.getVar('SUMMARY', True) or localdata.getVar('DESCRIPTION', True) or "."
+                     description = localdata.getVar('DESCRIPTION', True) or "."
                      description = textwrap.dedent(description).strip()
                      ctrlfile.write('Description: %s\n' % unicode(summary))
                      ctrlfile.write('%s\n' % unicode(textwrap.fill(description, width=74, initial_indent=' ', subsequent_indent=' ')))
@@ -332,18 +332,18 @@ python do_package_deb () {
 
         bb.build.exec_func("mapping_rename_hook", localdata)
 
-        rdepends = bb.utils.explode_dep_versions(bb.data.getVar("RDEPENDS", localdata, True) or "")
+        rdepends = bb.utils.explode_dep_versions(localdata.getVar("RDEPENDS", True) or "")
         for dep in rdepends:
                 if '*' in dep:
                         del rdepends[dep]
-        rrecommends = bb.utils.explode_dep_versions(bb.data.getVar("RRECOMMENDS", localdata, True) or "")
+        rrecommends = bb.utils.explode_dep_versions(localdata.getVar("RRECOMMENDS", True) or "")
         for dep in rrecommends:
                 if '*' in dep:
                         del rrecommends[dep]
-        rsuggests = bb.utils.explode_dep_versions(bb.data.getVar("RSUGGESTS", localdata, True) or "")
-        rprovides = bb.utils.explode_dep_versions(bb.data.getVar("RPROVIDES", localdata, True) or "")
-        rreplaces = bb.utils.explode_dep_versions(bb.data.getVar("RREPLACES", localdata, True) or "")
-        rconflicts = bb.utils.explode_dep_versions(bb.data.getVar("RCONFLICTS", localdata, True) or "")
+        rsuggests = bb.utils.explode_dep_versions(localdata.getVar("RSUGGESTS", True) or "")
+        rprovides = bb.utils.explode_dep_versions(localdata.getVar("RPROVIDES", True) or "")
+        rreplaces = bb.utils.explode_dep_versions(localdata.getVar("RREPLACES", True) or "")
+        rconflicts = bb.utils.explode_dep_versions(localdata.getVar("RCONFLICTS", True) or "")
         if rdepends:
             ctrlfile.write("Depends: %s\n" % unicode(bb.utils.join_deps(rdepends)))
         if rsuggests:
@@ -359,7 +359,7 @@ python do_package_deb () {
         ctrlfile.close()
 
         for script in ["preinst", "postinst", "prerm", "postrm"]:
-            scriptvar = bb.data.getVar('pkg_%s' % script, localdata, True)
+            scriptvar = localdata.getVar('pkg_%s' % script, True)
             if not scriptvar:
                 continue
             try:
@@ -372,7 +372,7 @@ python do_package_deb () {
             scriptfile.close()
             os.chmod(os.path.join(controldir, script), 0755)
 
-        conffiles_str = bb.data.getVar("CONFFILES", localdata, True)
+        conffiles_str = localdata.getVar("CONFFILES", True)
         if conffiles_str:
             try:
                 conffiles = file(os.path.join(controldir, 'conffiles'), 'w')
@@ -384,7 +384,7 @@ python do_package_deb () {
             conffiles.close()
 
         os.chdir(basedir)
-        ret = os.system("PATH=\"%s\" dpkg-deb -b %s %s" % (bb.data.getVar("PATH", localdata, True), root, pkgoutdir))
+        ret = os.system("PATH=\"%s\" dpkg-deb -b %s %s" % (localdata.getVar("PATH", True), root, pkgoutdir))
         if ret != 0:
             bb.utils.prunedir(controldir)
             bb.utils.unlockfile(lf)
@@ -405,17 +405,17 @@ python do_package_write_deb_setscene () {
 addtask do_package_write_deb_setscene
 
 python () {
-    if bb.data.getVar('PACKAGES', d, True) != '':
-        deps = (bb.data.getVarFlag('do_package_write_deb', 'depends', d) or "").split()
+    if d.getVar('PACKAGES', True) != '':
+        deps = (d.getVarFlag('do_package_write_deb', 'depends') or "").split()
         deps.append('dpkg-native:do_populate_sysroot')
         deps.append('virtual/fakeroot-native:do_populate_sysroot')
         bb.data.setVarFlag('do_package_write_deb', 'depends', " ".join(deps), d)
-        bb.data.setVarFlag('do_package_write_deb', 'fakeroot', "1", d)
-        bb.data.setVarFlag('do_package_write_deb_setscene', 'fakeroot', "1", d)
+        d.setVarFlag('do_package_write_deb', 'fakeroot', "1")
+        d.setVarFlag('do_package_write_deb_setscene', 'fakeroot', "1")
 
     # Map TARGET_ARCH to Debian's ideas about architectures
-    if bb.data.getVar('DPKG_ARCH', d, True) in ["x86", "i486", "i586", "i686", "pentium"]:
-        bb.data.setVar('DPKG_ARCH', 'i386', d)
+    if d.getVar('DPKG_ARCH', True) in ["x86", "i486", "i586", "i686", "pentium"]:
+        d.setVar('DPKG_ARCH', 'i386')
 }
 
 python do_package_write_deb () {

+ 41 - 41
meta/classes/package_ipk.bbclass

@@ -11,16 +11,16 @@ PKGWRITEDIRIPK = "${WORKDIR}/deploy-ipks"
 OPKGBUILDCMD ??= "opkg-build"
 
 python package_ipk_fn () {
-	bb.data.setVar('PKGFN', bb.data.getVar('PKG',d), d)
+	bb.data.setVar('PKGFN', d.getVar('PKG'), d)
 }
 
 python package_ipk_install () {
-	pkg = bb.data.getVar('PKG', d, 1)
-	pkgfn = bb.data.getVar('PKGFN', d, 1)
-	rootfs = bb.data.getVar('IMAGE_ROOTFS', d, 1)
-	ipkdir = bb.data.getVar('DEPLOY_DIR_IPK', d, 1)
-	stagingdir = bb.data.getVar('STAGING_DIR', d, 1)
-	tmpdir = bb.data.getVar('TMPDIR', d, 1)
+	pkg = d.getVar('PKG', 1)
+	pkgfn = d.getVar('PKGFN', 1)
+	rootfs = d.getVar('IMAGE_ROOTFS', 1)
+	ipkdir = d.getVar('DEPLOY_DIR_IPK', 1)
+	stagingdir = d.getVar('STAGING_DIR', 1)
+	tmpdir = d.getVar('TMPDIR', 1)
 
 	if None in (pkg,pkgfn,rootfs):
 		raise bb.build.FuncFailed("missing variables (one or more of PKG, PKGFN, IMAGEROOTFS)")
@@ -36,7 +36,7 @@ python package_ipk_install () {
 	# Generate ipk.conf if it or the stamp doesnt exist
 	conffile = os.path.join(stagingdir,"ipkg.conf")
 	if not os.access(conffile, os.R_OK):
-		ipkg_archs = bb.data.getVar('PACKAGE_ARCHS',d)
+		ipkg_archs = d.getVar('PACKAGE_ARCHS')
 		if ipkg_archs is None:
 			bb.error("PACKAGE_ARCHS missing")
 			raise FuncFailed
@@ -259,15 +259,15 @@ python do_package_ipk () {
 	import re, copy
 	import textwrap
 
-	workdir = bb.data.getVar('WORKDIR', d, True)
-	outdir = bb.data.getVar('PKGWRITEDIRIPK', d, True)
-	tmpdir = bb.data.getVar('TMPDIR', d, True)
-	pkgdest = bb.data.getVar('PKGDEST', d, True)
+	workdir = d.getVar('WORKDIR', True)
+	outdir = d.getVar('PKGWRITEDIRIPK', True)
+	tmpdir = d.getVar('TMPDIR', True)
+	pkgdest = d.getVar('PKGDEST', True)
 	if not workdir or not outdir or not tmpdir:
 		bb.error("Variables incorrectly set, unable to package")
 		return
 
-	packages = bb.data.getVar('PACKAGES', d, True)
+	packages = d.getVar('PACKAGES', True)
 	if not packages or packages == '':
 		bb.debug(1, "No packages; nothing to do")
 		return
@@ -283,18 +283,18 @@ python do_package_ipk () {
 
 		lf = bb.utils.lockfile(root + ".lock")
 
-		bb.data.setVar('ROOT', '', localdata)
-		bb.data.setVar('ROOT_%s' % pkg, root, localdata)
-		pkgname = bb.data.getVar('PKG_%s' % pkg, localdata, 1)
+		localdata.setVar('ROOT', '')
+		localdata.setVar('ROOT_%s' % pkg, root)
+		pkgname = localdata.getVar('PKG_%s' % pkg, 1)
 		if not pkgname:
 			pkgname = pkg
-		bb.data.setVar('PKG', pkgname, localdata)
+		localdata.setVar('PKG', pkgname)
 
-		bb.data.setVar('OVERRIDES', pkg, localdata)
+		localdata.setVar('OVERRIDES', pkg)
 
 		bb.data.update_data(localdata)
 		basedir = os.path.join(os.path.dirname(root))
-		arch = bb.data.getVar('PACKAGE_ARCH', localdata, 1)
+		arch = localdata.getVar('PACKAGE_ARCH', 1)
 		pkgoutdir = "%s/%s" % (outdir, arch)
 		bb.mkdirhier(pkgoutdir)
 		os.chdir(root)
@@ -305,8 +305,8 @@ python do_package_ipk () {
 			del g[g.index('./CONTROL')]
 		except ValueError:
 			pass
-		if not g and bb.data.getVar('ALLOW_EMPTY', localdata) != "1":
-			bb.note("Not creating empty archive for %s-%s-%s" % (pkg, bb.data.getVar('PKGV', localdata, 1), bb.data.getVar('PKGR', localdata, 1)))
+		if not g and localdata.getVar('ALLOW_EMPTY') != "1":
+			bb.note("Not creating empty archive for %s-%s-%s" % (pkg, localdata.getVar('PKGV', 1), localdata.getVar('PKGR', 1)))
 			bb.utils.unlockfile(lf)
 			continue
 
@@ -319,7 +319,7 @@ python do_package_ipk () {
 			raise bb.build.FuncFailed("unable to open control file for writing.")
 
 		fields = []
-		pe = bb.data.getVar('PKGE', d, 1)
+		pe = d.getVar('PKGE', 1)
 		if pe and int(pe) > 0:
 			fields.append(["Version: %s:%s-%s\n", ['PKGE', 'PKGV', 'PKGR']])
 		else:
@@ -336,7 +336,7 @@ python do_package_ipk () {
 		def pullData(l, d):
 			l2 = []
 			for i in l:
-				l2.append(bb.data.getVar(i, d, 1))
+				l2.append(d.getVar(i, 1))
 			return l2
 
 		ctrlfile.write("Package: %s\n" % pkgname)
@@ -344,12 +344,12 @@ python do_package_ipk () {
 		try:
 			for (c, fs) in fields:
 				for f in fs:
-					if bb.data.getVar(f, localdata) is None:
+					if localdata.getVar(f) is None:
 						raise KeyError(f)
 				# Special behavior for description...
 				if 'DESCRIPTION' in fs:
-					summary = bb.data.getVar('SUMMARY', localdata, True) or bb.data.getVar('DESCRIPTION', localdata, True) or "."
-					description = bb.data.getVar('DESCRIPTION', localdata, True) or "."
+					summary = localdata.getVar('SUMMARY', True) or localdata.getVar('DESCRIPTION', True) or "."
+					description = localdata.getVar('DESCRIPTION', True) or "."
 					description = textwrap.dedent(description).strip()
 					ctrlfile.write('Description: %s\n' % summary)
 					ctrlfile.write('%s\n' % textwrap.fill(description, width=74, initial_indent=' ', subsequent_indent=' '))
@@ -365,12 +365,12 @@ python do_package_ipk () {
 
 		bb.build.exec_func("mapping_rename_hook", localdata)
 
-		rdepends = bb.utils.explode_dep_versions(bb.data.getVar("RDEPENDS", localdata, 1) or "")
-		rrecommends = bb.utils.explode_dep_versions(bb.data.getVar("RRECOMMENDS", localdata, 1) or "")
-		rsuggests = bb.utils.explode_dep_versions(bb.data.getVar("RSUGGESTS", localdata, 1) or "")
-		rprovides = bb.utils.explode_dep_versions(bb.data.getVar("RPROVIDES", localdata, 1) or "")
-		rreplaces = bb.utils.explode_dep_versions(bb.data.getVar("RREPLACES", localdata, 1) or "")
-		rconflicts = bb.utils.explode_dep_versions(bb.data.getVar("RCONFLICTS", localdata, 1) or "")
+		rdepends = bb.utils.explode_dep_versions(localdata.getVar("RDEPENDS", 1) or "")
+		rrecommends = bb.utils.explode_dep_versions(localdata.getVar("RRECOMMENDS", 1) or "")
+		rsuggests = bb.utils.explode_dep_versions(localdata.getVar("RSUGGESTS", 1) or "")
+		rprovides = bb.utils.explode_dep_versions(localdata.getVar("RPROVIDES", 1) or "")
+		rreplaces = bb.utils.explode_dep_versions(localdata.getVar("RREPLACES", 1) or "")
+		rconflicts = bb.utils.explode_dep_versions(localdata.getVar("RCONFLICTS", 1) or "")
 
 		if rdepends:
 			ctrlfile.write("Depends: %s\n" % bb.utils.join_deps(rdepends))
@@ -384,14 +384,14 @@ python do_package_ipk () {
 			ctrlfile.write("Replaces: %s\n" % bb.utils.join_deps(rreplaces))
 		if rconflicts:
 			ctrlfile.write("Conflicts: %s\n" % bb.utils.join_deps(rconflicts))
-		src_uri = bb.data.getVar("SRC_URI", localdata, 1)
+		src_uri = localdata.getVar("SRC_URI", 1)
 		if src_uri:
 			src_uri = re.sub("\s+", " ", src_uri)
 			ctrlfile.write("Source: %s\n" % " ".join(src_uri.split()))
 		ctrlfile.close()
 
 		for script in ["preinst", "postinst", "prerm", "postrm"]:
-			scriptvar = bb.data.getVar('pkg_%s' % script, localdata, 1)
+			scriptvar = localdata.getVar('pkg_%s' % script, 1)
 			if not scriptvar:
 				continue
 			try:
@@ -403,7 +403,7 @@ python do_package_ipk () {
 			scriptfile.close()
 			os.chmod(os.path.join(controldir, script), 0755)
 
-		conffiles_str = bb.data.getVar("CONFFILES", localdata, 1)
+		conffiles_str = localdata.getVar("CONFFILES", 1)
 		if conffiles_str:
 			try:
 				conffiles = file(os.path.join(controldir, 'conffiles'), 'w')
@@ -415,8 +415,8 @@ python do_package_ipk () {
 			conffiles.close()
 
 		os.chdir(basedir)
-		ret = os.system("PATH=\"%s\" %s %s %s" % (bb.data.getVar("PATH", localdata, 1), 
-                                                          bb.data.getVar("OPKGBUILDCMD",d,1), pkg, pkgoutdir))
+		ret = os.system("PATH=\"%s\" %s %s %s" % (localdata.getVar("PATH", 1), 
+                                                          d.getVar("OPKGBUILDCMD",1), pkg, pkgoutdir))
 		if ret != 0:
 			bb.utils.unlockfile(lf)
 			raise bb.build.FuncFailed("opkg-build execution failed")
@@ -437,13 +437,13 @@ python do_package_write_ipk_setscene () {
 addtask do_package_write_ipk_setscene
 
 python () {
-    if bb.data.getVar('PACKAGES', d, True) != '':
-        deps = (bb.data.getVarFlag('do_package_write_ipk', 'depends', d) or "").split()
+    if d.getVar('PACKAGES', True) != '':
+        deps = (d.getVarFlag('do_package_write_ipk', 'depends') or "").split()
         deps.append('opkg-utils-native:do_populate_sysroot')
         deps.append('virtual/fakeroot-native:do_populate_sysroot')
         bb.data.setVarFlag('do_package_write_ipk', 'depends', " ".join(deps), d)
-        bb.data.setVarFlag('do_package_write_ipk', 'fakeroot', "1", d)
-        bb.data.setVarFlag('do_package_write_ipk_setscene', 'fakeroot', "1", d)
+        d.setVarFlag('do_package_write_ipk', 'fakeroot', "1")
+        d.setVarFlag('do_package_write_ipk_setscene', 'fakeroot', "1")
 }
 
 python do_package_write_ipk () {

+ 68 - 68
meta/classes/package_rpm.bbclass

@@ -8,7 +8,7 @@ RPMBUILD="rpmbuild"
 PKGWRITEDIRRPM = "${WORKDIR}/deploy-rpms"
 
 python package_rpm_fn () {
-	bb.data.setVar('PKGFN', bb.data.getVar('PKG',d), d)
+	bb.data.setVar('PKGFN', d.getVar('PKG'), d)
 }
 
 python package_rpm_install () {
@@ -406,7 +406,7 @@ python write_specfile () {
 				name = "".join(name.split(eext[1] + '-'))
 		return name
 
-#		ml = bb.data.getVar("MLPREFIX", d, True)
+#		ml = d.getVar("MLPREFIX", True)
 #		if ml and name and len(ml) != 0 and name.find(ml) == 0:
 #			return ml.join(name.split(ml, 1)[1:])
 #		return name
@@ -426,7 +426,7 @@ python write_specfile () {
 	# after renaming we cannot look up the dependencies in the packagedata
 	# store.
 	def translate_vers(varname, d):
-		depends = bb.data.getVar(varname, d, True)
+		depends = d.getVar(varname, True)
 		if depends:
 			depends_dict = bb.utils.explode_dep_versions(depends)
 			newdeps_dict = {}
@@ -481,34 +481,34 @@ python write_specfile () {
 		scr = scr[:pos] + 'if [ "$1" = "0" ] ; then\n' + scr[pos:] + '\nfi'
 		return scr
 
-	packages = bb.data.getVar('PACKAGES', d, True)
+	packages = d.getVar('PACKAGES', True)
 	if not packages or packages == '':
 		bb.debug(1, "No packages; nothing to do")
 		return
 
-	pkgdest = bb.data.getVar('PKGDEST', d, True)
+	pkgdest = d.getVar('PKGDEST', True)
 	if not pkgdest:
 		bb.fatal("No PKGDEST")
 		return
 
-	outspecfile = bb.data.getVar('OUTSPECFILE', d, True)
+	outspecfile = d.getVar('OUTSPECFILE', True)
 	if not outspecfile:
 		bb.fatal("No OUTSPECFILE")
 		return
 
 	# Construct the SPEC file...
-	srcname    = strip_multilib(bb.data.getVar('PN', d, True), d)
-	srcsummary = (bb.data.getVar('SUMMARY', d, True) or bb.data.getVar('DESCRIPTION', d, True) or ".")
-	srcversion = bb.data.getVar('PKGV', d, True).replace('-', '+')
-	srcrelease = bb.data.getVar('PKGR', d, True)
-	srcepoch   = (bb.data.getVar('PKGE', d, True) or "")
-	srclicense = bb.data.getVar('LICENSE', d, True)
-	srcsection = bb.data.getVar('SECTION', d, True)
-	srcmaintainer  = bb.data.getVar('MAINTAINER', d, True)
-	srchomepage    = bb.data.getVar('HOMEPAGE', d, True)
-	srcdescription = bb.data.getVar('DESCRIPTION', d, True) or "."
-
-	srcdepends     = strip_multilib(bb.data.getVar('DEPENDS', d, True), d)
+	srcname    = strip_multilib(d.getVar('PN', True), d)
+	srcsummary = (d.getVar('SUMMARY', True) or d.getVar('DESCRIPTION', True) or ".")
+	srcversion = d.getVar('PKGV', True).replace('-', '+')
+	srcrelease = d.getVar('PKGR', True)
+	srcepoch   = (d.getVar('PKGE', True) or "")
+	srclicense = d.getVar('LICENSE', True)
+	srcsection = d.getVar('SECTION', True)
+	srcmaintainer  = d.getVar('MAINTAINER', True)
+	srchomepage    = d.getVar('HOMEPAGE', True)
+	srcdescription = d.getVar('DESCRIPTION', True) or "."
+
+	srcdepends     = strip_multilib(d.getVar('DEPENDS', True), d)
 	srcrdepends    = []
 	srcrrecommends = []
 	srcrsuggests   = []
@@ -538,28 +538,28 @@ python write_specfile () {
 
 		lf = bb.utils.lockfile(root + ".lock")
 
-		bb.data.setVar('ROOT', '', localdata)
-		bb.data.setVar('ROOT_%s' % pkg, root, localdata)
-		pkgname = bb.data.getVar('PKG_%s' % pkg, localdata, 1)
+		localdata.setVar('ROOT', '')
+		localdata.setVar('ROOT_%s' % pkg, root)
+		pkgname = localdata.getVar('PKG_%s' % pkg, 1)
 		if not pkgname:
 			pkgname = pkg
-		bb.data.setVar('PKG', pkgname, localdata)
+		localdata.setVar('PKG', pkgname)
 
-		bb.data.setVar('OVERRIDES', pkg, localdata)
+		localdata.setVar('OVERRIDES', pkg)
 
 		bb.data.update_data(localdata)
 
-		conffiles = (bb.data.getVar('CONFFILES', localdata, True) or "").split()
+		conffiles = (localdata.getVar('CONFFILES', True) or "").split()
 
 		splitname    = strip_multilib(pkgname, d)
 
-		splitsummary = (bb.data.getVar('SUMMARY', localdata, True) or bb.data.getVar('DESCRIPTION', localdata, True) or ".")
-		splitversion = (bb.data.getVar('PKGV', localdata, True) or "").replace('-', '+')
-		splitrelease = (bb.data.getVar('PKGR', localdata, True) or "")
-		splitepoch   = (bb.data.getVar('PKGE', localdata, True) or "")
-		splitlicense = (bb.data.getVar('LICENSE', localdata, True) or "")
-		splitsection = (bb.data.getVar('SECTION', localdata, True) or "")
-		splitdescription = (bb.data.getVar('DESCRIPTION', localdata, True) or ".")
+		splitsummary = (localdata.getVar('SUMMARY', True) or localdata.getVar('DESCRIPTION', True) or ".")
+		splitversion = (localdata.getVar('PKGV', True) or "").replace('-', '+')
+		splitrelease = (localdata.getVar('PKGR', True) or "")
+		splitepoch   = (localdata.getVar('PKGE', True) or "")
+		splitlicense = (localdata.getVar('LICENSE', True) or "")
+		splitsection = (localdata.getVar('SECTION', True) or "")
+		splitdescription = (localdata.getVar('DESCRIPTION', True) or ".")
 
 		translate_vers('RDEPENDS', localdata)
 		translate_vers('RRECOMMENDS', localdata)
@@ -571,12 +571,12 @@ python write_specfile () {
 		# Map the dependencies into their final form
 		bb.build.exec_func("mapping_rename_hook", localdata)
 
-		splitrdepends    = strip_multilib(bb.data.getVar('RDEPENDS', localdata, True), d) or ""
-		splitrrecommends = strip_multilib(bb.data.getVar('RRECOMMENDS', localdata, True), d) or ""
-		splitrsuggests   = strip_multilib(bb.data.getVar('RSUGGESTS', localdata, True), d) or ""
-		splitrprovides   = strip_multilib(bb.data.getVar('RPROVIDES', localdata, True), d) or ""
-		splitrreplaces   = strip_multilib(bb.data.getVar('RREPLACES', localdata, True), d) or ""
-		splitrconflicts  = strip_multilib(bb.data.getVar('RCONFLICTS', localdata, True), d) or ""
+		splitrdepends    = strip_multilib(localdata.getVar('RDEPENDS', True), d) or ""
+		splitrrecommends = strip_multilib(localdata.getVar('RRECOMMENDS', True), d) or ""
+		splitrsuggests   = strip_multilib(localdata.getVar('RSUGGESTS', True), d) or ""
+		splitrprovides   = strip_multilib(localdata.getVar('RPROVIDES', True), d) or ""
+		splitrreplaces   = strip_multilib(localdata.getVar('RREPLACES', True), d) or ""
+		splitrconflicts  = strip_multilib(localdata.getVar('RCONFLICTS', True), d) or ""
 		splitrobsoletes  = []
 
 		# For now we need to manually supplement RPROVIDES with any update-alternatives links
@@ -592,14 +592,14 @@ python write_specfile () {
 			srcrreplaces   = splitrreplaces
 			srcrconflicts  = splitrconflicts
 
-			srcpreinst  = bb.data.getVar('pkg_preinst', localdata, True)
-			srcpostinst = bb.data.getVar('pkg_postinst', localdata, True)
-			srcprerm    = bb.data.getVar('pkg_prerm', localdata, True)
-			srcpostrm   = bb.data.getVar('pkg_postrm', localdata, True)
+			srcpreinst  = localdata.getVar('pkg_preinst', True)
+			srcpostinst = localdata.getVar('pkg_postinst', True)
+			srcprerm    = localdata.getVar('pkg_prerm', True)
+			srcpostrm   = localdata.getVar('pkg_postrm', True)
 
 			file_list = []
 			walk_files(root, file_list, conffiles)
-			if not file_list and bb.data.getVar('ALLOW_EMPTY', localdata) != "1":
+			if not file_list and localdata.getVar('ALLOW_EMPTY') != "1":
 				bb.note("Not creating empty RPM package for %s" % splitname)
 			else:
 				bb.note("Creating RPM package for %s" % splitname)
@@ -672,7 +672,7 @@ python write_specfile () {
 
 		# Now process scriptlets
 		for script in ["preinst", "postinst", "prerm", "postrm"]:
-			scriptvar = bb.data.getVar('pkg_%s' % script, localdata, True)
+			scriptvar = localdata.getVar('pkg_%s' % script, True)
 			if not scriptvar:
 				continue
 			if script == 'preinst':
@@ -691,7 +691,7 @@ python write_specfile () {
 		# Now process files
 		file_list = []
 		walk_files(root, file_list, conffiles)
-		if not file_list and bb.data.getVar('ALLOW_EMPTY', localdata) != "1":
+		if not file_list and localdata.getVar('ALLOW_EMPTY') != "1":
 			bb.note("Not creating empty RPM package for %s" % splitname)
 		else:
 			spec_files_bottom.append('%%files -n %s' % splitname)
@@ -813,29 +813,29 @@ python do_package_rpm () {
 	# We need a simple way to remove the MLPREFIX from the package name,
 	# and dependency information...
 	def strip_multilib(name, d):
-		ml = bb.data.getVar("MLPREFIX", d, True)
+		ml = d.getVar("MLPREFIX", True)
 		if ml and name and len(ml) != 0 and name.find(ml) >= 0:
 			return "".join(name.split(ml))
 		return name
 
-	workdir = bb.data.getVar('WORKDIR', d, True)
-	outdir = bb.data.getVar('DEPLOY_DIR_IPK', d, True)
-	tmpdir = bb.data.getVar('TMPDIR', d, True)
-	pkgd = bb.data.getVar('PKGD', d, True)
-	pkgdest = bb.data.getVar('PKGDEST', d, True)
+	workdir = d.getVar('WORKDIR', True)
+	outdir = d.getVar('DEPLOY_DIR_IPK', True)
+	tmpdir = d.getVar('TMPDIR', True)
+	pkgd = d.getVar('PKGD', True)
+	pkgdest = d.getVar('PKGDEST', True)
 	if not workdir or not outdir or not pkgd or not tmpdir:
 		bb.error("Variables incorrectly set, unable to package")
 		return
 
-	packages = bb.data.getVar('PACKAGES', d, True)
+	packages = d.getVar('PACKAGES', True)
 	if not packages or packages == '':
 		bb.debug(1, "No packages; nothing to do")
 		return
 
 	# Construct the spec file...
-	srcname    = strip_multilib(bb.data.getVar('PN', d, True), d)
+	srcname    = strip_multilib(d.getVar('PN', True), d)
 	outspecfile = workdir + "/" + srcname + ".spec"
-	bb.data.setVar('OUTSPECFILE', outspecfile, d)
+	d.setVar('OUTSPECFILE', outspecfile)
 	bb.build.exec_func('write_specfile', d)
 
 	# Construct per file dependencies file
@@ -844,10 +844,10 @@ python do_package_rpm () {
 		outfile.write("\n# Dependency table\n")
 		for pkg in packages.split():
 			dependsflist_key = 'FILE' + varname + 'FLIST' + "_" + pkg
-			dependsflist = (bb.data.getVar(dependsflist_key, d, True) or "")
+			dependsflist = (d.getVar(dependsflist_key, True) or "")
 			for dfile in dependsflist.split():
 				key = "FILE" + varname + "_" + dfile + "_" + pkg
-				depends_dict = bb.utils.explode_dep_versions(bb.data.getVar(key, d, True) or "")
+				depends_dict = bb.utils.explode_dep_versions(d.getVar(key, True) or "")
 				file = dfile.replace("@underscore@", "_")
 				file = file.replace("@closebrace@", "]")
 				file = file.replace("@openbrace@", "[")
@@ -899,15 +899,15 @@ python do_package_rpm () {
 	os.chmod(outprovides, 0755)
 
 	# Setup the rpmbuild arguments...
-	rpmbuild = bb.data.getVar('RPMBUILD', d, True)
-	targetsys = bb.data.getVar('TARGET_SYS', d, True)
-	targetvendor = bb.data.getVar('TARGET_VENDOR', d, True)
-	package_arch = bb.data.getVar('PACKAGE_ARCH', d, True) or ""
+	rpmbuild = d.getVar('RPMBUILD', True)
+	targetsys = d.getVar('TARGET_SYS', True)
+	targetvendor = d.getVar('TARGET_VENDOR', True)
+	package_arch = d.getVar('PACKAGE_ARCH', True) or ""
 	if package_arch not in "all any noarch".split():
-		ml_prefix = (bb.data.getVar('MLPREFIX', d, True) or "").replace("-", "_")
-		bb.data.setVar('PACKAGE_ARCH_EXTEND', ml_prefix + package_arch, d)
+		ml_prefix = (d.getVar('MLPREFIX', True) or "").replace("-", "_")
+		d.setVar('PACKAGE_ARCH_EXTEND', ml_prefix + package_arch)
 	else:
-		bb.data.setVar('PACKAGE_ARCH_EXTEND', package_arch, d)
+		d.setVar('PACKAGE_ARCH_EXTEND', package_arch)
 	pkgwritedir = bb.data.expand('${PKGWRITEDIRRPM}/${PACKAGE_ARCH_EXTEND}', d)
 	pkgarch = bb.data.expand('${PACKAGE_ARCH_EXTEND}${TARGET_VENDOR}-${TARGET_OS}', d)
 	magicfile = bb.data.expand('${STAGING_DIR_NATIVE}/usr/share/misc/magic.mgc', d)
@@ -927,19 +927,19 @@ python do_package_rpm () {
 	cmd = cmd + " -bb " + outspecfile
 
 	# Build the rpm package!
-	bb.data.setVar('BUILDSPEC', cmd + "\n", d)
-	bb.data.setVarFlag('BUILDSPEC', 'func', '1', d)
+	d.setVar('BUILDSPEC', cmd + "\n")
+	d.setVarFlag('BUILDSPEC', 'func', '1')
 	bb.build.exec_func('BUILDSPEC', d)
 }
 
 python () {
-    if bb.data.getVar('PACKAGES', d, True) != '':
-        deps = (bb.data.getVarFlag('do_package_write_rpm', 'depends', d) or "").split()
+    if d.getVar('PACKAGES', True) != '':
+        deps = (d.getVarFlag('do_package_write_rpm', 'depends') or "").split()
         deps.append('rpm-native:do_populate_sysroot')
         deps.append('virtual/fakeroot-native:do_populate_sysroot')
         bb.data.setVarFlag('do_package_write_rpm', 'depends', " ".join(deps), d)
-        bb.data.setVarFlag('do_package_write_rpm', 'fakeroot', 1, d)
-        bb.data.setVarFlag('do_package_write_rpm_setscene', 'fakeroot', 1, d)
+        d.setVarFlag('do_package_write_rpm', 'fakeroot', 1)
+        d.setVarFlag('do_package_write_rpm_setscene', 'fakeroot', 1)
 }
 
 SSTATETASKS += "do_package_write_rpm"

+ 19 - 19
meta/classes/package_tar.bbclass

@@ -3,15 +3,15 @@ inherit package
 IMAGE_PKGTYPE ?= "tar"
 
 python package_tar_fn () {
-	fn = os.path.join(bb.data.getVar('DEPLOY_DIR_TAR', d), "%s-%s-%s.tar.gz" % (bb.data.getVar('PKG', d), bb.data.getVar('PKGV', d), bb.data.getVar('PKGR', d)))
+	fn = os.path.join(d.getVar('DEPLOY_DIR_TAR'), "%s-%s-%s.tar.gz" % (d.getVar('PKG'), d.getVar('PKGV'), d.getVar('PKGR')))
 	fn = bb.data.expand(fn, d)
-	bb.data.setVar('PKGFN', fn, d)
+	d.setVar('PKGFN', fn)
 }
 
 python package_tar_install () {
-	pkg = bb.data.getVar('PKG', d, 1)
-	pkgfn = bb.data.getVar('PKGFN', d, 1)
-	rootfs = bb.data.getVar('IMAGE_ROOTFS', d, 1)
+	pkg = d.getVar('PKG', 1)
+	pkgfn = d.getVar('PKGFN', 1)
+	rootfs = d.getVar('IMAGE_ROOTFS', 1)
 
 	if None in (pkg,pkgfn,rootfs):
 		bb.error("missing variables (one or more of PKG, PKGFN, IMAGEROOTFS)")
@@ -35,24 +35,24 @@ python package_tar_install () {
 }
 
 python do_package_tar () {
-	workdir = bb.data.getVar('WORKDIR', d, 1)
+	workdir = d.getVar('WORKDIR', 1)
 	if not workdir:
 		bb.error("WORKDIR not defined, unable to package")
 		return
 
-	outdir = bb.data.getVar('DEPLOY_DIR_TAR', d, 1)
+	outdir = d.getVar('DEPLOY_DIR_TAR', 1)
 	if not outdir:
 		bb.error("DEPLOY_DIR_TAR not defined, unable to package")
 		return
 	bb.mkdirhier(outdir)
 
-	dvar = bb.data.getVar('D', d, 1)
+	dvar = d.getVar('D', 1)
 	if not dvar:
 		bb.error("D not defined, unable to package")
 		return
 	bb.mkdirhier(dvar)
 
-	packages = bb.data.getVar('PACKAGES', d, 1)
+	packages = d.getVar('PACKAGES', 1)
 	if not packages:
 		bb.debug(1, "PACKAGES not defined, nothing to package")
 		return
@@ -61,11 +61,11 @@ python do_package_tar () {
 		localdata = bb.data.createCopy(d)
 		root = "%s/install/%s" % (workdir, pkg)
 
-		bb.data.setVar('ROOT', '', localdata)
-		bb.data.setVar('ROOT_%s' % pkg, root, localdata)
-		bb.data.setVar('PKG', pkg, localdata)
+		localdata.setVar('ROOT', '')
+		localdata.setVar('ROOT_%s' % pkg, root)
+		localdata.setVar('PKG', pkg)
 
-		overrides = bb.data.getVar('OVERRIDES', localdata)
+		overrides = localdata.getVar('OVERRIDES')
 		if not overrides:
 			raise bb.build.FuncFailed('OVERRIDES not defined')
 		overrides = bb.data.expand(overrides, localdata)
@@ -73,17 +73,17 @@ python do_package_tar () {
 
 		bb.data.update_data(localdata)
 
-		root = bb.data.getVar('ROOT', localdata)
+		root = localdata.getVar('ROOT')
 		bb.mkdirhier(root)
 		basedir = os.path.dirname(root)
 		pkgoutdir = outdir
 		bb.mkdirhier(pkgoutdir)
 		bb.build.exec_func('package_tar_fn', localdata)
-		tarfn = bb.data.getVar('PKGFN', localdata, 1)
+		tarfn = localdata.getVar('PKGFN', 1)
 		os.chdir(root)
 		from glob import glob
 		if not glob('*'):
-			bb.note("Not creating empty archive for %s-%s-%s" % (pkg, bb.data.getVar('PKGV', localdata, 1), bb.data.getVar('PKGR', localdata, 1)))
+			bb.note("Not creating empty archive for %s-%s-%s" % (pkg, localdata.getVar('PKGV', 1), localdata.getVar('PKGR', 1)))
 			continue
 		ret = os.system("tar -czf %s %s" % (tarfn, '.'))
 		if ret != 0:
@@ -91,12 +91,12 @@ python do_package_tar () {
 }
 
 python () {
-    if bb.data.getVar('PACKAGES', d, True) != '':
-        deps = (bb.data.getVarFlag('do_package_write_tar', 'depends', d) or "").split()
+    if d.getVar('PACKAGES', True) != '':
+        deps = (d.getVarFlag('do_package_write_tar', 'depends') or "").split()
         deps.append('tar-native:do_populate_sysroot')
         deps.append('virtual/fakeroot-native:do_populate_sysroot')
         bb.data.setVarFlag('do_package_write_tar', 'depends', " ".join(deps), d)
-        bb.data.setVarFlag('do_package_write_ipk', 'fakeroot', "1", d)
+        d.setVarFlag('do_package_write_ipk', 'fakeroot', "1")
 }
 
 

+ 4 - 4
meta/classes/packagedata.bbclass

@@ -1,13 +1,13 @@
 python read_subpackage_metadata () {
 	import oe.packagedata
 
-	data = oe.packagedata.read_pkgdata(bb.data.getVar('PN', d, 1), d)
+	data = oe.packagedata.read_pkgdata(d.getVar('PN', 1), d)
 
 	for key in data.keys():
-		bb.data.setVar(key, data[key], d)
+		d.setVar(key, data[key])
 
-	for pkg in bb.data.getVar('PACKAGES', d, 1).split():
+	for pkg in d.getVar('PACKAGES', 1).split():
 		sdata = oe.packagedata.read_subpkgdata(pkg, d)
 		for key in sdata.keys():
-			bb.data.setVar(key, sdata[key], d)
+			d.setVar(key, sdata[key])
 }

+ 4 - 4
meta/classes/packagehistory.bbclass

@@ -10,8 +10,8 @@ PKGHIST_DIR = "${TMPDIR}/pkghistory/${BASEPKG_TARGET_SYS}/"
 # for comparision when writing future packages
 #
 python emit_pkghistory() {
-	packages = bb.data.getVar('PACKAGES', d, True)
-	pkghistdir = bb.data.getVar('PKGHIST_DIR', d, True)
+	packages = d.getVar('PACKAGES', True)
+	pkghistdir = d.getVar('PKGHIST_DIR', True)
 
 
 	# Should check PACKAGES here to see if anything removed
@@ -72,14 +72,14 @@ def check_pkghistory(pkg, pe, pv, pr, lastversion):
 def write_pkghistory(pkg, pe, pv, pr, d):
 	bb.debug(2, "Writing package history")
 
-	pkghistdir = bb.data.getVar('PKGHIST_DIR', d, True)
+	pkghistdir = d.getVar('PKGHIST_DIR', True)
 
 	verpath = os.path.join(pkghistdir, pkg, pe, pv, pr)
 	if not os.path.exists(verpath):
 		os.makedirs(verpath)
 
 def write_latestlink(pkg, pe, pv, pr, d):
-	pkghistdir = bb.data.getVar('PKGHIST_DIR', d, True)
+	pkghistdir = d.getVar('PKGHIST_DIR', True)
 
 	def rm_link(path):
 		try: 

+ 14 - 14
meta/classes/patch.bbclass

@@ -10,7 +10,7 @@ inherit terminal
 python patch_do_patch() {
 	import oe.patch
 
-	src_uri = (bb.data.getVar('SRC_URI', d, 1) or '').split()
+	src_uri = (d.getVar('SRC_URI', 1) or '').split()
 	if not src_uri:
 		return
 
@@ -20,23 +20,23 @@ python patch_do_patch() {
 		"git": oe.patch.GitApplyTree,
 	}
 
-	cls = patchsetmap[bb.data.getVar('PATCHTOOL', d, 1) or 'quilt']
+	cls = patchsetmap[d.getVar('PATCHTOOL', 1) or 'quilt']
 
 	resolvermap = {
 		"noop": oe.patch.NOOPResolver,
 		"user": oe.patch.UserResolver,
 	}
 
-	rcls = resolvermap[bb.data.getVar('PATCHRESOLVE', d, 1) or 'user']
+	rcls = resolvermap[d.getVar('PATCHRESOLVE', 1) or 'user']
 
-	s = bb.data.getVar('S', d, 1)
+	s = d.getVar('S', 1)
 
 	path = os.getenv('PATH')
-	os.putenv('PATH', bb.data.getVar('PATH', d, 1))
+	os.putenv('PATH', d.getVar('PATH', 1))
 
 	classes = {}
 
-	workdir = bb.data.getVar('WORKDIR', d, 1)
+	workdir = d.getVar('WORKDIR', 1)
 	for url in src_uri:
 		(type, host, path, user, pswd, parm) = bb.decodeurl(url)
 
@@ -76,13 +76,13 @@ python patch_do_patch() {
 			pname = os.path.basename(local)
 
 		if "mindate" in parm or "maxdate" in parm:
-			pn = bb.data.getVar('PN', d, 1)
-			srcdate = bb.data.getVar('SRCDATE_%s' % pn, d, 1)
+			pn = d.getVar('PN', 1)
+			srcdate = d.getVar('SRCDATE_%s' % pn, 1)
 			if not srcdate:
-				srcdate = bb.data.getVar('SRCDATE', d, 1)
+				srcdate = d.getVar('SRCDATE', 1)
 
 			if srcdate == "now":
-				srcdate = bb.data.getVar('DATE', d, 1)
+				srcdate = d.getVar('DATE', 1)
 
 			if "maxdate" in parm and parm["maxdate"] < srcdate:
 				bb.note("Patch '%s' is outdated" % pname)
@@ -94,25 +94,25 @@ python patch_do_patch() {
 
 
 		if "minrev" in parm:
-			srcrev = bb.data.getVar('SRCREV', d, 1)
+			srcrev = d.getVar('SRCREV', 1)
 			if srcrev and srcrev < parm["minrev"]:
 				bb.note("Patch '%s' applies to later revisions" % pname)
 				continue
 
 		if "maxrev" in parm:
-			srcrev = bb.data.getVar('SRCREV', d, 1)		
+			srcrev = d.getVar('SRCREV', 1)		
 			if srcrev and srcrev > parm["maxrev"]:
 				bb.note("Patch '%s' applies to earlier revisions" % pname)
 				continue
 
 		if "rev" in parm:
-			srcrev = bb.data.getVar('SRCREV', d, 1)		
+			srcrev = d.getVar('SRCREV', 1)		
 			if srcrev and parm["rev"] not in srcrev:
 				bb.note("Patch '%s' doesn't apply to revision" % pname)
 				continue
 
 		if "notrev" in parm:
-			srcrev = bb.data.getVar('SRCREV', d, 1)		
+			srcrev = d.getVar('SRCREV', 1)		
 			if srcrev and parm["notrev"] in srcrev:
 				bb.note("Patch '%s' doesn't apply to revision" % pname)
 				continue

+ 1 - 1
meta/classes/pkg_distribute.bbclass

@@ -1,6 +1,6 @@
 PKG_DISTRIBUTECOMMAND[func] = "1"
 python do_distribute_packages () {
-	cmd = bb.data.getVar('PKG_DISTRIBUTECOMMAND', d, 1)
+	cmd = d.getVar('PKG_DISTRIBUTECOMMAND', 1)
 	if not cmd:
 		raise bb.build.FuncFailed("Unable to distribute packages, PKG_DISTRIBUTECOMMAND not defined")
 	bb.build.exec_func('PKG_DISTRIBUTECOMMAND', d)

+ 6 - 6
meta/classes/pkg_metainfo.bbclass

@@ -1,5 +1,5 @@
 python do_pkg_write_metainfo () {
-	deploydir = bb.data.getVar('DEPLOY_DIR', d, 1)
+	deploydir = d.getVar('DEPLOY_DIR', 1)
 	if not deploydir:
 		bb.error("DEPLOY_DIR not defined, unable to write package info")
 		return
@@ -9,11 +9,11 @@ python do_pkg_write_metainfo () {
 	except OSError:
 		raise bb.build.FuncFailed("unable to open package-info file for writing.")
 	
-	name = bb.data.getVar('PN', d, 1)
-	version = bb.data.getVar('PV', d, 1)
-	desc = bb.data.getVar('DESCRIPTION', d, 1)
-	page = bb.data.getVar('HOMEPAGE', d, 1)
-	lic = bb.data.getVar('LICENSE', d, 1)
+	name = d.getVar('PN', 1)
+	version = d.getVar('PV', 1)
+	desc = d.getVar('DESCRIPTION', 1)
+	page = d.getVar('HOMEPAGE', 1)
+	lic = d.getVar('LICENSE', 1)
 	
 	infofile.write("|| "+ name +" || "+ version + " || "+ desc +" || "+ page +" || "+ lic + " ||\n" ) 
 	infofile.close()

+ 2 - 2
meta/classes/populate_sdk_deb.bbclass

@@ -2,8 +2,8 @@ do_populate_sdk[depends] += "dpkg-native:do_populate_sysroot apt-native:do_popul
 do_populate_sdk[recrdeptask] += "do_package_write_deb"
 
 
-DEB_SDK_ARCH = "${@[bb.data.getVar('SDK_ARCH', d, 1), "i386"]\
-                [bb.data.getVar('SDK_ARCH', d, 1) in \
+DEB_SDK_ARCH = "${@[d.getVar('SDK_ARCH', 1), "i386"]\
+                [d.getVar('SDK_ARCH', 1) in \
                 ["x86", "i486", "i586", "i686", "pentium"]]}"
 
 populate_sdk_post_deb () {

+ 1 - 1
meta/classes/populate_sdk_rpm.bbclass

@@ -127,6 +127,6 @@ python () {
             localdata.setVar("DEFAULTTUNE", localdata.getVar("DEFAULTTUNE_virtclass-multilib-" + eext[1], False) or "")
             ml_package_archs += localdata.getVar("PACKAGE_ARCHS", True) or ""
             #bb.note("ML_PACKAGE_ARCHS %s %s %s" % (eext[1], localdata.getVar("PACKAGE_ARCHS", True) or "(none)", overrides))
-    bb.data.setVar('MULTILIB_PACKAGE_ARCHS', ml_package_archs, d)
+    d.setVar('MULTILIB_PACKAGE_ARCHS', ml_package_archs)
 }
 

+ 1 - 1
meta/classes/qemu.bbclass

@@ -6,7 +6,7 @@
 def qemu_target_binary(data):
 	import bb
 
-	target_arch = bb.data.getVar("TARGET_ARCH", data, 1)
+	target_arch = data.getVar("TARGET_ARCH", 1)
 	if target_arch in ("i486", "i586", "i686"):
 		target_arch = "i386"
 	elif target_arch == "powerpc":

+ 1 - 1
meta/classes/qt4e.bbclass

@@ -1,4 +1,4 @@
-DEPENDS_prepend = "${@["qt4-embedded ", ""][(bb.data.getVar('PN', d, 1)[:12] == 'qt4-embedded')]}"
+DEPENDS_prepend = "${@["qt4-embedded ", ""][(d.getVar('PN', 1)[:12] == 'qt4-embedded')]}"
 
 inherit qmake2
 

+ 1 - 1
meta/classes/qt4x11.bbclass

@@ -1,4 +1,4 @@
-DEPENDS_prepend = "${@["qt4-x11-free ", ""][(bb.data.getVar('BPN', d, True)[:12] == 'qt4-x11-free')]}"
+DEPENDS_prepend = "${@["qt4-x11-free ", ""][(d.getVar('BPN', True)[:12] == 'qt4-x11-free')]}"
 
 inherit qmake2
 

+ 1 - 1
meta/classes/relocatable.bbclass

@@ -8,7 +8,7 @@ def process_dir (directory, d):
     import stat
 
     cmd = bb.data.expand('${CHRPATH_BIN}', d)
-    tmpdir = bb.data.getVar('TMPDIR', d)
+    tmpdir = d.getVar('TMPDIR')
     basedir = bb.data.expand('${base_prefix}', d)
 
     #bb.debug("Checking %s for binaries to process" % directory)

+ 5 - 5
meta/classes/rootfs_ipk.bbclass

@@ -185,13 +185,13 @@ ipk_insert_feed_uris () {
 
 python () {
 
-    if bb.data.getVar('BUILD_IMAGES_FROM_FEEDS', d, True):
-        flags = bb.data.getVarFlag('do_rootfs', 'recrdeptask', d)
+    if d.getVar('BUILD_IMAGES_FROM_FEEDS', True):
+        flags = d.getVarFlag('do_rootfs', 'recrdeptask')
         flags = flags.replace("do_package_write_ipk", "")
         flags = flags.replace("do_deploy", "")
         flags = flags.replace("do_populate_sysroot", "")
-        bb.data.setVarFlag('do_rootfs', 'recrdeptask', flags, d)
-        bb.data.setVar('OPKG_PREPROCESS_COMMANDS', "package_generate_archlist\nipk_insert_feed_uris", d)
-        bb.data.setVar('OPKG_POSTPROCESS_COMMANDS', '', d)
+        d.setVarFlag('do_rootfs', 'recrdeptask', flags)
+        d.setVar('OPKG_PREPROCESS_COMMANDS', "package_generate_archlist\nipk_insert_feed_uris")
+        d.setVar('OPKG_POSTPROCESS_COMMANDS', '')
 }
 

+ 7 - 7
meta/classes/rootfs_rpm.bbclass

@@ -200,14 +200,14 @@ install_all_locales() {
 }
 
 python () {
-    if bb.data.getVar('BUILD_IMAGES_FROM_FEEDS', d, True):
-        flags = bb.data.getVarFlag('do_rootfs', 'recrdeptask', d)
+    if d.getVar('BUILD_IMAGES_FROM_FEEDS', True):
+        flags = d.getVarFlag('do_rootfs', 'recrdeptask')
         flags = flags.replace("do_package_write_rpm", "")
         flags = flags.replace("do_deploy", "")
         flags = flags.replace("do_populate_sysroot", "")
-        bb.data.setVarFlag('do_rootfs', 'recrdeptask', flags, d)
-        bb.data.setVar('RPM_PREPROCESS_COMMANDS', '', d)
-        bb.data.setVar('RPM_POSTPROCESS_COMMANDS', '', d)
+        d.setVarFlag('do_rootfs', 'recrdeptask', flags)
+        d.setVar('RPM_PREPROCESS_COMMANDS', '')
+        d.setVar('RPM_POSTPROCESS_COMMANDS', '')
 
     ml_package_archs = ""
     ml_prefix_list = ""
@@ -224,6 +224,6 @@ python () {
             ml_package_archs += " " + package_archs
             ml_prefix_list += " " + eext[1]
             #bb.note("ML_PACKAGE_ARCHS %s %s %s" % (eext[1], localdata.getVar("PACKAGE_ARCHS", True) or "(none)", overrides))
-    bb.data.setVar('MULTILIB_PACKAGE_ARCHS', ml_package_archs, d)
-    bb.data.setVar('MULTILIB_PREFIX_LIST', ml_prefix_list, d)
+    d.setVar('MULTILIB_PACKAGE_ARCHS', ml_package_archs)
+    d.setVar('MULTILIB_PREFIX_LIST', ml_prefix_list)
 }

+ 5 - 5
meta/classes/sanity.bbclass

@@ -14,7 +14,7 @@ def raise_sanity_error(msg):
 def check_conf_exists(fn, data):
     bbpath = []
     fn = bb.data.expand(fn, data)
-    vbbpath = bb.data.getVar("BBPATH", data)
+    vbbpath = data.getVar("BBPATH")
     if vbbpath:
         bbpath += vbbpath.split(":")
     for p in bbpath:
@@ -87,12 +87,12 @@ def check_connectivity(d):
     # URI's to check can be set in the CONNECTIVITY_CHECK_URIS variable
     # using the same syntax as for SRC_URI. If the variable is not set
     # the check is skipped
-    test_uris = (bb.data.getVar('CONNECTIVITY_CHECK_URIS', d, True) or "").split()
+    test_uris = (d.getVar('CONNECTIVITY_CHECK_URIS', True) or "").split()
     retval = ""
 
     # Only check connectivity if network enabled and the
     # CONNECTIVITY_CHECK_URIS are set
-    network_enabled = not bb.data.getVar('BB_NO_NETWORK', d, True)
+    network_enabled = not d.getVar('BB_NO_NETWORK', True)
     check_enabled = len(test_uris)
     # Take a copy of the data store and unset MIRRORS and PREMIRROS
     data = bb.data.createCopy(d)
@@ -105,7 +105,7 @@ def check_connectivity(d):
         except Exception:
             # Allow the message to be configured so that users can be
             # pointed to a support mechanism.
-            msg = bb.data.getVar('CONNECTIVITY_CHECK_MSG', data, True) or ""
+            msg = data.getVar('CONNECTIVITY_CHECK_MSG', True) or ""
             if len(msg) == 0:
                 msg = "Failed to fetch test data from the network. Please ensure your network is configured correctly.\n"
             retval = msg
@@ -450,7 +450,7 @@ def check_sanity(e):
 
 addhandler check_sanity_eventhandler
 python check_sanity_eventhandler() {
-    if bb.event.getName(e) == "ConfigParsed" and bb.data.getVar("BB_WORKERCONTEXT", e.data, True) != "1":
+    if bb.event.getName(e) == "ConfigParsed" and e.data.getVar("BB_WORKERCONTEXT", True) != "1":
         check_sanity(e)
 
     return

+ 1 - 1
meta/classes/siteconfig.bbclass

@@ -2,7 +2,7 @@ python siteconfig_do_siteconfig () {
 	shared_state = sstate_state_fromvars(d)
 	if shared_state['name'] != 'populate-sysroot':
 		return
-	if not os.path.isdir(os.path.join(bb.data.getVar('FILE_DIRNAME', d, 1), 'site_config')):
+	if not os.path.isdir(os.path.join(d.getVar('FILE_DIRNAME', 1), 'site_config')):
 		bb.debug(1, "No site_config directory, skipping do_siteconfig")
 		return
 	bb.build.exec_func('do_siteconfig_gencache', d)

+ 1 - 1
meta/classes/siteinfo.bbclass

@@ -130,7 +130,7 @@ def siteinfo_get_files(d, no_cache = False):
     if no_cache: return sitefiles
 
     # Now check for siteconfig cache files
-    path_siteconfig = bb.data.getVar('SITECONFIG_SYSROOTCACHE', d, 1)
+    path_siteconfig = d.getVar('SITECONFIG_SYSROOTCACHE', 1)
     if os.path.isdir(path_siteconfig):
         for i in os.listdir(path_siteconfig):
             filename = os.path.join(path_siteconfig, i)

+ 6 - 6
meta/classes/sourcepkg.bbclass

@@ -6,12 +6,12 @@ DISTRO ?= "openembedded"
 
 def get_src_tree(d):
 
-	workdir = bb.data.getVar('WORKDIR', d, 1)
+	workdir = d.getVar('WORKDIR', 1)
 	if not workdir:
 		bb.error("WORKDIR not defined, unable to find source tree.")
 		return
 
-	s = bb.data.getVar('S', d, 0)
+	s = d.getVar('S', 0)
 	if not s:
 		bb.error("S not defined, unable to find source tree.")
 		return
@@ -55,8 +55,8 @@ sourcepkg_do_archive_bb() {
 
 python sourcepkg_do_dumpdata() {
 
-	workdir = bb.data.getVar('WORKDIR', d, 1)
-	distro = bb.data.getVar('DISTRO', d, 1)
+	workdir = d.getVar('WORKDIR', 1)
+	distro = d.getVar('DISTRO', 1)
 	s_tree = get_src_tree(d)
 	openembeddeddir = os.path.join(workdir, s_tree, distro)
 	dumpfile = os.path.join(openembeddeddir, bb.data.expand("${P}-${PR}.showdata.dump",d))
@@ -73,8 +73,8 @@ python sourcepkg_do_dumpdata() {
         bb.data.emit_env(f, d, True)
 	# emit the metadata which isnt valid shell
 	for e in d.keys():
-		if bb.data.getVarFlag(e, 'python', d):
-			f.write("\npython %s () {\n%s}\n" % (e, bb.data.getVar(e, d, 1)))
+		if d.getVarFlag(e, 'python'):
+			f.write("\npython %s () {\n%s}\n" % (e, d.getVar(e, 1)))
 	f.close()
 }
 

+ 8 - 8
meta/classes/src_distribute.bbclass

@@ -3,12 +3,12 @@ python do_distribute_sources () {
 	l = bb.data.createCopy(d)
 	bb.data.update_data(l)
 
-	sources_dir = bb.data.getVar('SRC_DISTRIBUTEDIR', d, 1)
-	src_uri = bb.data.getVar('SRC_URI', d, 1).split()
+	sources_dir = d.getVar('SRC_DISTRIBUTEDIR', 1)
+	src_uri = d.getVar('SRC_URI', 1).split()
 	fetcher = bb.fetch2.Fetch(src_uri, d)
 	ud = fetcher.ud
 
-	licenses = bb.data.getVar('LICENSE', d, 1).replace('&', '|')
+	licenses = d.getVar('LICENSE', 1).replace('&', '|')
 	licenses = licenses.replace('(', '').replace(')', '')
 	clean_licenses = ""
 	for x in licenses.split():
@@ -20,20 +20,20 @@ python do_distribute_sources () {
 
 	for license in clean_licenses.split('|'):
 		for url in ud.values():
-			cmd = bb.data.getVar('SRC_DISTRIBUTECOMMAND', d, 1)
+			cmd = d.getVar('SRC_DISTRIBUTECOMMAND', 1)
 			if not cmd:
 				raise bb.build.FuncFailed("Unable to distribute sources, SRC_DISTRIBUTECOMMAND not defined")
 			url.setup_localpath(d)
-			bb.data.setVar('SRC', url.localpath, d)
+			d.setVar('SRC', url.localpath)
 			if url.type == 'file':
 				if url.basename == '*':
 					import os.path
 					dest_dir = os.path.basename(os.path.dirname(os.path.abspath(url.localpath)))
-					bb.data.setVar('DEST', "%s_%s/" % (bb.data.getVar('PF', d, 1), dest_dir), d)
+					bb.data.setVar('DEST', "%s_%s/" % (d.getVar('PF', 1), dest_dir), d)
 				else:
-					bb.data.setVar('DEST', "%s_%s" % (bb.data.getVar('PF', d, 1), url.basename), d)
+					bb.data.setVar('DEST', "%s_%s" % (d.getVar('PF', 1), url.basename), d)
 			else:
-				bb.data.setVar('DEST', '', d)
+				d.setVar('DEST', '')
 
 			bb.data.setVar('SRC_DISTRIBUTEDIR', "%s/%s" % (sources_dir, license), d)
 			bb.build.exec_func('SRC_DISTRIBUTECOMMAND', d)

+ 47 - 47
meta/classes/sstate.bbclass

@@ -20,7 +20,7 @@ SSTATEPOSTINSTFUNCS ?= ""
 
 python () {
     if bb.data.inherits_class('native', d):
-        bb.data.setVar('SSTATE_PKGARCH', bb.data.getVar('BUILD_ARCH', d), d)
+        bb.data.setVar('SSTATE_PKGARCH', d.getVar('BUILD_ARCH'), d)
     elif bb.data.inherits_class('cross', d):
         bb.data.setVar('SSTATE_PKGARCH', bb.data.expand("${BUILD_ARCH}_${TUNE_PKGARCH}", d), d)
         bb.data.setVar('SSTATE_MANMACH', bb.data.expand("${BUILD_ARCH}_${MACHINE}", d), d)
@@ -37,19 +37,19 @@ python () {
     # reused if we manipulate the paths
     if bb.data.inherits_class('native', d) or bb.data.inherits_class('cross', d) or bb.data.inherits_class('sdk', d) or bb.data.inherits_class('crosssdk', d):
         scan_cmd = "grep -Irl ${STAGING_DIR} ${SSTATE_BUILDDIR}"
-        bb.data.setVar('SSTATE_SCAN_CMD', scan_cmd, d)
+        d.setVar('SSTATE_SCAN_CMD', scan_cmd)
 
-    unique_tasks = set((bb.data.getVar('SSTATETASKS', d, True) or "").split())
+    unique_tasks = set((d.getVar('SSTATETASKS', True) or "").split())
     d.setVar('SSTATETASKS', " ".join(unique_tasks))
     namemap = []
     for task in unique_tasks:
-        namemap.append(bb.data.getVarFlag(task, 'sstate-name', d))
-        funcs = bb.data.getVarFlag(task, 'prefuncs', d) or ""
+        namemap.append(d.getVarFlag(task, 'sstate-name'))
+        funcs = d.getVarFlag(task, 'prefuncs') or ""
         funcs = "sstate_task_prefunc " + funcs
-        bb.data.setVarFlag(task, 'prefuncs', funcs, d)
-        funcs = bb.data.getVarFlag(task, 'postfuncs', d) or ""
+        d.setVarFlag(task, 'prefuncs', funcs)
+        funcs = d.getVarFlag(task, 'postfuncs') or ""
         funcs = funcs + " sstate_task_postfunc"
-        bb.data.setVarFlag(task, 'postfuncs', funcs, d)
+        d.setVarFlag(task, 'postfuncs', funcs)
     d.setVar('SSTATETASKNAMES', " ".join(namemap))
 }
 
@@ -65,18 +65,18 @@ def sstate_init(name, task, d):
 
 def sstate_state_fromvars(d, task = None):
     if task is None:
-        task = bb.data.getVar('BB_CURRENTTASK', d, True)
+        task = d.getVar('BB_CURRENTTASK', True)
         if not task:
             bb.fatal("sstate code running without task context?!")
         task = task.replace("_setscene", "")
 
-    name = bb.data.expand(bb.data.getVarFlag("do_" + task, 'sstate-name', d), d)
-    inputs = (bb.data.expand(bb.data.getVarFlag("do_" + task, 'sstate-inputdirs', d) or "", d)).split()
-    outputs = (bb.data.expand(bb.data.getVarFlag("do_" + task, 'sstate-outputdirs', d) or "", d)).split()
-    plaindirs = (bb.data.expand(bb.data.getVarFlag("do_" + task, 'sstate-plaindirs', d) or "", d)).split()
-    lockfiles = (bb.data.expand(bb.data.getVarFlag("do_" + task, 'sstate-lockfile', d) or "", d)).split()
-    lockfilesshared = (bb.data.expand(bb.data.getVarFlag("do_" + task, 'sstate-lockfile-shared', d) or "", d)).split()
-    interceptfuncs = (bb.data.expand(bb.data.getVarFlag("do_" + task, 'sstate-interceptfuncs', d) or "", d)).split()
+    name = bb.data.expand(d.getVarFlag("do_" + task, 'sstate-name'), d)
+    inputs = (bb.data.expand(d.getVarFlag("do_" + task, 'sstate-inputdirs') or "", d)).split()
+    outputs = (bb.data.expand(d.getVarFlag("do_" + task, 'sstate-outputdirs') or "", d)).split()
+    plaindirs = (bb.data.expand(d.getVarFlag("do_" + task, 'sstate-plaindirs') or "", d)).split()
+    lockfiles = (bb.data.expand(d.getVarFlag("do_" + task, 'sstate-lockfile') or "", d)).split()
+    lockfilesshared = (bb.data.expand(d.getVarFlag("do_" + task, 'sstate-lockfile-shared') or "", d)).split()
+    interceptfuncs = (bb.data.expand(d.getVarFlag("do_" + task, 'sstate-interceptfuncs') or "", d)).split()
     if not name or len(inputs) != len(outputs):
         bb.fatal("sstate variables not setup correctly?!")
 
@@ -139,7 +139,7 @@ def sstate_install(ss, d):
         f.write(di + "\n")
     f.close()
 
-    for postinst in (bb.data.getVar('SSTATEPOSTINSTFUNCS', d, True) or '').split():
+    for postinst in (d.getVar('SSTATEPOSTINSTFUNCS', True) or '').split():
         bb.build.exec_func(postinst, d)
 
     for lock in locks:
@@ -156,7 +156,7 @@ def sstate_installpkg(ss, d):
         oe.path.remove(dir)
 
     sstateinst = bb.data.expand("${WORKDIR}/sstate-install-%s/" % ss['name'], d)
-    sstatepkg = bb.data.getVar('SSTATE_PKG', d, True) + '_' + ss['name'] + ".tgz"
+    sstatepkg = d.getVar('SSTATE_PKG', True) + '_' + ss['name'] + ".tgz"
 
     if not os.path.exists(sstatepkg):
        pstaging_fetch(sstatepkg, d)
@@ -167,16 +167,16 @@ def sstate_installpkg(ss, d):
 
     sstate_clean(ss, d)
 
-    bb.data.setVar('SSTATE_INSTDIR', sstateinst, d)
-    bb.data.setVar('SSTATE_PKG', sstatepkg, d)
+    d.setVar('SSTATE_INSTDIR', sstateinst)
+    d.setVar('SSTATE_PKG', sstatepkg)
     bb.build.exec_func('sstate_unpack_package', d)
 
     # Fixup hardcoded paths
     fixmefn =  sstateinst + "fixmepath"
     if os.path.isfile(fixmefn):
-        staging = bb.data.getVar('STAGING_DIR', d, True)
-        staging_target = bb.data.getVar('STAGING_DIR_TARGET', d, True)
-        staging_host = bb.data.getVar('STAGING_DIR_HOST', d, True)
+        staging = d.getVar('STAGING_DIR', True)
+        staging_target = d.getVar('STAGING_DIR_TARGET', True)
+        staging_host = d.getVar('STAGING_DIR_HOST', True)
         fixmefd = open(fixmefn, "r")
         fixmefiles = fixmefd.readlines()
         fixmefd.close()
@@ -206,13 +206,13 @@ def sstate_installpkg(ss, d):
 def sstate_clean_cachefile(ss, d):
     import oe.path
 
-    sstatepkgdir = bb.data.getVar('SSTATE_DIR', d, True)
-    sstatepkgfile = sstatepkgdir + '/' + bb.data.getVar('SSTATE_PKGSPEC', d, True) + "*_" + ss['name'] + ".tgz*"
+    sstatepkgdir = d.getVar('SSTATE_DIR', True)
+    sstatepkgfile = sstatepkgdir + '/' + d.getVar('SSTATE_PKGSPEC', True) + "*_" + ss['name'] + ".tgz*"
     bb.note("Removing %s" % sstatepkgfile)
     oe.path.remove(sstatepkgfile)
 
 def sstate_clean_cachefiles(d):
-    for task in (bb.data.getVar('SSTATETASKS', d, True) or "").split():
+    for task in (d.getVar('SSTATETASKS', True) or "").split():
         ss = sstate_state_fromvars(d, task[3:])
         sstate_clean_cachefile(ss, d)
 
@@ -274,10 +274,10 @@ CLEANFUNCS += "sstate_cleanall"
 python sstate_cleanall() {
     import fnmatch
 
-    bb.note("Removing shared state for package %s" % bb.data.getVar('PN', d, True))
+    bb.note("Removing shared state for package %s" % d.getVar('PN', True))
 
-    manifest_dir = bb.data.getVar('SSTATE_MANIFESTS', d, True)
-    manifest_prefix = bb.data.getVar("SSTATE_MANFILEPREFIX", d, True)
+    manifest_dir = d.getVar('SSTATE_MANIFESTS', True)
+    manifest_prefix = d.getVar("SSTATE_MANFILEPREFIX", True)
     manifest_pattern = os.path.basename(manifest_prefix) + ".*"
 
     if not os.path.exists(manifest_dir):
@@ -298,7 +298,7 @@ python sstate_cleanall() {
 def sstate_hardcode_path(d):
 	# Need to remove hardcoded paths and fix these when we install the
 	# staging packages.
-	sstate_scan_cmd = bb.data.getVar('SSTATE_SCAN_CMD', d, True)
+	sstate_scan_cmd = d.getVar('SSTATE_SCAN_CMD', True)
 	p = os.popen("%s" % sstate_scan_cmd)
 	file_list = p.read()
 
@@ -306,10 +306,10 @@ def sstate_hardcode_path(d):
 		p.close()
 		return
 
-	staging = bb.data.getVar('STAGING_DIR', d, True)
-	staging_target = bb.data.getVar('STAGING_DIR_TARGET', d, True)
-	staging_host = bb.data.getVar('STAGING_DIR_HOST', d, True)
-	sstate_builddir = bb.data.getVar('SSTATE_BUILDDIR', d, True)
+	staging = d.getVar('STAGING_DIR', True)
+	staging_target = d.getVar('STAGING_DIR_TARGET', True)
+	staging_host = d.getVar('STAGING_DIR_HOST', True)
+	sstate_builddir = d.getVar('SSTATE_BUILDDIR', True)
 
 	for i in file_list.split('\n'):
 		if not i:
@@ -349,10 +349,10 @@ def sstate_package(ss, d):
         os.remove(path)
         os.symlink(base, path)
 
-    tmpdir = bb.data.getVar('TMPDIR', d, True)
+    tmpdir = d.getVar('TMPDIR', True)
 
     sstatebuild = bb.data.expand("${WORKDIR}/sstate-build-%s/" % ss['name'], d)
-    sstatepkg = bb.data.getVar('SSTATE_PKG', d, True) + '_'+ ss['name'] + ".tgz"
+    sstatepkg = d.getVar('SSTATE_PKG', True) + '_'+ ss['name'] + ".tgz"
     bb.mkdirhier(sstatebuild)
     bb.mkdirhier(os.path.dirname(sstatepkg))
     for state in ss['dirs']:
@@ -369,15 +369,15 @@ def sstate_package(ss, d):
         bb.debug(2, "Preparing tree %s for packaging at %s" % (state[1], sstatebuild + state[0]))
         oe.path.copytree(state[1], sstatebuild + state[0])
 
-    workdir = bb.data.getVar('WORKDIR', d, True)
+    workdir = d.getVar('WORKDIR', True)
     for plain in ss['plaindirs']:
         pdir = plain.replace(workdir, sstatebuild)
         bb.mkdirhier(plain)
         bb.mkdirhier(pdir)
         oe.path.copytree(plain, pdir)
 
-    bb.data.setVar('SSTATE_BUILDDIR', sstatebuild, d)
-    bb.data.setVar('SSTATE_PKG', sstatepkg, d)
+    d.setVar('SSTATE_BUILDDIR', sstatebuild)
+    d.setVar('SSTATE_PKG', sstatepkg)
     sstate_hardcode_path(d)
     bb.build.exec_func('sstate_create_package', d)
     
@@ -389,7 +389,7 @@ def pstaging_fetch(sstatepkg, d):
     import bb.fetch2
 
     # Only try and fetch if the user has configured a mirror
-    mirrors = bb.data.getVar('SSTATE_MIRRORS', d, True)
+    mirrors = d.getVar('SSTATE_MIRRORS', True)
     if not mirrors:
         return
 
@@ -402,9 +402,9 @@ def pstaging_fetch(sstatepkg, d):
 
     bb.mkdirhier(dldir)
 
-    bb.data.setVar('DL_DIR', dldir, localdata)
-    bb.data.setVar('PREMIRRORS', mirrors, localdata)
-    bb.data.setVar('SRC_URI', srcuri, localdata)
+    localdata.setVar('DL_DIR', dldir)
+    localdata.setVar('PREMIRRORS', mirrors)
+    localdata.setVar('SRC_URI', srcuri)
 
     # Try a fetch from the sstate mirror, if it fails just return and
     # we will build the package
@@ -493,15 +493,15 @@ def sstate_checkhashes(sq_fn, sq_task, sq_hash, sq_hashfn, d):
         else:
             bb.debug(2, "SState: Looked for but didn't find file %s" % sstatefile)
 
-    mirrors = bb.data.getVar("SSTATE_MIRRORS", d, True)
+    mirrors = d.getVar("SSTATE_MIRRORS", True)
     if mirrors:
         # Copy the data object and override DL_DIR and SRC_URI
         localdata = bb.data.createCopy(d)
         bb.data.update_data(localdata)
 
         dldir = bb.data.expand("${SSTATE_DIR}", localdata)
-        bb.data.setVar('DL_DIR', dldir, localdata)
-        bb.data.setVar('PREMIRRORS', mirrors, localdata)
+        localdata.setVar('DL_DIR', dldir)
+        localdata.setVar('PREMIRRORS', mirrors)
 
         bb.debug(2, "SState using premirror of: %s" % mirrors)
 
@@ -513,7 +513,7 @@ def sstate_checkhashes(sq_fn, sq_task, sq_hash, sq_hashfn, d):
             sstatefile = sstatefile.replace("${BB_TASKHASH}", sq_hash[task])
 
             srcuri = "file://" + os.path.basename(sstatefile)
-            bb.data.setVar('SRC_URI', srcuri, localdata)
+            localdata.setVar('SRC_URI', srcuri)
             bb.debug(2, "SState: Attempting to fetch %s" % srcuri)
 
             try:

+ 3 - 3
meta/classes/staging.bbclass

@@ -84,7 +84,7 @@ python do_populate_sysroot () {
     #
 
     bb.build.exec_func("sysroot_stage_all", d)
-    for f in (bb.data.getVar('SYSROOT_PREPROCESS_FUNCS', d, True) or '').split():
+    for f in (d.getVar('SYSROOT_PREPROCESS_FUNCS', True) or '').split():
         bb.build.exec_func(f, d)
 }
 
@@ -100,8 +100,8 @@ python do_populate_sysroot_setscene () {
 addtask do_populate_sysroot_setscene
 
 python () {
-    if bb.data.getVar('do_stage', d, True) is not None:
-        bb.fatal("Legacy staging found for %s as it has a do_stage function. This will need conversion to a do_install or often simply removal to work with Poky" % bb.data.getVar("FILE", d, True))
+    if d.getVar('do_stage', True) is not None:
+        bb.fatal("Legacy staging found for %s as it has a do_stage function. This will need conversion to a do_install or often simply removal to work with Poky" % d.getVar("FILE", True))
 }
 
 

+ 17 - 17
meta/classes/syslinux.bbclass

@@ -8,12 +8,12 @@ python build_syslinux_menu () {
 	import copy
 	import sys
 
-	workdir = bb.data.getVar('WORKDIR', d, 1)
+	workdir = d.getVar('WORKDIR', 1)
 	if not workdir:
 		bb.error("WORKDIR is not defined")
 		return
 		
-	labels = bb.data.getVar('LABELS', d, 1)
+	labels = d.getVar('LABELS', 1)
 	if not labels:
 		bb.debug(1, "LABELS not defined, nothing to do")
 		return
@@ -22,7 +22,7 @@ python build_syslinux_menu () {
 		bb.debug(1, "No labels, nothing to do")
 		return
 
-	cfile = bb.data.getVar('SYSLINUXMENU', d, 1)
+	cfile = d.getVar('SYSLINUXMENU', 1)
 	if not cfile:
 		raise bb.build.FuncFailed('Unable to read SYSLINUXMENU')
 
@@ -45,15 +45,15 @@ python build_syslinux_menu () {
 		from copy import deepcopy
 		localdata = deepcopy(d)
 
-		overrides = bb.data.getVar('OVERRIDES', localdata)
+		overrides = localdata.getVar('OVERRIDES')
 		if not overrides:
 			raise bb.build.FuncFailed('OVERRIDES not defined')
 		overrides = bb.data.expand(overrides, localdata)
 	
-		bb.data.setVar('OVERRIDES', label + ':' + overrides, localdata)
+		localdata.setVar('OVERRIDES', label + ':' + overrides)
 		bb.data.update_data(localdata)
 
-		usage = bb.data.getVar('USAGE', localdata, 1)
+		usage = localdata.getVar('USAGE', 1)
 		cfgfile.write('  \x0F\x30\x3E%16s\x0F\x30\x37: ' % (label))
 		cfgfile.write('%s\n' % (usage))
 
@@ -67,12 +67,12 @@ python build_syslinux_cfg () {
 	import copy
 	import sys
 
-	workdir = bb.data.getVar('WORKDIR', d, 1)
+	workdir = d.getVar('WORKDIR', 1)
 	if not workdir:
 		bb.error("WORKDIR not defined, unable to package")
 		return
 		
-	labels = bb.data.getVar('LABELS', d, 1)
+	labels = d.getVar('LABELS', 1)
 	if not labels:
 		bb.debug(1, "LABELS not defined, nothing to do")
 		return
@@ -81,7 +81,7 @@ python build_syslinux_cfg () {
 		bb.debug(1, "No labels, nothing to do")
 		return
 
-	cfile = bb.data.getVar('SYSLINUXCFG', d, 1)
+	cfile = d.getVar('SYSLINUXCFG', 1)
 	if not cfile:
 		raise bb.build.FuncFailed('Unable to read SYSLINUXCFG')
 
@@ -98,7 +98,7 @@ python build_syslinux_cfg () {
 
 	cfgfile.write('# Automatically created by OE\n')
 
-	opts = bb.data.getVar('SYSLINUX_OPTS', d, 1)
+	opts = d.getVar('SYSLINUX_OPTS', 1)
 
 	if opts:
 		for opt in opts.split(';'):
@@ -107,7 +107,7 @@ python build_syslinux_cfg () {
 	cfgfile.write('ALLOWOPTIONS 1\n');
 	cfgfile.write('DEFAULT %s\n' % (labels.split()[0]))
 
-	timeout = bb.data.getVar('SYSLINUX_TIMEOUT', d, 1)
+	timeout = d.getVar('SYSLINUX_TIMEOUT', 1)
 
 	if timeout:
 		cfgfile.write('TIMEOUT %s\n' % timeout)
@@ -116,29 +116,29 @@ python build_syslinux_cfg () {
 
 	cfgfile.write('PROMPT 1\n')
 
-	menu = bb.data.getVar('AUTO_SYSLINUXMENU', d, 1)
+	menu = d.getVar('AUTO_SYSLINUXMENU', 1)
 
 	# This is ugly.  My bad.
 
 	if menu:
 		bb.build.exec_func('build_syslinux_menu', d)
-		mfile = bb.data.getVar('SYSLINUXMENU', d, 1)
+		mfile = d.getVar('SYSLINUXMENU', 1)
 		cfgfile.write('DISPLAY %s\n' % (mfile.split('/')[-1]) )
 	
 	for label in labels.split():
 		localdata = bb.data.createCopy(d)
 
-		overrides = bb.data.getVar('OVERRIDES', localdata, True)
+		overrides = localdata.getVar('OVERRIDES', True)
 		if not overrides:
 			raise bb.build.FuncFailed('OVERRIDES not defined')
 	
-		bb.data.setVar('OVERRIDES', label + ':' + overrides, localdata)
+		localdata.setVar('OVERRIDES', label + ':' + overrides)
 		bb.data.update_data(localdata)
 	
 		cfgfile.write('LABEL %s\nKERNEL vmlinuz\n' % (label))
 
-		append = bb.data.getVar('APPEND', localdata, 1)
-		initrd = bb.data.getVar('INITRD', localdata, 1)
+		append = localdata.getVar('APPEND', 1)
+		initrd = localdata.getVar('INITRD', 1)
 
 		if append:
 			cfgfile.write('APPEND ')

+ 1 - 1
meta/classes/task.bbclass

@@ -17,7 +17,7 @@ PACKAGE_ARCH = "all"
 # to the list. Their dependencies (RRECOMMENDS) are handled as usual
 # by package_depchains in a following step.
 python () {
-    packages = bb.data.getVar('PACKAGES', d, 1).split()
+    packages = d.getVar('PACKAGES', 1).split()
     genpackages = []
     for pkg in packages:
         for postfix in ['-dbg', '-dev']:

+ 3 - 3
meta/classes/toolchain-scripts.bbclass

@@ -137,8 +137,8 @@ toolchain_create_sdk_version () {
 }
 
 python __anonymous () {
-    deps = bb.data.getVarFlag('do_configure', 'depends', d) or ""
-    for dep in (bb.data.getVar('TOOLCHAIN_NEED_CONFIGSITE_CACHE', d, True) or "").split():
+    deps = d.getVarFlag('do_configure', 'depends') or ""
+    for dep in (d.getVar('TOOLCHAIN_NEED_CONFIGSITE_CACHE', True) or "").split():
         deps += " %s:do_populate_sysroot" % dep
-    bb.data.setVarFlag('do_configure', 'depends', deps, d)
+    d.setVarFlag('do_configure', 'depends', deps)
 }

+ 19 - 19
meta/classes/update-alternatives.bbclass

@@ -78,38 +78,38 @@ fi
 }
 
 def update_alternatives_after_parse(d):
-    if bb.data.getVar('ALTERNATIVE_LINKS', d) != None:
-        doinstall = bb.data.getVar('do_install', d, 0)
-        doinstall += bb.data.getVar('update_alternatives_batch_doinstall', d, 0)
-        bb.data.setVar('do_install', doinstall, d)
+    if d.getVar('ALTERNATIVE_LINKS') != None:
+        doinstall = d.getVar('do_install', 0)
+        doinstall += d.getVar('update_alternatives_batch_doinstall', 0)
+        d.setVar('do_install', doinstall)
         return
 
-    if bb.data.getVar('ALTERNATIVE_NAME', d) == None:
-        raise bb.build.FuncFailed, "%s inherits update-alternatives but doesn't set ALTERNATIVE_NAME" % bb.data.getVar('FILE', d)
-    if bb.data.getVar('ALTERNATIVE_PATH', d) == None:
-        raise bb.build.FuncFailed, "%s inherits update-alternatives but doesn't set ALTERNATIVE_PATH" % bb.data.getVar('FILE', d)
+    if d.getVar('ALTERNATIVE_NAME') == None:
+        raise bb.build.FuncFailed, "%s inherits update-alternatives but doesn't set ALTERNATIVE_NAME" % d.getVar('FILE')
+    if d.getVar('ALTERNATIVE_PATH') == None:
+        raise bb.build.FuncFailed, "%s inherits update-alternatives but doesn't set ALTERNATIVE_PATH" % d.getVar('FILE')
 
 python __anonymous() {
     update_alternatives_after_parse(d)
 }
 
 python populate_packages_prepend () {
-	pkg = bb.data.getVar('PN', d, 1)
+	pkg = d.getVar('PN', 1)
 	bb.note('adding update-alternatives calls to postinst/postrm for %s' % pkg)
-	postinst = bb.data.getVar('pkg_postinst_%s' % pkg, d, 1) or bb.data.getVar('pkg_postinst', d, 1)
+	postinst = d.getVar('pkg_postinst_%s' % pkg, 1) or d.getVar('pkg_postinst', 1)
 	if not postinst:
 		postinst = '#!/bin/sh\n'
-	if bb.data.getVar('ALTERNATIVE_LINKS', d) != None:
-		postinst += bb.data.getVar('update_alternatives_batch_postinst', d, 1)
+	if d.getVar('ALTERNATIVE_LINKS') != None:
+		postinst += d.getVar('update_alternatives_batch_postinst', 1)
 	else:
-		postinst += bb.data.getVar('update_alternatives_postinst', d, 1)
-	bb.data.setVar('pkg_postinst_%s' % pkg, postinst, d)
-	postrm = bb.data.getVar('pkg_postrm_%s' % pkg, d, 1) or bb.data.getVar('pkg_postrm', d, 1)
+		postinst += d.getVar('update_alternatives_postinst', 1)
+	d.setVar('pkg_postinst_%s' % pkg, postinst)
+	postrm = d.getVar('pkg_postrm_%s' % pkg, 1) or d.getVar('pkg_postrm', 1)
 	if not postrm:
 		postrm = '#!/bin/sh\n'
-	if bb.data.getVar('ALTERNATIVE_LINKS', d) != None:
-		postrm += bb.data.getVar('update_alternatives_batch_postrm', d, 1)
+	if d.getVar('ALTERNATIVE_LINKS') != None:
+		postrm += d.getVar('update_alternatives_batch_postrm', 1)
 	else:
-		postrm += bb.data.getVar('update_alternatives_postrm', d, 1)
-	bb.data.setVar('pkg_postrm_%s' % pkg, postrm, d)
+		postrm += d.getVar('update_alternatives_postrm', 1)
+	d.setVar('pkg_postrm_%s' % pkg, postrm)
 }

+ 18 - 18
meta/classes/update-rc.d.bbclass

@@ -30,11 +30,11 @@ update-rc.d $D ${INITSCRIPT_NAME} remove
 
 
 def update_rc_after_parse(d):
-    if bb.data.getVar('INITSCRIPT_PACKAGES', d) == None:
-        if bb.data.getVar('INITSCRIPT_NAME', d) == None:
-            raise bb.build.FuncFailed, "%s inherits update-rc.d but doesn't set INITSCRIPT_NAME" % bb.data.getVar('FILE', d)
-        if bb.data.getVar('INITSCRIPT_PARAMS', d) == None:
-            raise bb.build.FuncFailed, "%s inherits update-rc.d but doesn't set INITSCRIPT_PARAMS" % bb.data.getVar('FILE', d)
+    if d.getVar('INITSCRIPT_PACKAGES') == None:
+        if d.getVar('INITSCRIPT_NAME') == None:
+            raise bb.build.FuncFailed, "%s inherits update-rc.d but doesn't set INITSCRIPT_NAME" % d.getVar('FILE')
+        if d.getVar('INITSCRIPT_PARAMS') == None:
+            raise bb.build.FuncFailed, "%s inherits update-rc.d but doesn't set INITSCRIPT_PARAMS" % d.getVar('FILE')
 
 python __anonymous() {
     update_rc_after_parse(d)
@@ -44,7 +44,7 @@ python populate_packages_prepend () {
 	def update_rcd_package(pkg):
 		bb.debug(1, 'adding update-rc.d calls to postinst/postrm for %s' % pkg)
 		localdata = bb.data.createCopy(d)
-		overrides = bb.data.getVar("OVERRIDES", localdata, 1)
+		overrides = localdata.getVar("OVERRIDES", 1)
 		bb.data.setVar("OVERRIDES", "%s:%s" % (pkg, overrides), localdata)
 		bb.data.update_data(localdata)
 
@@ -53,28 +53,28 @@ python populate_packages_prepend () {
 		execute on the target. Not doing so may cause update_rc.d postinst invoked
 		twice to cause unwanted warnings.
 		""" 
-		postinst = bb.data.getVar('pkg_postinst', localdata, 1)
+		postinst = localdata.getVar('pkg_postinst', 1)
 		if not postinst:
 			postinst = '#!/bin/sh\n'
-		postinst += bb.data.getVar('updatercd_postinst', localdata, 1)
-		bb.data.setVar('pkg_postinst_%s' % pkg, postinst, d)
+		postinst += localdata.getVar('updatercd_postinst', 1)
+		d.setVar('pkg_postinst_%s' % pkg, postinst)
 
-		prerm = bb.data.getVar('pkg_prerm', localdata, 1)
+		prerm = localdata.getVar('pkg_prerm', 1)
 		if not prerm:
 			prerm = '#!/bin/sh\n'
-		prerm += bb.data.getVar('updatercd_prerm', localdata, 1)
-		bb.data.setVar('pkg_prerm_%s' % pkg, prerm, d)
+		prerm += localdata.getVar('updatercd_prerm', 1)
+		d.setVar('pkg_prerm_%s' % pkg, prerm)
 
-	        postrm = bb.data.getVar('pkg_postrm', localdata, 1)
+	        postrm = localdata.getVar('pkg_postrm', 1)
 	        if not postrm:
 	                postrm = '#!/bin/sh\n'
-                postrm += bb.data.getVar('updatercd_postrm', localdata, 1)
-		bb.data.setVar('pkg_postrm_%s' % pkg, postrm, d)
+                postrm += localdata.getVar('updatercd_postrm', 1)
+		d.setVar('pkg_postrm_%s' % pkg, postrm)
 
-	pkgs = bb.data.getVar('INITSCRIPT_PACKAGES', d, 1)
+	pkgs = d.getVar('INITSCRIPT_PACKAGES', 1)
 	if pkgs == None:
-		pkgs = bb.data.getVar('UPDATERCPN', d, 1)
-		packages = (bb.data.getVar('PACKAGES', d, 1) or "").split()
+		pkgs = d.getVar('UPDATERCPN', 1)
+		packages = (d.getVar('PACKAGES', 1) or "").split()
 		if not pkgs in packages and packages != []:
 			pkgs = packages[0]
 	for pkg in pkgs.split():

+ 4 - 4
meta/classes/useradd.bbclass

@@ -107,11 +107,11 @@ def update_useradd_after_parse(d):
 	useradd_packages = d.getVar('USERADD_PACKAGES', True)
 
 	if not useradd_packages:
-		raise bb.build.FuncFailed, "%s inherits useradd but doesn't set USERADD_PACKAGES" % bb.data.getVar('FILE', d)
+		raise bb.build.FuncFailed, "%s inherits useradd but doesn't set USERADD_PACKAGES" % d.getVar('FILE')
 
 	for pkg in useradd_packages.split():
 		if not d.getVar('USERADD_PARAM_%s' % pkg, True) and not d.getVar('GROUPADD_PARAM_%s' % pkg, True):
-			raise bb.build.FuncFailed, "%s inherits useradd but doesn't set USERADD_PARAM or GROUPADD_PARAM for package %s" % (bb.data.getVar('FILE', d), pkg)
+			raise bb.build.FuncFailed, "%s inherits useradd but doesn't set USERADD_PARAM or GROUPADD_PARAM for package %s" % (d.getVar('FILE'), pkg)
 
 python __anonymous() {
 	update_useradd_after_parse(d)
@@ -147,12 +147,12 @@ fakeroot python populate_packages_prepend () {
 		if not preinst:
 			preinst = '#!/bin/sh\n'
 		preinst += d.getVar('useradd_preinst', True)
-		bb.data.setVar('pkg_preinst_%s' % pkg, preinst, d)
+		d.setVar('pkg_preinst_%s' % pkg, preinst)
 
 		# RDEPENDS setup
 		rdepends = d.getVar("RDEPENDS_%s" % pkg, True) or ""
 		rdepends += " base-passwd shadow"
-		bb.data.setVar("RDEPENDS_%s" % pkg, rdepends, d)
+		d.setVar("RDEPENDS_%s" % pkg, rdepends)
 
 	# Add the user/group preinstall scripts and RDEPENDS requirements
 	# to packages specified by USERADD_PACKAGES

+ 4 - 4
meta/classes/utility-tasks.bbclass

@@ -6,7 +6,7 @@ python do_listtasks() {
 	#bb.data.emit_env(sys.__stdout__, d)
 	# emit the metadata which isnt valid shell
 	for e in d.keys():
-		if bb.data.getVarFlag(e, 'task', d):
+		if d.getVarFlag(e, 'task'):
 			bb.plain("%s" % e)
 }
 
@@ -20,18 +20,18 @@ python do_clean() {
 	bb.note("Removing " + dir)
 	oe.path.remove(dir)
 
-	dir = "%s.*" % bb.data.expand(bb.data.getVar('STAMP', d), d)
+	dir = "%s.*" % bb.data.expand(d.getVar('STAMP'), d)
 	bb.note("Removing " + dir)
 	oe.path.remove(dir)
 
-	for f in (bb.data.getVar('CLEANFUNCS', d, 1) or '').split():
+	for f in (d.getVar('CLEANFUNCS', 1) or '').split():
 		bb.build.exec_func(f, d)
 }
 
 addtask checkuri
 do_checkuri[nostamp] = "1"
 python do_checkuri() {
-	src_uri = (bb.data.getVar('SRC_URI', d, True) or "").split()
+	src_uri = (d.getVar('SRC_URI', True) or "").split()
 	if len(src_uri) == 0:
 		return
 

+ 2 - 2
meta/classes/utils.bbclass

@@ -331,12 +331,12 @@ def explode_deps(s):
 
 def base_set_filespath(path, d):
 	filespath = []
-	extrapaths = (bb.data.getVar("FILESEXTRAPATHS", d, True) or "")
+	extrapaths = (d.getVar("FILESEXTRAPATHS", True) or "")
 	# Don't prepend empty strings to the path list
 	if extrapaths != "":
 		path = extrapaths.split(":") + path
 	# The ":" ensures we have an 'empty' override
-	overrides = (bb.data.getVar("OVERRIDES", d, 1) or "") + ":"
+	overrides = (d.getVar("OVERRIDES", 1) or "") + ":"
 	for p in path:
 		if p != "": 
 			for o in overrides.split(":"):

+ 18 - 18
meta/conf/bitbake.conf

@@ -99,7 +99,7 @@ ABIEXTENSION ??= ""
 TARGET_ARCH = "${TUNE_ARCH}"
 TARGET_OS = "linux${LIBCEXTENSION}${ABIEXTENSION}"
 TARGET_VENDOR = "-oe"
-TARGET_SYS = "${TARGET_ARCH}${TARGET_VENDOR}${@['-' + bb.data.getVar('TARGET_OS', d, 1), ''][bb.data.getVar('TARGET_OS', d, 1) == ('' or 'custom')]}"
+TARGET_SYS = "${TARGET_ARCH}${TARGET_VENDOR}${@['-' + d.getVar('TARGET_OS', 1), ''][d.getVar('TARGET_OS', 1) == ('' or 'custom')]}"
 TARGET_PREFIX = "${TARGET_SYS}-"
 TARGET_CC_ARCH = "${TUNE_CCARGS}"
 TARGET_LD_ARCH = "${TUNE_LDARGS}"
@@ -108,7 +108,7 @@ TARGET_AS_ARCH = "${TUNE_ASARGS}"
 SDK_ARCH = "${BUILD_ARCH}"
 SDK_OS = "${BUILD_OS}"
 SDK_VENDOR = "-oesdk"
-SDK_SYS = "${SDK_ARCH}${SDK_VENDOR}${@['-' + bb.data.getVar('SDK_OS', d, 1), ''][bb.data.getVar('SDK_OS', d, 1) == ('' or 'custom')]}"
+SDK_SYS = "${SDK_ARCH}${SDK_VENDOR}${@['-' + d.getVar('SDK_OS', 1), ''][d.getVar('SDK_OS', 1) == ('' or 'custom')]}"
 SDK_PREFIX = "${SDK_SYS}-"
 SDK_CC_ARCH = "${BUILD_CC_ARCH}"
 SDK_PACKAGE_ARCHS = "all any noarch ${SDK_ARCH}-nativesdk"
@@ -116,7 +116,7 @@ SDK_LD_ARCH = "${BUILD_LD_ARCH}"
 SDK_AS_ARCH = "${BUILD_AS_ARCH}"
 
 PACKAGE_ARCH = "${TUNE_PKGARCH}"
-MACHINE_ARCH = "${@[bb.data.getVar('TUNE_PKGARCH', d, 1), bb.data.getVar('MACHINE', d, 1)][bool(bb.data.getVar('MACHINE', d, 1))].replace('-', '_')}"
+MACHINE_ARCH = "${@[d.getVar('TUNE_PKGARCH', 1), d.getVar('MACHINE', 1)][bool(d.getVar('MACHINE', 1))].replace('-', '_')}"
 PACKAGE_EXTRA_ARCHS ??= "${PACKAGE_EXTRA_ARCHS_tune-${DEFAULTTUNE}}"
 PACKAGE_ARCHS = "all any noarch ${PACKAGE_EXTRA_ARCHS} ${MACHINE_ARCH}"
 # MACHINE_ARCH shouldn't be included here as a variable dependency
@@ -167,33 +167,33 @@ ASSUME_PROVIDED = "\
 # Package default variables.
 ##################################################################
 
-PN = "${@bb.parse.BBHandler.vars_from_file(bb.data.getVar('FILE',d),d)[0] or 'defaultpkgname'}"
-PV = "${@bb.parse.BBHandler.vars_from_file(bb.data.getVar('FILE',d),d)[1] or '1.0'}"
-PR = "${@bb.parse.BBHandler.vars_from_file(bb.data.getVar('FILE',d),d)[2] or 'r0'}"
+PN = "${@bb.parse.BBHandler.vars_from_file(d.getVar('FILE'),d)[0] or 'defaultpkgname'}"
+PV = "${@bb.parse.BBHandler.vars_from_file(d.getVar('FILE'),d)[1] or '1.0'}"
+PR = "${@bb.parse.BBHandler.vars_from_file(d.getVar('FILE'),d)[2] or 'r0'}"
 PF = "${PN}-${EXTENDPE}${PV}-${PR}"
-EXTENDPE = "${@['','${PE\x7d_'][bb.data.getVar('PE',d,1) > 0]}"
+EXTENDPE = "${@['','${PE\x7d_'][d.getVar('PE',1) > 0]}"
 P = "${PN}-${PV}"
 
-EXTENDPRAUTO = "${@['.${PRAUTO\x7d',''][bb.data.getVar('PRAUTO',d,1) is None]}"
+EXTENDPRAUTO = "${@['.${PRAUTO\x7d',''][d.getVar('PRAUTO',1) is None]}"
 PRAUTOINX = "${PF}"
 
 PKGV ?= "${PV}"
 PKGR ?= "${PR}${EXTENDPRAUTO}"
-PKGE ?= "${@['','${PE\x7d'][bb.data.getVar('PE',d,1) > 0]}"
-EXTENDPKGEVER = "${@['','${PKGE\x7d:'][bb.data.getVar('PKGE',d,1).strip() != '']}"
+PKGE ?= "${@['','${PE\x7d'][d.getVar('PE',1) > 0]}"
+EXTENDPKGEVER = "${@['','${PKGE\x7d:'][d.getVar('PKGE',1).strip() != '']}"
 EXTENDPKGV ?= "${EXTENDPKGEVER}${PKGV}-${PKGR}"
 
 # Base package name
 # Automatically derives "foo" from "foo-native", "foo-cross" or "foo-initial"
 # otherwise it is the same as PN and P
 SPECIAL_PKGSUFFIX = "-native -cross -initial -intermediate -nativesdk -crosssdk -cross-canadian"
-BPN = "${@base_prune_suffix(bb.data.getVar('PN', d, True), bb.data.getVar('SPECIAL_PKGSUFFIX', d, True).split(), d)}"
+BPN = "${@base_prune_suffix(d.getVar('PN', True), d.getVar('SPECIAL_PKGSUFFIX', True).split(), d)}"
 BP = "${BPN}-${PV}"
 
 #
 # network based PR service
 #
-USE_PR_SERV = "${@[1,0][(bb.data.getVar('PRSERV_HOST',d,1) is None) or (bb.data.getVar('PRSERV_PORT',d,1) is None)]}"
+USE_PR_SERV = "${@[1,0][(d.getVar('PRSERV_HOST',1) is None) or (d.getVar('PRSERV_PORT',1) is None)]}"
 
 # Package info.
 
@@ -288,7 +288,7 @@ DOTDEBUG-dbg = "${bindir}/.debug ${sbindir}/.debug ${libexecdir}/.debug ${libdir
 
 DEBUGFILEDIRECTORY-dbg = "/usr/lib/debug /usr/src/debug"
 
-FILES_${PN}-dbg = "${@bb.data.getVar(['DOTDEBUG-dbg', 'DEBUGFILEDIRECTORY-dbg'][bb.data.getVar('PACKAGE_DEBUG_SPLIT_STYLE', d, 1) == 'debug-file-directory'], d, 1)}"
+FILES_${PN}-dbg = "${@bb.data.getVar(['DOTDEBUG-dbg', 'DEBUGFILEDIRECTORY-dbg'][d.getVar('PACKAGE_DEBUG_SPLIT_STYLE', 1) == 'debug-file-directory'], d, 1)}"
 
 SECTION_${PN}-dbg = "devel"
 ALLOW_EMPTY_${PN}-dbg = "1"
@@ -298,17 +298,17 @@ FILES_${PN}-locale = "${datadir}/locale"
 
 # File manifest
 
-FILE_DIRNAME = "${@os.path.dirname(bb.data.getVar('FILE', d))}"
+FILE_DIRNAME = "${@os.path.dirname(d.getVar('FILE'))}"
 # FILESPATH is set in base.bbclass
 #FILESPATH = "${FILE_DIRNAME}/${PF}:${FILE_DIRNAME}/${P}:${FILE_DIRNAME}/${PN}:${FILE_DIRNAME}/${BP}:${FILE_DIRNAME}/${BPN}:${FILE_DIRNAME}/files:${FILE_DIRNAME}"
-FILESDIR = "${@bb.which(bb.data.getVar('FILESPATH', d, 1), '.')}"
+FILESDIR = "${@bb.which(d.getVar('FILESPATH', 1), '.')}"
 
 ##################################################################
 # General work and output directories for the build system.
 ##################################################################
 
 TMPDIR ?= "${TOPDIR}/tmp"
-CACHE = "${TMPDIR}/cache${@['', '/' + str(bb.data.getVar('MACHINE', d, 1))][bool(bb.data.getVar('MACHINE', d, 1))]}${@['', '/' + str(bb.data.getVar('SDKMACHINE', d, 1))][bool(bb.data.getVar('SDKMACHINE', d, 1))]}"
+CACHE = "${TMPDIR}/cache${@['', '/' + str(d.getVar('MACHINE', 1))][bool(d.getVar('MACHINE', 1))]}${@['', '/' + str(d.getVar('SDKMACHINE', 1))][bool(d.getVar('SDKMACHINE', 1))]}"
 # The persistent cache should be shared by all builds
 PERSISTENT_DIR = "${TMPDIR}/cache"
 LOG_DIR = "${TMPDIR}/log"
@@ -403,7 +403,7 @@ export PATH
 # Build utility info.
 ##################################################################
 
-CCACHE = "${@bb.which(bb.data.getVar('PATH', d, 1), 'ccache') and 'ccache '}"
+CCACHE = "${@bb.which(d.getVar('PATH', 1), 'ccache') and 'ccache '}"
 TOOLCHAIN_OPTIONS = " --sysroot=${STAGING_DIR_TARGET}"
 
 export CCACHE_DIR = "${TMPDIR}/ccache/${MULTIMACH_HOST_SYS}/${PN}"
@@ -505,7 +505,7 @@ DEBUG_FLAGS ?= "-g -feliminate-unused-debug-types"
 # Disabled until the option works properly -feliminate-dwarf2-dups
 FULL_OPTIMIZATION = "-O2 -pipe ${DEBUG_FLAGS}"
 DEBUG_OPTIMIZATION = "-O -fno-omit-frame-pointer ${DEBUG_FLAGS} -pipe"
-SELECTED_OPTIMIZATION = "${@bb.data.getVar(['FULL_OPTIMIZATION', 'DEBUG_OPTIMIZATION'][bb.data.getVar('DEBUG_BUILD', d, 1) == '1'], d, 1)}"
+SELECTED_OPTIMIZATION = "${@bb.data.getVar(['FULL_OPTIMIZATION', 'DEBUG_OPTIMIZATION'][d.getVar('DEBUG_BUILD', 1) == '1'], d, 1)}"
 BUILD_OPTIMIZATION = "-O2 -pipe"
 
 ##################################################################

+ 1 - 1
meta/conf/distro/defaultsetup.conf

@@ -13,7 +13,7 @@ require conf/distro/include/tclibc-${TCLIBC}.inc
 TCLIBCAPPEND ?= "-${TCLIBC}"
 TMPDIR .= "${TCLIBCAPPEND}"
 
-CACHE = "${TMPDIR}/cache/${TCMODE}-${TCLIBC}${@['', '/' + str(bb.data.getVar('MACHINE', d, 1))][bool(bb.data.getVar('MACHINE', d, 1))]}${@['', '/' + str(bb.data.getVar('SDKMACHINE', d, 1))][bool(bb.data.getVar('SDKMACHINE', d, 1))]}"
+CACHE = "${TMPDIR}/cache/${TCMODE}-${TCLIBC}${@['', '/' + str(d.getVar('MACHINE', 1))][bool(d.getVar('MACHINE', 1))]}${@['', '/' + str(d.getVar('SDKMACHINE', 1))][bool(d.getVar('SDKMACHINE', 1))]}"
 
 USER_CLASSES ?= ""
 PACKAGE_CLASSES ?= "package_ipk"

+ 1 - 1
meta/conf/machine/include/arm/feature-arm-thumb.inc

@@ -5,7 +5,7 @@
 # but requires more instructions (140% for 70% smaller code) so may be
 # slower.
 TUNEVALID[thumb] = "Use thumb instructions instead of ARM"
-ARM_THUMB_M_OPT = "${@['-mno-thumb', '-mthumb'][bb.data.getVar('ARM_INSTRUCTION_SET', d, 1) == 'thumb']}"
+ARM_THUMB_M_OPT = "${@['-mno-thumb', '-mthumb'][d.getVar('ARM_INSTRUCTION_SET', 1) == 'thumb']}"
 TUNE_CCARGS += "${@bb.utils.contains("TUNE_FEATURES", "thumb", "${ARM_THUMB_M_OPT}", "", d)}"
 OVERRIDES .= "${@bb.utils.contains("TUNE_FEATURES", "thumb", ":thumb", "", d)}"
 

+ 4 - 4
meta/conf/machine/include/tune-thumb.inc

@@ -16,15 +16,15 @@ THUMB_INTERWORK ?= "yes"
 #    arm system and vice versa.  It is strongly recommended that DISTROs not
 #    turn this off - the actual cost is very small.
 
-OVERRIDE_THUMB = "${@['', ':thumb'][bb.data.getVar('ARM_INSTRUCTION_SET', d, 1) == 'thumb']}"
-OVERRIDE_INTERWORK = "${@['', ':thumb-interwork'][bb.data.getVar('THUMB_INTERWORK', d, 1) == 'yes']}"
+OVERRIDE_THUMB = "${@['', ':thumb'][d.getVar('ARM_INSTRUCTION_SET', 1) == 'thumb']}"
+OVERRIDE_INTERWORK = "${@['', ':thumb-interwork'][d.getVar('THUMB_INTERWORK', 1) == 'yes']}"
 OVERRIDES .= "${OVERRIDE_THUMB}${OVERRIDE_INTERWORK}"
 
 #    Compiler and linker options for application code and kernel code.  These
 #    options ensure that the compiler has the correct settings for the selected
 #    instruction set and interworking.
-ARM_INTERWORK_M_OPT = "${@['-mno-thumb-interwork', '-mthumb-interwork'][bb.data.getVar('THUMB_INTERWORK', d, 1) == 'yes']}"
-ARM_THUMB_M_OPT = "${@['-mno-thumb', '-mthumb'][bb.data.getVar('ARM_INSTRUCTION_SET', d, 1) == 'thumb']}"
+ARM_INTERWORK_M_OPT = "${@['-mno-thumb-interwork', '-mthumb-interwork'][d.getVar('THUMB_INTERWORK', 1) == 'yes']}"
+ARM_THUMB_M_OPT = "${@['-mno-thumb', '-mthumb'][d.getVar('ARM_INSTRUCTION_SET', 1) == 'thumb']}"
 
 #
 TUNE_CCARGS +=  "${ARM_INTERWORK_M_OPT} ${ARM_THUMB_M_OPT}"

+ 11 - 11
meta/lib/oe/distro_check.py

@@ -276,32 +276,32 @@ def compare_in_distro_packages_list(distro_check_dir, d):
     localdata = bb.data.createCopy(d)
     pkglst_dir = os.path.join(distro_check_dir, "package_lists")
     matching_distros = []
-    pn = bb.data.getVar('PN', d, True)
-    recipe_name = bb.data.getVar('PN', d, True)
+    pn = d.getVar('PN', True)
+    recipe_name = d.getVar('PN', True)
     bb.note("Checking: %s" % pn)
 
     trim_dict = dict({"-native":"-native", "-cross":"-cross", "-initial":"-initial"})
 
     if pn.find("-native") != -1:
         pnstripped = pn.split("-native")
-        bb.data.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + bb.data.getVar('OVERRIDES', d, True), localdata)
+        bb.data.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + d.getVar('OVERRIDES', True), localdata)
         bb.data.update_data(localdata)
         recipe_name = pnstripped[0]
 
     if pn.find("-cross") != -1:
         pnstripped = pn.split("-cross")
-        bb.data.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + bb.data.getVar('OVERRIDES', d, True), localdata)
+        bb.data.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + d.getVar('OVERRIDES', True), localdata)
         bb.data.update_data(localdata)
         recipe_name = pnstripped[0]
 
     if pn.find("-initial") != -1:
         pnstripped = pn.split("-initial")
-        bb.data.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + bb.data.getVar('OVERRIDES', d, True), localdata)
+        bb.data.setVar('OVERRIDES', "pn-" + pnstripped[0] + ":" + d.getVar('OVERRIDES', True), localdata)
         bb.data.update_data(localdata)
         recipe_name = pnstripped[0]
 
     bb.note("Recipe: %s" % recipe_name)
-    tmp = bb.data.getVar('DISTRO_PN_ALIAS', localdata, True)
+    tmp = localdata.getVar('DISTRO_PN_ALIAS', True)
 
     distro_exceptions = dict({"OE-Core":'OE-Core', "OpenedHand":'OpenedHand', "Intel":'Intel', "Upstream":'Upstream', "Windriver":'Windriver', "OSPDT":'OSPDT Approved', "Poky":'poky'})
 
@@ -343,23 +343,23 @@ def compare_in_distro_packages_list(distro_check_dir, d):
     return matching_distros
 
 def create_log_file(d, logname):
-    logpath = bb.data.getVar('LOG_DIR', d, True)
+    logpath = d.getVar('LOG_DIR', True)
     bb.utils.mkdirhier(logpath)
     logfn, logsuffix = os.path.splitext(logname)
-    logfile = os.path.join(logpath, "%s.%s%s" % (logfn, bb.data.getVar('DATETIME', d, True), logsuffix))
+    logfile = os.path.join(logpath, "%s.%s%s" % (logfn, d.getVar('DATETIME', True), logsuffix))
     if not os.path.exists(logfile):
             slogfile = os.path.join(logpath, logname)
             if os.path.exists(slogfile):
                     os.remove(slogfile)
             os.system("touch %s" % logfile)
             os.symlink(logfile, slogfile)
-            bb.data.setVar('LOG_FILE', logfile, d)
+            d.setVar('LOG_FILE', logfile)
     return logfile
 
 
 def save_distro_check_result(result, datetime, result_file, d):
-    pn = bb.data.getVar('PN', d, True)
-    logdir = bb.data.getVar('LOG_DIR', d, True)
+    pn = d.getVar('PN', True)
+    logdir = d.getVar('LOG_DIR', True)
     if not logdir:
         bb.error("LOG_DIR variable is not defined, can't write the distro_check results")
         return

+ 2 - 2
meta/lib/oe/patch.py

@@ -179,7 +179,7 @@ class GitApplyTree(PatchTree):
 
 class QuiltTree(PatchSet):
     def _runcmd(self, args, run = True):
-        quiltrc = bb.data.getVar('QUILTRCFILE', self.d, 1)
+        quiltrc = self.d.getVar('QUILTRCFILE', 1)
         if not run:
             return ["quilt"] + ["--quiltrc"] + [quiltrc] + args
         runcmd(["quilt"] + ["--quiltrc"] + [quiltrc] + args, self.dir)
@@ -357,7 +357,7 @@ class UserResolver(Resolver):
             # Patch application failed
             patchcmd = self.patchset.Push(True, False, False)
 
-            t = bb.data.getVar('T', self.patchset.d, 1)
+            t = self.patchset.d.getVar('T', 1)
             if not t:
                 bb.msg.fatal("Build", "T not set")
             bb.utils.mkdirhier(t)

+ 4 - 4
meta/lib/oe/utils.py

@@ -16,19 +16,19 @@ def ifelse(condition, iftrue = True, iffalse = False):
         return iffalse
 
 def conditional(variable, checkvalue, truevalue, falsevalue, d):
-    if bb.data.getVar(variable,d,1) == checkvalue:
+    if d.getVar(variable,1) == checkvalue:
         return truevalue
     else:
         return falsevalue
 
 def less_or_equal(variable, checkvalue, truevalue, falsevalue, d):
-    if float(bb.data.getVar(variable,d,1)) <= float(checkvalue):
+    if float(d.getVar(variable,1)) <= float(checkvalue):
         return truevalue
     else:
         return falsevalue
 
 def version_less_or_equal(variable, checkvalue, truevalue, falsevalue, d):
-    result = bb.vercmp(bb.data.getVar(variable,d,True), checkvalue)
+    result = bb.vercmp(d.getVar(variable,True), checkvalue)
     if result <= 0:
         return truevalue
     else:
@@ -48,7 +48,7 @@ def contains(variable, checkvalues, truevalue, falsevalue, d):
     return falsevalue
 
 def both_contain(variable1, variable2, checkvalue, d):
-    if bb.data.getVar(variable1,d,1).find(checkvalue) != -1 and bb.data.getVar(variable2,d,1).find(checkvalue) != -1:
+    if d.getVar(variable1,1).find(checkvalue) != -1 and d.getVar(variable2,1).find(checkvalue) != -1:
         return checkvalue
     else:
         return ""

+ 1 - 1
meta/recipes-bsp/grub/grub_0.97.bb

@@ -23,7 +23,7 @@ inherit autotools
 
 python __anonymous () {
     import re
-    host = bb.data.getVar('HOST_SYS', d, 1)
+    host = d.getVar('HOST_SYS', 1)
     if not re.match('i.86.*-linux', host):
         raise bb.parse.SkipPackage("incompatible with host %s" % host)
 }

+ 1 - 1
meta/recipes-bsp/grub/grub_1.99.bb

@@ -33,7 +33,7 @@ do_configure() {
 
 python __anonymous () {
     import re
-    host = bb.data.getVar('HOST_SYS', d, 1)
+    host = d.getVar('HOST_SYS', 1)
     if not re.match('x86.64.*-linux', host) and not re.match('i.86.*-linux', host):
         raise bb.parse.SkipPackage("incompatible with host %s" % host)
 }

+ 1 - 1
meta/recipes-bsp/uboot/u-boot_2011.03.bb

@@ -10,7 +10,7 @@ LICENSE = "GPLv2+"
 LIC_FILES_CHKSUM = "file://COPYING;md5=1707d6db1d42237583f50183a5651ecb \
                     file://README;beginline=1;endline=22;md5=3a00ef51d3fc96e9d6c1bc4708ccd3b5"
 
-FILESDIR = "${@os.path.dirname(bb.data.getVar('FILE',d,1))}/u-boot-git/${MACHINE}"
+FILESDIR = "${@os.path.dirname(d.getVar('FILE',1))}/u-boot-git/${MACHINE}"
 
 # This revision corresponds to the tag "v2011.03"
 # We use the revision in order to avoid having to fetch it from the repo during parse

+ 1 - 1
meta/recipes-bsp/uboot/u-boot_2011.06.bb

@@ -10,7 +10,7 @@ LICENSE = "GPLv2+"
 LIC_FILES_CHKSUM = "file://COPYING;md5=1707d6db1d42237583f50183a5651ecb \
                     file://README;beginline=1;endline=22;md5=5ba4218ac89af7846802d0348df3fb90"
 
-FILESDIR = "${@os.path.dirname(bb.data.getVar('FILE',d,1))}/u-boot-git/${MACHINE}"
+FILESDIR = "${@os.path.dirname(d.getVar('FILE',1))}/u-boot-git/${MACHINE}"
 
 # This revision corresponds to the tag "v2011.06"
 # We use the revision in order to avoid having to fetch it from the repo during parse

+ 1 - 1
meta/recipes-bsp/x-load/x-load_git.bb

@@ -1,6 +1,6 @@
 require x-load.inc
 
-FILESDIR = "${@os.path.dirname(bb.data.getVar('FILE',d,1))}/x-load-git/${MACHINE}"
+FILESDIR = "${@os.path.dirname(d.getVar('FILE',1))}/x-load-git/${MACHINE}"
 
 LICENSE = "GPLv2+"
 LIC_FILES_CHKSUM = "file://README;beginline=1;endline=25;md5=ef08d08cb99057bbb5b9d6d0c5a4396f"

+ 1 - 1
meta/recipes-connectivity/connman/connman.inc

@@ -64,5 +64,5 @@ python populate_packages_prepend() {
 		plugintype = package.split( '-' )[-1]
 		if plugintype in depmap:
 			bb.note( "Adding rdependency on %s to package %s" % ( depmap[plugintype], package ) )
-			bb.data.setVar("RDEPENDS_%s" % package, depmap[plugintype], d)
+			d.setVar("RDEPENDS_%s" % package, depmap[plugintype])
 }

+ 3 - 3
meta/recipes-core/busybox/busybox.inc

@@ -49,8 +49,8 @@ def busybox_cfg(feature, features, tokens, cnf, rem):
 # Map distro and machine features to config settings
 def features_to_busybox_settings(d):
 	cnf, rem = ([], [])
-	distro_features = bb.data.getVar('DISTRO_FEATURES', d, True).split()
-	machine_features = bb.data.getVar('MACHINE_FEATURES', d, True).split()
+	distro_features = d.getVar('DISTRO_FEATURES', True).split()
+	machine_features = d.getVar('MACHINE_FEATURES', True).split()
 	busybox_cfg('ipv6', distro_features, 'CONFIG_FEATURE_IPV6', cnf, rem)
 	busybox_cfg('largefile', distro_features, 'CONFIG_LFS', cnf, rem)
 	busybox_cfg('largefile', distro_features, 'CONFIG_FDISK_SUPPORT_LARGE_DISKS', cnf, rem)
@@ -79,7 +79,7 @@ DO_IPv6 := ${@base_contains('DISTRO_FEATURES', 'ipv6', 1, 0, d)}
 
 python () {
   if "${OE_DEL}":
-    bb.data.setVar('configmangle_append', "${OE_DEL}" + "\n", d)
+    d.setVar('configmangle_append', "${OE_DEL}" + "\n")
   if "${OE_FEATURES}":
     bb.data.setVar('configmangle_append',
                    "/^### DISTRO FEATURES$/a\\\n%s\n\n" %

+ 3 - 3
meta/recipes-core/eglibc/eglibc-options.inc

@@ -83,7 +83,7 @@ def distro_features_check_deps(distro_features):
 # Map distro features to eglibc options settings
 def features_to_eglibc_settings(d):
         cnf = ([])
-        distro_features = (bb.data.getVar('DISTRO_FEATURES', d, True) or '').split()
+        distro_features = (d.getVar('DISTRO_FEATURES', True) or '').split()
 
         distro_features_check_deps(distro_features)
 
@@ -128,8 +128,8 @@ def features_to_eglibc_settings(d):
 
         # try to fix disable charsets/locales/locale-code compile fail
         if 'libc-charsets' in distro_features and 'libc-locales' in distro_features and 'libc-locale-code' in distro_features:
-            bb.data.setVar('PACKAGE_NO_GCONV', '0', d)
+            d.setVar('PACKAGE_NO_GCONV', '0')
         else:
-            bb.data.setVar('PACKAGE_NO_GCONV', '1', d)
+            d.setVar('PACKAGE_NO_GCONV', '1')
 
         return "\n".join(cnf)

+ 2 - 2
meta/recipes-core/eglibc/eglibc-package.inc

@@ -8,10 +8,10 @@
 
 python __anonymous () {
     import bb, re
-    uc_os = (re.match('.*uclibc*', bb.data.getVar('TARGET_OS', d, 1)) != None)
+    uc_os = (re.match('.*uclibc*', d.getVar('TARGET_OS', 1)) != None)
     if uc_os:
         raise bb.parse.SkipPackage("incompatible with target %s" %
-                                   bb.data.getVar('TARGET_OS', d, 1))
+                                   d.getVar('TARGET_OS', 1))
 }
 
 # Set this to zero if you don't want ldconfig in the output package

+ 2 - 2
meta/recipes-core/eglibc/eglibc_2.13.bb

@@ -53,10 +53,10 @@ FILESPATH = "${@base_set_filespath([ '${FILE_DIRNAME}/eglibc-${PV}', '${FILE_DIR
 
 python __anonymous () {
     import bb, re
-    uc_os = (re.match('.*uclibc$', bb.data.getVar('TARGET_OS', d, 1)) != None)
+    uc_os = (re.match('.*uclibc$', d.getVar('TARGET_OS', 1)) != None)
     if uc_os:
         raise bb.parse.SkipPackage("incompatible with target %s" %
-                                   bb.data.getVar('TARGET_OS', d, 1))
+                                   d.getVar('TARGET_OS', 1))
 }
 
 export libc_cv_slibdir = "${base_libdir}"

+ 2 - 2
meta/recipes-core/eglibc/eglibc_2.14.bb

@@ -54,10 +54,10 @@ FILESPATH = "${@base_set_filespath([ '${FILE_DIRNAME}/eglibc-${PV}', '${FILE_DIR
 
 python __anonymous () {
     import bb, re
-    uc_os = (re.match('.*uclibc$', bb.data.getVar('TARGET_OS', d, 1)) != None)
+    uc_os = (re.match('.*uclibc$', d.getVar('TARGET_OS', 1)) != None)
     if uc_os:
         raise bb.parse.SkipPackage("incompatible with target %s" %
-                                   bb.data.getVar('TARGET_OS', d, 1))
+                                   d.getVar('TARGET_OS', 1))
 }
 
 export libc_cv_slibdir = "${base_libdir}"

+ 1 - 1
meta/recipes-core/glib-2.0/glib-2.0_2.30.0.bb

@@ -7,7 +7,7 @@ DEPENDS += "libffi python-argparse-native"
 DEPENDS_virtclass-native += "libffi-native python-argparse-native"
 DEPENDS_virtclass-nativesdk += "libffi-nativesdk python-argparse-native zlib-nativesdk"
 
-SHRT_VER = "${@bb.data.getVar('PV',d,1).split('.')[0]}.${@bb.data.getVar('PV',d,1).split('.')[1]}"
+SHRT_VER = "${@d.getVar('PV',1).split('.')[0]}.${@d.getVar('PV',1).split('.')[1]}"
 
 QSORT_PATCH = "file://remove.test.for.qsort_r.patch"
 QSORT_PATCH_virtclass-native = ""

+ 2 - 2
meta/recipes-core/libxml/libxml2.inc

@@ -33,8 +33,8 @@ export LDFLAGS += "-ldl"
 
 python populate_packages_prepend () {
 	# autonamer would call this libxml2-2, but we don't want that
-	if bb.data.getVar('DEBIAN_NAMES', d, 1):
-		bb.data.setVar('PKG_libxml2', '${MLPREFIX}libxml2', d)
+	if d.getVar('DEBIAN_NAMES', 1):
+		d.setVar('PKG_libxml2', '${MLPREFIX}libxml2')
 }
 
 PACKAGES += "${PN}-utils"

+ 5 - 5
meta/recipes-core/tasks/task-base.bb

@@ -126,17 +126,17 @@ python __anonymous () {
 
     import bb
 
-    distro_features = set(bb.data.getVar("DISTRO_FEATURES", d, 1).split())
-    machine_features= set(bb.data.getVar("MACHINE_FEATURES", d, 1).split())
+    distro_features = set(d.getVar("DISTRO_FEATURES", 1).split())
+    machine_features= set(d.getVar("MACHINE_FEATURES", 1).split())
 
     if "bluetooth" in distro_features and not "bluetooth" in machine_features and ("pcmcia" in machine_features or "pci" in machine_features or "usbhost" in machine_features):
-	bb.data.setVar("ADD_BT", "task-base-bluetooth", d)
+	d.setVar("ADD_BT", "task-base-bluetooth")
 
     if "wifi" in distro_features and not "wifi" in machine_features and ("pcmcia" in machine_features or "pci" in machine_features or "usbhost" in machine_features):
-	bb.data.setVar("ADD_WIFI", "task-base-wifi", d)
+	d.setVar("ADD_WIFI", "task-base-wifi")
 
     if "3g" in distro_features and not "3g" in machine_features and ("pcmcia" in machine_features or "pci" in machine_features or "usbhost" in machine_features):
-	bb.data.setVar("ADD_3G", "task-base-3g", d)
+	d.setVar("ADD_3G", "task-base-3g")
 }
 
 #

+ 3 - 3
meta/recipes-core/tasks/task-core-sdk.bb

@@ -50,7 +50,7 @@ RDEPENDS_task-core-sdk = "\
 
 #python generate_sdk_pkgs () {
 #    poky_pkgs = read_pkgdata('task-core', d)['PACKAGES']
-#    pkgs = bb.data.getVar('PACKAGES', d, 1).split()
+#    pkgs = d.getVar('PACKAGES', 1).split()
 #    for pkg in poky_pkgs.split():
 #        newpkg = pkg.replace('task-core', 'task-core-sdk')
 #
@@ -79,9 +79,9 @@ RDEPENDS_task-core-sdk = "\
 #                    if packaged('%s-dev' % name, d):
 #                        rreclist.append('%s-dev' % name)
 #
-#            oldrrec = bb.data.getVar('RRECOMMENDS_%s' % newpkg, d) or ''
+#            oldrrec = d.getVar('RRECOMMENDS_%s' % newpkg) or ''
 #            bb.data.setVar('RRECOMMENDS_%s' % newpkg, oldrrec + ' ' + ' '.join(rreclist), d)
-#            # bb.note('RRECOMMENDS_%s = "%s"' % (newpkg, bb.data.getVar('RRECOMMENDS_%s' % newpkg, d)))
+#            # bb.note('RRECOMMENDS_%s = "%s"' % (newpkg, d.getVar('RRECOMMENDS_%s' % newpkg)))
 #
 #    # bb.note('pkgs is %s' % pkgs)
 #    bb.data.setVar('PACKAGES', ' '.join(pkgs), d)

+ 8 - 8
meta/recipes-core/uclibc/uclibc-config.inc

@@ -35,7 +35,7 @@ def map_uclibc_arch(a, d):
 	"""Return the uClibc architecture for the given TARGET_ARCH."""
 	import re
 
-	valid_archs = bb.data.getVar('valid_archs', d, 1).split()
+	valid_archs = d.getVar('valid_archs', 1).split()
 
 	if   re.match('^(arm|sa110).*', a):	return 'arm'
 	elif re.match('^(i.86|athlon)$', a):	return 'i386'
@@ -50,14 +50,14 @@ def map_uclibc_arch(a, d):
 	else:
 		bb.error("cannot map '%s' to a uClibc architecture" % a)
 
-export UCLIBC_ARCH = "${@map_uclibc_arch(bb.data.getVar('TARGET_ARCH', d, 1), d)}"
+export UCLIBC_ARCH = "${@map_uclibc_arch(d.getVar('TARGET_ARCH', 1), d)}"
 
 def map_uclibc_abi(o, d):
 	"""Return the uClibc ABI for the given TARGET_OS."""
 	import re
 
-	arch = bb.data.getVar('TARGET_ARCH', d, 1)
-	if   map_uclibc_arch(bb.data.getVar('TARGET_ARCH', d, 1), d) == "arm":
+	arch = d.getVar('TARGET_ARCH', 1)
+	if   map_uclibc_arch(d.getVar('TARGET_ARCH', 1), d) == "arm":
 		if re.match('.*eabi$', o): return 'ARM_EABI'
 		else:                    return 'ARM_OABI'
 	# FIXME: This is inaccurate! Handle o32, n32, n64
@@ -65,7 +65,7 @@ def map_uclibc_abi(o, d):
 	elif re.match('^mips.*', arch):    return 'MIPS_O32_ABI'
 	return ""
 
-export UCLIBC_ABI = "${@map_uclibc_abi(bb.data.getVar('TARGET_OS', d, 1), d)}"
+export UCLIBC_ABI = "${@map_uclibc_abi(d.getVar('TARGET_OS', 1), d)}"
 
 def map_uclibc_endian(a, d):
 	"""Return the uClibc endianess for the given TARGET_ARCH."""
@@ -79,7 +79,7 @@ def map_uclibc_endian(a, d):
 		return 'BIG'
 	return 'LITTLE'
 
-export UCLIBC_ENDIAN = "${@map_uclibc_endian(bb.data.getVar('TARGET_ARCH', d, 1), d)}"
+export UCLIBC_ENDIAN = "${@map_uclibc_endian(d.getVar('TARGET_ARCH', 1), d)}"
 
 # internal helper
 def uclibc_cfg(feature, features, tokens, cnf, rem):
@@ -94,8 +94,8 @@ def uclibc_cfg(feature, features, tokens, cnf, rem):
 # Map distro and machine features to config settings
 def features_to_uclibc_settings(d):
 	cnf, rem = ([], [])
-	distro_features = bb.data.getVar('DISTRO_FEATURES', d, True).split()
-	machine_features = bb.data.getVar('MACHINE_FEATURES', d, True).split()
+	distro_features = d.getVar('DISTRO_FEATURES', True).split()
+	machine_features = d.getVar('MACHINE_FEATURES', True).split()
 	uclibc_cfg('ipv4',      distro_features, 'UCLIBC_HAS_IPV4', cnf, rem)
 	uclibc_cfg('ipv6',      distro_features, 'UCLIBC_HAS_IPV6', cnf, rem)
 	uclibc_cfg('largefile', distro_features, 'UCLIBC_HAS_LFS', cnf, rem)

+ 5 - 5
meta/recipes-core/uclibc/uclibc.inc

@@ -125,9 +125,9 @@ configmangle = '/^KERNEL_HEADERS/d; \
                 /^SHARED_LIB_LOADER_PREFIX/d; \
                 /^UCLIBC_EXTRA_CFLAGS/d; \
                 s,.*UCLIBC_HAS_WCHAR.*,UCLIBC_HAS_WCHAR=y,g; \
-                ${@["","s,.*COMPILE_IN_THUMB_MODE.*,COMPILE_IN_THUMB_MODE=y,;"][bb.data.getVar("ARM_INSTRUCTION_SET", d, 1) != "arm"]} \
-                ${@["","s,.*UCLIBC_HAS_LOCALE.*,UCLIBC_HAS_LOCALE=y,;"][bb.data.getVar("USE_NLS", d, 1) == "yes"]} \
-                ${@["","s,.*LDSO_GNU_HASH_SUPPORT.*,# LDSO_GNU_HASH_SUPPORT is not set,;"][bb.data.getVar("TARGET_ARCH", d, 1) in ['mips', 'mipsel', 'mips64', 'mips64el', 'avr32']]} \
+                ${@["","s,.*COMPILE_IN_THUMB_MODE.*,COMPILE_IN_THUMB_MODE=y,;"][d.getVar("ARM_INSTRUCTION_SET", 1) != "arm"]} \
+                ${@["","s,.*UCLIBC_HAS_LOCALE.*,UCLIBC_HAS_LOCALE=y,;"][d.getVar("USE_NLS", 1) == "yes"]} \
+                ${@["","s,.*LDSO_GNU_HASH_SUPPORT.*,# LDSO_GNU_HASH_SUPPORT is not set,;"][d.getVar("TARGET_ARCH", 1) in ['mips', 'mipsel', 'mips64', 'mips64el', 'avr32']]} \
                 /^CROSS/d; \
                 /^TARGET_ARCH=/d; \
                 /^TARGET_/s,^\([^=]*\).*,# \1 is not set,g; \
@@ -139,7 +139,7 @@ OE_FEATURES := "${@features_to_uclibc_conf(d)}"
 OE_DEL      := "${@features_to_uclibc_del(d)}"
 python () {
   if "${OE_DEL}":
-    bb.data.setVar('configmangle_append', "${OE_DEL}" + "\n", d)
+    d.setVar('configmangle_append', "${OE_DEL}" + "\n")
   if "${OE_FEATURES}":
     bb.data.setVar('configmangle_append',
                    "/^### DISTRO FEATURES$/a\\\n%s\n\n" %
@@ -161,7 +161,7 @@ python () {
                         ("${UCLIBC_ARCH}", "${UCLIBC_ARCH}"),
                  d)
   bb.data.setVar('configmangle_append',
-    "/^### FPU$/a\\\n%s\n\n" % (["UCLIBC_HAS_FPU=y","# UCLIBC_HAS_FPU is not set"][bb.data.getVar('TARGET_FPU', d, 1) in [ 'soft' ]]), d)
+    "/^### FPU$/a\\\n%s\n\n" % (["UCLIBC_HAS_FPU=y","# UCLIBC_HAS_FPU is not set"][d.getVar('TARGET_FPU', 1) in [ 'soft' ]]), d)
   if "${UCLIBC_ENDIAN}":
     bb.data.setVar('configmangle_append',
                    "/^### ABI$/a\\\nARCH_WANTS_%s_ENDIAN=y\n\n" % ("${UCLIBC_ENDIAN}"),

+ 2 - 2
meta/recipes-devtools/apt/apt-native.inc

@@ -13,14 +13,14 @@ python do_install () {
 }
 
 python do_install_config () {
-	indir = os.path.dirname(bb.data.getVar('FILE',d,1))
+	indir = os.path.dirname(d.getVar('FILE',1))
 	infile = file(os.path.join(indir, 'files', 'apt.conf'), 'r')
 	data = infile.read()
 	infile.close()
 
 	data = bb.data.expand(data, d)
 
-	outdir = os.path.join(bb.data.getVar('D', d, 1), bb.data.getVar('sysconfdir', d, 1), 'apt')
+	outdir = os.path.join(d.getVar('D', 1), d.getVar('sysconfdir', 1), 'apt')
 	if not os.path.exists(outdir):
 		os.makedirs(outdir)
 	outpath = os.path.join(outdir, 'apt.conf.sample')

+ 4 - 4
meta/recipes-devtools/apt/apt-package.inc

@@ -59,15 +59,15 @@ FILES_${PN} = "${bindir}/apt-cdrom ${bindir}/apt-get \
 	       ${localstatedir} ${sysconfdir} \
 	       ${libdir}/dpkg"
 FILES_${PN}-utils = "${bindir}/apt-sortpkgs ${bindir}/apt-extracttemplates"
-FILES_${PN}-doc = "${@get_files_apt_doc(d, bb, bb.data.getVar('apt-manpages', d, 1))} \
+FILES_${PN}-doc = "${@get_files_apt_doc(d, bb, d.getVar('apt-manpages', 1))} \
 		   ${docdir}/apt"
-FILES_${PN}-utils-doc = "${@get_files_apt_doc(d, bb, bb.data.getVar('apt-utils-manpages', d, 1))}"
+FILES_${PN}-utils-doc = "${@get_files_apt_doc(d, bb, d.getVar('apt-utils-manpages', 1))}"
 FILES_${PN}-dev = "${libdir}/libapt*.so ${includedir}"
 
 do_install () {
 	set -x
-	${@get_commands_apt_doc(d, bb, bb.data.getVar('apt-manpages', d, 1))}
-	${@get_commands_apt_doc(d, bb, bb.data.getVar('apt-utils-manpages', d, 1))}
+	${@get_commands_apt_doc(d, bb, d.getVar('apt-manpages', 1))}
+	${@get_commands_apt_doc(d, bb, d.getVar('apt-utils-manpages', 1))}
 	install -d ${D}${bindir}
 	install -m 0755 bin/apt-cdrom ${D}${bindir}/
 	install -m 0755 bin/apt-get ${D}${bindir}/

+ 1 - 1
meta/recipes-devtools/automake/automake.inc

@@ -10,6 +10,6 @@ SRC_URI = "${GNU_MIRROR}/automake/automake-${PV}.tar.bz2 "
 
 inherit autotools
 
-export AUTOMAKE = "${@bb.which('automake', bb.data.getVar('PATH', d, 1))}"
+export AUTOMAKE = "${@bb.which('automake', d.getVar('PATH', 1))}"
 
 FILES_${PN} += "${datadir}/automake* ${datadir}/aclocal*"

+ 1 - 1
meta/recipes-devtools/cmake/cmake.inc

@@ -11,7 +11,7 @@ LIC_FILES_CHKSUM = "file://Copyright.txt;md5=f372516292ff7c33337bf16a74a5f9a8 \
 
 INC_PR = "r1"
 
-CMAKE_MAJOR_VERSION = "${@'.'.join(bb.data.getVar('PV',d,1).split('.')[0:2])}"
+CMAKE_MAJOR_VERSION = "${@'.'.join(d.getVar('PV',1).split('.')[0:2])}"
 
 SRC_URI = "http://www.cmake.org/files/v${CMAKE_MAJOR_VERSION}/cmake-${PV}.tar.gz \
            file://support-oe-qt4-tools-names.patch"

+ 3 - 3
meta/recipes-devtools/cmake/cmake_2.8.5.bb

@@ -13,8 +13,8 @@ SRC_URI[sha256sum] = "5e18bff75f01656c64f553412a8905527e1b85efaf3163c6fb81ea5aac
 
 # Strip ${prefix} from ${docdir}, set result into docdir_stripped
 python () {
-    prefix=bb.data.getVar("prefix", d, 1)
-    docdir=bb.data.getVar("docdir", d, 1)
+    prefix=d.getVar("prefix", 1)
+    docdir=d.getVar("docdir", 1)
 
     if not docdir.startswith(prefix):
 	raise bb.build.FuncFailed('docdir must contain prefix as its prefix')
@@ -23,7 +23,7 @@ python () {
     if len(docdir_stripped) > 0 and docdir_stripped[0] == '/':
 	docdir_stripped = docdir_stripped[1:]
 
-    bb.data.setVar("docdir_stripped", docdir_stripped, d)
+    d.setVar("docdir_stripped", docdir_stripped)
 }
 
 EXTRA_OECMAKE=" \

+ 4 - 4
meta/recipes-devtools/gcc/gcc-common.inc

@@ -7,17 +7,17 @@ NATIVEDEPS = ""
 
 inherit autotools gettext
 
-FILESDIR = "${@os.path.dirname(bb.data.getVar('FILE',d,1))}/gcc-${PV}"
+FILESDIR = "${@os.path.dirname(d.getVar('FILE',1))}/gcc-${PV}"
 
 def get_gcc_fpu_setting(bb, d):
-    if bb.data.getVar('TARGET_FPU', d, 1) in [ 'soft' ]:
+    if d.getVar('TARGET_FPU', 1) in [ 'soft' ]:
         return "--with-float=soft"
-    if bb.data.getVar('TARGET_FPU', d, 1) in [ 'ppc-efd' ]:
+    if d.getVar('TARGET_FPU', 1) in [ 'ppc-efd' ]:
         return "--enable-e500_double"
     return ""
 
 def get_gcc_mips_plt_setting(bb, d):
-    if bb.data.getVar('TARGET_ARCH', d, 1) in [ 'mips', 'mipsel' ] and 'mplt' in bb.data.getVar('DISTRO_FEATURES',d,1).split() :
+    if d.getVar('TARGET_ARCH', 1) in [ 'mips', 'mipsel' ] and 'mplt' in d.getVar('DISTRO_FEATURES',1).split() :
         return "--with-mips-plt"
     return ""
 

+ 1 - 1
meta/recipes-devtools/gcc/gcc-configure-common.inc

@@ -27,7 +27,7 @@ EXTRA_OECONF_INTERMEDIATE ?= ""
 
 GCCMULTILIB = "--disable-multilib"
 
-EXTRA_OECONF = "${@['--enable-clocale=generic', ''][bb.data.getVar('USE_NLS', d, 1) != 'no']} \
+EXTRA_OECONF = "${@['--enable-clocale=generic', ''][d.getVar('USE_NLS', 1) != 'no']} \
                 --with-gnu-ld \
                 --enable-shared \
                 --enable-languages=${LANGUAGES} \

+ 1 - 1
meta/recipes-devtools/gnu-config/gnu-config_20080123.bb

@@ -8,7 +8,7 @@ DEPENDS_virtclass-native = "perl-native-runtime"
 
 INHIBIT_DEFAULT_DEPS = "1"
 
-FIXEDSRCDATE = "${@bb.data.getVar('FILE', d, 1).split('_')[-1].split('.')[0]}"
+FIXEDSRCDATE = "${@d.getVar('FILE', 1).split('_')[-1].split('.')[0]}"
 PV = "0.1+cvs${FIXEDSRCDATE}"
 PR = "r4"
 

+ 1 - 1
meta/recipes-devtools/intltool/intltool.inc

@@ -2,7 +2,7 @@ DESCRIPTION = "Utility scripts for internationalizing XML"
 SECTION = "devel"
 LICENSE = "GPLv2"
 
-URLV="${@'.'.join(bb.data.getVar('PV',d,1).split('.')[0:2])}"
+URLV="${@'.'.join(d.getVar('PV',1).split('.')[0:2])}"
 SRC_URI = "${GNOME_MIRROR}/intltool/${URLV}/intltool-${PV}.tar.bz2"
 S = "${WORKDIR}/intltool-${PV}"
 

+ 1 - 1
meta/recipes-devtools/opkg/opkg.inc

@@ -12,7 +12,7 @@ DEPENDS_virtclass-nativesdk = "curl-nativesdk"
 
 PE = "1"
 
-FILESDIR = "${@os.path.dirname(bb.data.getVar('FILE',d,1))}/opkg"
+FILESDIR = "${@os.path.dirname(d.getVar('FILE',1))}/opkg"
 
 # Werror gives all kinds bounds issuses with gcc 4.3.3
 do_configure_prepend() {

+ 1 - 1
meta/recipes-devtools/perl/perl_5.14.2.bb

@@ -285,7 +285,7 @@ FILES_perl-module-unicore += "${libdir}/perl/${PV}/unicore"
 # packages (actually the non modules packages and not created too)
 ALLOW_EMPTY_perl-modules = "1"
 PACKAGES_append = " perl-modules "
-RRECOMMENDS_perl-modules = "${@bb.data.getVar('PACKAGES', d, 1).replace('${PN}-modules ', '').replace('${PN}-dbg ', '').replace('${PN}-misc ', '').replace('${PN}-dev ', '').replace('${PN}-pod ', '').replace('${PN}-doc ', '')}"
+RRECOMMENDS_perl-modules = "${@d.getVar('PACKAGES', 1).replace('${PN}-modules ', '').replace('${PN}-dbg ', '').replace('${PN}-misc ', '').replace('${PN}-dev ', '').replace('${PN}-pod ', '').replace('${PN}-doc ', '')}"
 
 python populate_packages_prepend () {
         libdir = bb.data.expand('${libdir}/perl/${PV}', d)

+ 1 - 1
meta/recipes-devtools/python/python-pygobject_2.27.91.bb

@@ -7,7 +7,7 @@ DEPENDS_virtclass-native = "glib-2.0-native"
 RDEPENDS_virtclass-native = ""
 PR = "r3"
 
-MAJ_VER = "${@bb.data.getVar('PV',d,1).split('.')[0]}.${@bb.data.getVar('PV',d,1).split('.')[1]}"
+MAJ_VER = "${@d.getVar('PV',1).split('.')[0]}.${@d.getVar('PV',1).split('.')[1]}"
 
 SRC_URI = "${GNOME_MIRROR}/pygobject/${MAJ_VER}/pygobject-${PV}.tar.bz2"
 

Some files were not shown because too many files changed in this diff