mirror of
https://github.com/liberatedsystems/openCom-Companion.git
synced 2024-11-25 14:50:37 +01:00
1516 lines
58 KiB
Python
1516 lines
58 KiB
Python
|
from os.path import join
|
||
|
|
||
|
from os.path import basename, dirname, exists, isdir, isfile, join, realpath, split
|
||
|
import glob
|
||
|
|
||
|
import hashlib
|
||
|
from re import match
|
||
|
|
||
|
import sh
|
||
|
import shutil
|
||
|
import fnmatch
|
||
|
import zipfile
|
||
|
import urllib.request
|
||
|
from urllib.request import urlretrieve
|
||
|
from os import listdir, unlink, environ, curdir, walk
|
||
|
from sys import stdout
|
||
|
from wheel.wheelfile import WheelFile
|
||
|
from wheel.cli.tags import tags as wheel_tags
|
||
|
import time
|
||
|
try:
|
||
|
from urlparse import urlparse
|
||
|
except ImportError:
|
||
|
from urllib.parse import urlparse
|
||
|
|
||
|
import packaging.version
|
||
|
|
||
|
from pythonforandroid.logger import (
|
||
|
logger, info, warning, debug, shprint, info_main, error)
|
||
|
from pythonforandroid.util import (
|
||
|
current_directory, ensure_dir, BuildInterruptingException, rmdir, move,
|
||
|
touch)
|
||
|
from pythonforandroid.util import load_source as import_recipe
|
||
|
|
||
|
|
||
|
url_opener = urllib.request.build_opener()
|
||
|
url_orig_headers = url_opener.addheaders
|
||
|
urllib.request.install_opener(url_opener)
|
||
|
|
||
|
|
||
|
class RecipeMeta(type):
|
||
|
def __new__(cls, name, bases, dct):
|
||
|
if name != 'Recipe':
|
||
|
if 'url' in dct:
|
||
|
dct['_url'] = dct.pop('url')
|
||
|
if 'version' in dct:
|
||
|
dct['_version'] = dct.pop('version')
|
||
|
|
||
|
return super().__new__(cls, name, bases, dct)
|
||
|
|
||
|
|
||
|
class Recipe(metaclass=RecipeMeta):
|
||
|
_url = None
|
||
|
'''The address from which the recipe may be downloaded. This is not
|
||
|
essential, it may be omitted if the source is available some other
|
||
|
way, such as via the :class:`IncludedFilesBehaviour` mixin.
|
||
|
|
||
|
If the url includes the version, you may (and probably should)
|
||
|
replace this with ``{version}``, which will automatically be
|
||
|
replaced by the :attr:`version` string during download.
|
||
|
|
||
|
.. note:: Methods marked (internal) are used internally and you
|
||
|
probably don't need to call them, but they are available
|
||
|
if you want.
|
||
|
'''
|
||
|
|
||
|
_version = None
|
||
|
'''A string giving the version of the software the recipe describes,
|
||
|
e.g. ``2.0.3`` or ``master``.'''
|
||
|
|
||
|
md5sum = None
|
||
|
'''The md5sum of the source from the :attr:`url`. Non-essential, but
|
||
|
you should try to include this, it is used to check that the download
|
||
|
finished correctly.
|
||
|
'''
|
||
|
|
||
|
sha512sum = None
|
||
|
'''The sha512sum of the source from the :attr:`url`. Non-essential, but
|
||
|
you should try to include this, it is used to check that the download
|
||
|
finished correctly.
|
||
|
'''
|
||
|
|
||
|
blake2bsum = None
|
||
|
'''The blake2bsum of the source from the :attr:`url`. Non-essential, but
|
||
|
you should try to include this, it is used to check that the download
|
||
|
finished correctly.
|
||
|
'''
|
||
|
|
||
|
depends = []
|
||
|
'''A list containing the names of any recipes that this recipe depends on.
|
||
|
'''
|
||
|
|
||
|
conflicts = []
|
||
|
'''A list containing the names of any recipes that are known to be
|
||
|
incompatible with this one.'''
|
||
|
|
||
|
opt_depends = []
|
||
|
'''A list of optional dependencies, that must be built before this
|
||
|
recipe if they are built at all, but whose presence is not essential.'''
|
||
|
|
||
|
patches = []
|
||
|
'''A list of patches to apply to the source. Values can be either a string
|
||
|
referring to the patch file relative to the recipe dir, or a tuple of the
|
||
|
string patch file and a callable, which will receive the kwargs `arch` and
|
||
|
`recipe`, which should return True if the patch should be applied.'''
|
||
|
|
||
|
python_depends = []
|
||
|
'''A list of pure-Python packages that this package requires. These
|
||
|
packages will NOT be available at build time, but will be added to the
|
||
|
list of pure-Python packages to install via pip. If you need these packages
|
||
|
at build time, you must create a recipe.'''
|
||
|
|
||
|
archs = ['armeabi'] # Not currently implemented properly
|
||
|
|
||
|
built_libraries = {}
|
||
|
"""Each recipe that builds a system library (e.g.:libffi, openssl, etc...)
|
||
|
should contain a dict holding the relevant information of the library. The
|
||
|
keys should be the generated libraries and the values the relative path of
|
||
|
the library inside his build folder. This dict will be used to perform
|
||
|
different operations:
|
||
|
- copy the library into the right location, depending on if it's shared
|
||
|
or static)
|
||
|
- check if we have to rebuild the library
|
||
|
|
||
|
Here an example of how it would look like for `libffi` recipe:
|
||
|
|
||
|
- `built_libraries = {'libffi.so': '.libs'}`
|
||
|
|
||
|
.. note:: in case that the built library resides in recipe's build
|
||
|
directory, you can set the following values for the relative
|
||
|
path: `'.', None or ''`
|
||
|
"""
|
||
|
|
||
|
need_stl_shared = False
|
||
|
'''Some libraries or python packages may need the c++_shared in APK.
|
||
|
We can automatically do this for any recipe if we set this property to
|
||
|
`True`'''
|
||
|
|
||
|
stl_lib_name = 'c++_shared'
|
||
|
'''
|
||
|
The default STL shared lib to use: `c++_shared`.
|
||
|
|
||
|
.. note:: Android NDK version > 17 only supports 'c++_shared', because
|
||
|
starting from NDK r18 the `gnustl_shared` lib has been deprecated.
|
||
|
'''
|
||
|
|
||
|
def get_stl_library(self, arch):
|
||
|
return join(
|
||
|
arch.ndk_lib_dir,
|
||
|
'lib{name}.so'.format(name=self.stl_lib_name),
|
||
|
)
|
||
|
|
||
|
def install_stl_lib(self, arch):
|
||
|
if not self.ctx.has_lib(
|
||
|
arch.arch, 'lib{name}.so'.format(name=self.stl_lib_name)
|
||
|
):
|
||
|
self.install_libs(arch, self.get_stl_library(arch))
|
||
|
|
||
|
@property
|
||
|
def version(self):
|
||
|
key = 'VERSION_' + self.name
|
||
|
return environ.get(key, self._version)
|
||
|
|
||
|
@property
|
||
|
def url(self):
|
||
|
key = 'URL_' + self.name
|
||
|
return environ.get(key, self._url)
|
||
|
|
||
|
@property
|
||
|
def versioned_url(self):
|
||
|
'''A property returning the url of the recipe with ``{version}``
|
||
|
replaced by the :attr:`url`. If accessing the url, you should use this
|
||
|
property, *not* access the url directly.'''
|
||
|
if self.url is None:
|
||
|
return None
|
||
|
return self.url.format(version=self.version)
|
||
|
|
||
|
def download_file(self, url, target, cwd=None):
|
||
|
"""
|
||
|
(internal) Download an ``url`` to a ``target``.
|
||
|
"""
|
||
|
if not url:
|
||
|
return
|
||
|
|
||
|
info('Downloading {} from {}'.format(self.name, url))
|
||
|
|
||
|
if cwd:
|
||
|
target = join(cwd, target)
|
||
|
|
||
|
parsed_url = urlparse(url)
|
||
|
if parsed_url.scheme in ('http', 'https'):
|
||
|
def report_hook(index, blksize, size):
|
||
|
if size <= 0:
|
||
|
progression = '{0} bytes'.format(index * blksize)
|
||
|
else:
|
||
|
progression = '{0:.2f}%'.format(
|
||
|
index * blksize * 100. / float(size))
|
||
|
if "CI" not in environ:
|
||
|
stdout.write('- Download {}\r'.format(progression))
|
||
|
stdout.flush()
|
||
|
|
||
|
if exists(target):
|
||
|
unlink(target)
|
||
|
|
||
|
# Download item with multiple attempts (for bad connections):
|
||
|
attempts = 0
|
||
|
seconds = 1
|
||
|
while True:
|
||
|
try:
|
||
|
# jqueryui.com returns a 403 w/ the default user agent
|
||
|
# Mozilla/5.0 doesnt handle redirection for liblzma
|
||
|
url_opener.addheaders = [('User-agent', 'Wget/1.0')]
|
||
|
urlretrieve(url, target, report_hook)
|
||
|
except OSError as e:
|
||
|
attempts += 1
|
||
|
if attempts >= 5:
|
||
|
raise
|
||
|
stdout.write('Download failed: {}; retrying in {} second(s)...'.format(e, seconds))
|
||
|
time.sleep(seconds)
|
||
|
seconds *= 2
|
||
|
continue
|
||
|
finally:
|
||
|
url_opener.addheaders = url_orig_headers
|
||
|
break
|
||
|
return target
|
||
|
elif parsed_url.scheme in ('git', 'git+file', 'git+ssh', 'git+http', 'git+https'):
|
||
|
if not isdir(target):
|
||
|
if url.startswith('git+'):
|
||
|
url = url[4:]
|
||
|
# if 'version' is specified, do a shallow clone
|
||
|
if self.version:
|
||
|
ensure_dir(target)
|
||
|
with current_directory(target):
|
||
|
shprint(sh.git, 'init')
|
||
|
shprint(sh.git, 'remote', 'add', 'origin', url)
|
||
|
else:
|
||
|
shprint(sh.git, 'clone', '--recursive', url, target)
|
||
|
with current_directory(target):
|
||
|
if self.version:
|
||
|
shprint(sh.git, 'fetch', '--tags', '--depth', '1')
|
||
|
shprint(sh.git, 'checkout', self.version)
|
||
|
branch = sh.git('branch', '--show-current')
|
||
|
if branch:
|
||
|
shprint(sh.git, 'pull')
|
||
|
shprint(sh.git, 'pull', '--recurse-submodules')
|
||
|
shprint(sh.git, 'submodule', 'update', '--recursive', '--init', '--depth', '1')
|
||
|
return target
|
||
|
|
||
|
def apply_patch(self, filename, arch, build_dir=None):
|
||
|
"""
|
||
|
Apply a patch from the current recipe directory into the current
|
||
|
build directory.
|
||
|
|
||
|
.. versionchanged:: 0.6.0
|
||
|
Add ability to apply patch from any dir via kwarg `build_dir`'''
|
||
|
"""
|
||
|
info("Applying patch {}".format(filename))
|
||
|
build_dir = build_dir if build_dir else self.get_build_dir(arch)
|
||
|
filename = join(self.get_recipe_dir(), filename)
|
||
|
shprint(sh.patch, "-t", "-d", build_dir, "-p1",
|
||
|
"-i", filename, _tail=10)
|
||
|
|
||
|
def copy_file(self, filename, dest):
|
||
|
info("Copy {} to {}".format(filename, dest))
|
||
|
filename = join(self.get_recipe_dir(), filename)
|
||
|
dest = join(self.build_dir, dest)
|
||
|
shutil.copy(filename, dest)
|
||
|
|
||
|
def append_file(self, filename, dest):
|
||
|
info("Append {} to {}".format(filename, dest))
|
||
|
filename = join(self.get_recipe_dir(), filename)
|
||
|
dest = join(self.build_dir, dest)
|
||
|
with open(filename, "rb") as fd:
|
||
|
data = fd.read()
|
||
|
with open(dest, "ab") as fd:
|
||
|
fd.write(data)
|
||
|
|
||
|
@property
|
||
|
def name(self):
|
||
|
'''The name of the recipe, the same as the folder containing it.'''
|
||
|
modname = self.__class__.__module__
|
||
|
return modname.split(".", 2)[-1]
|
||
|
|
||
|
@property
|
||
|
def filtered_archs(self):
|
||
|
'''Return archs of self.ctx that are valid build archs
|
||
|
for the Recipe.'''
|
||
|
result = []
|
||
|
for arch in self.ctx.archs:
|
||
|
if not self.archs or (arch.arch in self.archs):
|
||
|
result.append(arch)
|
||
|
return result
|
||
|
|
||
|
def check_recipe_choices(self):
|
||
|
'''Checks what recipes are being built to see which of the alternative
|
||
|
and optional dependencies are being used,
|
||
|
and returns a list of these.'''
|
||
|
recipes = []
|
||
|
built_recipes = self.ctx.recipe_build_order
|
||
|
for recipe in self.depends:
|
||
|
if isinstance(recipe, (tuple, list)):
|
||
|
for alternative in recipe:
|
||
|
if alternative in built_recipes:
|
||
|
recipes.append(alternative)
|
||
|
break
|
||
|
for recipe in self.opt_depends:
|
||
|
if recipe in built_recipes:
|
||
|
recipes.append(recipe)
|
||
|
return sorted(recipes)
|
||
|
|
||
|
def get_opt_depends_in_list(self, recipes):
|
||
|
'''Given a list of recipe names, returns those that are also in
|
||
|
self.opt_depends.
|
||
|
'''
|
||
|
return [recipe for recipe in recipes if recipe in self.opt_depends]
|
||
|
|
||
|
def get_build_container_dir(self, arch):
|
||
|
'''Given the arch name, returns the directory where it will be
|
||
|
built.
|
||
|
|
||
|
This returns a different directory depending on what
|
||
|
alternative or optional dependencies are being built.
|
||
|
'''
|
||
|
dir_name = self.get_dir_name()
|
||
|
return join(self.ctx.build_dir, 'other_builds',
|
||
|
dir_name, '{}__ndk_target_{}'.format(arch, self.ctx.ndk_api))
|
||
|
|
||
|
def get_dir_name(self):
|
||
|
choices = self.check_recipe_choices()
|
||
|
dir_name = '-'.join([self.name] + choices)
|
||
|
return dir_name
|
||
|
|
||
|
def get_build_dir(self, arch):
|
||
|
'''Given the arch name, returns the directory where the
|
||
|
downloaded/copied package will be built.'''
|
||
|
|
||
|
return join(self.get_build_container_dir(arch), self.name)
|
||
|
|
||
|
def get_recipe_dir(self):
|
||
|
"""
|
||
|
Returns the local recipe directory or defaults to the core recipe
|
||
|
directory.
|
||
|
"""
|
||
|
if self.ctx.local_recipes is not None:
|
||
|
local_recipe_dir = join(self.ctx.local_recipes, self.name)
|
||
|
if exists(local_recipe_dir):
|
||
|
return local_recipe_dir
|
||
|
return join(self.ctx.root_dir, 'recipes', self.name)
|
||
|
|
||
|
# Public Recipe API to be subclassed if needed
|
||
|
|
||
|
def download_if_necessary(self):
|
||
|
info_main('Downloading {}'.format(self.name))
|
||
|
user_dir = environ.get('P4A_{}_DIR'.format(self.name.lower()))
|
||
|
if user_dir is not None:
|
||
|
info('P4A_{}_DIR is set, skipping download for {}'.format(
|
||
|
self.name, self.name))
|
||
|
return
|
||
|
self.download()
|
||
|
|
||
|
def download(self):
|
||
|
if self.url is None:
|
||
|
info('Skipping {} download as no URL is set'.format(self.name))
|
||
|
return
|
||
|
|
||
|
url = self.versioned_url
|
||
|
expected_digests = {}
|
||
|
for alg in set(hashlib.algorithms_guaranteed) | set(('md5', 'sha512', 'blake2b')):
|
||
|
expected_digest = getattr(self, alg + 'sum') if hasattr(self, alg + 'sum') else None
|
||
|
ma = match(u'^(.+)#' + alg + u'=([0-9a-f]{32,})$', url)
|
||
|
if ma: # fragmented URL?
|
||
|
if expected_digest:
|
||
|
raise ValueError(
|
||
|
('Received {}sum from both the {} recipe '
|
||
|
'and its url').format(alg, self.name))
|
||
|
url = ma.group(1)
|
||
|
expected_digest = ma.group(2)
|
||
|
if expected_digest:
|
||
|
expected_digests[alg] = expected_digest
|
||
|
|
||
|
ensure_dir(join(self.ctx.packages_path, self.name))
|
||
|
|
||
|
with current_directory(join(self.ctx.packages_path, self.name)):
|
||
|
filename = shprint(sh.basename, url).stdout[:-1].decode('utf-8')
|
||
|
|
||
|
do_download = True
|
||
|
marker_filename = '.mark-{}'.format(filename)
|
||
|
if exists(filename) and isfile(filename):
|
||
|
if not exists(marker_filename):
|
||
|
shprint(sh.rm, filename)
|
||
|
else:
|
||
|
for alg, expected_digest in expected_digests.items():
|
||
|
current_digest = algsum(alg, filename)
|
||
|
if current_digest != expected_digest:
|
||
|
debug('* Generated {}sum: {}'.format(alg,
|
||
|
current_digest))
|
||
|
debug('* Expected {}sum: {}'.format(alg,
|
||
|
expected_digest))
|
||
|
raise ValueError(
|
||
|
('Generated {0}sum does not match expected {0}sum '
|
||
|
'for {1} recipe').format(alg, self.name))
|
||
|
do_download = False
|
||
|
|
||
|
# If we got this far, we will download
|
||
|
if do_download:
|
||
|
debug('Downloading {} from {}'.format(self.name, url))
|
||
|
|
||
|
shprint(sh.rm, '-f', marker_filename)
|
||
|
self.download_file(self.versioned_url, filename)
|
||
|
touch(marker_filename)
|
||
|
|
||
|
if exists(filename) and isfile(filename):
|
||
|
for alg, expected_digest in expected_digests.items():
|
||
|
current_digest = algsum(alg, filename)
|
||
|
if current_digest != expected_digest:
|
||
|
debug('* Generated {}sum: {}'.format(alg,
|
||
|
current_digest))
|
||
|
debug('* Expected {}sum: {}'.format(alg,
|
||
|
expected_digest))
|
||
|
raise ValueError(
|
||
|
('Generated {0}sum does not match expected {0}sum '
|
||
|
'for {1} recipe').format(alg, self.name))
|
||
|
else:
|
||
|
info('{} download already cached, skipping'.format(self.name))
|
||
|
|
||
|
def unpack(self, arch):
|
||
|
info_main('Unpacking {} for {}'.format(self.name, arch))
|
||
|
|
||
|
build_dir = self.get_build_container_dir(arch)
|
||
|
|
||
|
user_dir = environ.get('P4A_{}_DIR'.format(self.name.lower()))
|
||
|
if user_dir is not None:
|
||
|
info('P4A_{}_DIR exists, symlinking instead'.format(
|
||
|
self.name.lower()))
|
||
|
if exists(self.get_build_dir(arch)):
|
||
|
return
|
||
|
rmdir(build_dir)
|
||
|
ensure_dir(build_dir)
|
||
|
shprint(sh.cp, '-a', user_dir, self.get_build_dir(arch))
|
||
|
return
|
||
|
|
||
|
if self.url is None:
|
||
|
info('Skipping {} unpack as no URL is set'.format(self.name))
|
||
|
return
|
||
|
|
||
|
filename = shprint(
|
||
|
sh.basename, self.versioned_url).stdout[:-1].decode('utf-8')
|
||
|
ma = match(u'^(.+)#[a-z0-9_]{3,}=([0-9a-f]{32,})$', filename)
|
||
|
if ma: # fragmented URL?
|
||
|
filename = ma.group(1)
|
||
|
|
||
|
with current_directory(build_dir):
|
||
|
directory_name = self.get_build_dir(arch)
|
||
|
|
||
|
if not exists(directory_name) or not isdir(directory_name):
|
||
|
extraction_filename = join(
|
||
|
self.ctx.packages_path, self.name, filename)
|
||
|
if isfile(extraction_filename):
|
||
|
if extraction_filename.endswith(('.zip', '.whl')):
|
||
|
try:
|
||
|
sh.unzip(extraction_filename)
|
||
|
except (sh.ErrorReturnCode_1, sh.ErrorReturnCode_2):
|
||
|
# return code 1 means unzipping had
|
||
|
# warnings but did complete,
|
||
|
# apparently happens sometimes with
|
||
|
# github zips
|
||
|
pass
|
||
|
fileh = zipfile.ZipFile(extraction_filename, 'r')
|
||
|
root_directory = fileh.filelist[0].filename.split('/')[0]
|
||
|
if root_directory != basename(directory_name):
|
||
|
move(root_directory, directory_name)
|
||
|
elif extraction_filename.endswith(
|
||
|
('.tar.gz', '.tgz', '.tar.bz2', '.tbz2', '.tar.xz', '.txz')):
|
||
|
sh.tar('xf', extraction_filename)
|
||
|
root_directory = sh.tar('tf', extraction_filename).stdout.decode(
|
||
|
'utf-8').split('\n')[0].split('/')[0]
|
||
|
if root_directory != basename(directory_name):
|
||
|
move(root_directory, directory_name)
|
||
|
else:
|
||
|
raise Exception(
|
||
|
'Could not extract {} download, it must be .zip, '
|
||
|
'.tar.gz or .tar.bz2 or .tar.xz'.format(extraction_filename))
|
||
|
elif isdir(extraction_filename):
|
||
|
ensure_dir(directory_name)
|
||
|
for entry in listdir(extraction_filename):
|
||
|
# Previously we filtered out the .git folder, but during the build process for some recipes
|
||
|
# (e.g. when version is parsed by `setuptools_scm`) that may be needed.
|
||
|
shprint(sh.cp, '-Rv',
|
||
|
join(extraction_filename, entry),
|
||
|
directory_name)
|
||
|
else:
|
||
|
raise Exception(
|
||
|
'Given path is neither a file nor a directory: {}'
|
||
|
.format(extraction_filename))
|
||
|
|
||
|
else:
|
||
|
info('{} is already unpacked, skipping'.format(self.name))
|
||
|
|
||
|
def get_recipe_env(self, arch=None, with_flags_in_cc=True):
|
||
|
"""Return the env specialized for the recipe
|
||
|
"""
|
||
|
if arch is None:
|
||
|
arch = self.filtered_archs[0]
|
||
|
env = arch.get_env(with_flags_in_cc=with_flags_in_cc)
|
||
|
return env
|
||
|
|
||
|
def prebuild_arch(self, arch):
|
||
|
'''Run any pre-build tasks for the Recipe. By default, this checks if
|
||
|
any prebuild_archname methods exist for the archname of the current
|
||
|
architecture, and runs them if so.'''
|
||
|
prebuild = "prebuild_{}".format(arch.arch.replace('-', '_'))
|
||
|
if hasattr(self, prebuild):
|
||
|
getattr(self, prebuild)()
|
||
|
else:
|
||
|
info('{} has no {}, skipping'.format(self.name, prebuild))
|
||
|
|
||
|
def is_patched(self, arch):
|
||
|
build_dir = self.get_build_dir(arch.arch)
|
||
|
return exists(join(build_dir, '.patched'))
|
||
|
|
||
|
def apply_patches(self, arch, build_dir=None):
|
||
|
'''Apply any patches for the Recipe.
|
||
|
|
||
|
.. versionchanged:: 0.6.0
|
||
|
Add ability to apply patches from any dir via kwarg `build_dir`'''
|
||
|
if self.patches:
|
||
|
info_main('Applying patches for {}[{}]'
|
||
|
.format(self.name, arch.arch))
|
||
|
|
||
|
if self.is_patched(arch):
|
||
|
info_main('{} already patched, skipping'.format(self.name))
|
||
|
return
|
||
|
|
||
|
build_dir = build_dir if build_dir else self.get_build_dir(arch.arch)
|
||
|
for patch in self.patches:
|
||
|
if isinstance(patch, (tuple, list)):
|
||
|
patch, patch_check = patch
|
||
|
if not patch_check(arch=arch, recipe=self):
|
||
|
continue
|
||
|
|
||
|
self.apply_patch(
|
||
|
patch.format(version=self.version, arch=arch.arch),
|
||
|
arch.arch, build_dir=build_dir)
|
||
|
|
||
|
touch(join(build_dir, '.patched'))
|
||
|
|
||
|
def should_build(self, arch):
|
||
|
'''Should perform any necessary test and return True only if it needs
|
||
|
building again. Per default we implement a library test, in case that
|
||
|
we detect so.
|
||
|
|
||
|
'''
|
||
|
if self.built_libraries:
|
||
|
return not all(
|
||
|
exists(lib) for lib in self.get_libraries(arch.arch)
|
||
|
)
|
||
|
return True
|
||
|
|
||
|
def build_arch(self, arch):
|
||
|
'''Run any build tasks for the Recipe. By default, this checks if
|
||
|
any build_archname methods exist for the archname of the current
|
||
|
architecture, and runs them if so.'''
|
||
|
build = "build_{}".format(arch.arch)
|
||
|
if hasattr(self, build):
|
||
|
getattr(self, build)()
|
||
|
|
||
|
def install_libraries(self, arch):
|
||
|
'''This method is always called after `build_arch`. In case that we
|
||
|
detect a library recipe, defined by the class attribute
|
||
|
`built_libraries`, we will copy all defined libraries into the
|
||
|
right location.
|
||
|
'''
|
||
|
if not self.built_libraries:
|
||
|
return
|
||
|
shared_libs = [
|
||
|
lib for lib in self.get_libraries(arch) if lib.endswith(".so")
|
||
|
]
|
||
|
self.install_libs(arch, *shared_libs)
|
||
|
|
||
|
def postbuild_arch(self, arch):
|
||
|
'''Run any post-build tasks for the Recipe. By default, this checks if
|
||
|
any postbuild_archname methods exist for the archname of the
|
||
|
current architecture, and runs them if so.
|
||
|
'''
|
||
|
postbuild = "postbuild_{}".format(arch.arch)
|
||
|
if hasattr(self, postbuild):
|
||
|
getattr(self, postbuild)()
|
||
|
|
||
|
if self.need_stl_shared:
|
||
|
self.install_stl_lib(arch)
|
||
|
|
||
|
def prepare_build_dir(self, arch):
|
||
|
'''Copies the recipe data into a build dir for the given arch. By
|
||
|
default, this unpacks a downloaded recipe. You should override
|
||
|
it (or use a Recipe subclass with different behaviour) if you
|
||
|
want to do something else.
|
||
|
'''
|
||
|
self.unpack(arch)
|
||
|
|
||
|
def clean_build(self, arch=None):
|
||
|
'''Deletes all the build information of the recipe.
|
||
|
|
||
|
If arch is not None, only this arch dir is deleted. Otherwise
|
||
|
(the default) all builds for all archs are deleted.
|
||
|
|
||
|
By default, this just deletes the main build dir. If the
|
||
|
recipe has e.g. object files biglinked, or .so files stored
|
||
|
elsewhere, you should override this method.
|
||
|
|
||
|
This method is intended for testing purposes, it may have
|
||
|
strange results. Rebuild everything if this seems to happen.
|
||
|
|
||
|
'''
|
||
|
if arch is None:
|
||
|
base_dir = join(self.ctx.build_dir, 'other_builds', self.name)
|
||
|
else:
|
||
|
base_dir = self.get_build_container_dir(arch)
|
||
|
dirs = glob.glob(base_dir + '-*')
|
||
|
if exists(base_dir):
|
||
|
dirs.append(base_dir)
|
||
|
if not dirs:
|
||
|
warning('Attempted to clean build for {} but found no existing '
|
||
|
'build dirs'.format(self.name))
|
||
|
|
||
|
for directory in dirs:
|
||
|
rmdir(directory)
|
||
|
|
||
|
# Delete any Python distributions to ensure the recipe build
|
||
|
# doesn't persist in site-packages
|
||
|
rmdir(self.ctx.python_installs_dir)
|
||
|
|
||
|
def install_libs(self, arch, *libs):
|
||
|
libs_dir = self.ctx.get_libs_dir(arch.arch)
|
||
|
if not libs:
|
||
|
warning('install_libs called with no libraries to install!')
|
||
|
return
|
||
|
args = libs + (libs_dir,)
|
||
|
shprint(sh.cp, *args)
|
||
|
|
||
|
def has_libs(self, arch, *libs):
|
||
|
return all(map(lambda lib: self.ctx.has_lib(arch.arch, lib), libs))
|
||
|
|
||
|
def get_libraries(self, arch_name, in_context=False):
|
||
|
"""Return the full path of the library depending on the architecture.
|
||
|
Per default, the build library path it will be returned, unless
|
||
|
`get_libraries` has been called with kwarg `in_context` set to
|
||
|
True.
|
||
|
|
||
|
.. note:: this method should be used for library recipes only
|
||
|
"""
|
||
|
recipe_libs = set()
|
||
|
if not self.built_libraries:
|
||
|
return recipe_libs
|
||
|
for lib, rel_path in self.built_libraries.items():
|
||
|
if not in_context:
|
||
|
abs_path = join(self.get_build_dir(arch_name), rel_path, lib)
|
||
|
if rel_path in {".", "", None}:
|
||
|
abs_path = join(self.get_build_dir(arch_name), lib)
|
||
|
else:
|
||
|
abs_path = join(self.ctx.get_libs_dir(arch_name), lib)
|
||
|
recipe_libs.add(abs_path)
|
||
|
return recipe_libs
|
||
|
|
||
|
@classmethod
|
||
|
def recipe_dirs(cls, ctx):
|
||
|
recipe_dirs = []
|
||
|
if ctx.local_recipes is not None:
|
||
|
recipe_dirs.append(realpath(ctx.local_recipes))
|
||
|
if ctx.storage_dir:
|
||
|
recipe_dirs.append(join(ctx.storage_dir, 'recipes'))
|
||
|
recipe_dirs.append(join(ctx.root_dir, "recipes"))
|
||
|
return recipe_dirs
|
||
|
|
||
|
@classmethod
|
||
|
def list_recipes(cls, ctx):
|
||
|
forbidden_dirs = ('__pycache__', )
|
||
|
for recipes_dir in cls.recipe_dirs(ctx):
|
||
|
if recipes_dir and exists(recipes_dir):
|
||
|
for name in listdir(recipes_dir):
|
||
|
if name in forbidden_dirs:
|
||
|
continue
|
||
|
fn = join(recipes_dir, name)
|
||
|
if isdir(fn):
|
||
|
yield name
|
||
|
|
||
|
@classmethod
|
||
|
def get_recipe(cls, name, ctx):
|
||
|
'''Returns the Recipe with the given name, if it exists.'''
|
||
|
name = name.lower()
|
||
|
if not hasattr(cls, "recipes"):
|
||
|
cls.recipes = {}
|
||
|
if name in cls.recipes:
|
||
|
return cls.recipes[name]
|
||
|
|
||
|
recipe_file = None
|
||
|
for recipes_dir in cls.recipe_dirs(ctx):
|
||
|
if not exists(recipes_dir):
|
||
|
continue
|
||
|
# Find matching folder (may differ in case):
|
||
|
for subfolder in listdir(recipes_dir):
|
||
|
if subfolder.lower() == name:
|
||
|
recipe_file = join(recipes_dir, subfolder, '__init__.py')
|
||
|
if exists(recipe_file):
|
||
|
name = subfolder # adapt to actual spelling
|
||
|
break
|
||
|
recipe_file = None
|
||
|
if recipe_file is not None:
|
||
|
break
|
||
|
|
||
|
else:
|
||
|
raise ValueError('Recipe does not exist: {}'.format(name))
|
||
|
|
||
|
mod = import_recipe('pythonforandroid.recipes.{}'.format(name), recipe_file)
|
||
|
if len(logger.handlers) > 1:
|
||
|
logger.removeHandler(logger.handlers[1])
|
||
|
recipe = mod.recipe
|
||
|
recipe.ctx = ctx
|
||
|
cls.recipes[name.lower()] = recipe
|
||
|
return recipe
|
||
|
|
||
|
|
||
|
class IncludedFilesBehaviour(object):
|
||
|
'''Recipe mixin class that will automatically unpack files included in
|
||
|
the recipe directory.'''
|
||
|
src_filename = None
|
||
|
|
||
|
def prepare_build_dir(self, arch):
|
||
|
if self.src_filename is None:
|
||
|
raise BuildInterruptingException(
|
||
|
'IncludedFilesBehaviour failed: no src_filename specified')
|
||
|
rmdir(self.get_build_dir(arch))
|
||
|
shprint(sh.cp, '-a', join(self.get_recipe_dir(), self.src_filename),
|
||
|
self.get_build_dir(arch))
|
||
|
|
||
|
|
||
|
class BootstrapNDKRecipe(Recipe):
|
||
|
'''A recipe class for recipes built in an Android project jni dir with
|
||
|
an Android.mk. These are not cached separatly, but built in the
|
||
|
bootstrap's own building directory.
|
||
|
|
||
|
To build an NDK project which is not part of the bootstrap, see
|
||
|
:class:`~pythonforandroid.recipe.NDKRecipe`.
|
||
|
|
||
|
To link with python, call the method :meth:`get_recipe_env`
|
||
|
with the kwarg *with_python=True*.
|
||
|
'''
|
||
|
|
||
|
dir_name = None # The name of the recipe build folder in the jni dir
|
||
|
|
||
|
def get_build_container_dir(self, arch):
|
||
|
return self.get_jni_dir()
|
||
|
|
||
|
def get_build_dir(self, arch):
|
||
|
if self.dir_name is None:
|
||
|
raise ValueError('{} recipe doesn\'t define a dir_name, but '
|
||
|
'this is necessary'.format(self.name))
|
||
|
return join(self.get_build_container_dir(arch), self.dir_name)
|
||
|
|
||
|
def get_jni_dir(self):
|
||
|
return join(self.ctx.bootstrap.build_dir, 'jni')
|
||
|
|
||
|
def get_recipe_env(self, arch=None, with_flags_in_cc=True, with_python=False):
|
||
|
env = super().get_recipe_env(arch, with_flags_in_cc)
|
||
|
if not with_python:
|
||
|
return env
|
||
|
|
||
|
env['PYTHON_INCLUDE_ROOT'] = self.ctx.python_recipe.include_root(arch.arch)
|
||
|
env['PYTHON_LINK_ROOT'] = self.ctx.python_recipe.link_root(arch.arch)
|
||
|
env['EXTRA_LDLIBS'] = ' -lpython{}'.format(
|
||
|
self.ctx.python_recipe.link_version)
|
||
|
return env
|
||
|
|
||
|
|
||
|
class NDKRecipe(Recipe):
|
||
|
'''A recipe class for any NDK project not included in the bootstrap.'''
|
||
|
|
||
|
generated_libraries = []
|
||
|
|
||
|
def should_build(self, arch):
|
||
|
lib_dir = self.get_lib_dir(arch)
|
||
|
|
||
|
for lib in self.generated_libraries:
|
||
|
if not exists(join(lib_dir, lib)):
|
||
|
return True
|
||
|
|
||
|
return False
|
||
|
|
||
|
def get_lib_dir(self, arch):
|
||
|
return join(self.get_build_dir(arch.arch), 'obj', 'local', arch.arch)
|
||
|
|
||
|
def get_jni_dir(self, arch):
|
||
|
return join(self.get_build_dir(arch.arch), 'jni')
|
||
|
|
||
|
def build_arch(self, arch, *extra_args):
|
||
|
super().build_arch(arch)
|
||
|
|
||
|
env = self.get_recipe_env(arch)
|
||
|
with current_directory(self.get_build_dir(arch.arch)):
|
||
|
shprint(
|
||
|
sh.Command(join(self.ctx.ndk_dir, "ndk-build")),
|
||
|
'V=1',
|
||
|
'NDK_DEBUG=' + ("1" if self.ctx.build_as_debuggable else "0"),
|
||
|
'APP_PLATFORM=android-' + str(self.ctx.ndk_api),
|
||
|
'APP_ABI=' + arch.arch,
|
||
|
*extra_args, _env=env
|
||
|
)
|
||
|
|
||
|
|
||
|
class PythonRecipe(Recipe):
|
||
|
site_packages_name = None
|
||
|
'''The name of the module's folder when installed in the Python
|
||
|
site-packages (e.g. for pyjnius it is 'jnius')'''
|
||
|
|
||
|
call_hostpython_via_targetpython = True
|
||
|
'''If True, tries to install the module using the hostpython binary
|
||
|
copied to the target (normally arm) python build dir. However, this
|
||
|
will fail if the module tries to import e.g. _io.so. Set this to False
|
||
|
to call hostpython from its own build dir, installing the module in
|
||
|
the right place via arguments to setup.py. However, this may not set
|
||
|
the environment correctly and so False is not the default.'''
|
||
|
|
||
|
install_in_hostpython = False
|
||
|
'''If True, additionally installs the module in the hostpython build
|
||
|
dir. This will make it available to other recipes if
|
||
|
call_hostpython_via_targetpython is False.
|
||
|
'''
|
||
|
|
||
|
install_in_targetpython = True
|
||
|
'''If True, installs the module in the targetpython installation dir.
|
||
|
This is almost always what you want to do.'''
|
||
|
|
||
|
setup_extra_args = []
|
||
|
'''List of extra arguments to pass to setup.py'''
|
||
|
|
||
|
depends = ['python3']
|
||
|
'''
|
||
|
.. note:: it's important to keep this depends as a class attribute outside
|
||
|
`__init__` because sometimes we only initialize the class, so the
|
||
|
`__init__` call won't be called and the deps would be missing
|
||
|
(which breaks the dependency graph computation)
|
||
|
|
||
|
.. warning:: don't forget to call `super().__init__()` in any recipe's
|
||
|
`__init__`, or otherwise it may not be ensured that it depends
|
||
|
on python2 or python3 which can break the dependency graph
|
||
|
'''
|
||
|
|
||
|
hostpython_prerequisites = []
|
||
|
'''List of hostpython packages required to build a recipe'''
|
||
|
|
||
|
def __init__(self, *args, **kwargs):
|
||
|
super().__init__(*args, **kwargs)
|
||
|
if 'python3' not in self.depends:
|
||
|
# We ensure here that the recipe depends on python even it overrode
|
||
|
# `depends`. We only do this if it doesn't already depend on any
|
||
|
# python, since some recipes intentionally don't depend on/work
|
||
|
# with all python variants
|
||
|
depends = self.depends
|
||
|
depends.append('python3')
|
||
|
depends = list(set(depends))
|
||
|
self.depends = depends
|
||
|
|
||
|
def clean_build(self, arch=None):
|
||
|
super().clean_build(arch=arch)
|
||
|
name = self.folder_name
|
||
|
python_install_dirs = glob.glob(join(self.ctx.python_installs_dir, '*'))
|
||
|
for python_install in python_install_dirs:
|
||
|
site_packages_dir = glob.glob(join(python_install, 'lib', 'python*',
|
||
|
'site-packages'))
|
||
|
if site_packages_dir:
|
||
|
build_dir = join(site_packages_dir[0], name)
|
||
|
if exists(build_dir):
|
||
|
info('Deleted {}'.format(build_dir))
|
||
|
rmdir(build_dir)
|
||
|
|
||
|
@property
|
||
|
def real_hostpython_location(self):
|
||
|
host_name = 'host{}'.format(self.ctx.python_recipe.name)
|
||
|
if host_name == 'hostpython3':
|
||
|
python_recipe = Recipe.get_recipe(host_name, self.ctx)
|
||
|
return python_recipe.python_exe
|
||
|
else:
|
||
|
python_recipe = self.ctx.python_recipe
|
||
|
return 'python{}'.format(python_recipe.version)
|
||
|
|
||
|
@property
|
||
|
def hostpython_location(self):
|
||
|
if not self.call_hostpython_via_targetpython:
|
||
|
return self.real_hostpython_location
|
||
|
return self.ctx.hostpython
|
||
|
|
||
|
@property
|
||
|
def folder_name(self):
|
||
|
'''The name of the build folders containing this recipe.'''
|
||
|
name = self.site_packages_name
|
||
|
if name is None:
|
||
|
name = self.name
|
||
|
return name
|
||
|
|
||
|
def get_recipe_env(self, arch=None, with_flags_in_cc=True):
|
||
|
env = super().get_recipe_env(arch, with_flags_in_cc)
|
||
|
env['PYTHONNOUSERSITE'] = '1'
|
||
|
# Set the LANG, this isn't usually important but is a better default
|
||
|
# as it occasionally matters how Python e.g. reads files
|
||
|
env['LANG'] = "en_GB.UTF-8"
|
||
|
# Binaries made by packages installed by pip
|
||
|
env["PATH"] = join(self.hostpython_site_dir, "bin") + ":" + env["PATH"]
|
||
|
|
||
|
if not self.call_hostpython_via_targetpython:
|
||
|
env['CFLAGS'] += ' -I{}'.format(
|
||
|
self.ctx.python_recipe.include_root(arch.arch)
|
||
|
)
|
||
|
env['LDFLAGS'] += ' -L{} -lpython{}'.format(
|
||
|
self.ctx.python_recipe.link_root(arch.arch),
|
||
|
self.ctx.python_recipe.link_version,
|
||
|
)
|
||
|
|
||
|
hppath = []
|
||
|
hppath.append(join(dirname(self.hostpython_location), 'Lib'))
|
||
|
hppath.append(join(hppath[0], 'site-packages'))
|
||
|
builddir = join(dirname(self.hostpython_location), 'build')
|
||
|
if exists(builddir):
|
||
|
hppath += [join(builddir, d) for d in listdir(builddir)
|
||
|
if isdir(join(builddir, d))]
|
||
|
if len(hppath) > 0:
|
||
|
if 'PYTHONPATH' in env:
|
||
|
env['PYTHONPATH'] = ':'.join(hppath + [env['PYTHONPATH']])
|
||
|
else:
|
||
|
env['PYTHONPATH'] = ':'.join(hppath)
|
||
|
return env
|
||
|
|
||
|
def should_build(self, arch):
|
||
|
name = self.folder_name
|
||
|
if self.ctx.has_package(name, arch):
|
||
|
info('Python package already exists in site-packages')
|
||
|
return False
|
||
|
info('{} apparently isn\'t already in site-packages'.format(name))
|
||
|
return True
|
||
|
|
||
|
def build_arch(self, arch):
|
||
|
'''Install the Python module by calling setup.py install with
|
||
|
the target Python dir.'''
|
||
|
self.install_hostpython_prerequisites()
|
||
|
super().build_arch(arch)
|
||
|
self.install_python_package(arch)
|
||
|
|
||
|
def install_python_package(self, arch, name=None, env=None, is_dir=True):
|
||
|
'''Automate the installation of a Python package (or a cython
|
||
|
package where the cython components are pre-built).'''
|
||
|
# arch = self.filtered_archs[0] # old kivy-ios way
|
||
|
if name is None:
|
||
|
name = self.name
|
||
|
if env is None:
|
||
|
env = self.get_recipe_env(arch)
|
||
|
|
||
|
info('Installing {} into site-packages'.format(self.name))
|
||
|
|
||
|
hostpython = sh.Command(self.hostpython_location)
|
||
|
hpenv = env.copy()
|
||
|
with current_directory(self.get_build_dir(arch.arch)):
|
||
|
shprint(hostpython, 'setup.py', 'install', '-O2',
|
||
|
'--root={}'.format(self.ctx.get_python_install_dir(arch.arch)),
|
||
|
'--install-lib=.',
|
||
|
_env=hpenv, *self.setup_extra_args)
|
||
|
|
||
|
# If asked, also install in the hostpython build dir
|
||
|
if self.install_in_hostpython:
|
||
|
self.install_hostpython_package(arch)
|
||
|
|
||
|
def get_hostrecipe_env(self, arch):
|
||
|
env = environ.copy()
|
||
|
env['PYTHONPATH'] = self.hostpython_site_dir
|
||
|
return env
|
||
|
|
||
|
@property
|
||
|
def hostpython_site_dir(self):
|
||
|
return join(dirname(self.real_hostpython_location), 'Lib', 'site-packages')
|
||
|
|
||
|
def install_hostpython_package(self, arch):
|
||
|
env = self.get_hostrecipe_env(arch)
|
||
|
real_hostpython = sh.Command(self.real_hostpython_location)
|
||
|
shprint(real_hostpython, 'setup.py', 'install', '-O2',
|
||
|
'--root={}'.format(dirname(self.real_hostpython_location)),
|
||
|
'--install-lib=Lib/site-packages',
|
||
|
_env=env, *self.setup_extra_args)
|
||
|
|
||
|
@property
|
||
|
def python_major_minor_version(self):
|
||
|
parsed_version = packaging.version.parse(self.ctx.python_recipe.version)
|
||
|
return f"{parsed_version.major}.{parsed_version.minor}"
|
||
|
|
||
|
def install_hostpython_prerequisites(self, packages=None, force_upgrade=True):
|
||
|
if not packages:
|
||
|
packages = self.hostpython_prerequisites
|
||
|
|
||
|
if len(packages) == 0:
|
||
|
return
|
||
|
|
||
|
pip_options = [
|
||
|
"install",
|
||
|
*packages,
|
||
|
"--target", self.hostpython_site_dir, "--python-version",
|
||
|
self.ctx.python_recipe.version,
|
||
|
# Don't use sources, instead wheels
|
||
|
"--only-binary=:all:",
|
||
|
]
|
||
|
if force_upgrade:
|
||
|
pip_options.append("--upgrade")
|
||
|
# Use system's pip
|
||
|
shprint(sh.pip, *pip_options)
|
||
|
|
||
|
def restore_hostpython_prerequisites(self, packages):
|
||
|
_packages = []
|
||
|
for package in packages:
|
||
|
original_version = Recipe.get_recipe(package, self.ctx).version
|
||
|
_packages.append(package + "==" + original_version)
|
||
|
self.install_hostpython_prerequisites(packages=_packages)
|
||
|
|
||
|
|
||
|
class CompiledComponentsPythonRecipe(PythonRecipe):
|
||
|
pre_build_ext = False
|
||
|
|
||
|
build_cmd = 'build_ext'
|
||
|
|
||
|
def build_arch(self, arch):
|
||
|
'''Build any cython components, then install the Python module by
|
||
|
calling setup.py install with the target Python dir.
|
||
|
'''
|
||
|
Recipe.build_arch(self, arch)
|
||
|
self.install_hostpython_prerequisites()
|
||
|
self.build_compiled_components(arch)
|
||
|
self.install_python_package(arch)
|
||
|
|
||
|
def build_compiled_components(self, arch):
|
||
|
info('Building compiled components in {}'.format(self.name))
|
||
|
|
||
|
env = self.get_recipe_env(arch)
|
||
|
hostpython = sh.Command(self.hostpython_location)
|
||
|
with current_directory(self.get_build_dir(arch.arch)):
|
||
|
if self.install_in_hostpython:
|
||
|
shprint(hostpython, 'setup.py', 'clean', '--all', _env=env)
|
||
|
shprint(hostpython, 'setup.py', self.build_cmd, '-v',
|
||
|
_env=env, *self.setup_extra_args)
|
||
|
build_dir = glob.glob('build/lib.*')[0]
|
||
|
shprint(sh.find, build_dir, '-name', '"*.o"', '-exec',
|
||
|
env['STRIP'], '{}', ';', _env=env)
|
||
|
|
||
|
def install_hostpython_package(self, arch):
|
||
|
env = self.get_hostrecipe_env(arch)
|
||
|
self.rebuild_compiled_components(arch, env)
|
||
|
super().install_hostpython_package(arch)
|
||
|
|
||
|
def rebuild_compiled_components(self, arch, env):
|
||
|
info('Rebuilding compiled components in {}'.format(self.name))
|
||
|
|
||
|
hostpython = sh.Command(self.real_hostpython_location)
|
||
|
shprint(hostpython, 'setup.py', 'clean', '--all', _env=env)
|
||
|
shprint(hostpython, 'setup.py', self.build_cmd, '-v', _env=env,
|
||
|
*self.setup_extra_args)
|
||
|
|
||
|
|
||
|
class CppCompiledComponentsPythonRecipe(CompiledComponentsPythonRecipe):
|
||
|
""" Extensions that require the cxx-stl """
|
||
|
call_hostpython_via_targetpython = False
|
||
|
need_stl_shared = True
|
||
|
|
||
|
|
||
|
class CythonRecipe(PythonRecipe):
|
||
|
pre_build_ext = False
|
||
|
cythonize = True
|
||
|
cython_args = []
|
||
|
call_hostpython_via_targetpython = False
|
||
|
|
||
|
def build_arch(self, arch):
|
||
|
'''Build any cython components, then install the Python module by
|
||
|
calling setup.py install with the target Python dir.
|
||
|
'''
|
||
|
Recipe.build_arch(self, arch)
|
||
|
self.build_cython_components(arch)
|
||
|
self.install_python_package(arch)
|
||
|
|
||
|
def build_cython_components(self, arch):
|
||
|
info('Cythonizing anything necessary in {}'.format(self.name))
|
||
|
|
||
|
env = self.get_recipe_env(arch)
|
||
|
|
||
|
with current_directory(self.get_build_dir(arch.arch)):
|
||
|
hostpython = sh.Command(self.ctx.hostpython)
|
||
|
shprint(hostpython, '-c', 'import sys; print(sys.path)', _env=env)
|
||
|
debug('cwd is {}'.format(realpath(curdir)))
|
||
|
info('Trying first build of {} to get cython files: this is '
|
||
|
'expected to fail'.format(self.name))
|
||
|
|
||
|
manually_cythonise = False
|
||
|
try:
|
||
|
shprint(hostpython, 'setup.py', 'build_ext', '-v', _env=env,
|
||
|
*self.setup_extra_args)
|
||
|
except sh.ErrorReturnCode_1:
|
||
|
print()
|
||
|
info('{} first build failed (as expected)'.format(self.name))
|
||
|
manually_cythonise = True
|
||
|
|
||
|
if manually_cythonise:
|
||
|
self.cythonize_build(env=env)
|
||
|
shprint(hostpython, 'setup.py', 'build_ext', '-v', _env=env,
|
||
|
_tail=20, _critical=True, *self.setup_extra_args)
|
||
|
else:
|
||
|
info('First build appeared to complete correctly, skipping manual'
|
||
|
'cythonising.')
|
||
|
|
||
|
if not self.ctx.with_debug_symbols:
|
||
|
self.strip_object_files(arch, env)
|
||
|
|
||
|
def strip_object_files(self, arch, env, build_dir=None):
|
||
|
if build_dir is None:
|
||
|
build_dir = self.get_build_dir(arch.arch)
|
||
|
with current_directory(build_dir):
|
||
|
info('Stripping object files')
|
||
|
shprint(sh.find, '.', '-iname', '*.so', '-exec',
|
||
|
'/usr/bin/echo', '{}', ';', _env=env)
|
||
|
shprint(sh.find, '.', '-iname', '*.so', '-exec',
|
||
|
env['STRIP'].split(' ')[0], '--strip-unneeded',
|
||
|
# '/usr/bin/strip', '--strip-unneeded',
|
||
|
'{}', ';', _env=env)
|
||
|
|
||
|
def cythonize_file(self, env, build_dir, filename):
|
||
|
short_filename = filename
|
||
|
if filename.startswith(build_dir):
|
||
|
short_filename = filename[len(build_dir) + 1:]
|
||
|
info(u"Cythonize {}".format(short_filename))
|
||
|
cyenv = env.copy()
|
||
|
if 'CYTHONPATH' in cyenv:
|
||
|
cyenv['PYTHONPATH'] = cyenv['CYTHONPATH']
|
||
|
elif 'PYTHONPATH' in cyenv:
|
||
|
del cyenv['PYTHONPATH']
|
||
|
if 'PYTHONNOUSERSITE' in cyenv:
|
||
|
cyenv.pop('PYTHONNOUSERSITE')
|
||
|
python_command = sh.Command("python{}".format(
|
||
|
self.ctx.python_recipe.major_minor_version_string.split(".")[0]
|
||
|
))
|
||
|
shprint(python_command, "-c"
|
||
|
"import sys; from Cython.Compiler.Main import setuptools_main; sys.exit(setuptools_main());",
|
||
|
filename, *self.cython_args, _env=cyenv)
|
||
|
|
||
|
def cythonize_build(self, env, build_dir="."):
|
||
|
if not self.cythonize:
|
||
|
info('Running cython cancelled per recipe setting')
|
||
|
return
|
||
|
info('Running cython where appropriate')
|
||
|
for root, dirnames, filenames in walk("."):
|
||
|
for filename in fnmatch.filter(filenames, "*.pyx"):
|
||
|
self.cythonize_file(env, build_dir, join(root, filename))
|
||
|
|
||
|
def get_recipe_env(self, arch, with_flags_in_cc=True):
|
||
|
env = super().get_recipe_env(arch, with_flags_in_cc)
|
||
|
env['LDFLAGS'] = env['LDFLAGS'] + ' -L{} '.format(
|
||
|
self.ctx.get_libs_dir(arch.arch) +
|
||
|
' -L{} '.format(self.ctx.libs_dir) +
|
||
|
' -L{}'.format(join(self.ctx.bootstrap.build_dir, 'obj', 'local',
|
||
|
arch.arch)))
|
||
|
|
||
|
env['LDSHARED'] = env['CC'] + ' -shared'
|
||
|
# shprint(sh.whereis, env['LDSHARED'], _env=env)
|
||
|
env['LIBLINK'] = 'NOTNONE'
|
||
|
if self.ctx.copy_libs:
|
||
|
env['COPYLIBS'] = '1'
|
||
|
|
||
|
# Every recipe uses its own liblink path, object files are
|
||
|
# collected and biglinked later
|
||
|
liblink_path = join(self.get_build_container_dir(arch.arch),
|
||
|
'objects_{}'.format(self.name))
|
||
|
env['LIBLINK_PATH'] = liblink_path
|
||
|
ensure_dir(liblink_path)
|
||
|
|
||
|
return env
|
||
|
|
||
|
|
||
|
class PyProjectRecipe(PythonRecipe):
|
||
|
'''Recipe for projects which containes `pyproject.toml`'''
|
||
|
|
||
|
# Extra args to pass to `python -m build ...`
|
||
|
extra_build_args = []
|
||
|
call_hostpython_via_targetpython = False
|
||
|
|
||
|
def get_recipe_env(self, arch, **kwargs):
|
||
|
# Custom hostpython
|
||
|
self.ctx.python_recipe.python_exe = join(
|
||
|
self.ctx.python_recipe.get_build_dir(arch), "android-build", "python3")
|
||
|
env = super().get_recipe_env(arch, **kwargs)
|
||
|
build_dir = self.get_build_dir(arch)
|
||
|
ensure_dir(build_dir)
|
||
|
build_opts = join(build_dir, "build-opts.cfg")
|
||
|
|
||
|
with open(build_opts, "w") as file:
|
||
|
file.write("[bdist_wheel]\nplat-name={}".format(
|
||
|
self.get_wheel_platform_tag(arch)
|
||
|
))
|
||
|
file.close()
|
||
|
|
||
|
env["DIST_EXTRA_CONFIG"] = build_opts
|
||
|
return env
|
||
|
|
||
|
def get_wheel_platform_tag(self, arch):
|
||
|
return "android_" + {
|
||
|
"armeabi-v7a": "arm",
|
||
|
"arm64-v8a": "aarch64",
|
||
|
"x86_64": "x86_64",
|
||
|
"x86": "i686",
|
||
|
}[arch.arch]
|
||
|
|
||
|
def install_wheel(self, arch, built_wheels):
|
||
|
_wheel = built_wheels[0]
|
||
|
built_wheel_dir = dirname(_wheel)
|
||
|
# Fix wheel platform tag
|
||
|
wheel_tag = wheel_tags(
|
||
|
_wheel,
|
||
|
platform_tags=self.get_wheel_platform_tag(arch),
|
||
|
remove=True,
|
||
|
)
|
||
|
selected_wheel = join(built_wheel_dir, wheel_tag)
|
||
|
|
||
|
_dev_wheel_dir = environ.get("P4A_WHEEL_DIR", False)
|
||
|
if _dev_wheel_dir:
|
||
|
ensure_dir(_dev_wheel_dir)
|
||
|
shprint(sh.cp, selected_wheel, _dev_wheel_dir)
|
||
|
|
||
|
info(f"Installing built wheel: {wheel_tag}")
|
||
|
destination = self.ctx.get_python_install_dir(arch.arch)
|
||
|
with WheelFile(selected_wheel) as wf:
|
||
|
for zinfo in wf.filelist:
|
||
|
wf.extract(zinfo, destination)
|
||
|
wf.close()
|
||
|
|
||
|
def build_arch(self, arch):
|
||
|
self.install_hostpython_prerequisites(
|
||
|
packages=["build[virtualenv]", "pip"] + self.hostpython_prerequisites
|
||
|
)
|
||
|
build_dir = self.get_build_dir(arch.arch)
|
||
|
env = self.get_recipe_env(arch, with_flags_in_cc=True)
|
||
|
# make build dir separatly
|
||
|
sub_build_dir = join(build_dir, "p4a_android_build")
|
||
|
ensure_dir(sub_build_dir)
|
||
|
# copy hostpython to built python to ensure correct selection of libs and includes
|
||
|
shprint(sh.cp, self.real_hostpython_location, self.ctx.python_recipe.python_exe)
|
||
|
|
||
|
build_args = [
|
||
|
"-m",
|
||
|
"build",
|
||
|
"--wheel",
|
||
|
"--config-setting",
|
||
|
"builddir={}".format(sub_build_dir),
|
||
|
] + self.extra_build_args
|
||
|
|
||
|
built_wheels = []
|
||
|
with current_directory(build_dir):
|
||
|
shprint(
|
||
|
sh.Command(self.ctx.python_recipe.python_exe), *build_args, _env=env
|
||
|
)
|
||
|
built_wheels = [realpath(whl) for whl in glob.glob("dist/*.whl")]
|
||
|
self.install_wheel(arch, built_wheels)
|
||
|
|
||
|
|
||
|
class MesonRecipe(PyProjectRecipe):
|
||
|
'''Recipe for projects which uses meson as build system'''
|
||
|
|
||
|
meson_version = "1.4.0"
|
||
|
ninja_version = "1.11.1.1"
|
||
|
|
||
|
def sanitize_flags(self, *flag_strings):
|
||
|
return " ".join(flag_strings).strip().split(" ")
|
||
|
|
||
|
def get_recipe_meson_options(self, arch):
|
||
|
env = self.get_recipe_env(arch, with_flags_in_cc=True)
|
||
|
return {
|
||
|
"binaries": {
|
||
|
"c": arch.get_clang_exe(with_target=True),
|
||
|
"cpp": arch.get_clang_exe(with_target=True, plus_plus=True),
|
||
|
"ar": self.ctx.ndk.llvm_ar,
|
||
|
"strip": self.ctx.ndk.llvm_strip,
|
||
|
},
|
||
|
"built-in options": {
|
||
|
"c_args": self.sanitize_flags(env["CFLAGS"], env["CPPFLAGS"]),
|
||
|
"cpp_args": self.sanitize_flags(env["CXXFLAGS"], env["CPPFLAGS"]),
|
||
|
"c_link_args": self.sanitize_flags(env["LDFLAGS"]),
|
||
|
"cpp_link_args": self.sanitize_flags(env["LDFLAGS"]),
|
||
|
},
|
||
|
"properties": {
|
||
|
"needs_exe_wrapper": True,
|
||
|
"sys_root": self.ctx.ndk.sysroot
|
||
|
},
|
||
|
"host_machine": {
|
||
|
"cpu_family": {
|
||
|
"arm64-v8a": "aarch64",
|
||
|
"armeabi-v7a": "arm",
|
||
|
"x86_64": "x86_64",
|
||
|
"x86": "x86"
|
||
|
}[arch.arch],
|
||
|
"cpu": {
|
||
|
"arm64-v8a": "aarch64",
|
||
|
"armeabi-v7a": "armv7",
|
||
|
"x86_64": "x86_64",
|
||
|
"x86": "i686"
|
||
|
}[arch.arch],
|
||
|
"endian": "little",
|
||
|
"system": "android",
|
||
|
}
|
||
|
}
|
||
|
|
||
|
def write_build_options(self, arch):
|
||
|
"""Writes python dict to meson config file"""
|
||
|
option_data = ""
|
||
|
build_options = self.get_recipe_meson_options(arch)
|
||
|
for key in build_options.keys():
|
||
|
data_chunk = "[{}]".format(key)
|
||
|
for subkey in build_options[key].keys():
|
||
|
value = build_options[key][subkey]
|
||
|
if isinstance(value, int):
|
||
|
value = str(value)
|
||
|
elif isinstance(value, str):
|
||
|
value = "'{}'".format(value)
|
||
|
elif isinstance(value, bool):
|
||
|
value = "true" if value else "false"
|
||
|
elif isinstance(value, list):
|
||
|
value = "['" + "', '".join(value) + "']"
|
||
|
data_chunk += "\n" + subkey + " = " + value
|
||
|
option_data += data_chunk + "\n\n"
|
||
|
return option_data
|
||
|
|
||
|
def ensure_args(self, *args):
|
||
|
for arg in args:
|
||
|
if arg not in self.extra_build_args:
|
||
|
self.extra_build_args.append(arg)
|
||
|
|
||
|
def build_arch(self, arch):
|
||
|
cross_file = join("/tmp", "android.meson.cross")
|
||
|
info("Writing cross file at: {}".format(cross_file))
|
||
|
# write cross config file
|
||
|
with open(cross_file, "w") as file:
|
||
|
file.write(self.write_build_options(arch))
|
||
|
file.close()
|
||
|
# set cross file
|
||
|
self.ensure_args('-Csetup-args=--cross-file', '-Csetup-args={}'.format(cross_file))
|
||
|
# ensure ninja and meson
|
||
|
for dep in [
|
||
|
"ninja=={}".format(self.ninja_version),
|
||
|
"meson=={}".format(self.meson_version),
|
||
|
]:
|
||
|
if dep not in self.hostpython_prerequisites:
|
||
|
self.hostpython_prerequisites.append(dep)
|
||
|
super().build_arch(arch)
|
||
|
|
||
|
|
||
|
class RustCompiledComponentsRecipe(PyProjectRecipe):
|
||
|
# Rust toolchain codes
|
||
|
# https://doc.rust-lang.org/nightly/rustc/platform-support.html
|
||
|
RUST_ARCH_CODES = {
|
||
|
"arm64-v8a": "aarch64-linux-android",
|
||
|
"armeabi-v7a": "armv7-linux-androideabi",
|
||
|
"x86_64": "x86_64-linux-android",
|
||
|
"x86": "i686-linux-android",
|
||
|
}
|
||
|
|
||
|
call_hostpython_via_targetpython = False
|
||
|
|
||
|
def get_recipe_env(self, arch, **kwargs):
|
||
|
env = super().get_recipe_env(arch, **kwargs)
|
||
|
|
||
|
# Set rust build target
|
||
|
build_target = self.RUST_ARCH_CODES[arch.arch]
|
||
|
cargo_linker_name = "CARGO_TARGET_{}_LINKER".format(
|
||
|
build_target.upper().replace("-", "_")
|
||
|
)
|
||
|
env["CARGO_BUILD_TARGET"] = build_target
|
||
|
env[cargo_linker_name] = join(
|
||
|
self.ctx.ndk.llvm_prebuilt_dir,
|
||
|
"bin",
|
||
|
"{}{}-clang".format(
|
||
|
# NDK's Clang format
|
||
|
build_target.replace("7", "7a")
|
||
|
if build_target.startswith("armv7")
|
||
|
else build_target,
|
||
|
self.ctx.ndk_api,
|
||
|
),
|
||
|
)
|
||
|
realpython_dir = self.ctx.python_recipe.get_build_dir(arch.arch)
|
||
|
|
||
|
env["RUSTFLAGS"] = "-Clink-args=-L{} -L{}".format(
|
||
|
self.ctx.get_libs_dir(arch.arch), join(realpython_dir, "android-build")
|
||
|
)
|
||
|
|
||
|
env["PYO3_CROSS_LIB_DIR"] = realpath(glob.glob(join(
|
||
|
realpython_dir, "android-build", "build",
|
||
|
"lib.linux-*-{}/".format(self.python_major_minor_version),
|
||
|
))[0])
|
||
|
|
||
|
info_main("Ensuring rust build toolchain")
|
||
|
shprint(sh.rustup, "target", "add", build_target)
|
||
|
|
||
|
# Add host python to PATH
|
||
|
env["PATH"] = ("{hostpython_dir}:{old_path}").format(
|
||
|
hostpython_dir=Recipe.get_recipe(
|
||
|
"hostpython3", self.ctx
|
||
|
).get_path_to_python(),
|
||
|
old_path=env["PATH"],
|
||
|
)
|
||
|
return env
|
||
|
|
||
|
def check_host_deps(self):
|
||
|
if not hasattr(sh, "rustup"):
|
||
|
error(
|
||
|
"`rustup` was not found on host system."
|
||
|
"Please install it using :"
|
||
|
"\n`curl https://sh.rustup.rs -sSf | sh`\n"
|
||
|
)
|
||
|
exit(1)
|
||
|
|
||
|
def build_arch(self, arch):
|
||
|
self.check_host_deps()
|
||
|
super().build_arch(arch)
|
||
|
|
||
|
|
||
|
class TargetPythonRecipe(Recipe):
|
||
|
'''Class for target python recipes. Sets ctx.python_recipe to point to
|
||
|
itself, so as to know later what kind of Python was built or used.'''
|
||
|
|
||
|
def __init__(self, *args, **kwargs):
|
||
|
self._ctx = None
|
||
|
super().__init__(*args, **kwargs)
|
||
|
|
||
|
def prebuild_arch(self, arch):
|
||
|
super().prebuild_arch(arch)
|
||
|
self.ctx.python_recipe = self
|
||
|
|
||
|
def include_root(self, arch):
|
||
|
'''The root directory from which to include headers.'''
|
||
|
raise NotImplementedError('Not implemented in TargetPythonRecipe')
|
||
|
|
||
|
def link_root(self):
|
||
|
raise NotImplementedError('Not implemented in TargetPythonRecipe')
|
||
|
|
||
|
@property
|
||
|
def major_minor_version_string(self):
|
||
|
parsed_version = packaging.version.parse(self.version)
|
||
|
return f"{parsed_version.major}.{parsed_version.minor}"
|
||
|
|
||
|
def create_python_bundle(self, dirn, arch):
|
||
|
"""
|
||
|
Create a packaged python bundle in the target directory, by
|
||
|
copying all the modules and standard library to the right
|
||
|
place.
|
||
|
"""
|
||
|
raise NotImplementedError('{} does not implement create_python_bundle'.format(self))
|
||
|
|
||
|
def reduce_object_file_names(self, dirn):
|
||
|
"""Recursively renames all files named XXX.cpython-...-linux-gnu.so"
|
||
|
to "XXX.so", i.e. removing the erroneous architecture name
|
||
|
coming from the local system.
|
||
|
"""
|
||
|
py_so_files = shprint(sh.find, dirn, '-iname', '*.so')
|
||
|
filens = py_so_files.stdout.decode('utf-8').split('\n')[:-1]
|
||
|
for filen in filens:
|
||
|
file_dirname, file_basename = split(filen)
|
||
|
parts = file_basename.split('.')
|
||
|
if len(parts) <= 2:
|
||
|
continue
|
||
|
# PySide6 libraries end with .abi3.so
|
||
|
if parts[1] == "abi3":
|
||
|
continue
|
||
|
move(filen, join(file_dirname, parts[0] + '.so'))
|
||
|
|
||
|
|
||
|
def algsum(alg, filen):
|
||
|
'''Calculate the digest of a file.
|
||
|
'''
|
||
|
with open(filen, 'rb') as fileh:
|
||
|
digest = getattr(hashlib, alg)(fileh.read())
|
||
|
|
||
|
return digest.hexdigest()
|
||
|
|
||
|
|
||
|
class FFPyPlayerRecipe(PyProjectRecipe):
|
||
|
version = 'v4.5.1'
|
||
|
url = 'https://github.com/matham/ffpyplayer/archive/{version}.zip'
|
||
|
depends = ['python3', 'sdl2', 'ffmpeg']
|
||
|
patches = ["setup.py.patch"]
|
||
|
opt_depends = ['openssl', 'ffpyplayer_codecs']
|
||
|
|
||
|
def get_recipe_env(self, arch, with_flags_in_cc=True):
|
||
|
env = super().get_recipe_env(arch)
|
||
|
|
||
|
build_dir = Recipe.get_recipe('ffmpeg', self.ctx).get_build_dir(arch.arch)
|
||
|
env["FFMPEG_INCLUDE_DIR"] = join(build_dir, "include")
|
||
|
env["FFMPEG_LIB_DIR"] = join(build_dir, "lib")
|
||
|
|
||
|
env["SDL_INCLUDE_DIR"] = join(self.ctx.bootstrap.build_dir, 'jni', 'SDL', 'include')
|
||
|
env["SDL_LIB_DIR"] = join(self.ctx.bootstrap.build_dir, 'libs', arch.arch)
|
||
|
|
||
|
env["USE_SDL2_MIXER"] = '1'
|
||
|
|
||
|
# ffpyplayer does not allow to pass more than one include dir for sdl2_mixer (and ATM is
|
||
|
# not needed), so we only pass the first one.
|
||
|
sdl2_mixer_recipe = self.get_recipe('sdl2_mixer', self.ctx)
|
||
|
env["SDL2_MIXER_INCLUDE_DIR"] = sdl2_mixer_recipe.get_include_dirs(arch)[0]
|
||
|
|
||
|
# NDKPLATFORM and LIBLINK are our switches for detecting Android platform, so can't be empty
|
||
|
# FIXME: We may want to introduce a cleaner approach to this?
|
||
|
env['NDKPLATFORM'] = "NOTNONE"
|
||
|
env['LIBLINK'] = 'NOTNONE'
|
||
|
|
||
|
# ffmpeg recipe enables GPL components only if ffpyplayer_codecs recipe used.
|
||
|
# Therefor we need to disable libpostproc if skipped.
|
||
|
if 'ffpyplayer_codecs' not in self.ctx.recipe_build_order:
|
||
|
env["CONFIG_POSTPROC"] = '0'
|
||
|
|
||
|
return env
|
||
|
|
||
|
|
||
|
recipe = FFPyPlayerRecipe()
|