Sideband/recipes/ffpyplayer/__init__.py

1516 lines
58 KiB
Python

from os.path import join
from os.path import basename, dirname, exists, isdir, isfile, join, realpath, split
import glob
import hashlib
from re import match
import sh
import shutil
import fnmatch
import zipfile
import urllib.request
from urllib.request import urlretrieve
from os import listdir, unlink, environ, curdir, walk
from sys import stdout
from wheel.wheelfile import WheelFile
from wheel.cli.tags import tags as wheel_tags
import time
try:
from urlparse import urlparse
except ImportError:
from urllib.parse import urlparse
import packaging.version
from pythonforandroid.logger import (
logger, info, warning, debug, shprint, info_main, error)
from pythonforandroid.util import (
current_directory, ensure_dir, BuildInterruptingException, rmdir, move,
touch)
from pythonforandroid.util import load_source as import_recipe
url_opener = urllib.request.build_opener()
url_orig_headers = url_opener.addheaders
urllib.request.install_opener(url_opener)
class RecipeMeta(type):
def __new__(cls, name, bases, dct):
if name != 'Recipe':
if 'url' in dct:
dct['_url'] = dct.pop('url')
if 'version' in dct:
dct['_version'] = dct.pop('version')
return super().__new__(cls, name, bases, dct)
class Recipe(metaclass=RecipeMeta):
_url = None
'''The address from which the recipe may be downloaded. This is not
essential, it may be omitted if the source is available some other
way, such as via the :class:`IncludedFilesBehaviour` mixin.
If the url includes the version, you may (and probably should)
replace this with ``{version}``, which will automatically be
replaced by the :attr:`version` string during download.
.. note:: Methods marked (internal) are used internally and you
probably don't need to call them, but they are available
if you want.
'''
_version = None
'''A string giving the version of the software the recipe describes,
e.g. ``2.0.3`` or ``master``.'''
md5sum = None
'''The md5sum of the source from the :attr:`url`. Non-essential, but
you should try to include this, it is used to check that the download
finished correctly.
'''
sha512sum = None
'''The sha512sum of the source from the :attr:`url`. Non-essential, but
you should try to include this, it is used to check that the download
finished correctly.
'''
blake2bsum = None
'''The blake2bsum of the source from the :attr:`url`. Non-essential, but
you should try to include this, it is used to check that the download
finished correctly.
'''
depends = []
'''A list containing the names of any recipes that this recipe depends on.
'''
conflicts = []
'''A list containing the names of any recipes that are known to be
incompatible with this one.'''
opt_depends = []
'''A list of optional dependencies, that must be built before this
recipe if they are built at all, but whose presence is not essential.'''
patches = []
'''A list of patches to apply to the source. Values can be either a string
referring to the patch file relative to the recipe dir, or a tuple of the
string patch file and a callable, which will receive the kwargs `arch` and
`recipe`, which should return True if the patch should be applied.'''
python_depends = []
'''A list of pure-Python packages that this package requires. These
packages will NOT be available at build time, but will be added to the
list of pure-Python packages to install via pip. If you need these packages
at build time, you must create a recipe.'''
archs = ['armeabi'] # Not currently implemented properly
built_libraries = {}
"""Each recipe that builds a system library (e.g.:libffi, openssl, etc...)
should contain a dict holding the relevant information of the library. The
keys should be the generated libraries and the values the relative path of
the library inside his build folder. This dict will be used to perform
different operations:
- copy the library into the right location, depending on if it's shared
or static)
- check if we have to rebuild the library
Here an example of how it would look like for `libffi` recipe:
- `built_libraries = {'libffi.so': '.libs'}`
.. note:: in case that the built library resides in recipe's build
directory, you can set the following values for the relative
path: `'.', None or ''`
"""
need_stl_shared = False
'''Some libraries or python packages may need the c++_shared in APK.
We can automatically do this for any recipe if we set this property to
`True`'''
stl_lib_name = 'c++_shared'
'''
The default STL shared lib to use: `c++_shared`.
.. note:: Android NDK version > 17 only supports 'c++_shared', because
starting from NDK r18 the `gnustl_shared` lib has been deprecated.
'''
def get_stl_library(self, arch):
return join(
arch.ndk_lib_dir,
'lib{name}.so'.format(name=self.stl_lib_name),
)
def install_stl_lib(self, arch):
if not self.ctx.has_lib(
arch.arch, 'lib{name}.so'.format(name=self.stl_lib_name)
):
self.install_libs(arch, self.get_stl_library(arch))
@property
def version(self):
key = 'VERSION_' + self.name
return environ.get(key, self._version)
@property
def url(self):
key = 'URL_' + self.name
return environ.get(key, self._url)
@property
def versioned_url(self):
'''A property returning the url of the recipe with ``{version}``
replaced by the :attr:`url`. If accessing the url, you should use this
property, *not* access the url directly.'''
if self.url is None:
return None
return self.url.format(version=self.version)
def download_file(self, url, target, cwd=None):
"""
(internal) Download an ``url`` to a ``target``.
"""
if not url:
return
info('Downloading {} from {}'.format(self.name, url))
if cwd:
target = join(cwd, target)
parsed_url = urlparse(url)
if parsed_url.scheme in ('http', 'https'):
def report_hook(index, blksize, size):
if size <= 0:
progression = '{0} bytes'.format(index * blksize)
else:
progression = '{0:.2f}%'.format(
index * blksize * 100. / float(size))
if "CI" not in environ:
stdout.write('- Download {}\r'.format(progression))
stdout.flush()
if exists(target):
unlink(target)
# Download item with multiple attempts (for bad connections):
attempts = 0
seconds = 1
while True:
try:
# jqueryui.com returns a 403 w/ the default user agent
# Mozilla/5.0 doesnt handle redirection for liblzma
url_opener.addheaders = [('User-agent', 'Wget/1.0')]
urlretrieve(url, target, report_hook)
except OSError as e:
attempts += 1
if attempts >= 5:
raise
stdout.write('Download failed: {}; retrying in {} second(s)...'.format(e, seconds))
time.sleep(seconds)
seconds *= 2
continue
finally:
url_opener.addheaders = url_orig_headers
break
return target
elif parsed_url.scheme in ('git', 'git+file', 'git+ssh', 'git+http', 'git+https'):
if not isdir(target):
if url.startswith('git+'):
url = url[4:]
# if 'version' is specified, do a shallow clone
if self.version:
ensure_dir(target)
with current_directory(target):
shprint(sh.git, 'init')
shprint(sh.git, 'remote', 'add', 'origin', url)
else:
shprint(sh.git, 'clone', '--recursive', url, target)
with current_directory(target):
if self.version:
shprint(sh.git, 'fetch', '--tags', '--depth', '1')
shprint(sh.git, 'checkout', self.version)
branch = sh.git('branch', '--show-current')
if branch:
shprint(sh.git, 'pull')
shprint(sh.git, 'pull', '--recurse-submodules')
shprint(sh.git, 'submodule', 'update', '--recursive', '--init', '--depth', '1')
return target
def apply_patch(self, filename, arch, build_dir=None):
"""
Apply a patch from the current recipe directory into the current
build directory.
.. versionchanged:: 0.6.0
Add ability to apply patch from any dir via kwarg `build_dir`'''
"""
info("Applying patch {}".format(filename))
build_dir = build_dir if build_dir else self.get_build_dir(arch)
filename = join(self.get_recipe_dir(), filename)
shprint(sh.patch, "-t", "-d", build_dir, "-p1",
"-i", filename, _tail=10)
def copy_file(self, filename, dest):
info("Copy {} to {}".format(filename, dest))
filename = join(self.get_recipe_dir(), filename)
dest = join(self.build_dir, dest)
shutil.copy(filename, dest)
def append_file(self, filename, dest):
info("Append {} to {}".format(filename, dest))
filename = join(self.get_recipe_dir(), filename)
dest = join(self.build_dir, dest)
with open(filename, "rb") as fd:
data = fd.read()
with open(dest, "ab") as fd:
fd.write(data)
@property
def name(self):
'''The name of the recipe, the same as the folder containing it.'''
modname = self.__class__.__module__
return modname.split(".", 2)[-1]
@property
def filtered_archs(self):
'''Return archs of self.ctx that are valid build archs
for the Recipe.'''
result = []
for arch in self.ctx.archs:
if not self.archs or (arch.arch in self.archs):
result.append(arch)
return result
def check_recipe_choices(self):
'''Checks what recipes are being built to see which of the alternative
and optional dependencies are being used,
and returns a list of these.'''
recipes = []
built_recipes = self.ctx.recipe_build_order
for recipe in self.depends:
if isinstance(recipe, (tuple, list)):
for alternative in recipe:
if alternative in built_recipes:
recipes.append(alternative)
break
for recipe in self.opt_depends:
if recipe in built_recipes:
recipes.append(recipe)
return sorted(recipes)
def get_opt_depends_in_list(self, recipes):
'''Given a list of recipe names, returns those that are also in
self.opt_depends.
'''
return [recipe for recipe in recipes if recipe in self.opt_depends]
def get_build_container_dir(self, arch):
'''Given the arch name, returns the directory where it will be
built.
This returns a different directory depending on what
alternative or optional dependencies are being built.
'''
dir_name = self.get_dir_name()
return join(self.ctx.build_dir, 'other_builds',
dir_name, '{}__ndk_target_{}'.format(arch, self.ctx.ndk_api))
def get_dir_name(self):
choices = self.check_recipe_choices()
dir_name = '-'.join([self.name] + choices)
return dir_name
def get_build_dir(self, arch):
'''Given the arch name, returns the directory where the
downloaded/copied package will be built.'''
return join(self.get_build_container_dir(arch), self.name)
def get_recipe_dir(self):
"""
Returns the local recipe directory or defaults to the core recipe
directory.
"""
if self.ctx.local_recipes is not None:
local_recipe_dir = join(self.ctx.local_recipes, self.name)
if exists(local_recipe_dir):
return local_recipe_dir
return join(self.ctx.root_dir, 'recipes', self.name)
# Public Recipe API to be subclassed if needed
def download_if_necessary(self):
info_main('Downloading {}'.format(self.name))
user_dir = environ.get('P4A_{}_DIR'.format(self.name.lower()))
if user_dir is not None:
info('P4A_{}_DIR is set, skipping download for {}'.format(
self.name, self.name))
return
self.download()
def download(self):
if self.url is None:
info('Skipping {} download as no URL is set'.format(self.name))
return
url = self.versioned_url
expected_digests = {}
for alg in set(hashlib.algorithms_guaranteed) | set(('md5', 'sha512', 'blake2b')):
expected_digest = getattr(self, alg + 'sum') if hasattr(self, alg + 'sum') else None
ma = match(u'^(.+)#' + alg + u'=([0-9a-f]{32,})$', url)
if ma: # fragmented URL?
if expected_digest:
raise ValueError(
('Received {}sum from both the {} recipe '
'and its url').format(alg, self.name))
url = ma.group(1)
expected_digest = ma.group(2)
if expected_digest:
expected_digests[alg] = expected_digest
ensure_dir(join(self.ctx.packages_path, self.name))
with current_directory(join(self.ctx.packages_path, self.name)):
filename = shprint(sh.basename, url).stdout[:-1].decode('utf-8')
do_download = True
marker_filename = '.mark-{}'.format(filename)
if exists(filename) and isfile(filename):
if not exists(marker_filename):
shprint(sh.rm, filename)
else:
for alg, expected_digest in expected_digests.items():
current_digest = algsum(alg, filename)
if current_digest != expected_digest:
debug('* Generated {}sum: {}'.format(alg,
current_digest))
debug('* Expected {}sum: {}'.format(alg,
expected_digest))
raise ValueError(
('Generated {0}sum does not match expected {0}sum '
'for {1} recipe').format(alg, self.name))
do_download = False
# If we got this far, we will download
if do_download:
debug('Downloading {} from {}'.format(self.name, url))
shprint(sh.rm, '-f', marker_filename)
self.download_file(self.versioned_url, filename)
touch(marker_filename)
if exists(filename) and isfile(filename):
for alg, expected_digest in expected_digests.items():
current_digest = algsum(alg, filename)
if current_digest != expected_digest:
debug('* Generated {}sum: {}'.format(alg,
current_digest))
debug('* Expected {}sum: {}'.format(alg,
expected_digest))
raise ValueError(
('Generated {0}sum does not match expected {0}sum '
'for {1} recipe').format(alg, self.name))
else:
info('{} download already cached, skipping'.format(self.name))
def unpack(self, arch):
info_main('Unpacking {} for {}'.format(self.name, arch))
build_dir = self.get_build_container_dir(arch)
user_dir = environ.get('P4A_{}_DIR'.format(self.name.lower()))
if user_dir is not None:
info('P4A_{}_DIR exists, symlinking instead'.format(
self.name.lower()))
if exists(self.get_build_dir(arch)):
return
rmdir(build_dir)
ensure_dir(build_dir)
shprint(sh.cp, '-a', user_dir, self.get_build_dir(arch))
return
if self.url is None:
info('Skipping {} unpack as no URL is set'.format(self.name))
return
filename = shprint(
sh.basename, self.versioned_url).stdout[:-1].decode('utf-8')
ma = match(u'^(.+)#[a-z0-9_]{3,}=([0-9a-f]{32,})$', filename)
if ma: # fragmented URL?
filename = ma.group(1)
with current_directory(build_dir):
directory_name = self.get_build_dir(arch)
if not exists(directory_name) or not isdir(directory_name):
extraction_filename = join(
self.ctx.packages_path, self.name, filename)
if isfile(extraction_filename):
if extraction_filename.endswith(('.zip', '.whl')):
try:
sh.unzip(extraction_filename)
except (sh.ErrorReturnCode_1, sh.ErrorReturnCode_2):
# return code 1 means unzipping had
# warnings but did complete,
# apparently happens sometimes with
# github zips
pass
fileh = zipfile.ZipFile(extraction_filename, 'r')
root_directory = fileh.filelist[0].filename.split('/')[0]
if root_directory != basename(directory_name):
move(root_directory, directory_name)
elif extraction_filename.endswith(
('.tar.gz', '.tgz', '.tar.bz2', '.tbz2', '.tar.xz', '.txz')):
sh.tar('xf', extraction_filename)
root_directory = sh.tar('tf', extraction_filename).stdout.decode(
'utf-8').split('\n')[0].split('/')[0]
if root_directory != basename(directory_name):
move(root_directory, directory_name)
else:
raise Exception(
'Could not extract {} download, it must be .zip, '
'.tar.gz or .tar.bz2 or .tar.xz'.format(extraction_filename))
elif isdir(extraction_filename):
ensure_dir(directory_name)
for entry in listdir(extraction_filename):
# Previously we filtered out the .git folder, but during the build process for some recipes
# (e.g. when version is parsed by `setuptools_scm`) that may be needed.
shprint(sh.cp, '-Rv',
join(extraction_filename, entry),
directory_name)
else:
raise Exception(
'Given path is neither a file nor a directory: {}'
.format(extraction_filename))
else:
info('{} is already unpacked, skipping'.format(self.name))
def get_recipe_env(self, arch=None, with_flags_in_cc=True):
"""Return the env specialized for the recipe
"""
if arch is None:
arch = self.filtered_archs[0]
env = arch.get_env(with_flags_in_cc=with_flags_in_cc)
return env
def prebuild_arch(self, arch):
'''Run any pre-build tasks for the Recipe. By default, this checks if
any prebuild_archname methods exist for the archname of the current
architecture, and runs them if so.'''
prebuild = "prebuild_{}".format(arch.arch.replace('-', '_'))
if hasattr(self, prebuild):
getattr(self, prebuild)()
else:
info('{} has no {}, skipping'.format(self.name, prebuild))
def is_patched(self, arch):
build_dir = self.get_build_dir(arch.arch)
return exists(join(build_dir, '.patched'))
def apply_patches(self, arch, build_dir=None):
'''Apply any patches for the Recipe.
.. versionchanged:: 0.6.0
Add ability to apply patches from any dir via kwarg `build_dir`'''
if self.patches:
info_main('Applying patches for {}[{}]'
.format(self.name, arch.arch))
if self.is_patched(arch):
info_main('{} already patched, skipping'.format(self.name))
return
build_dir = build_dir if build_dir else self.get_build_dir(arch.arch)
for patch in self.patches:
if isinstance(patch, (tuple, list)):
patch, patch_check = patch
if not patch_check(arch=arch, recipe=self):
continue
self.apply_patch(
patch.format(version=self.version, arch=arch.arch),
arch.arch, build_dir=build_dir)
touch(join(build_dir, '.patched'))
def should_build(self, arch):
'''Should perform any necessary test and return True only if it needs
building again. Per default we implement a library test, in case that
we detect so.
'''
if self.built_libraries:
return not all(
exists(lib) for lib in self.get_libraries(arch.arch)
)
return True
def build_arch(self, arch):
'''Run any build tasks for the Recipe. By default, this checks if
any build_archname methods exist for the archname of the current
architecture, and runs them if so.'''
build = "build_{}".format(arch.arch)
if hasattr(self, build):
getattr(self, build)()
def install_libraries(self, arch):
'''This method is always called after `build_arch`. In case that we
detect a library recipe, defined by the class attribute
`built_libraries`, we will copy all defined libraries into the
right location.
'''
if not self.built_libraries:
return
shared_libs = [
lib for lib in self.get_libraries(arch) if lib.endswith(".so")
]
self.install_libs(arch, *shared_libs)
def postbuild_arch(self, arch):
'''Run any post-build tasks for the Recipe. By default, this checks if
any postbuild_archname methods exist for the archname of the
current architecture, and runs them if so.
'''
postbuild = "postbuild_{}".format(arch.arch)
if hasattr(self, postbuild):
getattr(self, postbuild)()
if self.need_stl_shared:
self.install_stl_lib(arch)
def prepare_build_dir(self, arch):
'''Copies the recipe data into a build dir for the given arch. By
default, this unpacks a downloaded recipe. You should override
it (or use a Recipe subclass with different behaviour) if you
want to do something else.
'''
self.unpack(arch)
def clean_build(self, arch=None):
'''Deletes all the build information of the recipe.
If arch is not None, only this arch dir is deleted. Otherwise
(the default) all builds for all archs are deleted.
By default, this just deletes the main build dir. If the
recipe has e.g. object files biglinked, or .so files stored
elsewhere, you should override this method.
This method is intended for testing purposes, it may have
strange results. Rebuild everything if this seems to happen.
'''
if arch is None:
base_dir = join(self.ctx.build_dir, 'other_builds', self.name)
else:
base_dir = self.get_build_container_dir(arch)
dirs = glob.glob(base_dir + '-*')
if exists(base_dir):
dirs.append(base_dir)
if not dirs:
warning('Attempted to clean build for {} but found no existing '
'build dirs'.format(self.name))
for directory in dirs:
rmdir(directory)
# Delete any Python distributions to ensure the recipe build
# doesn't persist in site-packages
rmdir(self.ctx.python_installs_dir)
def install_libs(self, arch, *libs):
libs_dir = self.ctx.get_libs_dir(arch.arch)
if not libs:
warning('install_libs called with no libraries to install!')
return
args = libs + (libs_dir,)
shprint(sh.cp, *args)
def has_libs(self, arch, *libs):
return all(map(lambda lib: self.ctx.has_lib(arch.arch, lib), libs))
def get_libraries(self, arch_name, in_context=False):
"""Return the full path of the library depending on the architecture.
Per default, the build library path it will be returned, unless
`get_libraries` has been called with kwarg `in_context` set to
True.
.. note:: this method should be used for library recipes only
"""
recipe_libs = set()
if not self.built_libraries:
return recipe_libs
for lib, rel_path in self.built_libraries.items():
if not in_context:
abs_path = join(self.get_build_dir(arch_name), rel_path, lib)
if rel_path in {".", "", None}:
abs_path = join(self.get_build_dir(arch_name), lib)
else:
abs_path = join(self.ctx.get_libs_dir(arch_name), lib)
recipe_libs.add(abs_path)
return recipe_libs
@classmethod
def recipe_dirs(cls, ctx):
recipe_dirs = []
if ctx.local_recipes is not None:
recipe_dirs.append(realpath(ctx.local_recipes))
if ctx.storage_dir:
recipe_dirs.append(join(ctx.storage_dir, 'recipes'))
recipe_dirs.append(join(ctx.root_dir, "recipes"))
return recipe_dirs
@classmethod
def list_recipes(cls, ctx):
forbidden_dirs = ('__pycache__', )
for recipes_dir in cls.recipe_dirs(ctx):
if recipes_dir and exists(recipes_dir):
for name in listdir(recipes_dir):
if name in forbidden_dirs:
continue
fn = join(recipes_dir, name)
if isdir(fn):
yield name
@classmethod
def get_recipe(cls, name, ctx):
'''Returns the Recipe with the given name, if it exists.'''
name = name.lower()
if not hasattr(cls, "recipes"):
cls.recipes = {}
if name in cls.recipes:
return cls.recipes[name]
recipe_file = None
for recipes_dir in cls.recipe_dirs(ctx):
if not exists(recipes_dir):
continue
# Find matching folder (may differ in case):
for subfolder in listdir(recipes_dir):
if subfolder.lower() == name:
recipe_file = join(recipes_dir, subfolder, '__init__.py')
if exists(recipe_file):
name = subfolder # adapt to actual spelling
break
recipe_file = None
if recipe_file is not None:
break
else:
raise ValueError('Recipe does not exist: {}'.format(name))
mod = import_recipe('pythonforandroid.recipes.{}'.format(name), recipe_file)
if len(logger.handlers) > 1:
logger.removeHandler(logger.handlers[1])
recipe = mod.recipe
recipe.ctx = ctx
cls.recipes[name.lower()] = recipe
return recipe
class IncludedFilesBehaviour(object):
'''Recipe mixin class that will automatically unpack files included in
the recipe directory.'''
src_filename = None
def prepare_build_dir(self, arch):
if self.src_filename is None:
raise BuildInterruptingException(
'IncludedFilesBehaviour failed: no src_filename specified')
rmdir(self.get_build_dir(arch))
shprint(sh.cp, '-a', join(self.get_recipe_dir(), self.src_filename),
self.get_build_dir(arch))
class BootstrapNDKRecipe(Recipe):
'''A recipe class for recipes built in an Android project jni dir with
an Android.mk. These are not cached separatly, but built in the
bootstrap's own building directory.
To build an NDK project which is not part of the bootstrap, see
:class:`~pythonforandroid.recipe.NDKRecipe`.
To link with python, call the method :meth:`get_recipe_env`
with the kwarg *with_python=True*.
'''
dir_name = None # The name of the recipe build folder in the jni dir
def get_build_container_dir(self, arch):
return self.get_jni_dir()
def get_build_dir(self, arch):
if self.dir_name is None:
raise ValueError('{} recipe doesn\'t define a dir_name, but '
'this is necessary'.format(self.name))
return join(self.get_build_container_dir(arch), self.dir_name)
def get_jni_dir(self):
return join(self.ctx.bootstrap.build_dir, 'jni')
def get_recipe_env(self, arch=None, with_flags_in_cc=True, with_python=False):
env = super().get_recipe_env(arch, with_flags_in_cc)
if not with_python:
return env
env['PYTHON_INCLUDE_ROOT'] = self.ctx.python_recipe.include_root(arch.arch)
env['PYTHON_LINK_ROOT'] = self.ctx.python_recipe.link_root(arch.arch)
env['EXTRA_LDLIBS'] = ' -lpython{}'.format(
self.ctx.python_recipe.link_version)
return env
class NDKRecipe(Recipe):
'''A recipe class for any NDK project not included in the bootstrap.'''
generated_libraries = []
def should_build(self, arch):
lib_dir = self.get_lib_dir(arch)
for lib in self.generated_libraries:
if not exists(join(lib_dir, lib)):
return True
return False
def get_lib_dir(self, arch):
return join(self.get_build_dir(arch.arch), 'obj', 'local', arch.arch)
def get_jni_dir(self, arch):
return join(self.get_build_dir(arch.arch), 'jni')
def build_arch(self, arch, *extra_args):
super().build_arch(arch)
env = self.get_recipe_env(arch)
with current_directory(self.get_build_dir(arch.arch)):
shprint(
sh.Command(join(self.ctx.ndk_dir, "ndk-build")),
'V=1',
'NDK_DEBUG=' + ("1" if self.ctx.build_as_debuggable else "0"),
'APP_PLATFORM=android-' + str(self.ctx.ndk_api),
'APP_ABI=' + arch.arch,
*extra_args, _env=env
)
class PythonRecipe(Recipe):
site_packages_name = None
'''The name of the module's folder when installed in the Python
site-packages (e.g. for pyjnius it is 'jnius')'''
call_hostpython_via_targetpython = True
'''If True, tries to install the module using the hostpython binary
copied to the target (normally arm) python build dir. However, this
will fail if the module tries to import e.g. _io.so. Set this to False
to call hostpython from its own build dir, installing the module in
the right place via arguments to setup.py. However, this may not set
the environment correctly and so False is not the default.'''
install_in_hostpython = False
'''If True, additionally installs the module in the hostpython build
dir. This will make it available to other recipes if
call_hostpython_via_targetpython is False.
'''
install_in_targetpython = True
'''If True, installs the module in the targetpython installation dir.
This is almost always what you want to do.'''
setup_extra_args = []
'''List of extra arguments to pass to setup.py'''
depends = ['python3']
'''
.. note:: it's important to keep this depends as a class attribute outside
`__init__` because sometimes we only initialize the class, so the
`__init__` call won't be called and the deps would be missing
(which breaks the dependency graph computation)
.. warning:: don't forget to call `super().__init__()` in any recipe's
`__init__`, or otherwise it may not be ensured that it depends
on python2 or python3 which can break the dependency graph
'''
hostpython_prerequisites = []
'''List of hostpython packages required to build a recipe'''
def __init__(self, *args, **kwargs):
super().__init__(*args, **kwargs)
if 'python3' not in self.depends:
# We ensure here that the recipe depends on python even it overrode
# `depends`. We only do this if it doesn't already depend on any
# python, since some recipes intentionally don't depend on/work
# with all python variants
depends = self.depends
depends.append('python3')
depends = list(set(depends))
self.depends = depends
def clean_build(self, arch=None):
super().clean_build(arch=arch)
name = self.folder_name
python_install_dirs = glob.glob(join(self.ctx.python_installs_dir, '*'))
for python_install in python_install_dirs:
site_packages_dir = glob.glob(join(python_install, 'lib', 'python*',
'site-packages'))
if site_packages_dir:
build_dir = join(site_packages_dir[0], name)
if exists(build_dir):
info('Deleted {}'.format(build_dir))
rmdir(build_dir)
@property
def real_hostpython_location(self):
host_name = 'host{}'.format(self.ctx.python_recipe.name)
if host_name == 'hostpython3':
python_recipe = Recipe.get_recipe(host_name, self.ctx)
return python_recipe.python_exe
else:
python_recipe = self.ctx.python_recipe
return 'python{}'.format(python_recipe.version)
@property
def hostpython_location(self):
if not self.call_hostpython_via_targetpython:
return self.real_hostpython_location
return self.ctx.hostpython
@property
def folder_name(self):
'''The name of the build folders containing this recipe.'''
name = self.site_packages_name
if name is None:
name = self.name
return name
def get_recipe_env(self, arch=None, with_flags_in_cc=True):
env = super().get_recipe_env(arch, with_flags_in_cc)
env['PYTHONNOUSERSITE'] = '1'
# Set the LANG, this isn't usually important but is a better default
# as it occasionally matters how Python e.g. reads files
env['LANG'] = "en_GB.UTF-8"
# Binaries made by packages installed by pip
env["PATH"] = join(self.hostpython_site_dir, "bin") + ":" + env["PATH"]
if not self.call_hostpython_via_targetpython:
env['CFLAGS'] += ' -I{}'.format(
self.ctx.python_recipe.include_root(arch.arch)
)
env['LDFLAGS'] += ' -L{} -lpython{}'.format(
self.ctx.python_recipe.link_root(arch.arch),
self.ctx.python_recipe.link_version,
)
hppath = []
hppath.append(join(dirname(self.hostpython_location), 'Lib'))
hppath.append(join(hppath[0], 'site-packages'))
builddir = join(dirname(self.hostpython_location), 'build')
if exists(builddir):
hppath += [join(builddir, d) for d in listdir(builddir)
if isdir(join(builddir, d))]
if len(hppath) > 0:
if 'PYTHONPATH' in env:
env['PYTHONPATH'] = ':'.join(hppath + [env['PYTHONPATH']])
else:
env['PYTHONPATH'] = ':'.join(hppath)
return env
def should_build(self, arch):
name = self.folder_name
if self.ctx.has_package(name, arch):
info('Python package already exists in site-packages')
return False
info('{} apparently isn\'t already in site-packages'.format(name))
return True
def build_arch(self, arch):
'''Install the Python module by calling setup.py install with
the target Python dir.'''
self.install_hostpython_prerequisites()
super().build_arch(arch)
self.install_python_package(arch)
def install_python_package(self, arch, name=None, env=None, is_dir=True):
'''Automate the installation of a Python package (or a cython
package where the cython components are pre-built).'''
# arch = self.filtered_archs[0] # old kivy-ios way
if name is None:
name = self.name
if env is None:
env = self.get_recipe_env(arch)
info('Installing {} into site-packages'.format(self.name))
hostpython = sh.Command(self.hostpython_location)
hpenv = env.copy()
with current_directory(self.get_build_dir(arch.arch)):
shprint(hostpython, 'setup.py', 'install', '-O2',
'--root={}'.format(self.ctx.get_python_install_dir(arch.arch)),
'--install-lib=.',
_env=hpenv, *self.setup_extra_args)
# If asked, also install in the hostpython build dir
if self.install_in_hostpython:
self.install_hostpython_package(arch)
def get_hostrecipe_env(self, arch):
env = environ.copy()
env['PYTHONPATH'] = self.hostpython_site_dir
return env
@property
def hostpython_site_dir(self):
return join(dirname(self.real_hostpython_location), 'Lib', 'site-packages')
def install_hostpython_package(self, arch):
env = self.get_hostrecipe_env(arch)
real_hostpython = sh.Command(self.real_hostpython_location)
shprint(real_hostpython, 'setup.py', 'install', '-O2',
'--root={}'.format(dirname(self.real_hostpython_location)),
'--install-lib=Lib/site-packages',
_env=env, *self.setup_extra_args)
@property
def python_major_minor_version(self):
parsed_version = packaging.version.parse(self.ctx.python_recipe.version)
return f"{parsed_version.major}.{parsed_version.minor}"
def install_hostpython_prerequisites(self, packages=None, force_upgrade=True):
if not packages:
packages = self.hostpython_prerequisites
if len(packages) == 0:
return
pip_options = [
"install",
*packages,
"--target", self.hostpython_site_dir, "--python-version",
self.ctx.python_recipe.version,
# Don't use sources, instead wheels
"--only-binary=:all:",
]
if force_upgrade:
pip_options.append("--upgrade")
# Use system's pip
shprint(sh.pip, *pip_options)
def restore_hostpython_prerequisites(self, packages):
_packages = []
for package in packages:
original_version = Recipe.get_recipe(package, self.ctx).version
_packages.append(package + "==" + original_version)
self.install_hostpython_prerequisites(packages=_packages)
class CompiledComponentsPythonRecipe(PythonRecipe):
pre_build_ext = False
build_cmd = 'build_ext'
def build_arch(self, arch):
'''Build any cython components, then install the Python module by
calling setup.py install with the target Python dir.
'''
Recipe.build_arch(self, arch)
self.install_hostpython_prerequisites()
self.build_compiled_components(arch)
self.install_python_package(arch)
def build_compiled_components(self, arch):
info('Building compiled components in {}'.format(self.name))
env = self.get_recipe_env(arch)
hostpython = sh.Command(self.hostpython_location)
with current_directory(self.get_build_dir(arch.arch)):
if self.install_in_hostpython:
shprint(hostpython, 'setup.py', 'clean', '--all', _env=env)
shprint(hostpython, 'setup.py', self.build_cmd, '-v',
_env=env, *self.setup_extra_args)
build_dir = glob.glob('build/lib.*')[0]
shprint(sh.find, build_dir, '-name', '"*.o"', '-exec',
env['STRIP'], '{}', ';', _env=env)
def install_hostpython_package(self, arch):
env = self.get_hostrecipe_env(arch)
self.rebuild_compiled_components(arch, env)
super().install_hostpython_package(arch)
def rebuild_compiled_components(self, arch, env):
info('Rebuilding compiled components in {}'.format(self.name))
hostpython = sh.Command(self.real_hostpython_location)
shprint(hostpython, 'setup.py', 'clean', '--all', _env=env)
shprint(hostpython, 'setup.py', self.build_cmd, '-v', _env=env,
*self.setup_extra_args)
class CppCompiledComponentsPythonRecipe(CompiledComponentsPythonRecipe):
""" Extensions that require the cxx-stl """
call_hostpython_via_targetpython = False
need_stl_shared = True
class CythonRecipe(PythonRecipe):
pre_build_ext = False
cythonize = True
cython_args = []
call_hostpython_via_targetpython = False
def build_arch(self, arch):
'''Build any cython components, then install the Python module by
calling setup.py install with the target Python dir.
'''
Recipe.build_arch(self, arch)
self.build_cython_components(arch)
self.install_python_package(arch)
def build_cython_components(self, arch):
info('Cythonizing anything necessary in {}'.format(self.name))
env = self.get_recipe_env(arch)
with current_directory(self.get_build_dir(arch.arch)):
hostpython = sh.Command(self.ctx.hostpython)
shprint(hostpython, '-c', 'import sys; print(sys.path)', _env=env)
debug('cwd is {}'.format(realpath(curdir)))
info('Trying first build of {} to get cython files: this is '
'expected to fail'.format(self.name))
manually_cythonise = False
try:
shprint(hostpython, 'setup.py', 'build_ext', '-v', _env=env,
*self.setup_extra_args)
except sh.ErrorReturnCode_1:
print()
info('{} first build failed (as expected)'.format(self.name))
manually_cythonise = True
if manually_cythonise:
self.cythonize_build(env=env)
shprint(hostpython, 'setup.py', 'build_ext', '-v', _env=env,
_tail=20, _critical=True, *self.setup_extra_args)
else:
info('First build appeared to complete correctly, skipping manual'
'cythonising.')
if not self.ctx.with_debug_symbols:
self.strip_object_files(arch, env)
def strip_object_files(self, arch, env, build_dir=None):
if build_dir is None:
build_dir = self.get_build_dir(arch.arch)
with current_directory(build_dir):
info('Stripping object files')
shprint(sh.find, '.', '-iname', '*.so', '-exec',
'/usr/bin/echo', '{}', ';', _env=env)
shprint(sh.find, '.', '-iname', '*.so', '-exec',
env['STRIP'].split(' ')[0], '--strip-unneeded',
# '/usr/bin/strip', '--strip-unneeded',
'{}', ';', _env=env)
def cythonize_file(self, env, build_dir, filename):
short_filename = filename
if filename.startswith(build_dir):
short_filename = filename[len(build_dir) + 1:]
info(u"Cythonize {}".format(short_filename))
cyenv = env.copy()
if 'CYTHONPATH' in cyenv:
cyenv['PYTHONPATH'] = cyenv['CYTHONPATH']
elif 'PYTHONPATH' in cyenv:
del cyenv['PYTHONPATH']
if 'PYTHONNOUSERSITE' in cyenv:
cyenv.pop('PYTHONNOUSERSITE')
python_command = sh.Command("python{}".format(
self.ctx.python_recipe.major_minor_version_string.split(".")[0]
))
shprint(python_command, "-c"
"import sys; from Cython.Compiler.Main import setuptools_main; sys.exit(setuptools_main());",
filename, *self.cython_args, _env=cyenv)
def cythonize_build(self, env, build_dir="."):
if not self.cythonize:
info('Running cython cancelled per recipe setting')
return
info('Running cython where appropriate')
for root, dirnames, filenames in walk("."):
for filename in fnmatch.filter(filenames, "*.pyx"):
self.cythonize_file(env, build_dir, join(root, filename))
def get_recipe_env(self, arch, with_flags_in_cc=True):
env = super().get_recipe_env(arch, with_flags_in_cc)
env['LDFLAGS'] = env['LDFLAGS'] + ' -L{} '.format(
self.ctx.get_libs_dir(arch.arch) +
' -L{} '.format(self.ctx.libs_dir) +
' -L{}'.format(join(self.ctx.bootstrap.build_dir, 'obj', 'local',
arch.arch)))
env['LDSHARED'] = env['CC'] + ' -shared'
# shprint(sh.whereis, env['LDSHARED'], _env=env)
env['LIBLINK'] = 'NOTNONE'
if self.ctx.copy_libs:
env['COPYLIBS'] = '1'
# Every recipe uses its own liblink path, object files are
# collected and biglinked later
liblink_path = join(self.get_build_container_dir(arch.arch),
'objects_{}'.format(self.name))
env['LIBLINK_PATH'] = liblink_path
ensure_dir(liblink_path)
return env
class PyProjectRecipe(PythonRecipe):
'''Recipe for projects which containes `pyproject.toml`'''
# Extra args to pass to `python -m build ...`
extra_build_args = []
call_hostpython_via_targetpython = False
def get_recipe_env(self, arch, **kwargs):
# Custom hostpython
self.ctx.python_recipe.python_exe = join(
self.ctx.python_recipe.get_build_dir(arch), "android-build", "python3")
env = super().get_recipe_env(arch, **kwargs)
build_dir = self.get_build_dir(arch)
ensure_dir(build_dir)
build_opts = join(build_dir, "build-opts.cfg")
with open(build_opts, "w") as file:
file.write("[bdist_wheel]\nplat-name={}".format(
self.get_wheel_platform_tag(arch)
))
file.close()
env["DIST_EXTRA_CONFIG"] = build_opts
return env
def get_wheel_platform_tag(self, arch):
return "android_" + {
"armeabi-v7a": "arm",
"arm64-v8a": "aarch64",
"x86_64": "x86_64",
"x86": "i686",
}[arch.arch]
def install_wheel(self, arch, built_wheels):
_wheel = built_wheels[0]
built_wheel_dir = dirname(_wheel)
# Fix wheel platform tag
wheel_tag = wheel_tags(
_wheel,
platform_tags=self.get_wheel_platform_tag(arch),
remove=True,
)
selected_wheel = join(built_wheel_dir, wheel_tag)
_dev_wheel_dir = environ.get("P4A_WHEEL_DIR", False)
if _dev_wheel_dir:
ensure_dir(_dev_wheel_dir)
shprint(sh.cp, selected_wheel, _dev_wheel_dir)
info(f"Installing built wheel: {wheel_tag}")
destination = self.ctx.get_python_install_dir(arch.arch)
with WheelFile(selected_wheel) as wf:
for zinfo in wf.filelist:
wf.extract(zinfo, destination)
wf.close()
def build_arch(self, arch):
self.install_hostpython_prerequisites(
packages=["build[virtualenv]", "pip"] + self.hostpython_prerequisites
)
build_dir = self.get_build_dir(arch.arch)
env = self.get_recipe_env(arch, with_flags_in_cc=True)
# make build dir separatly
sub_build_dir = join(build_dir, "p4a_android_build")
ensure_dir(sub_build_dir)
# copy hostpython to built python to ensure correct selection of libs and includes
shprint(sh.cp, self.real_hostpython_location, self.ctx.python_recipe.python_exe)
build_args = [
"-m",
"build",
"--wheel",
"--config-setting",
"builddir={}".format(sub_build_dir),
] + self.extra_build_args
built_wheels = []
with current_directory(build_dir):
shprint(
sh.Command(self.ctx.python_recipe.python_exe), *build_args, _env=env
)
built_wheels = [realpath(whl) for whl in glob.glob("dist/*.whl")]
self.install_wheel(arch, built_wheels)
class MesonRecipe(PyProjectRecipe):
'''Recipe for projects which uses meson as build system'''
meson_version = "1.4.0"
ninja_version = "1.11.1.1"
def sanitize_flags(self, *flag_strings):
return " ".join(flag_strings).strip().split(" ")
def get_recipe_meson_options(self, arch):
env = self.get_recipe_env(arch, with_flags_in_cc=True)
return {
"binaries": {
"c": arch.get_clang_exe(with_target=True),
"cpp": arch.get_clang_exe(with_target=True, plus_plus=True),
"ar": self.ctx.ndk.llvm_ar,
"strip": self.ctx.ndk.llvm_strip,
},
"built-in options": {
"c_args": self.sanitize_flags(env["CFLAGS"], env["CPPFLAGS"]),
"cpp_args": self.sanitize_flags(env["CXXFLAGS"], env["CPPFLAGS"]),
"c_link_args": self.sanitize_flags(env["LDFLAGS"]),
"cpp_link_args": self.sanitize_flags(env["LDFLAGS"]),
},
"properties": {
"needs_exe_wrapper": True,
"sys_root": self.ctx.ndk.sysroot
},
"host_machine": {
"cpu_family": {
"arm64-v8a": "aarch64",
"armeabi-v7a": "arm",
"x86_64": "x86_64",
"x86": "x86"
}[arch.arch],
"cpu": {
"arm64-v8a": "aarch64",
"armeabi-v7a": "armv7",
"x86_64": "x86_64",
"x86": "i686"
}[arch.arch],
"endian": "little",
"system": "android",
}
}
def write_build_options(self, arch):
"""Writes python dict to meson config file"""
option_data = ""
build_options = self.get_recipe_meson_options(arch)
for key in build_options.keys():
data_chunk = "[{}]".format(key)
for subkey in build_options[key].keys():
value = build_options[key][subkey]
if isinstance(value, int):
value = str(value)
elif isinstance(value, str):
value = "'{}'".format(value)
elif isinstance(value, bool):
value = "true" if value else "false"
elif isinstance(value, list):
value = "['" + "', '".join(value) + "']"
data_chunk += "\n" + subkey + " = " + value
option_data += data_chunk + "\n\n"
return option_data
def ensure_args(self, *args):
for arg in args:
if arg not in self.extra_build_args:
self.extra_build_args.append(arg)
def build_arch(self, arch):
cross_file = join("/tmp", "android.meson.cross")
info("Writing cross file at: {}".format(cross_file))
# write cross config file
with open(cross_file, "w") as file:
file.write(self.write_build_options(arch))
file.close()
# set cross file
self.ensure_args('-Csetup-args=--cross-file', '-Csetup-args={}'.format(cross_file))
# ensure ninja and meson
for dep in [
"ninja=={}".format(self.ninja_version),
"meson=={}".format(self.meson_version),
]:
if dep not in self.hostpython_prerequisites:
self.hostpython_prerequisites.append(dep)
super().build_arch(arch)
class RustCompiledComponentsRecipe(PyProjectRecipe):
# Rust toolchain codes
# https://doc.rust-lang.org/nightly/rustc/platform-support.html
RUST_ARCH_CODES = {
"arm64-v8a": "aarch64-linux-android",
"armeabi-v7a": "armv7-linux-androideabi",
"x86_64": "x86_64-linux-android",
"x86": "i686-linux-android",
}
call_hostpython_via_targetpython = False
def get_recipe_env(self, arch, **kwargs):
env = super().get_recipe_env(arch, **kwargs)
# Set rust build target
build_target = self.RUST_ARCH_CODES[arch.arch]
cargo_linker_name = "CARGO_TARGET_{}_LINKER".format(
build_target.upper().replace("-", "_")
)
env["CARGO_BUILD_TARGET"] = build_target
env[cargo_linker_name] = join(
self.ctx.ndk.llvm_prebuilt_dir,
"bin",
"{}{}-clang".format(
# NDK's Clang format
build_target.replace("7", "7a")
if build_target.startswith("armv7")
else build_target,
self.ctx.ndk_api,
),
)
realpython_dir = self.ctx.python_recipe.get_build_dir(arch.arch)
env["RUSTFLAGS"] = "-Clink-args=-L{} -L{}".format(
self.ctx.get_libs_dir(arch.arch), join(realpython_dir, "android-build")
)
env["PYO3_CROSS_LIB_DIR"] = realpath(glob.glob(join(
realpython_dir, "android-build", "build",
"lib.linux-*-{}/".format(self.python_major_minor_version),
))[0])
info_main("Ensuring rust build toolchain")
shprint(sh.rustup, "target", "add", build_target)
# Add host python to PATH
env["PATH"] = ("{hostpython_dir}:{old_path}").format(
hostpython_dir=Recipe.get_recipe(
"hostpython3", self.ctx
).get_path_to_python(),
old_path=env["PATH"],
)
return env
def check_host_deps(self):
if not hasattr(sh, "rustup"):
error(
"`rustup` was not found on host system."
"Please install it using :"
"\n`curl https://sh.rustup.rs -sSf | sh`\n"
)
exit(1)
def build_arch(self, arch):
self.check_host_deps()
super().build_arch(arch)
class TargetPythonRecipe(Recipe):
'''Class for target python recipes. Sets ctx.python_recipe to point to
itself, so as to know later what kind of Python was built or used.'''
def __init__(self, *args, **kwargs):
self._ctx = None
super().__init__(*args, **kwargs)
def prebuild_arch(self, arch):
super().prebuild_arch(arch)
self.ctx.python_recipe = self
def include_root(self, arch):
'''The root directory from which to include headers.'''
raise NotImplementedError('Not implemented in TargetPythonRecipe')
def link_root(self):
raise NotImplementedError('Not implemented in TargetPythonRecipe')
@property
def major_minor_version_string(self):
parsed_version = packaging.version.parse(self.version)
return f"{parsed_version.major}.{parsed_version.minor}"
def create_python_bundle(self, dirn, arch):
"""
Create a packaged python bundle in the target directory, by
copying all the modules and standard library to the right
place.
"""
raise NotImplementedError('{} does not implement create_python_bundle'.format(self))
def reduce_object_file_names(self, dirn):
"""Recursively renames all files named XXX.cpython-...-linux-gnu.so"
to "XXX.so", i.e. removing the erroneous architecture name
coming from the local system.
"""
py_so_files = shprint(sh.find, dirn, '-iname', '*.so')
filens = py_so_files.stdout.decode('utf-8').split('\n')[:-1]
for filen in filens:
file_dirname, file_basename = split(filen)
parts = file_basename.split('.')
if len(parts) <= 2:
continue
# PySide6 libraries end with .abi3.so
if parts[1] == "abi3":
continue
move(filen, join(file_dirname, parts[0] + '.so'))
def algsum(alg, filen):
'''Calculate the digest of a file.
'''
with open(filen, 'rb') as fileh:
digest = getattr(hashlib, alg)(fileh.read())
return digest.hexdigest()
class FFPyPlayerRecipe(PyProjectRecipe):
version = 'v4.5.1'
url = 'https://github.com/matham/ffpyplayer/archive/{version}.zip'
depends = ['python3', 'sdl2', 'ffmpeg']
patches = ["setup.py.patch"]
opt_depends = ['openssl', 'ffpyplayer_codecs']
def get_recipe_env(self, arch, with_flags_in_cc=True):
env = super().get_recipe_env(arch)
build_dir = Recipe.get_recipe('ffmpeg', self.ctx).get_build_dir(arch.arch)
env["FFMPEG_INCLUDE_DIR"] = join(build_dir, "include")
env["FFMPEG_LIB_DIR"] = join(build_dir, "lib")
env["SDL_INCLUDE_DIR"] = join(self.ctx.bootstrap.build_dir, 'jni', 'SDL', 'include')
env["SDL_LIB_DIR"] = join(self.ctx.bootstrap.build_dir, 'libs', arch.arch)
env["USE_SDL2_MIXER"] = '1'
# ffpyplayer does not allow to pass more than one include dir for sdl2_mixer (and ATM is
# not needed), so we only pass the first one.
sdl2_mixer_recipe = self.get_recipe('sdl2_mixer', self.ctx)
env["SDL2_MIXER_INCLUDE_DIR"] = sdl2_mixer_recipe.get_include_dirs(arch)[0]
# NDKPLATFORM and LIBLINK are our switches for detecting Android platform, so can't be empty
# FIXME: We may want to introduce a cleaner approach to this?
env['NDKPLATFORM'] = "NOTNONE"
env['LIBLINK'] = 'NOTNONE'
# ffmpeg recipe enables GPL components only if ffpyplayer_codecs recipe used.
# Therefor we need to disable libpostproc if skipped.
if 'ffpyplayer_codecs' not in self.ctx.recipe_build_order:
env["CONFIG_POSTPROC"] = '0'
return env
recipe = FFPyPlayerRecipe()