mirror of
https://github.com/peridotbuild/pv2.git
synced 2024-11-24 06:01:25 +00:00
Mass Update
* Util Module * Provides: color class (for specialty stdout logging) * Provides: constants classes for rpm, errors, and mock * Provides: error classes for generic error handling and future fault handler * Provides: generic classes for generic, repeatable use cases * Provides: rpmutil with rpm utilities that range from basic to advanced metadata handling * Add mock module * Can generate a usable mock config based on input provided * Can generate mock plugin configuration as provided * cache related plugins are hardcoded as disabled * Supports plugins: chroot scanning, embedding files, bind mounts * Can generate basic dnf configs with repo information * (Currently limited) Error handler * Runs mock commands (such as build, buildsrpm, init, shell) * Add modularity module (very limited, doesn't really do much) * Add peridotpb example (does nothing, will likely be its own thing) * Add MIT license
This commit is contained in:
parent
771e79c637
commit
e48a54db3a
1
.gitignore
vendored
Normal file
1
.gitignore
vendored
Normal file
@ -0,0 +1 @@
|
|||||||
|
__pycache__/
|
18
LICENSE
Normal file
18
LICENSE
Normal file
@ -0,0 +1,18 @@
|
|||||||
|
Copyright 2023 Louis Abel <label@rockylinux.org>
|
||||||
|
|
||||||
|
Permission is hereby granted, free of charge, to any person obtaining a copy of
|
||||||
|
this software and associated documentation files (the “Software”), to deal in
|
||||||
|
the Software without restriction, including without limitation the rights to
|
||||||
|
use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of
|
||||||
|
the Software, and to permit persons to whom the Software is furnished to do so,
|
||||||
|
subject to the following conditions:
|
||||||
|
|
||||||
|
The above copyright notice and this permission notice shall be included in all
|
||||||
|
copies or substantial portions of the Software.
|
||||||
|
|
||||||
|
THE SOFTWARE IS PROVIDED “AS IS”, WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
||||||
|
IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS
|
||||||
|
FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR
|
||||||
|
COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER
|
||||||
|
IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
|
||||||
|
CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
|
106
README.md
106
README.md
@ -1,3 +1,109 @@
|
|||||||
# Platform POC
|
# Platform POC
|
||||||
|
|
||||||
A POC for builder nodes or developer purposes.
|
A POC for builder nodes or developer purposes.
|
||||||
|
|
||||||
|
## Examples of pv2.util
|
||||||
|
|
||||||
|
```
|
||||||
|
[label@sani buildsys]$ python3
|
||||||
|
Python 3.11.3 (main, Apr 5 2023, 00:00:00) [GCC 13.0.1 20230401 (Red Hat 13.0.1-0)] on linux
|
||||||
|
Type "help", "copyright", "credits" or "license" for more information.
|
||||||
|
>>> from pv2.util import rpmutil
|
||||||
|
>>> rpm_header = rpmutil.get_rpm_header('/tmp/golang-1.19.4-1.el9.src.rpm')
|
||||||
|
>>> generic = rpmutil.get_rpm_metadata_from_hdr(rpm_header)
|
||||||
|
>>> generic['excludearch']
|
||||||
|
[]
|
||||||
|
>>> generic['exclusivearch']
|
||||||
|
['x86_64', 'aarch64', 'ppc64le', 's390x']
|
||||||
|
|
||||||
|
# Or the actual definition itself to skip the above
|
||||||
|
>>> rpmutil.get_exclu_from_package(rpm_header)
|
||||||
|
{'ExcludeArch': [], 'ExclusiveArch': ['x86_64', 'aarch64', 'ppc64le', 's390x']}
|
||||||
|
```
|
||||||
|
|
||||||
|
```
|
||||||
|
[label@sani buildsys]$ python3
|
||||||
|
Python 3.11.3 (main, Apr 5 2023, 00:00:00) [GCC 13.0.1 20230401 (Red Hat 13.0.1-0)] on linux
|
||||||
|
Type "help", "copyright", "credits" or "license" for more information.
|
||||||
|
>>> from pv2.util import rpmutil
|
||||||
|
>>> rpm_header = rpmputil.get_rpm_header('/tmp/rocky-release-8.9-1.4.el8.noarch.rpm')
|
||||||
|
>>> generic = rpmutil.get_rpm_metadata_from_hdr(rpm_header)
|
||||||
|
>>> generic.keys()
|
||||||
|
dict_keys(['changelog_xml', 'files', 'obsoletes', 'provides', 'conflicts', 'requires', 'vendor', 'buildhost', 'filetime', 'description', 'license', 'nvr', 'nevra', 'name', 'version', 'release', 'epoch', 'arch', 'archivesize', 'packagesize'])
|
||||||
|
>>> generic['buildhost']
|
||||||
|
'ord1-prod-a64build003.svc.aws.rockylinux.org'
|
||||||
|
>>> generic['description']
|
||||||
|
'Rocky Linux release files.'
|
||||||
|
>>> generic['nvr']
|
||||||
|
'rocky-release-8.9-1.4.el8'
|
||||||
|
>>> generic['files']
|
||||||
|
['/etc/centos-release', '/etc/issue', '/etc/issue.net', '/etc/os-release', '/etc/redhat-release', '/etc/rocky-release', '/etc/rocky-release-upstream', '/etc/system-release', '/etc/system-release-cpe', '/usr/lib/os-release', '/usr/lib/rpm/macros.d/macros.dist', '/usr/lib/systemd/system-preset/85-display-manager.preset', '/usr/lib/systemd/system-preset/90-default.preset', '/usr/lib/systemd/system-preset/99-default-disable.preset', '/usr/share/doc/rocky-release/COMMUNITY-CHARTER', '/usr/share/doc/rocky-release/Contributors', '/usr/share/licenses/rocky-release/LICENSE', '/usr/share/man/man1/rocky.1.gz', '/usr/share/redhat-release', '/usr/share/rocky-release/EULA']
|
||||||
|
```
|
||||||
|
|
||||||
|
## Examples of pv2.mock
|
||||||
|
|
||||||
|
```
|
||||||
|
[label@sani buildsys]$ python3
|
||||||
|
Python 3.11.3 (main, Apr 5 2023, 00:00:00) [GCC 13.0.1 20230401 (Red Hat 13.0.1-0)] on linux
|
||||||
|
Type "help", "copyright", "credits" or "license" for more information.
|
||||||
|
>>> from pv2.mock.config import DnfConfig, DnfRepoConfig, MockConfig, MockPluginConfig, MockChrootFileConfig, MockMacroFileConfig
|
||||||
|
>>> repo = DnfRepoConfig(repoid='devel', name='baseos', priority='99', baseurl='http://dl.rockylinux.org/pub/rocky/9/devel/x86_64/os', enabled=True, gpgcheck=False)
|
||||||
|
>>> repo_list = [repo]
|
||||||
|
>>> dnf_base_config = DnfConfig(repositories=repo_list)
|
||||||
|
>>> mock_config = MockConfig(root='rocky-9-x86_64-example', target_arch='x86_64', dist='.el9', distribution='Rocky Linux', dnf_config=dnf_base_config, releasever='9')
|
||||||
|
>>> mock_config.export_mock_config('/tmp/ex.cfg')
|
||||||
|
|
||||||
|
[label@sani buildsys]$ cat /tmp/ex.cfg
|
||||||
|
config_opts["root"] = "rocky-9-x86_64-example"
|
||||||
|
config_opts["chroot_setup_cmd"] = "install bash bzip2 coreutils cpio diffutils findutils gawk glibc-minimal-langpack grep gzip info make patch redhat-rpm-config rpm-build sed shadow-utils system-release tar unzip util-linux which xz"
|
||||||
|
config_opts["dist"] = "el9"
|
||||||
|
config_opts["legal_host_arches"] = ('x86_64',)
|
||||||
|
config_opts["macros"]["%_host"] = "x86_64-redhat-linux-gnu"
|
||||||
|
config_opts["macros"]["%_host_cpu"] = "x86_64"
|
||||||
|
config_opts["macros"]["%_rpmfilename"] = "%%{NAME}-%%{VERSION}-%%{RELEASE}.%%{ARCH}.rpm"
|
||||||
|
config_opts["macros"]["%_vendor"] = "redhat"
|
||||||
|
config_opts["macros"]["%_vendor_host"] = "redhat"
|
||||||
|
config_opts["macros"]["%packager"] = "Default Packager <packager@noone.home>"
|
||||||
|
config_opts["macros"]["%vendor"] = "Default Vendor"
|
||||||
|
config_opts["print_main_output"] = True
|
||||||
|
config_opts["releasever"] = "9"
|
||||||
|
config_opts["rpmbuild_networking"] = False
|
||||||
|
config_opts["target_arch"] = "x86_64"
|
||||||
|
config_opts["use_host_resolv"] = False
|
||||||
|
config_opts["files"]["/etc/rpm/macros.xx"] = """
|
||||||
|
|
||||||
|
%dist %{!?distprefix0:%{?distprefix}}%{expand:%{lua:for i=0,9999 do print("%{?distprefix" .. i .."}") end}}.el9%{?distsuffix}%{?with_bootstrap:~bootstrap}
|
||||||
|
%distribution Rocky Linux
|
||||||
|
|
||||||
|
"""
|
||||||
|
|
||||||
|
config_opts["dnf.conf"] = """
|
||||||
|
[main]
|
||||||
|
assumeyes=1
|
||||||
|
best=1
|
||||||
|
debuglevel=1
|
||||||
|
gpgcheck=0
|
||||||
|
install_weak_deps=0
|
||||||
|
keepcache=1
|
||||||
|
logfile=/var/log/yum.log
|
||||||
|
mdpolicy=group:primary
|
||||||
|
metadata_expire=0
|
||||||
|
obsoletes=1
|
||||||
|
protected_packages=
|
||||||
|
reposdir=/dev/null
|
||||||
|
retries=20
|
||||||
|
rpm_verbosity=info
|
||||||
|
syslog_device=
|
||||||
|
syslog_ident=peridotbuilder
|
||||||
|
user_agent=peridotbuilder
|
||||||
|
|
||||||
|
[devel]
|
||||||
|
baseurl=http://dl.rockylinux.org/pub/rocky/9/devel/x86_64/os
|
||||||
|
enabled=1
|
||||||
|
gpgcheck=0
|
||||||
|
name=baseos
|
||||||
|
priority=99
|
||||||
|
repoid=devel
|
||||||
|
|
||||||
|
"""
|
||||||
|
```
|
||||||
|
13
mock/__init__.py
Normal file
13
mock/__init__.py
Normal file
@ -0,0 +1,13 @@
|
|||||||
|
# -*-:python; coding:utf-8; -*-
|
||||||
|
# author: Louis Abel <label@rockylinux.org>
|
||||||
|
"""
|
||||||
|
Mock and mock accessories
|
||||||
|
"""
|
||||||
|
|
||||||
|
# import all thingies here
|
||||||
|
from .config import (DnfConfig, DnfRepoConfig, MockConfig, MockPluginConfig,
|
||||||
|
MockBindMountPluginConfig, MockChrootFileConfig,
|
||||||
|
MockChrootScanPluginConfig, MockMacroConfig,
|
||||||
|
MockMacroFileConfig, MockShowrcPluginConfig)
|
||||||
|
from .error import MockErrorParser
|
||||||
|
from .runner import MockResult, MockRunner, MockErrorResulter
|
872
mock/config.py
Normal file
872
mock/config.py
Normal file
@ -0,0 +1,872 @@
|
|||||||
|
# -*- mode:python; coding:utf-8; -*-
|
||||||
|
# Louis Abel <label@rockylinux.org>
|
||||||
|
"""
|
||||||
|
Utility functions for mock configuration.
|
||||||
|
"""
|
||||||
|
|
||||||
|
import collections
|
||||||
|
import copy
|
||||||
|
import json
|
||||||
|
import re
|
||||||
|
import hashlib
|
||||||
|
from configparser import ConfigParser
|
||||||
|
from io import StringIO, IOBase
|
||||||
|
from pv2.util import error as err
|
||||||
|
from pv2.util import constants as const
|
||||||
|
from pv2.util import generic as generic_util
|
||||||
|
|
||||||
|
# List all classes in this module
|
||||||
|
__all__ = [
|
||||||
|
'DnfConfig',
|
||||||
|
'DnfRepoConfig',
|
||||||
|
'MockConfig',
|
||||||
|
'MockPluginConfig',
|
||||||
|
'MockBindMountPluginConfig',
|
||||||
|
'MockChrootScanPluginConfig',
|
||||||
|
'MockChrootFileConfig',
|
||||||
|
'MockConfigUtils',
|
||||||
|
'MockMacroConfig',
|
||||||
|
'MockMacroFileConfig',
|
||||||
|
]
|
||||||
|
|
||||||
|
# pylint: disable=too-few-public-methods
|
||||||
|
class MockConfigUtils:
|
||||||
|
"""
|
||||||
|
Mock config utilities. Provides basic things needed when making a mock
|
||||||
|
config.
|
||||||
|
"""
|
||||||
|
@staticmethod
|
||||||
|
def config_string(value):
|
||||||
|
"""
|
||||||
|
Converts a given value to a mock compatible string
|
||||||
|
|
||||||
|
Value should be:
|
||||||
|
* bool
|
||||||
|
* int
|
||||||
|
* string
|
||||||
|
* list
|
||||||
|
* tuple
|
||||||
|
* None
|
||||||
|
"""
|
||||||
|
|
||||||
|
# If the value being sent is none, a boolean, int, or tuple, just
|
||||||
|
# straight up return it as a string.
|
||||||
|
if value is None or isinstance(value, (bool, int, tuple)):
|
||||||
|
return str(value)
|
||||||
|
|
||||||
|
# If it's a string or a list, return it as a json string/list. We make
|
||||||
|
# sure we convert it properly and going through json makes sure it
|
||||||
|
# comes out right.
|
||||||
|
if isinstance(value, (str, list)):
|
||||||
|
return json.dumps(value)
|
||||||
|
|
||||||
|
# Error out if a value was sent that is not supported.
|
||||||
|
raise err.ProvidedValueError(f'{type(value)}: {value} is not supported.')
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def gen_config_string(name: str, status: bool) -> str:
|
||||||
|
"""
|
||||||
|
Generates a output string to enable a plugin
|
||||||
|
"""
|
||||||
|
config_name = copy.copy(name)
|
||||||
|
config_status = __class__.config_string(status)
|
||||||
|
output = f'config_opts["plugin_conf"]["{config_name}_enable"] = {config_status}\n'
|
||||||
|
return output
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def gen_config_string_with_opts(name: str, status: bool, opts: dict) -> str:
|
||||||
|
"""
|
||||||
|
Generates a output string to add options to an enabled plugin
|
||||||
|
"""
|
||||||
|
config_name = copy.copy(name)
|
||||||
|
config_status = __class__.config_string(status)
|
||||||
|
config_opts = copy.copy(opts)
|
||||||
|
output = f'config_opts["plugin_conf"]["{config_name}_enable"] = {config_status}\n'
|
||||||
|
if not status:
|
||||||
|
return output
|
||||||
|
|
||||||
|
output += f'config_opts["plugin_conf"]["{config_name}_opts"] = {{}}\n'
|
||||||
|
|
||||||
|
# If plugin options were provided, we try to go through and spit them
|
||||||
|
# out properly. Some documented plugins use nested dictionaries and the
|
||||||
|
# value being a string. This helps with that.
|
||||||
|
for key, option in sorted(config_opts):
|
||||||
|
key_config = __class__.config_string(key)
|
||||||
|
option_config = __class__.config_string(option)
|
||||||
|
# pylint: disable=line-too-long
|
||||||
|
output += f'config_opts["plugin_conf"]["{config_name}_opts"][{key_config}] = {option_config}\n'
|
||||||
|
|
||||||
|
return output
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def gen_config_option(option, value, append=False) -> str:
|
||||||
|
"""
|
||||||
|
Helps generate the 'config_opts' part of a mock configuration.
|
||||||
|
"""
|
||||||
|
outter = ''
|
||||||
|
option = __class__.config_string(option)
|
||||||
|
|
||||||
|
# If a dictionary, get all key value pairs and splay them out into
|
||||||
|
# strings (sending to config_string).
|
||||||
|
if isinstance(value, dict):
|
||||||
|
for key, val in sorted(value.items()):
|
||||||
|
key_name = __class__.config_string(key)
|
||||||
|
val_name = __class__.config_string(val)
|
||||||
|
outter += f'config_opts[{option}][{key_name}] = {val_name}\n'
|
||||||
|
# Some options/plugins use .append for whatever reason. Setting
|
||||||
|
# append to True will allow this portion to work and play it out into a
|
||||||
|
# string.
|
||||||
|
elif append:
|
||||||
|
value_str = __class__.config_string(value)
|
||||||
|
outter += f'config_opts[{option}].append({value_str})\n'
|
||||||
|
# Some options are just options in general, a key value string. This
|
||||||
|
# covers the rest.
|
||||||
|
else:
|
||||||
|
value_str = __class__.config_string(value)
|
||||||
|
# pylint: disable=consider-using-f-string
|
||||||
|
outter += f'config_opts[{option}] = {value_str}\n'
|
||||||
|
return outter
|
||||||
|
|
||||||
|
class DnfConfigurator:
|
||||||
|
"""
|
||||||
|
Base class for dnf configuration generation. Should only contain static
|
||||||
|
classes.
|
||||||
|
"""
|
||||||
|
@staticmethod
|
||||||
|
def gen_config_section(section, opts):
|
||||||
|
"""
|
||||||
|
Generate a config section using the config parser and data we're
|
||||||
|
receiving. This should be able to handle both [main] and repo sections.
|
||||||
|
"""
|
||||||
|
# A dnf configuration is key=value, sort of like an ini file.
|
||||||
|
# ConfigParser gets us close to that.
|
||||||
|
config = ConfigParser()
|
||||||
|
config.add_section(section)
|
||||||
|
for key, value in sorted(opts.items()):
|
||||||
|
|
||||||
|
# Continue if repositoryid was caught. We already added the section
|
||||||
|
# above.
|
||||||
|
if key == 'repositoryid':
|
||||||
|
continue
|
||||||
|
|
||||||
|
# Based on the key we received, we'll determine how the value will
|
||||||
|
# be presented. For example, for cases of the key/values being
|
||||||
|
# boolean options, regardless of what's received as the truthy
|
||||||
|
# value, we'll convert it to a string integer. The rest are
|
||||||
|
# strings in general.
|
||||||
|
if key in const.MockConstants.MOCK_DNF_BOOL_OPTIONS:
|
||||||
|
config.set(section, key, generic_util.gen_bool_option(value))
|
||||||
|
elif key in const.MockConstants.MOCK_DNF_STR_OPTIONS:
|
||||||
|
config.set(section, key, str(value))
|
||||||
|
elif key in const.MockConstants.MOCK_DNF_LIST_OPTIONS:
|
||||||
|
config.set(section, key, value.strip())
|
||||||
|
elif key == 'baseurl':
|
||||||
|
if isinstance(value, (list, tuple)):
|
||||||
|
value = "\n ".join(value)
|
||||||
|
config.set(section, key, value.strip())
|
||||||
|
else:
|
||||||
|
config.set(section, key, generic_util.trim_non_empty_string(key, value))
|
||||||
|
|
||||||
|
# Export the configuration we made into a file descriptor for use in
|
||||||
|
# DnfConfig.
|
||||||
|
file_descriptor = StringIO()
|
||||||
|
config.write(file_descriptor, space_around_delimiters=False)
|
||||||
|
file_descriptor.flush()
|
||||||
|
file_descriptor.seek(0)
|
||||||
|
return file_descriptor.read()
|
||||||
|
|
||||||
|
class DnfConfig(DnfConfigurator):
|
||||||
|
"""
|
||||||
|
This helps with the base configuration part of a mock config.
|
||||||
|
"""
|
||||||
|
# All these arguments are used. Everything made here is typically pushed
|
||||||
|
# into MockConfig.
|
||||||
|
# pylint: disable=too-many-locals,too-many-arguments,unused-argument
|
||||||
|
def __init__(
|
||||||
|
self,
|
||||||
|
debuglevel=1,
|
||||||
|
retries=20,
|
||||||
|
obsoletes=True,
|
||||||
|
gpgcheck=False,
|
||||||
|
assumeyes=True,
|
||||||
|
keepcache=True,
|
||||||
|
best=True,
|
||||||
|
syslog_ident='peridotbuilder',
|
||||||
|
syslog_device='',
|
||||||
|
metadata_expire=0,
|
||||||
|
install_weak_deps=False,
|
||||||
|
protected_packages='',
|
||||||
|
reposdir='/dev/null',
|
||||||
|
logfile='/var/log/yum.log',
|
||||||
|
mdpolicy='group:primary',
|
||||||
|
rpmverbosity='info',
|
||||||
|
repositories=None,
|
||||||
|
module_platform_id=None,
|
||||||
|
user_agent='peridotbuilder',
|
||||||
|
exclude=None,
|
||||||
|
):
|
||||||
|
if rpmverbosity not in const.MockConstants.MOCK_RPM_VERBOSITY:
|
||||||
|
raise err.ProvidedValueError(f'{rpmverbosity} is not set to a valid value')
|
||||||
|
# The repodata setup is a bit weird. What we do is we go through all
|
||||||
|
# "locals" for this class and build everything into a dictionary. We
|
||||||
|
# later send this and the repositories dictionary to gen_config_section.
|
||||||
|
self.__repodata = {}
|
||||||
|
for (key, value) in iter(list(locals().items())):
|
||||||
|
if key not in ['self', 'repositories'] and value is not None:
|
||||||
|
self.__repodata[key] = value
|
||||||
|
|
||||||
|
self.__repositories = {}
|
||||||
|
if repositories:
|
||||||
|
for repo in repositories:
|
||||||
|
self.add_repo_slot(repo)
|
||||||
|
|
||||||
|
def add_repo_slot(self, repo):
|
||||||
|
"""
|
||||||
|
Adds a repository as needed for mock.
|
||||||
|
|
||||||
|
DnfRepoConfig object is expected for repo.
|
||||||
|
"""
|
||||||
|
if not isinstance(repo, DnfRepoConfig):
|
||||||
|
raise err.ProvidedValueError(f'This type of repo is not supported: {type(repo)}')
|
||||||
|
if repo.name in self.__repositories:
|
||||||
|
raise err.ExistsValueError(f'Repository already added: {repo.name}')
|
||||||
|
self.__repositories[repo.name] = repo
|
||||||
|
|
||||||
|
def gen_config(self) -> str:
|
||||||
|
"""
|
||||||
|
Generates the configuration that will be used for mock.
|
||||||
|
|
||||||
|
Call this to generate the configuration.
|
||||||
|
"""
|
||||||
|
outter = 'config_opts["dnf.conf"] = """\n'
|
||||||
|
outter += self.gen_config_section('main', self.__repodata)
|
||||||
|
# Each "repo" instance as a gen_config() command as DnfRepoConfig has
|
||||||
|
# that method.
|
||||||
|
for repo_name in sorted(self.__repositories.keys()):
|
||||||
|
outter += self.__repositories[repo_name].gen_config()
|
||||||
|
outter += '"""\n'
|
||||||
|
return outter
|
||||||
|
|
||||||
|
class DnfRepoConfig(DnfConfigurator):
|
||||||
|
"""
|
||||||
|
This helps with the repo configs that would be in a mock config.
|
||||||
|
"""
|
||||||
|
# pylint: disable=too-many-arguments,unused-argument
|
||||||
|
def __init__(self,
|
||||||
|
repoid,
|
||||||
|
name,
|
||||||
|
priority,
|
||||||
|
baseurl=None,
|
||||||
|
enabled=True,
|
||||||
|
gpgcheck=None,
|
||||||
|
gpgkey=None,
|
||||||
|
sslverify=None,
|
||||||
|
module_hotfixes=None
|
||||||
|
):
|
||||||
|
"""
|
||||||
|
Basic dnf repo init, tailored for peridot usage. Mirror lists are *not*
|
||||||
|
supported in this class.
|
||||||
|
|
||||||
|
repoid: str
|
||||||
|
A unique name for the repository.
|
||||||
|
name: str
|
||||||
|
Human readable repo description
|
||||||
|
priority: str
|
||||||
|
Repository priority. Recommended to set if emulating koji tagging
|
||||||
|
and/or doing bootstrapping of some sort.
|
||||||
|
baseurl: str or list
|
||||||
|
A URL to the directory where the repo is located. repodata must be
|
||||||
|
there. Multiple URL's can be provided as a list.
|
||||||
|
enabled: bool or int
|
||||||
|
Enabled (True or 1) or disabled (False or 0) for this repository.
|
||||||
|
More than likely if you've added some extra repository, you want it
|
||||||
|
enabled. Otherwise, why are you adding it? For aesthetic reasons?
|
||||||
|
gpgcheck: bool or int
|
||||||
|
Perform a GPG check on packages if set to True/1.
|
||||||
|
gpgkey: str or None
|
||||||
|
Some URL or location of the repo gpg key
|
||||||
|
sslverify: str or None
|
||||||
|
Enable SSL certificate verification if set to 1.
|
||||||
|
"""
|
||||||
|
|
||||||
|
self.__repoconf = {}
|
||||||
|
for (key, value) in locals().items():
|
||||||
|
if key != 'self' and value is not None:
|
||||||
|
self.__repoconf[key] = value
|
||||||
|
|
||||||
|
def gen_config(self) -> str:
|
||||||
|
"""
|
||||||
|
Generates the dnf repo config
|
||||||
|
|
||||||
|
Returns a string
|
||||||
|
"""
|
||||||
|
section = generic_util.trim_non_empty_string(
|
||||||
|
'repoid',
|
||||||
|
self.__repoconf['repoid']
|
||||||
|
)
|
||||||
|
return self.gen_config_section(section, self.__repoconf)
|
||||||
|
|
||||||
|
@property
|
||||||
|
def name(self):
|
||||||
|
"""
|
||||||
|
Repo name
|
||||||
|
"""
|
||||||
|
return self.__repoconf['name']
|
||||||
|
|
||||||
|
|
||||||
|
# All mock classes
|
||||||
|
class MockConfig(MockConfigUtils):
|
||||||
|
"""
|
||||||
|
Mock configuration file generator
|
||||||
|
"""
|
||||||
|
# pylint: disable=too-many-locals,too-many-arguments,unused-argument
|
||||||
|
def __init__(
|
||||||
|
self,
|
||||||
|
target_arch,
|
||||||
|
root=None,
|
||||||
|
chroot_setup_cmd=None,
|
||||||
|
chroot_setup_cmd_pkgs=None,
|
||||||
|
dist=None,
|
||||||
|
releasever=None,
|
||||||
|
package_manager: str = 'dnf',
|
||||||
|
enable_networking: bool = False,
|
||||||
|
files=None,
|
||||||
|
macros=None,
|
||||||
|
dnf_config=None,
|
||||||
|
basedir=None,
|
||||||
|
print_main_output: bool = True,
|
||||||
|
target_vendor: str = 'redhat',
|
||||||
|
vendor: str = 'Default Vendor',
|
||||||
|
packager: str = 'Default Packager <packager@noone.home>',
|
||||||
|
distsuffix=None,
|
||||||
|
distribution=None,
|
||||||
|
**kwargs
|
||||||
|
):
|
||||||
|
"""
|
||||||
|
Mock config init
|
||||||
|
|
||||||
|
target_arch: string (config_opts['target_arch'])
|
||||||
|
files: list (optional)
|
||||||
|
dist: must be a string with starting characters . and alphanumeric character
|
||||||
|
macros: dict expected, key should start with '%'
|
||||||
|
target_vendor: typically 'redhat' and shouldn't be changed in most cases
|
||||||
|
vendor: packaging vendor, e.g. Rocky Enterprise Software Foundation
|
||||||
|
packager: the packager, e.g. Release Engineering <releng@rockylinux.org>
|
||||||
|
chroot_setup_cmd_pkgs: list of packages for the chroot
|
||||||
|
"""
|
||||||
|
|
||||||
|
# A dist value must be defined. This dist value is typically what we
|
||||||
|
# see as the %{dist} macro in RPM distributions. For EL and Fedora,
|
||||||
|
# they usually start with a "." and then continue with an alphanumeric
|
||||||
|
# character.
|
||||||
|
if not dist:
|
||||||
|
raise err.MissingValueError('The dist value is NOT defined')
|
||||||
|
if dist and not re.match(r'^\.[a-zA-Z0-9]', dist):
|
||||||
|
raise err.ProvidedValueError('The dist value does not start with a ' +
|
||||||
|
'. and alphanumeric character')
|
||||||
|
|
||||||
|
# A releasever value must be defined. This is basically the version of
|
||||||
|
# the EL we're building for.
|
||||||
|
if not releasever:
|
||||||
|
raise err.MissingValueError('The releasever value is NOT defined.')
|
||||||
|
if releasever and not re.match(r'^[0-9]+', releasever):
|
||||||
|
raise err.ProvidedValueError('The releasever value does not start ' +
|
||||||
|
'with a number.')
|
||||||
|
|
||||||
|
# Set chroot defaults if necessary. In the constants module, we have a
|
||||||
|
# list of the most basic package set required. In the event that
|
||||||
|
# someone is building a mock config to use, they can set the
|
||||||
|
# chroot_setup_cmd if they wish to something other than install
|
||||||
|
# (usually this is almost never the case). More importantly, the
|
||||||
|
# packages actually installed into the chroot can be set. Some projects
|
||||||
|
# in peridot can potentially dictate this to something other than the
|
||||||
|
# defaults.
|
||||||
|
if not chroot_setup_cmd:
|
||||||
|
chroot_setup_cmd = const.MockConstants.MOCK_DEFAULT_CHROOT_SETUP_CMD
|
||||||
|
if not chroot_setup_cmd_pkgs:
|
||||||
|
chroot_setup_cmd_pkgs = const.MockConstants.MOCK_DEFAULT_CHROOT_BUILD_PKGS
|
||||||
|
|
||||||
|
# Each mock chroot needs a name. We do not arbitrarily generate any.
|
||||||
|
# The admin must be specific on what they want.
|
||||||
|
if not root:
|
||||||
|
raise err.MissingValueError('The mock root name was not provided.')
|
||||||
|
|
||||||
|
# Here we are building the basic mock configuration. We push most of it
|
||||||
|
# into dictionaries and then later translate it all into strings.
|
||||||
|
legal_host_arches = self.determine_legal_host_arches(target_arch)
|
||||||
|
interpreted_dist = self.determine_dist_macro(dist)
|
||||||
|
chroot_pkgs = ' '.join(chroot_setup_cmd_pkgs)
|
||||||
|
chroot_setup_cmd_string = chroot_setup_cmd + ' ' + chroot_pkgs
|
||||||
|
default_macros = {
|
||||||
|
'%_rpmfilename': '%%{NAME}-%%{VERSION}-%%{RELEASE}.%%{ARCH}.rpm',
|
||||||
|
'%_host': f'{target_arch}-{target_vendor}-linux-gnu',
|
||||||
|
'%_host_cpu': target_arch,
|
||||||
|
'%_vendor': target_vendor,
|
||||||
|
'%_vendor_host': target_vendor,
|
||||||
|
'%vendor': vendor,
|
||||||
|
'%packager': packager,
|
||||||
|
}
|
||||||
|
self.__config_opts = {
|
||||||
|
'root': root,
|
||||||
|
'target_arch': target_arch,
|
||||||
|
'legal_host_arches': legal_host_arches,
|
||||||
|
'chroot_setup_cmd': chroot_setup_cmd_string,
|
||||||
|
'dist': dist.strip('.'),
|
||||||
|
'releasever': releasever,
|
||||||
|
'basedir': basedir,
|
||||||
|
'use_host_resolv': enable_networking,
|
||||||
|
'rpmbuild_networking': enable_networking,
|
||||||
|
'print_main_output': print_main_output,
|
||||||
|
'macros': default_macros,
|
||||||
|
}
|
||||||
|
self.__config_opts.update(**kwargs)
|
||||||
|
self.__extra_config_opts = collections.defaultdict(list)
|
||||||
|
self.__files = {}
|
||||||
|
self.__macros = {}
|
||||||
|
self.__plugins = {}
|
||||||
|
if files:
|
||||||
|
for chroot_file in files:
|
||||||
|
self.add_file(chroot_file)
|
||||||
|
|
||||||
|
# Set absolute default macros for each build. This is a partial carry
|
||||||
|
# over from peridot v1. We add these to an /etc/rpm/macros... file on
|
||||||
|
# purpose. Otherwise, if they are set as macros in config_opts, they
|
||||||
|
# are placed in /builddir/.rpmmacros, which cannot be overriden. Doing
|
||||||
|
# this ensures we can override these macros (e.g. for modules)
|
||||||
|
starter_macros = {
|
||||||
|
'%dist': interpreted_dist,
|
||||||
|
'%distribution': distribution,
|
||||||
|
}
|
||||||
|
self.add_macros(starter_macros, macro_file='/etc/rpm/macros.xx')
|
||||||
|
if macros:
|
||||||
|
self.add_macros(macros)
|
||||||
|
|
||||||
|
# Set the absolute disabled plugins for each build. These three are
|
||||||
|
# disabled on purpose. Do NOT alter these. Do NOT attempt to override
|
||||||
|
# them. There should never be a reason to ever have these enabled in a
|
||||||
|
# build system nor in development tools that use this module.
|
||||||
|
yum_cache_plugin = MockPluginConfig(name='yum_cache', enable=False)
|
||||||
|
root_cache_plugin = MockPluginConfig(name='root_cache', enable=False)
|
||||||
|
ccache_plugin = MockPluginConfig(name='ccache', enable=False)
|
||||||
|
self.add_plugin(yum_cache_plugin)
|
||||||
|
self.add_plugin(root_cache_plugin)
|
||||||
|
self.add_plugin(ccache_plugin)
|
||||||
|
|
||||||
|
self.__dnf_config = dnf_config
|
||||||
|
|
||||||
|
def add_file(self, chroot_file):
|
||||||
|
"""
|
||||||
|
Adds a chroot file to the configuration.
|
||||||
|
"""
|
||||||
|
if chroot_file.file in self.__files:
|
||||||
|
raise err.ProvidedValueError(f'file {chroot_file.file} is already added')
|
||||||
|
self.__files[chroot_file.file] = chroot_file
|
||||||
|
|
||||||
|
def add_macros(self, macro_set, macro_file='/etc/macros/macros.zz'):
|
||||||
|
"""
|
||||||
|
Adds a set of macros to a mock configuration. This generates a file
|
||||||
|
that will be placed into the mock chroot, rather than
|
||||||
|
/builddir/.rpmmacros made by config_opts.
|
||||||
|
"""
|
||||||
|
macro_data = ''
|
||||||
|
for key, value in macro_set.items():
|
||||||
|
if '%' not in key:
|
||||||
|
macro_name = f'%{key}'
|
||||||
|
else:
|
||||||
|
macro_name = key
|
||||||
|
|
||||||
|
if not value:
|
||||||
|
continue
|
||||||
|
|
||||||
|
macro_value = value
|
||||||
|
|
||||||
|
macro_data += f'{macro_name} {macro_value}\n'
|
||||||
|
|
||||||
|
macro_config = MockMacroFileConfig(content=macro_data, file=macro_file)
|
||||||
|
returned_content = macro_config.gen_config()
|
||||||
|
self.__macros[macro_file] = returned_content
|
||||||
|
|
||||||
|
def add_plugin(self, plugin):
|
||||||
|
"""
|
||||||
|
Adds a mock plugin to the configuration.
|
||||||
|
"""
|
||||||
|
if plugin.name in self.__plugins:
|
||||||
|
raise err.ProvidedValueError(f'plugin {plugin.name} is already configured')
|
||||||
|
self.__plugins[plugin.name] = plugin
|
||||||
|
|
||||||
|
def module_install(self, module_name):
|
||||||
|
"""
|
||||||
|
Adds a module to module_install
|
||||||
|
"""
|
||||||
|
if 'module_install' not in self.__config_opts:
|
||||||
|
self.__config_opts['module_install'] = []
|
||||||
|
|
||||||
|
if module_name in self.__config_opts['module_install']:
|
||||||
|
raise err.ExistsValueError(f'{module_name} is already provided in module_install')
|
||||||
|
|
||||||
|
self.__config_opts['module_install'].append(module_name)
|
||||||
|
|
||||||
|
def module_enable(self, module_name):
|
||||||
|
"""
|
||||||
|
Adds a module to module_enable
|
||||||
|
"""
|
||||||
|
if 'module_enable' not in self.__config_opts:
|
||||||
|
self.__config_opts['module_enable'] = []
|
||||||
|
|
||||||
|
if module_name in self.__config_opts['module_enable']:
|
||||||
|
raise err.ExistsValueError(f'{module_name} is already provided in module_enable')
|
||||||
|
|
||||||
|
self.__config_opts['module_enable'].append(module_name)
|
||||||
|
|
||||||
|
def add_config_opt(self, key: str, value: str):
|
||||||
|
"""
|
||||||
|
Use this to add additional options not covered by this module
|
||||||
|
"""
|
||||||
|
self.__extra_config_opts[key].append(value)
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def determine_dist_macro(dist: str) -> str:
|
||||||
|
"""
|
||||||
|
Return a string of the interpreted dist macro. This will typically
|
||||||
|
match current EL release packages.
|
||||||
|
"""
|
||||||
|
# We don't want a case where we are sending "~bootstrap" as the dist
|
||||||
|
# already. So we're stripping it and letting the build figure it out
|
||||||
|
# for itself. The macro with_bootstrap conditional should dictate it.
|
||||||
|
if "~bootstrap" in dist:
|
||||||
|
starting_dist = dist.replace('~bootstrap', '')
|
||||||
|
else:
|
||||||
|
starting_dist = dist
|
||||||
|
|
||||||
|
# This is the current dist value that is used in current EL's. It will
|
||||||
|
# likely change over time. This value is *also* provided in
|
||||||
|
# system-release, but having it here is to make sure it *is* here just
|
||||||
|
# in case. This is especially useful when bootstrapping from ELN or
|
||||||
|
# stream.
|
||||||
|
# pylint: disable=line-too-long,consider-using-f-string
|
||||||
|
dist_value = '%{{!?distprefix0:%{{?distprefix}}}}%{{expand:%{{lua:for i=0,9999 do print("%{{?distprefix" .. i .."}}") end}}}}{0}%{{?distsuffix}}%{{?with_bootstrap:~bootstrap}}'.format(starting_dist)
|
||||||
|
return dist_value
|
||||||
|
|
||||||
|
# pylint: disable=too-many-return-statements
|
||||||
|
@staticmethod
|
||||||
|
def determine_legal_host_arches(target_arch: str) -> tuple:
|
||||||
|
"""
|
||||||
|
Return a tuple of acceptable arches for a given architecture. This will
|
||||||
|
appear as a list in the final mock config.
|
||||||
|
"""
|
||||||
|
# The legal_host_arches is typically a tuple of supported arches for a
|
||||||
|
# given platform. Based on the target_arch sent, we'll set the legal
|
||||||
|
# arches.
|
||||||
|
|
||||||
|
# We can easily use "switch" here but we are accounting for python 3.9
|
||||||
|
# at this time, which does not have it.
|
||||||
|
if target_arch == "x86_64":
|
||||||
|
return const.MockConstants.MOCK_X86_64_LEGAL_ARCHES
|
||||||
|
|
||||||
|
if target_arch in ['i386', 'i486', 'i586', 'i686']:
|
||||||
|
return const.MockConstants.MOCK_I686_LEGAL_ARCHES
|
||||||
|
|
||||||
|
if target_arch in "aarch64":
|
||||||
|
return const.MockConstants.MOCK_AARCH64_LEGAL_ARCHES
|
||||||
|
|
||||||
|
if target_arch in "armv7hl":
|
||||||
|
return const.MockConstants.MOCK_ARMV7HL_LEGAL_ARCHES
|
||||||
|
|
||||||
|
if target_arch in "ppc64le":
|
||||||
|
return const.MockConstants.MOCK_PPC64LE_LEGAL_ARCHES
|
||||||
|
|
||||||
|
if target_arch in "s390x":
|
||||||
|
return const.MockConstants.MOCK_S390X_LEGAL_ARCHES
|
||||||
|
|
||||||
|
if target_arch in "riscv64":
|
||||||
|
return const.MockConstants.MOCK_RISCV64_LEGAL_ARCHES
|
||||||
|
|
||||||
|
# This shouldn't happen, but who knows
|
||||||
|
if target_arch in "noarch":
|
||||||
|
return const.MockConstants.MOCK_NOARCH_LEGAL_ARCHES
|
||||||
|
|
||||||
|
return err.ProvidedValueError(f'Legal arches not found for {target_arch}.')
|
||||||
|
|
||||||
|
def set_dnf_config(self, dnf_config):
|
||||||
|
"""
|
||||||
|
Adds a dnf config section
|
||||||
|
"""
|
||||||
|
self.__dnf_config = dnf_config
|
||||||
|
|
||||||
|
# Disabling until I can figure out a better way to handle this
|
||||||
|
# pylint: disable=too-many-branches
|
||||||
|
def export_mock_config(self, config_file, root=None):
|
||||||
|
"""
|
||||||
|
Exports the mock configuration to a file.
|
||||||
|
"""
|
||||||
|
if not root:
|
||||||
|
if self.__config_opts.get('root'):
|
||||||
|
root = self.__config_opts.get('root')
|
||||||
|
else:
|
||||||
|
raise err.MissingValueError('root value is missing. This should ' +
|
||||||
|
'not have happened and is likely the ' +
|
||||||
|
'result of this module being ' +
|
||||||
|
'modified and not tested.')
|
||||||
|
|
||||||
|
if not isinstance(config_file, str):
|
||||||
|
if isinstance(config_file, IOBase):
|
||||||
|
raise err.ProvidedValueError('config_file must be a string. it cannot ' \
|
||||||
|
'be an open file handle.')
|
||||||
|
raise err.ProvidedValueError('config_file must be a string.')
|
||||||
|
|
||||||
|
# This is where we'll write the file. We'll go through each
|
||||||
|
# configuration option, generate their configs as they're found, and
|
||||||
|
# write them. It should look close, if not identical to a typical mock
|
||||||
|
# configuration.
|
||||||
|
with open(config_file, 'w', encoding='utf-8') as file_descriptor:
|
||||||
|
try:
|
||||||
|
if root:
|
||||||
|
file_descriptor.write(self.gen_config_option('root', root))
|
||||||
|
for option, value in sorted(self.__config_opts.items()):
|
||||||
|
if option == 'root' or value is None:
|
||||||
|
continue
|
||||||
|
file_descriptor.write(self.gen_config_option(option, value))
|
||||||
|
for option, value_list in sorted(self.__extra_config_opts.items()):
|
||||||
|
for value in value_list:
|
||||||
|
file_descriptor.write(self.gen_config_option(option, value, append=True))
|
||||||
|
for plugin in self.__plugins.values():
|
||||||
|
file_descriptor.write(plugin.gen_config())
|
||||||
|
for macro_file in self.__macros.values():
|
||||||
|
file_descriptor.write(macro_file)
|
||||||
|
for chroot_file in self.__files.values():
|
||||||
|
file_descriptor.write(chroot_file.gen_config())
|
||||||
|
if self.__dnf_config:
|
||||||
|
file_descriptor.write(self.__dnf_config.gen_config())
|
||||||
|
except Exception as exc:
|
||||||
|
raise err.ConfigurationError('There was an error exporting the mock ' \
|
||||||
|
f'configuration: {exc}')
|
||||||
|
finally:
|
||||||
|
file_descriptor.close()
|
||||||
|
|
||||||
|
@property
|
||||||
|
def mock_config_hash(self):
|
||||||
|
"""
|
||||||
|
Creates a hash sum of the configuration. Could be used for tracking
|
||||||
|
and/or comparison purposes.
|
||||||
|
|
||||||
|
This may not currently work at this time.
|
||||||
|
"""
|
||||||
|
hasher = hashlib.sha256()
|
||||||
|
file_descriptor = StringIO()
|
||||||
|
self.export_mock_config(file_descriptor)
|
||||||
|
file_descriptor.seek(0)
|
||||||
|
hasher.update(file_descriptor.read().encode('utf-8'))
|
||||||
|
file_descriptor.close()
|
||||||
|
return hasher.hexdigest()
|
||||||
|
|
||||||
|
### Start Plugins
|
||||||
|
class MockPluginConfig(MockConfigUtils):
|
||||||
|
"""
|
||||||
|
Mock plugin configuration helper. For cases where some plugin doesn't have
|
||||||
|
some sort of class in this module.
|
||||||
|
"""
|
||||||
|
def __init__(
|
||||||
|
self,
|
||||||
|
name: str,
|
||||||
|
enable: bool,
|
||||||
|
**kwargs
|
||||||
|
):
|
||||||
|
"""
|
||||||
|
Plugin config init. Used to enable/disable plugins. Additional plugin
|
||||||
|
options can be defined in kwargs (may or may not work)
|
||||||
|
|
||||||
|
name: plugin name, string
|
||||||
|
enable: boolean
|
||||||
|
"""
|
||||||
|
self.name = copy.copy(name)
|
||||||
|
self.enable = enable
|
||||||
|
self.opts = copy.copy(kwargs)
|
||||||
|
|
||||||
|
def gen_config(self):
|
||||||
|
"""
|
||||||
|
Helps add a plugin configuration to mock
|
||||||
|
"""
|
||||||
|
plugin_name = self.name
|
||||||
|
config_string_status = self.enable
|
||||||
|
outter = self.gen_config_string_with_opts(
|
||||||
|
name=plugin_name,
|
||||||
|
status=config_string_status,
|
||||||
|
opts=self.opts
|
||||||
|
)
|
||||||
|
return outter
|
||||||
|
|
||||||
|
class MockBindMountPluginConfig(MockConfigUtils):
|
||||||
|
"""
|
||||||
|
Mock plugin configuration helper
|
||||||
|
"""
|
||||||
|
def __init__(
|
||||||
|
self,
|
||||||
|
enable: bool,
|
||||||
|
mounts: list
|
||||||
|
):
|
||||||
|
"""
|
||||||
|
Plugin config init. Used to enable/disable bind mount plugin.
|
||||||
|
|
||||||
|
enable: boolean
|
||||||
|
mounts: list of tuples
|
||||||
|
"""
|
||||||
|
self.name = 'bind_mount'
|
||||||
|
self.enable = enable
|
||||||
|
self.mounts = mounts
|
||||||
|
|
||||||
|
def gen_config(self):
|
||||||
|
"""
|
||||||
|
Helps add a plugin configuration to mock
|
||||||
|
"""
|
||||||
|
bind_config_status = self.config_string(self.enable)
|
||||||
|
|
||||||
|
# Documentation wants a ['dirs'] section added, so we're obliging.
|
||||||
|
outter = self.gen_config_string(name='bind_mount', status=bind_config_status)
|
||||||
|
|
||||||
|
if not self.enable or not self.mounts:
|
||||||
|
return outter
|
||||||
|
|
||||||
|
for local_path, mock_chroot_path in self.mounts:
|
||||||
|
# pylint: disable=line-too-long
|
||||||
|
outter += f'config_opts["plugin_conf"]["bind_mount_opts"]["dirs"].append(("{local_path}", "{mock_chroot_path}"))\n'
|
||||||
|
|
||||||
|
return outter
|
||||||
|
|
||||||
|
class MockChrootScanPluginConfig(MockConfigUtils):
|
||||||
|
"""
|
||||||
|
Helps setup the chroot scan plugin.
|
||||||
|
"""
|
||||||
|
|
||||||
|
def __init__(
|
||||||
|
self,
|
||||||
|
enable,
|
||||||
|
**kwargs
|
||||||
|
):
|
||||||
|
"""
|
||||||
|
Inits the plugin configuration.
|
||||||
|
|
||||||
|
enable: bool
|
||||||
|
kwargs: additional options can be sent in here
|
||||||
|
"""
|
||||||
|
self.name = 'chroot_scan'
|
||||||
|
self.enable = enable
|
||||||
|
self.opts = copy.copy(kwargs)
|
||||||
|
|
||||||
|
def gen_config(self):
|
||||||
|
"""
|
||||||
|
Helps add a plugin configuration to mock
|
||||||
|
"""
|
||||||
|
chroot_config_status = self.enable
|
||||||
|
|
||||||
|
# This one is weird. The documentation specifically wants a "dict" as a
|
||||||
|
# string... Not really clear why. But we'll roll with it.
|
||||||
|
outter = self.gen_config_string(
|
||||||
|
name='chroot_scan',
|
||||||
|
status=chroot_config_status
|
||||||
|
)
|
||||||
|
|
||||||
|
opts_dict = {}
|
||||||
|
for key, option in sorted(self.opts.items()):
|
||||||
|
opts_dict[key] = option
|
||||||
|
|
||||||
|
outter += f'config_opts["plugin_conf"]["chroot_scan_opts"] = {opts_dict}\n'
|
||||||
|
return outter
|
||||||
|
|
||||||
|
class MockShowrcPluginConfig(MockConfigUtils):
|
||||||
|
"""
|
||||||
|
Helps enable the showrc plugin. Useful for showing defined rpm macros for a
|
||||||
|
build.
|
||||||
|
"""
|
||||||
|
|
||||||
|
def __init__(self, enable):
|
||||||
|
"""
|
||||||
|
Inits the plugin configuration.
|
||||||
|
|
||||||
|
enable: bool
|
||||||
|
"""
|
||||||
|
self.name = 'showrc'
|
||||||
|
self.enable = enable
|
||||||
|
|
||||||
|
def gen_config(self):
|
||||||
|
"""
|
||||||
|
Helps add a plugin configuration to mock
|
||||||
|
"""
|
||||||
|
showrc_config_status = self.enable
|
||||||
|
outter = f'config_opts["plugin_conf"]["showrc_enable"] = {showrc_config_status}\n'
|
||||||
|
return outter
|
||||||
|
|
||||||
|
### End Plugins
|
||||||
|
|
||||||
|
class MockChrootFileConfig:
|
||||||
|
"""
|
||||||
|
Helps embed files into a mock chroot. May be useful to trick builds if
|
||||||
|
necessary but also could help with things like secureboot if needed.
|
||||||
|
"""
|
||||||
|
def __init__(
|
||||||
|
self,
|
||||||
|
file: str,
|
||||||
|
content=None
|
||||||
|
):
|
||||||
|
"""
|
||||||
|
Create a file to embed into the mock root
|
||||||
|
"""
|
||||||
|
|
||||||
|
if not content:
|
||||||
|
raise err.MissingValueError('Macro content was not provided')
|
||||||
|
|
||||||
|
self.file = file
|
||||||
|
self._content = content
|
||||||
|
|
||||||
|
def gen_config(self):
|
||||||
|
"""
|
||||||
|
Return a string to be added to mock config
|
||||||
|
"""
|
||||||
|
return f'config_opts["files"]["{self.file}"] = """{self._content}\n"""\n\n'
|
||||||
|
|
||||||
|
class MockMacroConfig:
|
||||||
|
"""
|
||||||
|
Helps add macros into a mock configuration. This is a typical staple of
|
||||||
|
builds. In most cases, you won't need this and instead will use
|
||||||
|
MockMacroFileConfig.
|
||||||
|
"""
|
||||||
|
def __init__(
|
||||||
|
self,
|
||||||
|
name: str,
|
||||||
|
value: str
|
||||||
|
):
|
||||||
|
"""
|
||||||
|
init the class
|
||||||
|
"""
|
||||||
|
self.name = name
|
||||||
|
self.value = value
|
||||||
|
|
||||||
|
def gen_config(self):
|
||||||
|
"""
|
||||||
|
Generate the macro option
|
||||||
|
"""
|
||||||
|
return f'config_opts["macros"]["{self.name}"] = "{self.value}"'
|
||||||
|
|
||||||
|
class MockMacroFileConfig:
|
||||||
|
"""
|
||||||
|
Helps add macros into a mock configuration into a file instead.
|
||||||
|
"""
|
||||||
|
def __init__(
|
||||||
|
self,
|
||||||
|
file: str = '/etc/rpm/macros.zz',
|
||||||
|
content=None
|
||||||
|
):
|
||||||
|
"""
|
||||||
|
Create a macro file to embed into the mock root
|
||||||
|
"""
|
||||||
|
|
||||||
|
if not content:
|
||||||
|
raise err.MissingValueError('Macro content was not provided')
|
||||||
|
|
||||||
|
self.file = file
|
||||||
|
self._content = content
|
||||||
|
|
||||||
|
def gen_config(self):
|
||||||
|
"""
|
||||||
|
Return a string to be added to mock config
|
||||||
|
"""
|
||||||
|
return f'config_opts["files"]["{self.file}"] = """\n\n{self._content}\n"""\n\n'
|
100
mock/error.py
Normal file
100
mock/error.py
Normal file
@ -0,0 +1,100 @@
|
|||||||
|
# -*-:python; coding:utf-8; -*-
|
||||||
|
# author: Louis Abel <label@rockylinux.org>
|
||||||
|
"""
|
||||||
|
Mock Error Classes (mainly for parsing, if we care)
|
||||||
|
"""
|
||||||
|
|
||||||
|
import os
|
||||||
|
import re
|
||||||
|
from pv2.util import constants as const
|
||||||
|
from pv2.util import generic as generic_util
|
||||||
|
|
||||||
|
# list every error class that's enabled
|
||||||
|
|
||||||
|
__all__ = [
|
||||||
|
'MockErrorParser'
|
||||||
|
]
|
||||||
|
|
||||||
|
class MockErrorChecks:
|
||||||
|
"""
|
||||||
|
Static methods of all error checks
|
||||||
|
"""
|
||||||
|
@staticmethod
|
||||||
|
def analyze_log(checks, log_file):
|
||||||
|
"""
|
||||||
|
Go through the list of checks and verify the log file
|
||||||
|
|
||||||
|
All checks are listed throughout the class below this one.
|
||||||
|
"""
|
||||||
|
log_file_name = os.path.basename(log_file)
|
||||||
|
result_dict = {}
|
||||||
|
with open(log_file_name, 'rb') as file_descriptor:
|
||||||
|
for line_number, line in enumerate(file_descriptor, 1):
|
||||||
|
for check in checks:
|
||||||
|
result = check(line)
|
||||||
|
if result:
|
||||||
|
error_code, error_message = result
|
||||||
|
result_dict = {
|
||||||
|
'error_code': error_code,
|
||||||
|
'error_message': error_message,
|
||||||
|
'file_name': log_file_name,
|
||||||
|
'line': line_number
|
||||||
|
}
|
||||||
|
|
||||||
|
return result_dict
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def check_error(regex, message, error_code, line):
|
||||||
|
"""
|
||||||
|
Does the actual regex verification
|
||||||
|
"""
|
||||||
|
result = re.search(regex, generic_util.to_unicode(line))
|
||||||
|
if result:
|
||||||
|
return error_code, message.format(*result.groups())
|
||||||
|
|
||||||
|
return None
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def unmet_dep(line):
|
||||||
|
"""
|
||||||
|
Searches for a dependency error in the root log
|
||||||
|
"""
|
||||||
|
regex = r'Error:\s+No\s+Package\s+found\s+for\s+(.*?)$'
|
||||||
|
message_template = 'No package(s) found for "{0}"'
|
||||||
|
verify_pattern = __class__.check_error(regex,
|
||||||
|
message_template,
|
||||||
|
const.MockConstants.MOCK_EXIT_DNF_ERROR,
|
||||||
|
line)
|
||||||
|
|
||||||
|
return verify_pattern
|
||||||
|
|
||||||
|
class MockErrorParser(MockErrorChecks):
|
||||||
|
"""
|
||||||
|
Helps provide checking definitions to find errors and report them. This
|
||||||
|
could be used in the case of having a generic error (like 1) from mock and
|
||||||
|
needing to find the real reason.
|
||||||
|
"""
|
||||||
|
def __init__(
|
||||||
|
self,
|
||||||
|
root_log,
|
||||||
|
build_log
|
||||||
|
):
|
||||||
|
"""
|
||||||
|
Initialize parser
|
||||||
|
"""
|
||||||
|
self._root_log = root_log
|
||||||
|
self._build_log = build_log
|
||||||
|
|
||||||
|
def check_for_error(self):
|
||||||
|
"""
|
||||||
|
Checks for errors
|
||||||
|
"""
|
||||||
|
# we'll get this eventually
|
||||||
|
#build_log_check = []
|
||||||
|
|
||||||
|
root_log_check = [
|
||||||
|
self.unmet_dep
|
||||||
|
]
|
||||||
|
|
||||||
|
# pylint: disable=line-too-long
|
||||||
|
return self.analyze_log(root_log_check, self._root_log)
|
423
mock/runner.py
Normal file
423
mock/runner.py
Normal file
@ -0,0 +1,423 @@
|
|||||||
|
# -*- mode:python; coding:utf-8; -*-
|
||||||
|
# Louis Abel <label@rockylinux.org>
|
||||||
|
"""
|
||||||
|
Mock runners and limited error handler
|
||||||
|
"""
|
||||||
|
|
||||||
|
import os
|
||||||
|
import re
|
||||||
|
import logging
|
||||||
|
from pv2.util import error as err
|
||||||
|
from pv2.util import fileutil
|
||||||
|
from pv2.util import constants as const
|
||||||
|
from pv2.util import processor
|
||||||
|
|
||||||
|
__all__ = [
|
||||||
|
'MockRunner',
|
||||||
|
'MockResult'
|
||||||
|
]
|
||||||
|
|
||||||
|
class MockRunner:
|
||||||
|
"""
|
||||||
|
Mock runner definitions
|
||||||
|
"""
|
||||||
|
def __init__(self, config_path: str):
|
||||||
|
"""
|
||||||
|
Initialize the runner
|
||||||
|
"""
|
||||||
|
self.logger = logging.getLogger(self.__module__)
|
||||||
|
self.config_path = config_path
|
||||||
|
|
||||||
|
def init(self, resultdir=None, quiet=None, isolation=None, foreground=False):
|
||||||
|
"""
|
||||||
|
Inits a mock root
|
||||||
|
"""
|
||||||
|
return self.__run_mock(mock_call='init', resultdir=resultdir,
|
||||||
|
quiet=quiet, isolation=isolation,
|
||||||
|
foreground=foreground)
|
||||||
|
|
||||||
|
# pylint: disable=too-many-arguments
|
||||||
|
def shell(
|
||||||
|
self,
|
||||||
|
command: str,
|
||||||
|
resultdir=None,
|
||||||
|
quiet=None,
|
||||||
|
isolation=None,
|
||||||
|
foreground=False
|
||||||
|
):
|
||||||
|
"""
|
||||||
|
Runs shell for a given mock root
|
||||||
|
"""
|
||||||
|
return self.__run_mock(mock_call='shell', mock_arg=command,
|
||||||
|
resultdir=resultdir, quiet=quiet,
|
||||||
|
isolation=isolation, foreground=foreground)
|
||||||
|
|
||||||
|
def clean(self, quiet=None, isolation=None, foreground=False):
|
||||||
|
"""
|
||||||
|
Clean up the mock root
|
||||||
|
"""
|
||||||
|
try:
|
||||||
|
self.__run_mock(mock_call='clean', quiet=quiet,
|
||||||
|
isolation=isolation, foreground=foreground)
|
||||||
|
except MockErrorResulter as exc:
|
||||||
|
self.logger.error('Unable to run clean on %s', self.config_path)
|
||||||
|
self.logger.error('Output:\n%s\n', exc)
|
||||||
|
|
||||||
|
self.__run_mock(mock_call='clean')
|
||||||
|
|
||||||
|
# pylint: disable=too-many-arguments
|
||||||
|
def buildsrpm(
|
||||||
|
self,
|
||||||
|
spec: str,
|
||||||
|
sources: str,
|
||||||
|
resultdir=None,
|
||||||
|
definitions=None,
|
||||||
|
timeout=None,
|
||||||
|
quiet=None,
|
||||||
|
isolation=None,
|
||||||
|
foreground=False
|
||||||
|
):
|
||||||
|
"""
|
||||||
|
Builds a source RPM, but does not actually build the package
|
||||||
|
"""
|
||||||
|
return self.__run_mock(
|
||||||
|
mock_call='buildsrpm',
|
||||||
|
spec=spec,
|
||||||
|
sources=sources,
|
||||||
|
resultdir=resultdir,
|
||||||
|
definitions=definitions,
|
||||||
|
rpmbuild_timeout=timeout,
|
||||||
|
quiet=quiet,
|
||||||
|
target='noarch',
|
||||||
|
isolation=isolation,
|
||||||
|
foreground=foreground
|
||||||
|
)
|
||||||
|
|
||||||
|
# pylint: disable=too-many-arguments
|
||||||
|
def build(
|
||||||
|
self,
|
||||||
|
srpm_path: str,
|
||||||
|
resultdir=None,
|
||||||
|
definitions=None,
|
||||||
|
timeout=None,
|
||||||
|
quiet=None,
|
||||||
|
isolation=None,
|
||||||
|
foreground=False
|
||||||
|
):
|
||||||
|
"""
|
||||||
|
Builds a given source package
|
||||||
|
"""
|
||||||
|
return self.__run_mock(
|
||||||
|
mock_call='rebuild',
|
||||||
|
mock_arg=srpm_path,
|
||||||
|
resultdir=resultdir,
|
||||||
|
rpmbuild_timeout=timeout,
|
||||||
|
definitions=definitions,
|
||||||
|
quiet=quiet,
|
||||||
|
isolation=isolation,
|
||||||
|
foreground=foreground
|
||||||
|
)
|
||||||
|
|
||||||
|
def __determine_resultdir(self):
|
||||||
|
"""
|
||||||
|
Receives no input. This should figure out where the resultdir
|
||||||
|
will ultimately be.
|
||||||
|
"""
|
||||||
|
|
||||||
|
mock_debug_args = [
|
||||||
|
'mock',
|
||||||
|
'--root', self.config_path,
|
||||||
|
'--debug-config-expanded'
|
||||||
|
]
|
||||||
|
|
||||||
|
mock_debug_run = processor.run_proc_no_output(command=mock_debug_args)
|
||||||
|
regex = r"^config_opts\['resultdir'\] = '(.*)'"
|
||||||
|
regex_search = re.search(regex, mock_debug_run.stdout, re.MULTILINE)
|
||||||
|
if regex_search:
|
||||||
|
return regex_search.group(1)
|
||||||
|
|
||||||
|
return None
|
||||||
|
|
||||||
|
# pylint: disable=too-many-locals,too-many-branches
|
||||||
|
def __run_mock(
|
||||||
|
self,
|
||||||
|
mock_call: str,
|
||||||
|
mock_arg: str = '',
|
||||||
|
resultdir=None,
|
||||||
|
foreground=False,
|
||||||
|
**kwargs
|
||||||
|
):
|
||||||
|
"""
|
||||||
|
Actually run mock.
|
||||||
|
|
||||||
|
mock_call should be the command being used (such as rebuild, shell, and
|
||||||
|
so on)
|
||||||
|
mock_arg is a string, and can be an additional argument (some mock
|
||||||
|
commands do not need an additional argument, thus default is an empty
|
||||||
|
string)
|
||||||
|
kwargs can be any set of additional arguments to add to mock as
|
||||||
|
key:value pairs. for example, lets say your function accepts an
|
||||||
|
argument like isolation and you set it to 'simple', the kwargs.items()
|
||||||
|
block will parse it as `--isolation simple`. if your function does not
|
||||||
|
require an argument, and it's not a matter of it being true or false,
|
||||||
|
you would send it as argument='' to ensure that an additional list item
|
||||||
|
is not added after the argument.
|
||||||
|
"""
|
||||||
|
# Note: You will notice that everything appears to be separate list
|
||||||
|
# items. This is on purpose to try to make sure subprocess is happy.
|
||||||
|
# Don't try to simplify it.
|
||||||
|
initial_args = [
|
||||||
|
'mock',
|
||||||
|
'--root', self.config_path,
|
||||||
|
f'--{mock_call}', mock_arg
|
||||||
|
]
|
||||||
|
|
||||||
|
if resultdir:
|
||||||
|
initial_args.append('--resultdir')
|
||||||
|
initial_args.append(resultdir)
|
||||||
|
|
||||||
|
# As you probably noticed, not all options being sent by the other
|
||||||
|
# methods are accounted for, so we are using kwargs to deal with them
|
||||||
|
# instead. This is because not all mock commands use the same options
|
||||||
|
# (or get the same effects out of them if they can be specified). But
|
||||||
|
# we are firm on on the ones that should be set.
|
||||||
|
for option, argument in kwargs.items():
|
||||||
|
if argument is None:
|
||||||
|
continue
|
||||||
|
|
||||||
|
# If we are sending mock specific macro definitions that are not in
|
||||||
|
# the config, this is how you do it. It's expected that definitions
|
||||||
|
# is a dict with only key value pairs.
|
||||||
|
if option == 'definitions':
|
||||||
|
for macro, value in argument.items():
|
||||||
|
initial_args.append('--define')
|
||||||
|
# Macro definitions require quotes between name and value.
|
||||||
|
# DO NOT UNDO THIS.
|
||||||
|
initial_args.append(f"'{macro} {value}'")
|
||||||
|
# "quiet" is a weird one because it doesn't accept a value in mock.
|
||||||
|
# We purposely set it to "None" so it gets passed over (way above).
|
||||||
|
# Setting to True will make this flag appear.
|
||||||
|
elif option == 'quiet':
|
||||||
|
initial_args.append('--quiet')
|
||||||
|
elif option == 'isolation':
|
||||||
|
if argument in ('simple', 'nspawn', 'simple'):
|
||||||
|
initial_args.append('--isolation')
|
||||||
|
initial_args.append(str(argument))
|
||||||
|
else:
|
||||||
|
raise err.ProvidedValueError(f'{argument} is an invalid isolation option.')
|
||||||
|
|
||||||
|
# If we're not covering the obvious ones above that we need special
|
||||||
|
# care for, this is where the rest happens. If an argument is sent
|
||||||
|
# with an empty string, it'll just show up as --option. Any
|
||||||
|
# argument will make it show up as --option argument.
|
||||||
|
else:
|
||||||
|
initial_args.append(f'--{option}')
|
||||||
|
if len(str(argument)) > 0:
|
||||||
|
initial_args.append(str(argument))
|
||||||
|
|
||||||
|
# Might not need this. This just makes sure our list is in order.
|
||||||
|
initial_args = [arg for arg in initial_args if arg]
|
||||||
|
mock_command = ' '.join(initial_args)
|
||||||
|
self.logger.info('The following mock command will be executed: %s', mock_command)
|
||||||
|
|
||||||
|
# If foreground is enabled, all output from mock will show up in the
|
||||||
|
# user's terminal (or wherever the output is being sent). This means
|
||||||
|
# stdout and stderr will NOT contain any data. It may be better to set
|
||||||
|
# "quiet" instead of foreground and then stream the actual log files
|
||||||
|
# themselves, but this requires you to be specific on the resultdir to
|
||||||
|
# find and stream them.
|
||||||
|
if foreground:
|
||||||
|
mock_run = processor.run_proc_foreground(command=initial_args)
|
||||||
|
else:
|
||||||
|
mock_run = processor.run_proc_no_output(command=initial_args)
|
||||||
|
|
||||||
|
# Assign vars based on what happened above.
|
||||||
|
mock_config = self.config_path
|
||||||
|
exit_code = mock_run.returncode
|
||||||
|
stdout = mock_run.stdout
|
||||||
|
stderr = mock_run.stderr
|
||||||
|
|
||||||
|
# If a resultdir wasn't presented, we try to look for it. We do this by
|
||||||
|
# running mock's debug commands to get the correct value and regex it
|
||||||
|
# out.
|
||||||
|
if not resultdir:
|
||||||
|
resultdir = self.__determine_resultdir()
|
||||||
|
|
||||||
|
if exit_code != 0:
|
||||||
|
raise MockErrorResulter(
|
||||||
|
mock_command,
|
||||||
|
exit_code,
|
||||||
|
resultdir)
|
||||||
|
|
||||||
|
return MockResult(
|
||||||
|
mock_command,
|
||||||
|
mock_config,
|
||||||
|
exit_code,
|
||||||
|
stdout,
|
||||||
|
stderr,
|
||||||
|
resultdir)
|
||||||
|
|
||||||
|
class MockResult:
|
||||||
|
"""
|
||||||
|
Mock result parser
|
||||||
|
"""
|
||||||
|
# pylint: disable=too-many-arguments
|
||||||
|
def __init__(
|
||||||
|
self,
|
||||||
|
mock_command,
|
||||||
|
mock_config,
|
||||||
|
exit_code,
|
||||||
|
stdout,
|
||||||
|
stderr,
|
||||||
|
resultdir=None
|
||||||
|
):
|
||||||
|
"""
|
||||||
|
Initialize the mock result parser
|
||||||
|
"""
|
||||||
|
self.mock_command = mock_command
|
||||||
|
self.mock_config = mock_config
|
||||||
|
self.exit_code = exit_code
|
||||||
|
self.__stdout = stdout
|
||||||
|
self.__stderr = stderr
|
||||||
|
self.resultdir = resultdir
|
||||||
|
|
||||||
|
@property
|
||||||
|
def srpm(self):
|
||||||
|
"""
|
||||||
|
Turns a string (or None) of the build source RPM package
|
||||||
|
"""
|
||||||
|
return next(iter(fileutil.filter_files(
|
||||||
|
self.resultdir,
|
||||||
|
lambda file: file.endswith('src.rpm'))),
|
||||||
|
None
|
||||||
|
)
|
||||||
|
|
||||||
|
@property
|
||||||
|
def rpms(self):
|
||||||
|
"""
|
||||||
|
Returns a list of RPM package paths in the resultdir.
|
||||||
|
"""
|
||||||
|
return fileutil.filter_files(
|
||||||
|
self.resultdir,
|
||||||
|
lambda file: re.search(r'(?<!\.src)\.rpm$', file)
|
||||||
|
)
|
||||||
|
|
||||||
|
@property
|
||||||
|
def logs(self):
|
||||||
|
"""
|
||||||
|
Returns a list of mock log files
|
||||||
|
"""
|
||||||
|
mock_log_files = fileutil.filter_files(self.resultdir,
|
||||||
|
lambda file: file.endswith('.log'))
|
||||||
|
|
||||||
|
# If we are using the chroot scan plugin, then let's search for other
|
||||||
|
# logs that we may have cared about in this build.
|
||||||
|
chroot_scan_dir = os.path.join(self.resultdir, 'chroot_scan')
|
||||||
|
if os.path.exists(chroot_scan_dir):
|
||||||
|
for dir_name, _, files in os.walk(chroot_scan_dir):
|
||||||
|
for file in files:
|
||||||
|
if file.endswith('.log'):
|
||||||
|
mock_log_files.append(os.path.join(os.path.abspath(dir_name), file))
|
||||||
|
|
||||||
|
return mock_log_files
|
||||||
|
|
||||||
|
@property
|
||||||
|
def stdout(self):
|
||||||
|
"""
|
||||||
|
Returns stdout
|
||||||
|
"""
|
||||||
|
return self.__stdout
|
||||||
|
|
||||||
|
@property
|
||||||
|
def stderr(self):
|
||||||
|
"""
|
||||||
|
Returns stdout
|
||||||
|
"""
|
||||||
|
return self.__stderr
|
||||||
|
|
||||||
|
# Is there a better way to do this?
|
||||||
|
# Note that this isn't in pv2.util.error because this *may* be used to parse
|
||||||
|
# logs at some point, and we do not want to add additional parsers to
|
||||||
|
# pv2.util.error or have it import mock modules if it's not actually required.
|
||||||
|
# I don't want to have to import more than needed in pv2.util.error.
|
||||||
|
class MockErrorResulter(Exception):
|
||||||
|
"""
|
||||||
|
Mock error result checker.
|
||||||
|
|
||||||
|
Takes in an exception and reports the exit code.
|
||||||
|
"""
|
||||||
|
def __init__(
|
||||||
|
self,
|
||||||
|
mock_command,
|
||||||
|
exit_code,
|
||||||
|
resultdir=None,
|
||||||
|
result_message=None
|
||||||
|
):
|
||||||
|
"""
|
||||||
|
Initialize the MockError class this way.
|
||||||
|
"""
|
||||||
|
|
||||||
|
# We probably don't need to do this, but it doesn't hurt. There should
|
||||||
|
# always be a resultdir to reference.
|
||||||
|
self.build_log = None
|
||||||
|
self.root_log = None
|
||||||
|
|
||||||
|
if resultdir:
|
||||||
|
self.build_log = os.path.join(resultdir, 'build.log')
|
||||||
|
self.root_log = os.path.join(resultdir, 'root.log')
|
||||||
|
|
||||||
|
if not result_message:
|
||||||
|
result_message = f'Command {mock_command} exited with code ' \
|
||||||
|
f'{exit_code}. Please review build.log and root.log ' \
|
||||||
|
f'located in the main root ({resultdir}) or bootstrap root.'
|
||||||
|
|
||||||
|
# This is awkward. I can't think of a better way to do this.
|
||||||
|
if exit_code == const.MockConstants.MOCK_EXIT_ERROR:
|
||||||
|
#error_object = errmock(self.root_log, self.build_log)
|
||||||
|
#error_dict = error_object.check_for_error()
|
||||||
|
|
||||||
|
#if len(error_dict) > 0:
|
||||||
|
# result_message = f'Command {mock_command} exited with code ' \
|
||||||
|
# '{error_dict["error_code"]}: {error_dict["error_message"]}'
|
||||||
|
|
||||||
|
# pylint: disable=non-parent-init-called
|
||||||
|
err.MockGenericError.__init__(self, result_message)
|
||||||
|
# Send to log parser to figure out what it actually is, and use the
|
||||||
|
# above to report it.
|
||||||
|
elif exit_code == const.MockConstants.MOCK_EXIT_SETUID:
|
||||||
|
# pylint: disable=non-parent-init-called
|
||||||
|
result_message = 'Either setuid/setgid is not available or ' \
|
||||||
|
'another error occurred (such as a bootstrap init failure). ' \
|
||||||
|
'Please review build.log or root.log, in the main root ' \
|
||||||
|
f'({resultdir}) or bootstrap root if applicable.'
|
||||||
|
err.MockGenericError.__init__(self, result_message)
|
||||||
|
|
||||||
|
elif exit_code == const.MockConstants.MOCK_EXIT_INVCONF:
|
||||||
|
# pylint: disable=non-parent-init-called
|
||||||
|
err.MockInvalidConfError.__init__(self, result_message)
|
||||||
|
|
||||||
|
elif exit_code == const.MockConstants.MOCK_EXIT_INVARCH:
|
||||||
|
# pylint: disable=non-parent-init-called
|
||||||
|
err.MockInvalidArchError.__init__(self, result_message)
|
||||||
|
|
||||||
|
elif exit_code in (const.MockConstants.MOCK_EXIT_DNF_ERROR,
|
||||||
|
const.MockConstants.MOCK_EXIT_EXTERNAL_DEP):
|
||||||
|
# pylint: disable=non-parent-init-called
|
||||||
|
err.MockDnfError.__init__(self, result_message)
|
||||||
|
|
||||||
|
elif exit_code == const.MockConstants.MOCK_EXIT_RESULTDIR_NOT_CREATED:
|
||||||
|
# pylint: disable=non-parent-init-called
|
||||||
|
err.MockResultdirError.__init__(self, result_message)
|
||||||
|
|
||||||
|
elif exit_code in (const.MockConstants.MOCK_EXIT_SIGHUP_RECEIVED,
|
||||||
|
const.MockConstants.MOCK_EXIT_SIGPIPE_RECEIVED,
|
||||||
|
const.MockConstants.MOCK_EXIT_SIGTERM_RECEIVED):
|
||||||
|
# pylint: disable=non-parent-init-called
|
||||||
|
err.MockSignalReceivedError.__init__(self, result_message)
|
||||||
|
|
||||||
|
else:
|
||||||
|
result_message = 'An unexpected mock error was caught. Review ' \
|
||||||
|
f'stdout/stderr or other logs to determine the issue. ' \
|
||||||
|
f'\n\nMock command: {mock_command}'
|
||||||
|
# pylint: disable=non-parent-init-called
|
||||||
|
err.MockUnexpectedError.__init__(self, result_message)
|
5
models/__init__.py
Normal file
5
models/__init__.py
Normal file
@ -0,0 +1,5 @@
|
|||||||
|
# -*-:python; coding:utf-8; -*-
|
||||||
|
# author: Louis Abel <label@rockylinux.org>
|
||||||
|
"""
|
||||||
|
Useful models. These may not be used and may be put elsewhere.
|
||||||
|
"""
|
8
modularity/__init__.py
Normal file
8
modularity/__init__.py
Normal file
@ -0,0 +1,8 @@
|
|||||||
|
# -*-:python; coding:utf-8; -*-
|
||||||
|
# author: Louis Abel <label@rockylinux.org>
|
||||||
|
"""
|
||||||
|
Mock and mock accessories
|
||||||
|
"""
|
||||||
|
|
||||||
|
# import all thingies here
|
||||||
|
from .util import GenericModule,ArtifactHandler,ModuleMangler
|
208
modularity/util.py
Normal file
208
modularity/util.py
Normal file
@ -0,0 +1,208 @@
|
|||||||
|
# -*- mode:python; coding:utf-8; -*-
|
||||||
|
# Louis Abel <label@rockylinux.org>
|
||||||
|
"""
|
||||||
|
Utility functions for Modularity
|
||||||
|
"""
|
||||||
|
import datetime
|
||||||
|
import hashlib
|
||||||
|
import gi
|
||||||
|
from pv2.util import error as err
|
||||||
|
from pv2.util import constants as const
|
||||||
|
from pv2.util import generic
|
||||||
|
from pv2.util import fileutil
|
||||||
|
|
||||||
|
gi.require_version('Modulemd', '2.0')
|
||||||
|
# Note: linter says this should be at the top. but then the linter says that
|
||||||
|
# everything else should be above it. it's fine here.
|
||||||
|
# pylint: disable=wrong-import-order,wrong-import-position
|
||||||
|
from gi.repository import Modulemd
|
||||||
|
|
||||||
|
__all__ = [
|
||||||
|
'GenericModuleHandler',
|
||||||
|
'ArtifactHandler',
|
||||||
|
'ModuleMangler'
|
||||||
|
]
|
||||||
|
|
||||||
|
class GenericModuleHandler:
|
||||||
|
"""
|
||||||
|
Generic module utility functions
|
||||||
|
"""
|
||||||
|
@staticmethod
|
||||||
|
def gen_stream_prefix(major: int, minor: int, patch: int) -> int:
|
||||||
|
"""
|
||||||
|
Generates a module stream prefix if one isn't provided by some other
|
||||||
|
means.
|
||||||
|
"""
|
||||||
|
major_version = str(major)
|
||||||
|
minor_version = str(minor) if len(str(minor)) > 1 else f'0{str(minor)}'
|
||||||
|
patch_version = str(patch) if len(str(patch)) > 1 else f'0{str(patch)}'
|
||||||
|
return int(f'{major_version}{minor_version}{patch_version}')
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def gen_stream_version(prefix: int) -> int:
|
||||||
|
"""
|
||||||
|
Generates a module stream version. Requires an initial prefix (like
|
||||||
|
90200 or similar).
|
||||||
|
"""
|
||||||
|
timestamp = datetime.datetime.utcnow().strftime('%Y%m%d%H%M%S')
|
||||||
|
return int(f'{prefix}{timestamp}')
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def gen_stream_dist_prefix(major: int, minor: int, patch: int) -> str:
|
||||||
|
"""
|
||||||
|
Generates a dist prefix (elX.Y.Z)
|
||||||
|
"""
|
||||||
|
major_version = str(major)
|
||||||
|
minor_version = str(minor)
|
||||||
|
patch_version = str(patch)
|
||||||
|
return f'el{major_version}.{minor_version}.{patch_version}'
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def gen_stream_dist_macro(
|
||||||
|
dist_prefix: str,
|
||||||
|
stream,
|
||||||
|
index=None,
|
||||||
|
scratch_build=False
|
||||||
|
) -> str:
|
||||||
|
"""
|
||||||
|
Generates a dist macro. stream should be a Modulemd.ModuleStreamV2 object
|
||||||
|
"""
|
||||||
|
# Fedora uses + it seems, while there are others who seem to use _.
|
||||||
|
# We'll just use +
|
||||||
|
# (Hopefully I did this better than in lazybuilder)
|
||||||
|
mod_prefix = 'module+'
|
||||||
|
|
||||||
|
# If this is a scratch build, change the initial prefix. Should be like
|
||||||
|
# what MBS does.
|
||||||
|
if scratch_build:
|
||||||
|
mod_prefix = 'scrmod+'
|
||||||
|
|
||||||
|
dist_string = '.'.join([
|
||||||
|
stream.get_module_name(),
|
||||||
|
stream.get_stream_name(),
|
||||||
|
str(stream.get_version()),
|
||||||
|
str(stream.get_context())
|
||||||
|
]
|
||||||
|
).encode('utf-8')
|
||||||
|
|
||||||
|
dist_hash = hashlib.sha1(dist_string, usedforsecurity=False).hexdigest()[:8]
|
||||||
|
template = f'.{mod_prefix}{dist_prefix}+{index}+{dist_hash}'
|
||||||
|
|
||||||
|
return template
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def gen_stream_build_deps():
|
||||||
|
"""
|
||||||
|
Gets a module stream's build deps
|
||||||
|
"""
|
||||||
|
return 'how'
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def gen_stream_runtime_deps():
|
||||||
|
"""
|
||||||
|
Gets a module stream's runtime deps
|
||||||
|
"""
|
||||||
|
return 'how'
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def gen_xmd_data(data: dict):
|
||||||
|
"""
|
||||||
|
Generates basic XMD information
|
||||||
|
"""
|
||||||
|
xmd = {'peridot': data}
|
||||||
|
return xmd
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def gen_module_defaults(name):
|
||||||
|
"""
|
||||||
|
Creates a modulemd default object
|
||||||
|
"""
|
||||||
|
return Modulemd.DefaultsV1.new(name)
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def merge_modules(module_a, module_b):
|
||||||
|
"""
|
||||||
|
Merges two module yamls together
|
||||||
|
"""
|
||||||
|
merge_object = Modulemd.ModuleIndexMerger.new()
|
||||||
|
merge_object.associate_index(module_b, 0)
|
||||||
|
merge_object.associate_index(module_a, 0)
|
||||||
|
return merge_object.resolve()
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def dump_to_yaml(stream):
|
||||||
|
"""
|
||||||
|
Dumps a module stream to YAML string
|
||||||
|
"""
|
||||||
|
module_index = Modulemd.ModuleIndex.new()
|
||||||
|
module_index.add_module_stream(stream)
|
||||||
|
return module_index.dump_to_string()
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def get_stream_metadata(module, stream):
|
||||||
|
"""
|
||||||
|
Gets a module's general information. Expects a Modulemd.Module object
|
||||||
|
and a Modulemd.ModuleStreamV2 object.
|
||||||
|
"""
|
||||||
|
module_dict = {
|
||||||
|
'name': stream.get_module_name(),
|
||||||
|
'stream': stream.get_stream_name(),
|
||||||
|
'arch': stream.get_arch(),
|
||||||
|
'version': stream.get_version(),
|
||||||
|
'context': stream.get_context(),
|
||||||
|
'summary': stream.get_summary(),
|
||||||
|
'is_default_stream': False,
|
||||||
|
'default_profiles': [],
|
||||||
|
'yaml_template': __class__.dump_to_yaml(stream)
|
||||||
|
}
|
||||||
|
defaults = module.get_defaults()
|
||||||
|
|
||||||
|
if not defaults:
|
||||||
|
return module_dict
|
||||||
|
|
||||||
|
default_stream = defaults.get_default_stream()
|
||||||
|
module_dict['is_default_stream'] = stream.get_stream_name() == default_stream
|
||||||
|
module_dict['default_profiles'] = defaults.get_default_profiles_for_stream(
|
||||||
|
stream.get_stream_name()
|
||||||
|
)
|
||||||
|
|
||||||
|
return module_dict
|
||||||
|
|
||||||
|
# pylint: disable=too-few-public-methods
|
||||||
|
class ArtifactHandler:
|
||||||
|
"""
|
||||||
|
Handles artifacts for a module. Typically RPMs
|
||||||
|
"""
|
||||||
|
# pylint: disable=too-many-arguments
|
||||||
|
def __init__(
|
||||||
|
self,
|
||||||
|
name: str,
|
||||||
|
version: str,
|
||||||
|
release: str,
|
||||||
|
arch: str,
|
||||||
|
epoch=None
|
||||||
|
):
|
||||||
|
"""
|
||||||
|
Initialize wrapper
|
||||||
|
"""
|
||||||
|
self.name = name
|
||||||
|
self.version = version
|
||||||
|
self.release = release
|
||||||
|
self.arch = arch
|
||||||
|
self.epoch = epoch
|
||||||
|
|
||||||
|
def return_artifact(self) -> str:
|
||||||
|
"""
|
||||||
|
Returns artifact string
|
||||||
|
"""
|
||||||
|
epoch = self.epoch if self.epoch else '0'
|
||||||
|
return f'{self.name}-{epoch}:{self.version}-{self.release}.{self.arch}'
|
||||||
|
|
||||||
|
class ModuleMangler:
|
||||||
|
"""
|
||||||
|
Specific functions for dealing with module yamls.
|
||||||
|
"""
|
||||||
|
def __init__(self):
|
||||||
|
"""
|
||||||
|
Initialize class
|
||||||
|
"""
|
5
peridotpb/__init__.py
Normal file
5
peridotpb/__init__.py
Normal file
@ -0,0 +1,5 @@
|
|||||||
|
# -*-:python; coding:utf-8; -*-
|
||||||
|
# author: Louis Abel <label@rockylinux.org>
|
||||||
|
"""
|
||||||
|
The peridotpb part of everything I suppose.
|
||||||
|
"""
|
3
util/README.md
Normal file
3
util/README.md
Normal file
@ -0,0 +1,3 @@
|
|||||||
|
# util module
|
||||||
|
|
||||||
|
This is for pv2 utilities.
|
5
util/__init__.py
Normal file
5
util/__init__.py
Normal file
@ -0,0 +1,5 @@
|
|||||||
|
# -*-:python; coding:utf-8; -*-
|
||||||
|
# author: Louis Abel <label@rockylinux.org>
|
||||||
|
"""
|
||||||
|
Start up the util module with no defaults
|
||||||
|
"""
|
30
util/color.py
Normal file
30
util/color.py
Normal file
@ -0,0 +1,30 @@
|
|||||||
|
# -*-:python; coding:utf-8; -*-
|
||||||
|
# author: Louis Abel <label@rockylinux.org>
|
||||||
|
# borrowed from empanadas
|
||||||
|
"""
|
||||||
|
Color classes
|
||||||
|
"""
|
||||||
|
# RPM utilities
|
||||||
|
__all__ = [
|
||||||
|
'Color',
|
||||||
|
]
|
||||||
|
|
||||||
|
# pylint: disable=too-few-public-methods
|
||||||
|
class Color:
|
||||||
|
"""
|
||||||
|
Supported colors
|
||||||
|
"""
|
||||||
|
RED = "\033[91m"
|
||||||
|
GREEN = "\033[92m"
|
||||||
|
PURPLE = "\033[95m"
|
||||||
|
CYAN = "\033[96m"
|
||||||
|
DARKCYAN = "\033[36m"
|
||||||
|
BLUE = "\033[94m"
|
||||||
|
YELLOW = "\033[93m"
|
||||||
|
UNDERLINE = "\033[4m"
|
||||||
|
BOLD = "\033[1m"
|
||||||
|
END = "\033[0m"
|
||||||
|
INFO = "[" + BOLD + GREEN + "INFO" + END + "] "
|
||||||
|
WARN = "[" + BOLD + YELLOW + "WARN" + END + "] "
|
||||||
|
FAIL = "[" + BOLD + RED + "FAIL" + END + "] "
|
||||||
|
STAT = "[" + BOLD + CYAN + "STAT" + END + "] "
|
168
util/constants.py
Normal file
168
util/constants.py
Normal file
@ -0,0 +1,168 @@
|
|||||||
|
# -*-:python; coding:utf-8; -*-
|
||||||
|
# author: Louis Abel <label@rockylinux.org>
|
||||||
|
"""
|
||||||
|
All constants
|
||||||
|
"""
|
||||||
|
__all__ = [
|
||||||
|
'RpmConstants',
|
||||||
|
'ErrorConstants',
|
||||||
|
'MockConstants'
|
||||||
|
]
|
||||||
|
|
||||||
|
# pylint: disable=too-few-public-methods
|
||||||
|
class RpmConstants:
|
||||||
|
"""
|
||||||
|
All RPM constants are here. These are used mainly in the rpm utility but
|
||||||
|
could be used elsewhere.
|
||||||
|
"""
|
||||||
|
RPM_HEADER_MAGIC = b'\xed\xab\xee\xdb'
|
||||||
|
RPM_TAG_HEADERSIGNATURES = 62
|
||||||
|
RPM_TAG_FILEDIGESTALGO = 5011
|
||||||
|
RPM_SIGTAG_DSA = 267
|
||||||
|
RPM_SIGTAG_RSA = 268
|
||||||
|
RPM_SIGTAG_PGP = 1002
|
||||||
|
RPM_SIGTAG_MD5 = 1004
|
||||||
|
RPM_SIGTAG_GPG = 1005
|
||||||
|
|
||||||
|
RPM_FILEDIGESTALGO_IDS = {
|
||||||
|
None: 'MD5',
|
||||||
|
1: 'MD5',
|
||||||
|
2: 'SHA1',
|
||||||
|
3: 'RIPEMD160',
|
||||||
|
8: 'SHA256',
|
||||||
|
9: 'SHA384',
|
||||||
|
10: 'SHA512',
|
||||||
|
11: 'SHA224'
|
||||||
|
}
|
||||||
|
|
||||||
|
MOCK_CLONE_DIRECTORY = '/var/peridot/peridot__rpmbuild_content'
|
||||||
|
|
||||||
|
# pylint: disable=too-few-public-methods
|
||||||
|
class ErrorConstants:
|
||||||
|
"""
|
||||||
|
All error codes as constants.
|
||||||
|
|
||||||
|
9000-9099: Generic Errors, this means not entirely specific to a process or
|
||||||
|
component.
|
||||||
|
|
||||||
|
9100-9199: Mock errors, any error that can occur in mock.
|
||||||
|
"""
|
||||||
|
# General errors
|
||||||
|
ERR_GENERAL = 9000
|
||||||
|
ERR_PROVIDED_VALUE = 9001
|
||||||
|
ERR_VALUE_EXISTS = 9002
|
||||||
|
ERR_MISSING_VALUE = 9003
|
||||||
|
ERR_CONFIGURATION = 9004
|
||||||
|
ERR_NOTFOUND = 9005
|
||||||
|
# Error in spec file
|
||||||
|
MOCK_ERR_SPEC = 9100
|
||||||
|
# Error trying to get dependencies for a build
|
||||||
|
MOCK_ERR_DEP = 9101
|
||||||
|
# Architecture is excluded - there should be no reason this appears normally.
|
||||||
|
MOCK_ERR_ARCH_EXCLUDED = 9102
|
||||||
|
# A build hung up during build
|
||||||
|
MOCK_ERR_BUILD_HUP = 9103
|
||||||
|
# Build ran out of disk space
|
||||||
|
MOCK_ERR_NO_SPACE = 9104
|
||||||
|
# Missing file error
|
||||||
|
MOCK_ERR_ENOENT = 9105
|
||||||
|
# Files were installed but not packaged
|
||||||
|
MOCK_ERR_UNPACKED_FILES = 9106
|
||||||
|
# Error in repository
|
||||||
|
MOCK_ERR_REPO = 9107
|
||||||
|
# Timeout
|
||||||
|
MOCK_ERR_ETIMEDOUT = 9108
|
||||||
|
# Changelog is not in chronological order
|
||||||
|
MOCK_ERR_CHLOG_CHRONO = 9109
|
||||||
|
# Invalid conf
|
||||||
|
MOCK_ERR_CONF_INVALID = 9110
|
||||||
|
# DNF Error
|
||||||
|
MOCK_ERR_DNF_ERROR = 9111
|
||||||
|
# Result dir generic error
|
||||||
|
MOCK_ERR_RESULTDIR_GENERIC = 9180
|
||||||
|
# Unexpected error
|
||||||
|
MOCK_ERR_UNEXPECTED = 9198
|
||||||
|
# Generic error
|
||||||
|
MOCK_ERR_GENERIC = 9199
|
||||||
|
|
||||||
|
# pylint: disable=too-few-public-methods
|
||||||
|
class MockConstants:
|
||||||
|
"""
|
||||||
|
All mock constants, usually for defaults
|
||||||
|
"""
|
||||||
|
# I'm aware this line is too long
|
||||||
|
MOCK_DEFAULT_CHROOT_BUILD_PKGS = [
|
||||||
|
'bash',
|
||||||
|
'bzip2',
|
||||||
|
'coreutils',
|
||||||
|
'cpio',
|
||||||
|
'diffutils',
|
||||||
|
'findutils',
|
||||||
|
'gawk',
|
||||||
|
'glibc-minimal-langpack',
|
||||||
|
'grep',
|
||||||
|
'gzip',
|
||||||
|
'info',
|
||||||
|
'make',
|
||||||
|
'patch',
|
||||||
|
'redhat-rpm-config',
|
||||||
|
'rpm-build',
|
||||||
|
'sed',
|
||||||
|
'shadow-utils',
|
||||||
|
'system-release',
|
||||||
|
'tar',
|
||||||
|
'unzip',
|
||||||
|
'util-linux',
|
||||||
|
'which',
|
||||||
|
'xz'
|
||||||
|
]
|
||||||
|
MOCK_DEFAULT_CHROOT_SRPM_PKGS = [
|
||||||
|
'bash',
|
||||||
|
"glibc-minimal-langpack",
|
||||||
|
"gnupg2",
|
||||||
|
"rpm-build",
|
||||||
|
"shadow-utils",
|
||||||
|
"system-release"
|
||||||
|
]
|
||||||
|
MOCK_DEFAULT_CHROOT_SETUP_CMD = 'install'
|
||||||
|
|
||||||
|
# Mock architecture related
|
||||||
|
MOCK_X86_64_LEGAL_ARCHES = ('x86_64',)
|
||||||
|
MOCK_I686_LEGAL_ARCHES = ('i386', 'i486', 'i586', 'i686', 'x86_64',)
|
||||||
|
MOCK_AARCH64_LEGAL_ARCHES = ('aarch64',)
|
||||||
|
MOCK_ARMV7HL_LEGAL_ARCHES = ('armv7hl',)
|
||||||
|
MOCK_PPC64LE_LEGAL_ARCHES = ('ppc64le',)
|
||||||
|
MOCK_S390X_LEGAL_ARCHES = ('s390x',)
|
||||||
|
MOCK_RISCV64_LEGAL_ARCHES = ('riscv64',)
|
||||||
|
# pylint: disable=line-too-long
|
||||||
|
MOCK_NOARCH_LEGAL_ARCHES = ('i386', 'i486', 'i586', 'i686', 'x86_64', 'aarch64', 'ppc64le', 's390x', 'noarch')
|
||||||
|
|
||||||
|
# Mock general config related
|
||||||
|
MOCK_DNF_BOOL_OPTIONS = ('assumeyes', 'best', 'enabled', 'gpgcheck',
|
||||||
|
'install_weak_deps', 'keepcache', 'module_hotfixes',
|
||||||
|
'obsoletes')
|
||||||
|
|
||||||
|
MOCK_DNF_STR_OPTIONS = ('debuglevel', 'retries', 'metadata_expire')
|
||||||
|
MOCK_DNF_LIST_OPTIONS = ('syslog_device', 'protected_packages')
|
||||||
|
MOCK_RPM_VERBOSITY = ('critical', 'debug', 'emergency', 'error', 'info', 'warn')
|
||||||
|
|
||||||
|
# Most mock error codes
|
||||||
|
MOCK_EXIT_SUCCESS = 0
|
||||||
|
MOCK_EXIT_ERROR = 1
|
||||||
|
MOCK_EXIT_SETUID = 2
|
||||||
|
MOCK_EXIT_INVCONF = 3
|
||||||
|
MOCK_EXIT_CMDLINE = 5
|
||||||
|
MOCK_EXIT_INVARCH = 6
|
||||||
|
MOCK_EXIT_BUILD_PROBLEM = 10
|
||||||
|
MOCK_EXIT_CMDTMOUT = 11
|
||||||
|
MOCK_EXIT_ERROR_IN_CHROOT = 20
|
||||||
|
MOCK_EXIT_DNF_ERROR = 30
|
||||||
|
MOCK_EXIT_EXTERNAL_DEP = 31
|
||||||
|
MOCK_EXIT_PKG_ERROR = 40
|
||||||
|
MOCK_EXIT_MOCK_CMDLINE = 50
|
||||||
|
MOCK_EXIT_BUILDROOT_LOCKED = 60
|
||||||
|
MOCK_EXIT_RESULTDIR_NOT_CREATED = 70
|
||||||
|
MOCK_EXIT_WEAK_DEP_NOT_INSTALLED = 120
|
||||||
|
MOCK_EXIT_SIGHUP_RECEIVED = 129
|
||||||
|
MOCK_EXIT_SIGPIPE_RECEIVED = 141
|
||||||
|
MOCK_EXIT_SIGTERM_RECEIVED = 143
|
13
util/cr.py
Normal file
13
util/cr.py
Normal file
@ -0,0 +1,13 @@
|
|||||||
|
"""
|
||||||
|
Parses repo metadata to get information. May be useful for getting general info
|
||||||
|
about a project's repository, like for generating a summary.
|
||||||
|
"""
|
||||||
|
|
||||||
|
#import os
|
||||||
|
#import createrepo_c as cr
|
||||||
|
|
||||||
|
__all__ = []
|
||||||
|
|
||||||
|
def _warning_cb(warning_type, message):
|
||||||
|
print(f"WARNING: {message}")
|
||||||
|
return True
|
118
util/error.py
Normal file
118
util/error.py
Normal file
@ -0,0 +1,118 @@
|
|||||||
|
# -*-:python; coding:utf-8; -*-
|
||||||
|
# author: Louis Abel <label@rockylinux.org>
|
||||||
|
"""
|
||||||
|
Generic Error Classes
|
||||||
|
"""
|
||||||
|
|
||||||
|
# needed imports
|
||||||
|
from pv2.util.constants import ErrorConstants as errconst
|
||||||
|
|
||||||
|
# list every error class that's enabled
|
||||||
|
|
||||||
|
__all__ = [
|
||||||
|
'GenericError',
|
||||||
|
'ProvidedValueError',
|
||||||
|
'ExistsValueError',
|
||||||
|
'MissingValueError',
|
||||||
|
'ConfigurationError',
|
||||||
|
'FileNotFound',
|
||||||
|
'MockGenericError',
|
||||||
|
'MockUnexpectedError',
|
||||||
|
'MockInvalidConfError',
|
||||||
|
'MockInvalidArchError',
|
||||||
|
'MockDnfError',
|
||||||
|
'MockResultdirError',
|
||||||
|
'MockSignalReceivedError',
|
||||||
|
]
|
||||||
|
|
||||||
|
|
||||||
|
# todo: find a way to logically use fault_code
|
||||||
|
class GenericError(Exception):
|
||||||
|
"""
|
||||||
|
Custom exceptions entrypoint
|
||||||
|
"""
|
||||||
|
fault_code = errconst.ERR_GENERAL
|
||||||
|
from_fault = False
|
||||||
|
def __str__(self):
|
||||||
|
try:
|
||||||
|
return str(self.args[0]['args'][0])
|
||||||
|
# pylint: disable=broad-exception-caught
|
||||||
|
except Exception:
|
||||||
|
try:
|
||||||
|
return str(self.args[0])
|
||||||
|
# pylint: disable=broad-exception-caught
|
||||||
|
except Exception:
|
||||||
|
return str(self.__dict__)
|
||||||
|
|
||||||
|
# Starting at this point is every error class that pv2 will deal with.
|
||||||
|
class ProvidedValueError(GenericError):
|
||||||
|
"""
|
||||||
|
What it says on the tin
|
||||||
|
"""
|
||||||
|
fault_code = errconst.ERR_PROVIDED_VALUE
|
||||||
|
|
||||||
|
class ExistsValueError(GenericError):
|
||||||
|
"""
|
||||||
|
Value being requested already exists
|
||||||
|
"""
|
||||||
|
fault_code = errconst.ERR_VALUE_EXISTS
|
||||||
|
|
||||||
|
class MissingValueError(GenericError):
|
||||||
|
"""
|
||||||
|
Value being requested already exists
|
||||||
|
"""
|
||||||
|
fault_code = errconst.ERR_MISSING_VALUE
|
||||||
|
|
||||||
|
class ConfigurationError(GenericError):
|
||||||
|
"""
|
||||||
|
Value being requested already exists
|
||||||
|
"""
|
||||||
|
fault_code = errconst.ERR_CONFIGURATION
|
||||||
|
|
||||||
|
class FileNotFound(GenericError):
|
||||||
|
"""
|
||||||
|
Value being requested already exists
|
||||||
|
"""
|
||||||
|
fault_code = errconst.ERR_NOTFOUND
|
||||||
|
|
||||||
|
class MockGenericError(GenericError):
|
||||||
|
"""
|
||||||
|
Mock error exceptions
|
||||||
|
"""
|
||||||
|
fault_code = errconst.MOCK_ERR_GENERIC
|
||||||
|
|
||||||
|
class MockUnexpectedError(MockGenericError):
|
||||||
|
"""
|
||||||
|
Mock (or the environment) experienced an unexpected error.
|
||||||
|
"""
|
||||||
|
fault_code = errconst.MOCK_ERR_UNEXPECTED
|
||||||
|
|
||||||
|
class MockInvalidConfError(MockGenericError):
|
||||||
|
"""
|
||||||
|
Mock (or the environment) experienced an error with the conf.
|
||||||
|
"""
|
||||||
|
fault_code = errconst.MOCK_ERR_CONF_INVALID
|
||||||
|
|
||||||
|
class MockInvalidArchError(MockGenericError):
|
||||||
|
"""
|
||||||
|
Mock (or the environment) didn't like the arch
|
||||||
|
"""
|
||||||
|
fault_code = errconst.MOCK_ERR_ARCH_EXCLUDED
|
||||||
|
|
||||||
|
class MockDnfError(MockGenericError):
|
||||||
|
"""
|
||||||
|
Mock (or the environment) had some kind of dnf error
|
||||||
|
"""
|
||||||
|
fault_code = errconst.MOCK_ERR_DNF_ERROR
|
||||||
|
|
||||||
|
class MockResultdirError(MockGenericError):
|
||||||
|
"""
|
||||||
|
Mock (or the environment) had some kind of error in the resultdir
|
||||||
|
"""
|
||||||
|
fault_code = errconst.MOCK_ERR_RESULTDIR_GENERIC
|
||||||
|
|
||||||
|
class MockSignalReceivedError(MockGenericError):
|
||||||
|
"""
|
||||||
|
Mock had a SIG received
|
||||||
|
"""
|
||||||
|
fault_code = errconst.MOCK_ERR_BUILD_HUP
|
55
util/fileutil.py
Normal file
55
util/fileutil.py
Normal file
@ -0,0 +1,55 @@
|
|||||||
|
"""
|
||||||
|
File functions
|
||||||
|
"""
|
||||||
|
|
||||||
|
import os
|
||||||
|
import hashlib
|
||||||
|
from pv2.util import error as err
|
||||||
|
|
||||||
|
# File utilities
|
||||||
|
__all__ = [
|
||||||
|
'filter_files',
|
||||||
|
'get_checksum'
|
||||||
|
]
|
||||||
|
|
||||||
|
def filter_files(directory_path: str, filter_filename: str) -> list:
|
||||||
|
"""
|
||||||
|
Filter out specified files
|
||||||
|
"""
|
||||||
|
# it's literally 101/100 ...
|
||||||
|
# pylint: disable=line-too-long
|
||||||
|
return_list = []
|
||||||
|
for file in os.listdir(directory_path):
|
||||||
|
if filter_filename(file):
|
||||||
|
return_list.append(os.path.join(directory_path, file))
|
||||||
|
|
||||||
|
return return_list
|
||||||
|
|
||||||
|
def get_checksum(file_path: str, hashtype: str = 'sha256') -> str:
|
||||||
|
"""
|
||||||
|
Generates a checksum from the provided path by doing things in chunks. This
|
||||||
|
reduces the time needed to make the hashes and avoids memory issues.
|
||||||
|
|
||||||
|
Borrowed from empanadas with some modifications
|
||||||
|
"""
|
||||||
|
# We shouldn't be using sha1 or md5.
|
||||||
|
if hashtype in ('sha', 'sha1', 'md5'):
|
||||||
|
raise err.ProvidedValueError(f'{hashtype} is not allowed.')
|
||||||
|
|
||||||
|
try:
|
||||||
|
checksum = hashlib.new(hashtype)
|
||||||
|
except ValueError as exc:
|
||||||
|
raise err.GenericError(f'hash type not available: {ValueError}') from exc
|
||||||
|
|
||||||
|
try:
|
||||||
|
with open(file_path, 'rb') as input_file:
|
||||||
|
while True:
|
||||||
|
chunk = input_file.read(8192)
|
||||||
|
if not chunk:
|
||||||
|
break
|
||||||
|
checksum.update(chunk)
|
||||||
|
|
||||||
|
input_file.close()
|
||||||
|
return checksum.hexdigest()
|
||||||
|
except IOError as exc:
|
||||||
|
raise err.GenericError(f'Could not open or process file {file_path}: {exc})')
|
78
util/generic.py
Normal file
78
util/generic.py
Normal file
@ -0,0 +1,78 @@
|
|||||||
|
"""
|
||||||
|
Generic functions
|
||||||
|
"""
|
||||||
|
import datetime
|
||||||
|
import hashlib
|
||||||
|
from pv2.util import error as err
|
||||||
|
|
||||||
|
# General utilities
|
||||||
|
__all__ = [
|
||||||
|
'ordered',
|
||||||
|
'conv_multibyte',
|
||||||
|
'to_unicode',
|
||||||
|
'convert_from_unix_time',
|
||||||
|
'trim_non_empty_string',
|
||||||
|
'gen_bool_option',
|
||||||
|
'generate_password_hash'
|
||||||
|
]
|
||||||
|
|
||||||
|
def to_unicode(string: str) -> str:
|
||||||
|
"""
|
||||||
|
Convert to unicode
|
||||||
|
"""
|
||||||
|
if isinstance(string, bytes):
|
||||||
|
return string.decode('utf8')
|
||||||
|
if isinstance(string, str):
|
||||||
|
return string
|
||||||
|
return str(string)
|
||||||
|
|
||||||
|
def conv_multibyte(data):
|
||||||
|
"""
|
||||||
|
Convert to multibytes
|
||||||
|
"""
|
||||||
|
potential_sum = 0
|
||||||
|
num = len(data)
|
||||||
|
for i in range(num):
|
||||||
|
potential_sum += data[i] << (8 * (num - i - 1))
|
||||||
|
return potential_sum
|
||||||
|
|
||||||
|
def ordered(data):
|
||||||
|
"""
|
||||||
|
Lazy ordering
|
||||||
|
"""
|
||||||
|
if isinstance(data, int):
|
||||||
|
return data
|
||||||
|
return ord(data)
|
||||||
|
|
||||||
|
def convert_from_unix_time(timestamp: int) -> str:
|
||||||
|
"""
|
||||||
|
Convert UNIX time to a timestamp
|
||||||
|
"""
|
||||||
|
return datetime.datetime.fromtimestamp(timestamp).strftime('%Y-%m-%dT%H:%M:%S')
|
||||||
|
|
||||||
|
def trim_non_empty_string(key, value) -> str:
|
||||||
|
"""
|
||||||
|
Verify that a given value is a non-empty string
|
||||||
|
"""
|
||||||
|
if not isinstance(value, str) or not value.strip():
|
||||||
|
raise err.ProvidedValueError(f'{key} must be a non-empty string')
|
||||||
|
return value
|
||||||
|
|
||||||
|
def gen_bool_option(value) -> str:
|
||||||
|
"""
|
||||||
|
Helps convert a value to how dnf and other configs interpret a boolean config value.
|
||||||
|
|
||||||
|
This should accept bool, string, or int and will act accordingly.
|
||||||
|
"""
|
||||||
|
return '1' if value and value != '0' else '0'
|
||||||
|
|
||||||
|
def generate_password_hash(password: str, salt: str, hashtype: str = 'sha256') -> str:
|
||||||
|
"""
|
||||||
|
Generates a password hash with a given hash type and salt
|
||||||
|
"""
|
||||||
|
if hashtype in ('sha', 'sha1', 'md5'):
|
||||||
|
raise err.ProvidedValueError(f'{hashtype} is not allowed.')
|
||||||
|
|
||||||
|
hasher = hashlib.new(hashtype)
|
||||||
|
hasher.update((salt + password).encode('utf-8'))
|
||||||
|
return str(hasher.hexdigest())
|
72
util/processor.py
Normal file
72
util/processor.py
Normal file
@ -0,0 +1,72 @@
|
|||||||
|
# -*-:python; coding:utf-8; -*-
|
||||||
|
# author: Louis Abel <label@rockylinux.org>
|
||||||
|
"""
|
||||||
|
Provides subprocess utilities
|
||||||
|
"""
|
||||||
|
|
||||||
|
import os
|
||||||
|
import sys
|
||||||
|
import subprocess
|
||||||
|
from pv2.util import error as err
|
||||||
|
|
||||||
|
# todo: remove python 3.6 checks. nodes won't be on el8.
|
||||||
|
|
||||||
|
def run_proc_foreground(command: list):
|
||||||
|
"""
|
||||||
|
Takes in the command in the form of a list and runs it via subprocess.
|
||||||
|
Everything should be in the foreground. The return is just for the exit
|
||||||
|
code.
|
||||||
|
"""
|
||||||
|
try:
|
||||||
|
processor = subprocess.run(args=command, check=False)
|
||||||
|
except Exception as exc:
|
||||||
|
raise err.GenericError(f'There was an error with your command: {exc}')
|
||||||
|
|
||||||
|
return processor
|
||||||
|
|
||||||
|
def run_proc_no_output(command: list):
|
||||||
|
"""
|
||||||
|
Output will be stored in stdout and stderr as needed.
|
||||||
|
"""
|
||||||
|
try:
|
||||||
|
if sys.version_info <= (3, 6):
|
||||||
|
processor = subprocess.run(args=command, check=False,
|
||||||
|
stdout=subprocess.PIPE,
|
||||||
|
stderr=subprocess.PIPE,
|
||||||
|
universal_newlines=True)
|
||||||
|
else:
|
||||||
|
processor = subprocess.run(args=command, check=False, capture_output=True,
|
||||||
|
text=True)
|
||||||
|
except Exception as exc:
|
||||||
|
raise err.GenericError(f'There was an error with your command: {exc}')
|
||||||
|
|
||||||
|
return processor
|
||||||
|
|
||||||
|
def popen_proc_no_output(command: list):
|
||||||
|
"""
|
||||||
|
This opens a process, but is non-blocking.
|
||||||
|
"""
|
||||||
|
try:
|
||||||
|
if sys.version_info <= (3, 6):
|
||||||
|
processor = subprocess.Popen(args=command, stdout=subprocess.PIPE,
|
||||||
|
universal_newlines=True)
|
||||||
|
else:
|
||||||
|
# pylint: disable=consider-using-with
|
||||||
|
processor = subprocess.Popen(args=command, stdout=subprocess.PIPE,
|
||||||
|
text=True)
|
||||||
|
except Exception as exc:
|
||||||
|
raise err.GenericError(f'There was an error with your command: {exc}')
|
||||||
|
|
||||||
|
return processor
|
||||||
|
|
||||||
|
def run_check_call(command: list) -> int:
|
||||||
|
"""
|
||||||
|
Runs subprocess check_call and returns an integer.
|
||||||
|
"""
|
||||||
|
env = os.environ
|
||||||
|
try:
|
||||||
|
subprocess.check_call(command, env=env)
|
||||||
|
except subprocess.CalledProcessError as exc:
|
||||||
|
sys.stderr.write(f'Run failed: {exc}\n')
|
||||||
|
return 1
|
||||||
|
return 0
|
357
util/rpmutil.py
Normal file
357
util/rpmutil.py
Normal file
@ -0,0 +1,357 @@
|
|||||||
|
# -*- mode:python; coding:utf-8; -*-
|
||||||
|
# Louis Abel <label@rockylinux.org>
|
||||||
|
"""
|
||||||
|
Utility functions for RPM's
|
||||||
|
"""
|
||||||
|
import re
|
||||||
|
import stat
|
||||||
|
import lxml.etree
|
||||||
|
from pv2.util import error as err
|
||||||
|
from pv2.util import generic
|
||||||
|
from pv2.util import processor
|
||||||
|
from pv2.util.constants import RpmConstants as rpmconst
|
||||||
|
|
||||||
|
# We should have the python rpm modules. Forcing `rpm` to be none should make
|
||||||
|
# that known to the admin that they did not setup their node correctly.
|
||||||
|
try:
|
||||||
|
import rpm
|
||||||
|
except ImportError:
|
||||||
|
rpm = None
|
||||||
|
|
||||||
|
__all__ = [
|
||||||
|
'is_debug_package',
|
||||||
|
'get_rpm_header',
|
||||||
|
'get_rpm_metadata_from_hdr',
|
||||||
|
'compare_rpms',
|
||||||
|
'is_rpm',
|
||||||
|
'get_files_from_package',
|
||||||
|
'get_exclu_from_package',
|
||||||
|
'get_rpm_hdr_size',
|
||||||
|
'split_rpm_by_header',
|
||||||
|
'get_all_rpm_header_keys'
|
||||||
|
]
|
||||||
|
|
||||||
|
# NOTES TO THOSE RUNNING PYLINT OR ANOTHER TOOL
|
||||||
|
#
|
||||||
|
# It is normal that your linter will say that "rpm" does not have some sort of
|
||||||
|
# RPMTAG member or otherwise. You will find when you run this module in normal
|
||||||
|
# circumstances, everything is returned as normal. You are free to ignore all
|
||||||
|
# linting errors.
|
||||||
|
|
||||||
|
def is_debug_package(file_name: str) -> bool:
|
||||||
|
"""
|
||||||
|
Quick utility to state if a package is a debug package
|
||||||
|
|
||||||
|
file_name: str, package filename
|
||||||
|
|
||||||
|
Returns: bool
|
||||||
|
"""
|
||||||
|
|
||||||
|
file_name_search_rpm_res = re.search(r'.*?\.rpm$', file_name, re.IGNORECASE)
|
||||||
|
file_name_search_srpm_res = re.search(r'.*?\.src\.rpm$', file_name, re.IGNORECASE)
|
||||||
|
|
||||||
|
if not file_name_search_rpm_res:
|
||||||
|
return False
|
||||||
|
if file_name_search_srpm_res:
|
||||||
|
return False
|
||||||
|
|
||||||
|
return bool(re.search(r'-debug(info|source)', file_name))
|
||||||
|
|
||||||
|
def get_rpm_header(file_name: str):
|
||||||
|
"""
|
||||||
|
Gets RPM header metadata. This is a vital component to getting RPM
|
||||||
|
information for usage later.
|
||||||
|
|
||||||
|
Returns: dict
|
||||||
|
"""
|
||||||
|
|
||||||
|
if rpm is None:
|
||||||
|
raise err.GenericError("You must have the rpm python bindings installed")
|
||||||
|
|
||||||
|
trans_set = rpm.TransactionSet()
|
||||||
|
# this is harmless.
|
||||||
|
# pylint: disable=protected-access
|
||||||
|
trans_set.setVSFlags(rpm._RPMVSF_NOSIGNATURES | rpm._RPMVSF_NODIGESTS)
|
||||||
|
with open(file_name, 'rb') as rpm_package:
|
||||||
|
hdr = trans_set.hdrFromFdno(rpm_package)
|
||||||
|
return hdr
|
||||||
|
|
||||||
|
# pylint: disable=too-many-locals
|
||||||
|
def get_rpm_metadata_from_hdr(hdr) -> dict:
|
||||||
|
"""
|
||||||
|
Asks for RPM header information and generates some basic metadata in the
|
||||||
|
form of a dict.
|
||||||
|
|
||||||
|
Currently the metadata returns the following information, and their
|
||||||
|
potential use cases:
|
||||||
|
|
||||||
|
* changelog_xml -> Provides the changelog, which could be parsed and
|
||||||
|
placed on to a build's summary page
|
||||||
|
|
||||||
|
* files -> List of all files in the package. Obtained from
|
||||||
|
get_files_from_package
|
||||||
|
|
||||||
|
* obsoletes -> Packages that this obsoletes
|
||||||
|
|
||||||
|
* provides -> Packages that this provides
|
||||||
|
|
||||||
|
* conflicts -> Packages that this conflicts with
|
||||||
|
|
||||||
|
* requires -> Packages that this requires
|
||||||
|
|
||||||
|
* vendor -> Package vendor
|
||||||
|
|
||||||
|
* buildhost -> Which system/container built it
|
||||||
|
|
||||||
|
* filetime -> When the package was built
|
||||||
|
|
||||||
|
* description -> Package description
|
||||||
|
|
||||||
|
* license -> License of the packaged software
|
||||||
|
|
||||||
|
* nvr -> NVR, excluding epoch and arch. This can be used as a build package
|
||||||
|
name, similar to how koji displays a particular build. For
|
||||||
|
example, bash-5.2.15-3.fc38
|
||||||
|
|
||||||
|
* nevra -> Full NEVRA. Could be used as a filing mechanism in a
|
||||||
|
database and/or could be used to be part of a list of what
|
||||||
|
architecture this package may belong to for a particular
|
||||||
|
build.
|
||||||
|
|
||||||
|
* name -> Package name
|
||||||
|
|
||||||
|
* version -> Package version
|
||||||
|
|
||||||
|
* release -> Package release
|
||||||
|
|
||||||
|
* epoch -> Package epoch
|
||||||
|
|
||||||
|
* arch -> Package arch
|
||||||
|
|
||||||
|
* archivesize -> Size of the archive
|
||||||
|
|
||||||
|
* packagesize -> Size of the package
|
||||||
|
"""
|
||||||
|
changelog_result = ''
|
||||||
|
header_data = hdr
|
||||||
|
file_stuff = get_files_from_package(header_data)
|
||||||
|
exclu_stuff = get_exclu_from_package(header_data)
|
||||||
|
change_logs = zip(
|
||||||
|
# pylint: disable=no-member
|
||||||
|
header_data[rpm.RPMTAG_CHANGELOGNAME],
|
||||||
|
header_data[rpm.RPMTAG_CHANGELOGTIME],
|
||||||
|
header_data[rpm.RPMTAG_CHANGELOGTEXT]
|
||||||
|
)
|
||||||
|
for name, time, text in reversed(list(change_logs)):
|
||||||
|
# I need to come back and address this
|
||||||
|
# pylint: disable=c-extension-no-member
|
||||||
|
change = lxml.etree.Element(
|
||||||
|
'changelog',
|
||||||
|
author=generic.to_unicode(name),
|
||||||
|
date=generic.to_unicode(time)
|
||||||
|
)
|
||||||
|
change.text = generic.to_unicode(text)
|
||||||
|
changelog_result += generic.to_unicode(lxml.etree.tostring(change, pretty_print=True))
|
||||||
|
|
||||||
|
# Source RPM's can be built on any given architecture, regardless of where
|
||||||
|
# they'll be built. There are also cases where an RPM may report some other
|
||||||
|
# architecture that may be multilib or not native to the system checking
|
||||||
|
# the headers. As a result, the RPM header may return erroneous information if we
|
||||||
|
# are trying to look at the metadata of a source package. So this is a hack
|
||||||
|
# to determine if we are dealing with a source package.
|
||||||
|
# pylint: disable=no-member
|
||||||
|
source_files = header_data[rpm.RPMTAG_SOURCE]
|
||||||
|
source_pkg = header_data[rpm.RPMTAG_SOURCERPM]
|
||||||
|
pkg_arch = generic.to_unicode(header_data[rpm.RPMTAG_ARCH])
|
||||||
|
|
||||||
|
if len(source_files) != 0 or not source_pkg:
|
||||||
|
pkg_arch = 'src'
|
||||||
|
|
||||||
|
# The NEVRA exhibits the same issue.
|
||||||
|
found_nevra = header_data[rpm.RPMTAG_NEVR] + '.' + pkg_arch
|
||||||
|
|
||||||
|
# This avoids epoch being None or 'None' in the dict.
|
||||||
|
found_epoch = header_data[rpm.RPMTAG_EPOCH]
|
||||||
|
if not found_epoch:
|
||||||
|
found_epoch = ''
|
||||||
|
|
||||||
|
metadata = {
|
||||||
|
'changelog_xml': changelog_result,
|
||||||
|
'files': file_stuff['file'],
|
||||||
|
'obsoletes': header_data[rpm.RPMTAG_OBSOLETENEVRS],
|
||||||
|
'provides': header_data[rpm.RPMTAG_PROVIDENEVRS],
|
||||||
|
'conflicts': header_data[rpm.RPMTAG_CONFLICTNEVRS],
|
||||||
|
'requires': header_data[rpm.RPMTAG_REQUIRENEVRS],
|
||||||
|
'vendor': generic.to_unicode(header_data[rpm.RPMTAG_VENDOR]),
|
||||||
|
'buildhost': generic.to_unicode(header_data[rpm.RPMTAG_BUILDHOST]),
|
||||||
|
'filetime': int(header_data[rpm.RPMTAG_BUILDTIME]),
|
||||||
|
'description': generic.to_unicode(header_data[rpm.RPMTAG_DESCRIPTION]),
|
||||||
|
'license': generic.to_unicode(header_data[rpm.RPMTAG_LICENSE]),
|
||||||
|
'exclusivearch': exclu_stuff['ExclusiveArch'],
|
||||||
|
'excludearch': exclu_stuff['ExcludeArch'],
|
||||||
|
'nvr': generic.to_unicode(header_data[rpm.RPMTAG_NEVR]),
|
||||||
|
'nevra': found_nevra,
|
||||||
|
'name': generic.to_unicode(header_data[rpm.RPMTAG_NAME]),
|
||||||
|
'version': generic.to_unicode(header_data[rpm.RPMTAG_VERSION]),
|
||||||
|
'release': generic.to_unicode(header_data[rpm.RPMTAG_RELEASE]),
|
||||||
|
'epoch': found_epoch,
|
||||||
|
'arch': pkg_arch,
|
||||||
|
}
|
||||||
|
for key, rpmkey, in (('archivesize', rpm.RPMTAG_ARCHIVESIZE),
|
||||||
|
('packagesize', rpm.RPMTAG_SIZE)):
|
||||||
|
value = header_data[rpmkey]
|
||||||
|
if value is not None:
|
||||||
|
value = int(value)
|
||||||
|
metadata[key] = value
|
||||||
|
return metadata
|
||||||
|
|
||||||
|
def compare_rpms(first_pkg, second_pkg) -> int:
|
||||||
|
"""
|
||||||
|
Compares package versions. Both arguments must be a dict.
|
||||||
|
|
||||||
|
Returns an int.
|
||||||
|
1 = first version is greater
|
||||||
|
0 = versions are equal
|
||||||
|
-1 = second version is greater
|
||||||
|
"""
|
||||||
|
# pylint: disable=no-member
|
||||||
|
return rpm.labelCompare(
|
||||||
|
(first_pkg['epoch'], first_pkg['version'], first_pkg['release']),
|
||||||
|
(second_pkg['epoch'], second_pkg['version'], second_pkg['release'])
|
||||||
|
)
|
||||||
|
|
||||||
|
def is_rpm(file_name: str, magic: bool = False) -> bool:
|
||||||
|
"""
|
||||||
|
Checks if a file is an RPM
|
||||||
|
"""
|
||||||
|
file_name_search_res = re.search(r'.*?\.rpm$', file_name, re.IGNORECASE)
|
||||||
|
if magic:
|
||||||
|
with open(file_name, 'rb') as file:
|
||||||
|
block = file.read(4)
|
||||||
|
file.close()
|
||||||
|
return bool(block == rpmconst.RPM_HEADER_MAGIC) and bool(file_name_search_res)
|
||||||
|
return bool(file_name_search_res)
|
||||||
|
|
||||||
|
def get_files_from_package(hdr) -> dict:
|
||||||
|
"""
|
||||||
|
hdr should be the header of the package.
|
||||||
|
|
||||||
|
returns a dict
|
||||||
|
"""
|
||||||
|
cache = {}
|
||||||
|
# pylint: disable=no-member
|
||||||
|
files = hdr[rpm.RPMTAG_FILENAMES]
|
||||||
|
fileflags = hdr[rpm.RPMTAG_FILEFLAGS]
|
||||||
|
filemodes = hdr[rpm.RPMTAG_FILEMODES]
|
||||||
|
filetuple = list(zip(files, filemodes, fileflags))
|
||||||
|
returned_files = {}
|
||||||
|
|
||||||
|
for (filename, mode, flag) in filetuple:
|
||||||
|
if mode is None or mode == '':
|
||||||
|
if 'file' not in returned_files:
|
||||||
|
returned_files['file'] = []
|
||||||
|
returned_files['file'].append(generic.to_unicode(filename))
|
||||||
|
continue
|
||||||
|
if mode not in cache:
|
||||||
|
cache[mode] = stat.S_ISDIR(mode)
|
||||||
|
filekey = 'file'
|
||||||
|
if cache[mode]:
|
||||||
|
filekey = 'dir'
|
||||||
|
elif flag is not None and (flag & 64):
|
||||||
|
filekey = 'ghost'
|
||||||
|
returned_files.setdefault(filekey, []).append(generic.to_unicode(filename))
|
||||||
|
return returned_files
|
||||||
|
|
||||||
|
def get_exclu_from_package(hdr) -> dict:
|
||||||
|
"""
|
||||||
|
Gets exclusivearch and excludedarch from an RPM's header. This mainly
|
||||||
|
applies to source packages.
|
||||||
|
"""
|
||||||
|
# pylint: disable=no-member
|
||||||
|
excluded_arches = hdr[rpm.RPMTAG_EXCLUDEARCH]
|
||||||
|
exclusive_arches = hdr[rpm.RPMTAG_EXCLUSIVEARCH]
|
||||||
|
|
||||||
|
exclu = {
|
||||||
|
'ExcludeArch': excluded_arches,
|
||||||
|
'ExclusiveArch': exclusive_arches
|
||||||
|
}
|
||||||
|
return exclu
|
||||||
|
|
||||||
|
def get_rpm_hdr_size(file_name: str, offset: int = 0, padding: bool = False) -> int:
|
||||||
|
"""
|
||||||
|
Returns the length of the rpm header in bytes
|
||||||
|
|
||||||
|
Accepts only a file name.
|
||||||
|
"""
|
||||||
|
with open(file_name, 'rb') as file_outer:
|
||||||
|
if offset is not None:
|
||||||
|
file_outer.seek(offset, 0)
|
||||||
|
magic = file_outer.read(4)
|
||||||
|
if magic != rpmconst.RPM_HEADER_MAGIC:
|
||||||
|
raise err.GenericError(f"RPM error: bad magic: {magic}")
|
||||||
|
|
||||||
|
# Skips magic, plus end of reserve (4 bytes)
|
||||||
|
file_outer.seek(offset + 8, 0)
|
||||||
|
|
||||||
|
data = [generic.ordered(x) for x in file_outer.read(8)]
|
||||||
|
start_length = generic.conv_multibyte(data[0:4])
|
||||||
|
end_length = generic.conv_multibyte(data[4:8])
|
||||||
|
|
||||||
|
hdrsize = 8 + 16 * start_length + end_length
|
||||||
|
|
||||||
|
if padding:
|
||||||
|
# signature headers are padded to a multiple of 8 bytes
|
||||||
|
hdrsize = hdrsize + (8 - (hdrsize % 8)) % 8
|
||||||
|
|
||||||
|
hdrsize = hdrsize + 8
|
||||||
|
file_outer.close()
|
||||||
|
|
||||||
|
return hdrsize
|
||||||
|
|
||||||
|
def split_rpm_by_header(hdr) -> tuple:
|
||||||
|
"""
|
||||||
|
Attempts to split an RPM name into parts. Relies on the RPM header. May
|
||||||
|
result in failures.
|
||||||
|
|
||||||
|
Only use this if you need simplicity.
|
||||||
|
|
||||||
|
Note: A package without an epoch turns None. We turn an empty string
|
||||||
|
instead.
|
||||||
|
|
||||||
|
Note: Splitting a source package will result in an erroneous "arch" field.
|
||||||
|
"""
|
||||||
|
|
||||||
|
# pylint: disable=no-member
|
||||||
|
name = hdr[rpm.RPMTAG_NAME]
|
||||||
|
version = hdr[rpm.RPMTAG_VERSION]
|
||||||
|
release = hdr[rpm.RPMTAG_RELEASE]
|
||||||
|
epoch = hdr[rpm.RPMTAG_EPOCH]
|
||||||
|
arch = hdr[rpm.RPMTAG_ARCH]
|
||||||
|
|
||||||
|
if not epoch:
|
||||||
|
epoch = ''
|
||||||
|
|
||||||
|
return name, version, release, epoch, arch
|
||||||
|
|
||||||
|
def get_all_rpm_header_keys(hdr) -> dict:
|
||||||
|
"""
|
||||||
|
Gets all applicable header keys from an RPM.
|
||||||
|
"""
|
||||||
|
returner = {}
|
||||||
|
# pylint: disable=no-member
|
||||||
|
fields = [rpm.tagnames[k] for k in hdr.keys()]
|
||||||
|
for field in fields:
|
||||||
|
hdr_key = getattr(rpm, f'RPMTAG_{field}', None)
|
||||||
|
returner[field] = hdr_key
|
||||||
|
|
||||||
|
return returner
|
||||||
|
|
||||||
|
def quick_bump(file_name: str, user: str, comment: str):
|
||||||
|
"""
|
||||||
|
Does a quick bump of a spec file. For dev purposes only.
|
||||||
|
|
||||||
|
Loosely borrowed from sig core toolkit mangler
|
||||||
|
"""
|
||||||
|
bumprel = ['rpmdev-bumpspec', '-D', '-u', user, '-c', comment, file_name]
|
||||||
|
success = processor.run_check_call(bumprel)
|
||||||
|
return success
|
5
util/srpmproc.py
Normal file
5
util/srpmproc.py
Normal file
@ -0,0 +1,5 @@
|
|||||||
|
# -*-:python; coding:utf-8; -*-
|
||||||
|
# author: Louis Abel <label@rockylinux.org>
|
||||||
|
"""
|
||||||
|
srpmproc handler. this may end up not being used at all.
|
||||||
|
"""
|
Loading…
Reference in New Issue
Block a user