Switch to black for formatting

This commit is contained in:
Mustafa Gezen 2022-08-04 05:01:32 +02:00
parent 9e91db39a1
commit 147aedf354
Signed by: mustafa
GPG Key ID: DCDF010D946438C1
7 changed files with 255 additions and 203 deletions

View File

@ -28,6 +28,7 @@
# POSSIBILITY OF SUCH DAMAGE.
import argparse
# noinspection PyPep8Naming
import xml.etree.ElementTree as ET
from xml.dom import minidom
@ -36,49 +37,47 @@ from group import Group, PackageReq, Environment, EnvGroup
def write_variant(groups, environments, categories, out):
root = ET.Element('comps')
root = ET.Element("comps")
for group in groups:
group_elem = ET.SubElement(root, 'group')
ET.SubElement(group_elem, 'id').text = group.id
group_elem = ET.SubElement(root, "group")
ET.SubElement(group_elem, "id").text = group.id
for lang in group.name:
name = ET.SubElement(group_elem, 'name')
name = ET.SubElement(group_elem, "name")
if lang != "":
name.set('xml:lang', lang)
name.set("xml:lang", lang)
name.text = group.name[lang]
for lang in group.description:
description = ET.SubElement(group_elem, 'description')
description = ET.SubElement(group_elem, "description")
if lang != "":
description.set('xml:lang', lang)
description.set("xml:lang", lang)
description.text = group.description[lang]
ET.SubElement(group_elem, 'default').text = str(group.default).lower()
ET.SubElement(group_elem, 'uservisible').text = str(
group.user_visible).lower()
package_list = ET.SubElement(group_elem, 'packagelist')
ET.SubElement(group_elem, "default").text = str(group.default).lower()
ET.SubElement(group_elem, "uservisible").text = str(group.user_visible).lower()
package_list = ET.SubElement(group_elem, "packagelist")
for package in group.packages:
package_elem = ET.SubElement(package_list, 'packagereq')
package_elem.set('type', package.type)
package_elem = ET.SubElement(package_list, "packagereq")
package_elem.set("type", package.type)
package_elem.text = package.name
for environment in environments:
env_elem = ET.SubElement(root, 'environment')
ET.SubElement(env_elem, 'id').text = environment.id
env_elem = ET.SubElement(root, "environment")
ET.SubElement(env_elem, "id").text = environment.id
for lang in environment.name:
name = ET.SubElement(env_elem, 'name')
name = ET.SubElement(env_elem, "name")
if lang != "":
name.set('xml:lang', lang)
name.set("xml:lang", lang)
name.text = environment.name[lang]
for lang in environment.description:
description = ET.SubElement(env_elem, 'description')
description = ET.SubElement(env_elem, "description")
if lang != "":
description.set('xml:lang', lang)
description.set("xml:lang", lang)
description.text = environment.description[lang]
ET.SubElement(env_elem, 'display_order').text = str(
environment.display_order)
group_list = ET.SubElement(env_elem, 'grouplist')
ET.SubElement(env_elem, "display_order").text = str(environment.display_order)
group_list = ET.SubElement(env_elem, "grouplist")
for group in environment.group_list:
ET.SubElement(group_list, 'groupid').text = group.name
option_list = ET.SubElement(env_elem, 'optionlist')
ET.SubElement(group_list, "groupid").text = group.name
option_list = ET.SubElement(env_elem, "optionlist")
for option in environment.option_list:
ET.SubElement(option_list, 'optionid').text = option.name
ET.SubElement(option_list, "optionid").text = option.name
for category_name in categories.keys():
category = categories[category_name]
new_group_list = []
@ -89,37 +88,41 @@ def write_variant(groups, environments, categories, out):
break
if len(new_group_list) == 0:
continue
category_elem = ET.SubElement(root, 'category')
ET.SubElement(category_elem, 'id').text = category_name
category_elem = ET.SubElement(root, "category")
ET.SubElement(category_elem, "id").text = category_name
for lang in category.name:
name = ET.SubElement(category_elem, 'name')
name = ET.SubElement(category_elem, "name")
if lang != "":
name.set('xml:lang', lang)
name.set("xml:lang", lang)
name.text = category.name[lang]
for lang in category.description:
description = ET.SubElement(category_elem, 'description')
description = ET.SubElement(category_elem, "description")
if lang != "":
description.set('xml:lang', lang)
description.set("xml:lang", lang)
description.text = category.description[lang]
ET.SubElement(category_elem, 'display_order').text = str(
category.display_order)
group_list = ET.SubElement(category_elem, 'grouplist')
ET.SubElement(category_elem, "display_order").text = str(category.display_order)
group_list = ET.SubElement(category_elem, "grouplist")
for group in new_group_list:
ET.SubElement(group_list, 'groupid').text = group.name
ET.ElementTree(root).write(out, encoding='utf-8', xml_declaration=False)
ET.SubElement(group_list, "groupid").text = group.name
ET.ElementTree(root).write(out, encoding="utf-8", xml_declaration=False)
with open(out, 'r') as f:
with open(out, "r") as f:
data = f.read()
with open(out, 'w') as f:
f.writelines("""<?xml version="1.0" encoding="UTF-8"?>
with open(out, "w") as f:
f.writelines(
"""<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE comps
PUBLIC '-//Red Hat, Inc.//DTD Comps info//EN'
'comps.dtd'>
""" + minidom.parseString(data).toprettyxml(indent=" ").replace('<?xml version="1.0" ?>\n', ''))
"""
+ minidom.parseString(data)
.toprettyxml(indent=" ")
.replace('<?xml version="1.0" ?>\n', "")
)
def main(comps_path: str, variants_path: str, output_path: str):
default_arches = ['x86_64', 'aarch64', 'ppc64le', 's390x']
default_arches = ["x86_64", "aarch64", "ppc64le", "s390x"]
variants = {}
environments = {}
categories = {}
@ -127,53 +130,55 @@ def main(comps_path: str, variants_path: str, output_path: str):
tree = ET.parse(comps_path)
root = tree.getroot()
for gchild in root:
if gchild.tag == 'group':
if gchild.tag == "group":
group_name = {}
group_desc = {}
group_id = ''
group_id = ""
is_default = False
is_visible = False
variant = ''
variant = ""
package_list_xml = None
if 'variant' in gchild.attrib:
variant = gchild.attrib['variant']
if 'arch' in gchild.attrib:
arches = gchild.attrib['arch'].split(',')
if "variant" in gchild.attrib:
variant = gchild.attrib["variant"]
if "arch" in gchild.attrib:
arches = gchild.attrib["arch"].split(",")
else:
arches = default_arches
for gattr in gchild:
if gattr.tag == 'id':
if gattr.tag == "id":
group_id = gattr.text
elif gattr.tag == 'name':
if '{http://www.w3.org/XML/1998/namespace}lang' in gattr.attrib:
group_name[gattr.attrib[
'{http://www.w3.org/XML/1998/namespace}lang']] = gattr.text
elif gattr.tag == "name":
if "{http://www.w3.org/XML/1998/namespace}lang" in gattr.attrib:
group_name[
gattr.attrib["{http://www.w3.org/XML/1998/namespace}lang"]
] = gattr.text
else:
group_name[""] = gattr.text
elif gattr.tag == 'description':
if '{http://www.w3.org/XML/1998/namespace}lang' in gattr.attrib:
group_desc[gattr.attrib[
'{http://www.w3.org/XML/1998/namespace}lang']] = gattr.text
elif gattr.tag == "description":
if "{http://www.w3.org/XML/1998/namespace}lang" in gattr.attrib:
group_desc[
gattr.attrib["{http://www.w3.org/XML/1998/namespace}lang"]
] = gattr.text
else:
group_desc[""] = gattr.text
elif gattr.tag == 'default':
is_default = gattr.text == 'true'
elif gattr.tag == 'uservisible':
is_visible = gattr.text == 'true'
elif gattr.tag == 'packagelist':
elif gattr.tag == "default":
is_default = gattr.text == "true"
elif gattr.tag == "uservisible":
is_visible = gattr.text == "true"
elif gattr.tag == "packagelist":
package_list_xml = gattr
package_list = {}
if variant != '':
if variant != "":
package_list[variant] = {}
for reqxml in package_list_xml:
req_variant = variant
req_type = 'default'
if 'variant' in reqxml.attrib:
req_variant = reqxml.attrib['variant']
if 'type' in reqxml.attrib:
req_type = reqxml.attrib['type']
if 'arch' in reqxml.attrib:
req_arches = reqxml.attrib['arch'].split(',')
req_type = "default"
if "variant" in reqxml.attrib:
req_variant = reqxml.attrib["variant"]
if "type" in reqxml.attrib:
req_type = reqxml.attrib["type"]
if "arch" in reqxml.attrib:
req_arches = reqxml.attrib["arch"].split(",")
else:
req_arches = arches
if req_variant not in package_list:
@ -182,7 +187,8 @@ def main(comps_path: str, variants_path: str, output_path: str):
if arch not in package_list[req_variant]:
package_list[req_variant][arch] = []
package_list[req_variant][arch].append(
PackageReq(reqxml.text, req_type, req_arches))
PackageReq(reqxml.text, req_type, req_arches)
)
for variant in package_list:
if variant not in variants:
variants[variant] = {}
@ -191,59 +197,62 @@ def main(comps_path: str, variants_path: str, output_path: str):
package_list[variant][arch] = []
if group_id not in variants[variant]:
variants[variant][group_id] = {}
variants[variant][group_id][arch] = Group(group_id,
variants[variant][group_id][arch] = Group(
group_id,
group_name,
group_desc,
is_default,
is_visible,
package_list[
variant][
arch])
elif gchild.tag == 'environment' or gchild.tag == 'category':
package_list[variant][arch],
)
elif gchild.tag == "environment" or gchild.tag == "category":
env_name = {}
env_desc = {}
env_id = ''
env_id = ""
display_order = 0
group_list = []
option_list = []
for gattr in gchild:
if gattr.tag == 'id':
if gattr.tag == "id":
env_id = gattr.text
elif gattr.tag == 'name':
if '{http://www.w3.org/XML/1998/namespace}lang' in gattr.attrib:
env_name[gattr.attrib[
'{http://www.w3.org/XML/1998/namespace}lang']] = gattr.text
elif gattr.tag == "name":
if "{http://www.w3.org/XML/1998/namespace}lang" in gattr.attrib:
env_name[
gattr.attrib["{http://www.w3.org/XML/1998/namespace}lang"]
] = gattr.text
else:
env_name[""] = gattr.text
elif gattr.tag == 'description':
if '{http://www.w3.org/XML/1998/namespace}lang' in gattr.attrib:
env_desc[gattr.attrib[
'{http://www.w3.org/XML/1998/namespace}lang']] = gattr.text
elif gattr.tag == "description":
if "{http://www.w3.org/XML/1998/namespace}lang" in gattr.attrib:
env_desc[
gattr.attrib["{http://www.w3.org/XML/1998/namespace}lang"]
] = gattr.text
else:
env_desc[""] = gattr.text
elif gattr.tag == 'display_order':
elif gattr.tag == "display_order":
display_order = gattr.text
elif gattr.tag == 'grouplist':
elif gattr.tag == "grouplist":
for group in gattr:
if 'arch' in group.attrib:
arches = group.attrib['arch'].split(',')
if "arch" in group.attrib:
arches = group.attrib["arch"].split(",")
else:
arches = default_arches
group_list.append(EnvGroup(group.text, arches))
elif gattr.tag == 'optionlist':
elif gattr.tag == "optionlist":
for group in gattr:
if 'arch' in group.attrib:
arches = group.attrib['arch'].split(',')
if "arch" in group.attrib:
arches = group.attrib["arch"].split(",")
else:
arches = default_arches
option_list.append(EnvGroup(group.text, arches))
new_env = Environment(env_id, env_name, env_desc, display_order,
group_list, option_list)
new_env = Environment(
env_id, env_name, env_desc, display_order, group_list, option_list
)
dictmap = categories
if gchild.tag == 'environment':
if gchild.tag == "environment":
dictmap = environments
if 'arch' in gchild.attrib:
arches = gchild.attrib['arch'].split(',')
if "arch" in gchild.attrib:
arches = gchild.attrib["arch"].split(",")
else:
arches = default_arches
for arch in arches:
@ -262,18 +271,18 @@ def main(comps_path: str, variants_path: str, output_path: str):
environment_arch_index = {}
pungi_variants_tree = ET.parse(variants_path).getroot()
for pungi_variant in pungi_variants_tree:
if pungi_variant.tag == 'variant':
if pungi_variant.attrib['type'] != 'variant':
if pungi_variant.tag == "variant":
if pungi_variant.attrib["type"] != "variant":
continue
arches = []
groups = {}
n_environments = {}
variant_id = pungi_variant.attrib['id']
variant_id = pungi_variant.attrib["id"]
for child in pungi_variant:
if child.tag == 'arches':
if child.tag == "arches":
for arch in child:
arches.append(arch.text)
elif child.tag == 'groups':
elif child.tag == "groups":
for group in child:
groupbase = variants[""]
if variant_id in variants:
@ -284,17 +293,20 @@ def main(comps_path: str, variants_path: str, output_path: str):
for arch_group in groupind.keys():
if arch_group not in groups:
groups[arch_group] = []
if 'default' in group.attrib:
groupind[arch_group].default = group.attrib['default'] == 'true'
if "default" in group.attrib:
groupind[arch_group].default = (
group.attrib["default"] == "true"
)
groups[arch_group].append(groupind[arch_group])
elif child.tag == 'environments':
elif child.tag == "environments":
for environment in child:
envind = environment_id_index[environment.text]
for arch_environment in envind.keys():
if arch_environment not in n_environments:
n_environments[arch_environment] = []
n_environments[arch_environment].append(
envind[arch_environment])
envind[arch_environment]
)
for arch in arches:
if arch in groups:
if arch not in variant_arch_index:
@ -308,25 +320,29 @@ def main(comps_path: str, variants_path: str, output_path: str):
if variant_id not in environment_arch_index[arch]:
environment_arch_index[arch][variant_id] = []
environment_arch_index[arch][variant_id].extend(
n_environments[arch])
n_environments[arch]
)
for arch in variant_arch_index.keys():
for variant in variant_arch_index[arch].keys():
write_variant(variant_arch_index[arch][variant] if variant in
variant_arch_index[
arch] else [],
environment_arch_index[arch][variant] if variant in
environment_arch_index[
arch] else [],
write_variant(
variant_arch_index[arch][variant]
if variant in variant_arch_index[arch]
else [],
environment_arch_index[arch][variant]
if variant in environment_arch_index[arch]
else [],
categories[arch].copy(),
f'{output_path}/{variant}-{arch}.xml')
f"{output_path}/{variant}-{arch}.xml",
)
if __name__ == '__main__':
if __name__ == "__main__":
parser = argparse.ArgumentParser(
description='Convert comps to Peridot compatible configuration.')
parser.add_argument('--comps-path', type=str, required=True)
parser.add_argument('--variants-path', type=str, required=True)
parser.add_argument('--output-path', type=str, default=".")
description="Convert comps to Peridot compatible configuration."
)
parser.add_argument("--comps-path", type=str, required=True)
parser.add_argument("--variants-path", type=str, required=True)
parser.add_argument("--output-path", type=str, default=".")
args = parser.parse_args()
main(args.comps_path, args.variants_path, args.output_path)

View File

@ -38,8 +38,12 @@ class PeridotCatalogSyncPackageType(str, Enum):
PACKAGE_TYPE_MODULE_FORK = "PACKAGE_TYPE_MODULE_FORK"
PACKAGE_TYPE_MODULE_FORK_COMPONENT = "PACKAGE_TYPE_MODULE_FORK_COMPONENT"
PACKAGE_TYPE_NORMAL_FORK_MODULE = "PACKAGE_TYPE_NORMAL_FORK_MODULE"
PACKAGE_TYPE_NORMAL_FORK_MODULE_COMPONENT = "PACKAGE_TYPE_NORMAL_FORK_MODULE_COMPONENT"
PACKAGE_TYPE_MODULE_FORK_MODULE_COMPONENT = "PACKAGE_TYPE_MODULE_FORK_MODULE_COMPONENT"
PACKAGE_TYPE_NORMAL_FORK_MODULE_COMPONENT = (
"PACKAGE_TYPE_NORMAL_FORK_MODULE_COMPONENT"
)
PACKAGE_TYPE_MODULE_FORK_MODULE_COMPONENT = (
"PACKAGE_TYPE_MODULE_FORK_MODULE_COMPONENT"
)
@dataclass
@ -49,12 +53,12 @@ class PeridotCatalogSyncRepository:
multilib: list[str]
def include_filter_to_prototxt(self):
return '\n' + '\n'.join(
[f" include_filter: \"{f}\"" for f in self.include_filter])
return "\n" + "\n".join(
[f' include_filter: "{f}"' for f in self.include_filter]
)
def multilib_to_prototxt(self):
return '\n' + '\n'.join(
[f" multilib: \"{f}\"" for f in self.multilib])
return "\n" + "\n".join([f' multilib: "{f}"' for f in self.multilib])
@dataclass
@ -65,19 +69,26 @@ class PeridotCatalogSyncPackage:
repositories: list[PeridotCatalogSyncRepository]
def mc_to_prototxt(self):
return '\n' + '\n'.join(
[f" module_component: \"{component}\"" for component in
self.module_components])
return "\n" + "\n".join(
[
f' module_component: "{component}"'
for component in self.module_components
]
)
def repos_to_prototxt(self):
return '\n'.join(
[f""" repository {{
return "\n".join(
[
f""" repository {{
name: \"{repo.name}\"{
repo.include_filter_to_prototxt() if repo.include_filter else ""
}{
repo.multilib_to_prototxt() if repo.multilib else ""
}
}}""" for repo in self.repositories])
}}"""
for repo in self.repositories
]
)
class PeridotCatalogSync:
@ -91,42 +102,53 @@ class PeridotCatalogSync:
self.packages.append(package)
def additional_multilib_to_prototxt(self):
return '\n'.join(
[f"additional_multilib: \"{f}\"" for f in
self.additional_multilib])
return "\n".join(
[f'additional_multilib: "{f}"' for f in self.additional_multilib]
)
def exclude_multilib_filter_to_prototxt(self):
return '\n' + '\n'.join(
[f"exclude_multilib_filter: \"{f}\"" for f in
self.exclude_multilib_filter])
return "\n" + "\n".join(
[f'exclude_multilib_filter: "{f}"' for f in self.exclude_multilib_filter]
)
def filter_arch_to_prototxt(self, arch: dict):
nl = '\n'
nl = "\n"
glob_match = {}
for k, v in arch.items():
glob_match[k] = [f" glob_match: \"{f}\"" for f in v]
glob_match[k] = [f' glob_match: "{f}"' for f in v]
for k in glob_match.keys():
if len(glob_match[k]) > 0:
glob_match[k][0] = '\n' + glob_match[k][0]
return '\n'.join([f""" arch {{
glob_match[k][0] = "\n" + glob_match[k][0]
return "\n".join(
[
f""" arch {{
key: \"{f}\"{nl.join(glob_match[f])}
}}""" for f in arch.keys()])
}}"""
for f in arch.keys()
]
)
def exclude_filter_to_prototxt(self):
return '\n' + '\n'.join(
[f"""exclude_filter {{
return "\n" + "\n".join(
[
f"""exclude_filter {{
repo_match: \"{f[0]}\"
{self.filter_arch_to_prototxt(f[1])}
}}""" for f in
self.exclude_filter])
}}"""
for f in self.exclude_filter
]
)
def include_filter_to_prototxt(self):
return '\n' + '\n'.join(
[f"""include_filter {{
return "\n" + "\n".join(
[
f"""include_filter {{
repo_match: \"{f[0]}\"
{self.filter_arch_to_prototxt(f[1])}
}}""" for f in
self.include_filter])
}}"""
for f in self.include_filter
]
)
def to_prototxt(self):
ret = f"""# kind: resf.peridot.v1.CatalogSync

View File

@ -32,8 +32,12 @@ import os
import kobo.conf
from catalog import PeridotCatalogSync, PeridotCatalogSyncPackage, \
PeridotCatalogSyncPackageType, PeridotCatalogSyncRepository
from catalog import (
PeridotCatalogSync,
PeridotCatalogSyncPackage,
PeridotCatalogSyncPackageType,
PeridotCatalogSyncRepository,
)
from scm import SCM
@ -51,10 +55,10 @@ def main(pungi_conf_path: str, output_path: str):
catalog = PeridotCatalogSync()
# Set multilib filters
catalog.additional_multilib.extend(
list(conf.get("multilib_whitelist").values())[0])
catalog.additional_multilib.extend(list(conf.get("multilib_whitelist").values())[0])
catalog.exclude_multilib_filter.extend(
list(conf.get("multilib_blacklist").values())[0])
list(conf.get("multilib_blacklist").values())[0]
)
# Set additional packages/filters
catalog.exclude_filter.extend(conf.get("filter_packages"))
@ -73,31 +77,40 @@ def main(pungi_conf_path: str, output_path: str):
if package not in package_index:
package_index[package] = {}
if repo not in package_index[package]:
package_index[package][repo] = {"include_filter": [],
"multilib": []}
package_index[package][repo] = {
"include_filter": [],
"multilib": [],
}
na_list = gpjson[repo][arch][package]
for na in na_list:
splitted = na.split(".")
arch_package = splitted[len(splitted) - 1]
if arch != arch_package and arch_package != "noarch":
if arch not in package_index[package][repo]["multilib"]:
package_index[package][repo]["multilib"].append(
arch)
package_index[package][repo]["multilib"].append(arch)
if na not in package_index[package][repo]["include_filter"]:
package_index[package][repo]["include_filter"].append(
na)
package_index[package][repo]["include_filter"].append(na)
arch_specific_excludes = {}
na_index = {}
for pkg in package_index.keys():
for repo in package_index[pkg].keys():
na_list = list(filter(lambda x: x.endswith('.noarch'), package_index[pkg][repo]["include_filter"]))
na_list = list(
filter(
lambda x: x.endswith(".noarch"),
package_index[pkg][repo]["include_filter"],
)
)
if not na_list:
continue
exclude_arches = {}
for na in na_list:
for arch in all_arches:
if arch not in gpjson[repo] or pkg not in gpjson[repo][arch] or na not in gpjson[repo][arch][pkg]:
if (
arch not in gpjson[repo]
or pkg not in gpjson[repo][arch]
or na not in gpjson[repo][arch][pkg]
):
if na not in exclude_arches:
exclude_arches[na] = []
exclude_arches[na].append(arch)
@ -132,7 +145,7 @@ def main(pungi_conf_path: str, output_path: str):
if arch not in filter_tuple:
filter_tuple[arch] = []
for na in repo_arch_index[repo][arch]:
na = na.removesuffix('.noarch')
na = na.removesuffix(".noarch")
if na not in filter_tuple[arch]:
filter_tuple[arch].append(na)
catalog.exclude_filter.append((repo_key, filter_tuple))
@ -141,12 +154,18 @@ def main(pungi_conf_path: str, output_path: str):
catalog.add_package(
PeridotCatalogSyncPackage(
package,
PeridotCatalogSyncPackageType.PACKAGE_TYPE_NORMAL_FORK if not package.startswith("rocky-") else PeridotCatalogSyncPackageType.PACKAGE_TYPE_NORMAL_SRC,
PeridotCatalogSyncPackageType.PACKAGE_TYPE_NORMAL_FORK
if not package.startswith("rocky-")
else PeridotCatalogSyncPackageType.PACKAGE_TYPE_NORMAL_SRC,
[],
[PeridotCatalogSyncRepository(x, package_index[package][x][
"include_filter"], package_index[package][x]["multilib"])
for
x in package_index[package].keys()]
[
PeridotCatalogSyncRepository(
x,
package_index[package][x]["include_filter"],
package_index[package][x]["multilib"],
)
for x in package_index[package].keys()
],
)
)
@ -159,9 +178,9 @@ def main(pungi_conf_path: str, output_path: str):
if __name__ == "__main__":
parser = argparse.ArgumentParser(
description='Convert Pungi configuration to Peridot compatible '
'catalogs.')
parser.add_argument('--pungi-conf-path', type=str, required=True)
parser.add_argument('--output-path', type=str, default="catalog.cfg")
description="Convert Pungi configuration to Peridot compatible " "catalogs."
)
parser.add_argument("--pungi-conf-path", type=str, required=True)
parser.add_argument("--output-path", type=str, default="catalog.cfg")
args = parser.parse_args()
main(args.pungi_conf_path, args.output_path)

View File

@ -27,16 +27,16 @@
# ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
# POSSIBILITY OF SUCH DAMAGE.
BASE_URL = 'https://peridot-api.build.resf.org/v1'
PROJECT_ID_PROD = '55b17281-bc54-4929-8aca-a8a11d628738'
BASE_URL = "https://peridot-api.build.resf.org/v1"
PROJECT_ID_PROD = "55b17281-bc54-4929-8aca-a8a11d628738"
def construct_url(path, project_id=PROJECT_ID_PROD):
return f'{BASE_URL}/projects/{project_id}{path}'
return f"{BASE_URL}/projects/{project_id}{path}"
def build_batches_url(batch_type, task_id, page, status,
project_id=PROJECT_ID_PROD):
def build_batches_url(batch_type, task_id, page, status, project_id=PROJECT_ID_PROD):
return construct_url(
f'/{batch_type}_batches/{task_id}?page={page}&limit=100&filter.status={status}',
project_id)
f"/{batch_type}_batches/{task_id}?page={page}&limit=100&filter.status={status}",
project_id,
)

View File

@ -38,7 +38,7 @@ from common import build_batches_url
def get_batch(batch_type, task_id, status, page):
r = requests.get(build_batches_url(batch_type, task_id, page, status))
return r.json()[f'{batch_type}s']
return r.json()[f"{batch_type}s"]
def process_batch(batch_type, task_id, status):
@ -52,18 +52,16 @@ def process_batch(batch_type, task_id, status):
page = page + 1
if __name__ == '__main__':
if __name__ == "__main__":
batch_type = sys.argv[1]
task_id = sys.argv[2]
batch_items = process_batch(batch_type, task_id, 4)
req = {}
key = f'{batch_type}s'
key = f"{batch_type}s"
req[key] = []
for item in batch_items:
req[key].append({
'package_name': item['name']
})
req[key].append({"package_name": item["name"]})
print(json.dumps(req))

View File

@ -38,13 +38,14 @@ from common import construct_url
def chunks(lst, n):
for i in range(0, len(lst), n):
yield lst[i:i + n]
yield lst[i : i + n]
def get_packages(page):
r = requests.get(
construct_url(f'/packages?limit=100&page={page}&filters.no_builds=1'))
return r.json()['packages']
construct_url(f"/packages?limit=100&page={page}&filters.no_builds=1")
)
return r.json()["packages"]
def process_packages():
@ -58,13 +59,11 @@ def process_packages():
page = page + 1
if __name__ == '__main__':
if __name__ == "__main__":
batch_items = process_packages()
builds = []
for item in batch_items:
builds.append({
'package_name': item['name']
})
builds.append({"package_name": item["name"]})
for chunk in chunks(builds, 400):
print(json.dumps({"builds": chunk}))

View File

@ -30,15 +30,13 @@
import sys
import json
if __name__ == '__main__':
if __name__ == "__main__":
build_type = sys.argv[1]
key = f'{build_type}s'
key = f"{build_type}s"
req = {}
req[key] = []
for line in sys.stdin:
req[key].append({
'package_name': line.strip()
})
req[key].append({"package_name": line.strip()})
print(json.dumps(req))