Complete Yocto mirror with license table for TQMa6UL (2038-compliance)

- 264 license table entries with exact download URLs (224/264 resolved)
- Complete sources/ directory with all BitBake recipes
- Build configuration: tqma6ul-multi-mba6ulx, spaetzle (musl)
- Full traceability for Softwarefreigabeantrag
- GCC 13.4.0, Linux 6.6.102, U-Boot 2023.04, musl 1.2.4
- License distribution: GPL-2.0 (24), MIT (23), GPL-2.0+ (18), BSD-3 (16)
This commit is contained in:
Siggi (OpenClaw Agent)
2026-03-01 20:58:18 +00:00
commit 16accb6b24
15086 changed files with 1292356 additions and 0 deletions

View File

@@ -0,0 +1,14 @@
# Running build tests
These tests are to test the running of builds and the data produced by the builds.
Your oe build environment must be sourced/initialised for these tests to run.
The simplest way to run the tests are the following commands:
$ . oe-init-build-env
$ cd bitbake/lib/toaster/ # path my vary but this is into toaster's directory
$ DJANGO_SETTINGS_MODULE='toastermain.settings_test' ./manage.py test tests.builds
Optional environment variables:
- TOASTER_DIR (where toaster keeps it's artifacts)
- TOASTER_CONF a path to the toasterconf.json file. This will need to be set if you don't execute the tests from toaster's own directory.

View File

@@ -0,0 +1,166 @@
#! /usr/bin/env python3
#
# BitBake Toaster Implementation
#
# Copyright (C) 2016 Intel Corporation
#
# SPDX-License-Identifier: GPL-2.0-only
#
import os
import sys
import time
import unittest
from orm.models import Project, Release, ProjectTarget, Build, ProjectVariable
from bldcontrol.models import BuildEnvironment
from bldcontrol.management.commands.runbuilds import Command\
as RunBuildsCommand
from django.core.management import call_command
import subprocess
import logging
logger = logging.getLogger("toaster")
# We use unittest.TestCase instead of django.test.TestCase because we don't
# want to wrap everything in a database transaction as an external process
# (bitbake needs access to the database)
def load_build_environment():
call_command('loaddata', 'settings.xml', app_label="orm")
call_command('loaddata', 'poky.xml', app_label="orm")
current_builddir = os.environ.get("BUILDDIR")
if current_builddir:
BuildTest.BUILDDIR = current_builddir
else:
# Setup a builddir based on default layout
# bitbake inside openebedded-core
oe_init_build_env_path = os.path.join(
os.path.dirname(os.path.abspath(__file__)),
os.pardir,
os.pardir,
os.pardir,
os.pardir,
os.pardir,
'oe-init-build-env'
)
if not os.path.exists(oe_init_build_env_path):
raise Exception("We had no BUILDDIR set and couldn't "
"find oe-init-build-env to set this up "
"ourselves please run oe-init-build-env "
"before running these tests")
oe_init_build_env_path = os.path.realpath(oe_init_build_env_path)
cmd = "bash -c 'source oe-init-build-env %s'" % BuildTest.BUILDDIR
p = subprocess.Popen(
cmd,
cwd=os.path.dirname(oe_init_build_env_path),
shell=True,
stdout=subprocess.PIPE,
stderr=subprocess.PIPE)
output, err = p.communicate()
p.wait()
logger.info("oe-init-build-env %s %s" % (output, err))
os.environ['BUILDDIR'] = BuildTest.BUILDDIR
# Setup the path to bitbake we know where to find this
bitbake_path = os.path.join(
os.path.dirname(os.path.abspath(__file__)),
os.pardir,
os.pardir,
os.pardir,
os.pardir,
'bin',
'bitbake')
if not os.path.exists(bitbake_path):
raise Exception("Could not find bitbake at the expected path %s"
% bitbake_path)
os.environ['BBBASEDIR'] = bitbake_path
class BuildTest(unittest.TestCase):
PROJECT_NAME = "Testbuild"
BUILDDIR = os.environ.get("BUILDDIR")
def build(self, target):
# So that the buildinfo helper uses the test database'
self.assertEqual(
os.environ.get('DJANGO_SETTINGS_MODULE', ''),
'toastermain.settings_test',
"Please initialise django with the tests settings: "
"DJANGO_SETTINGS_MODULE='toastermain.settings_test'")
built = self.target_already_built(target)
if built:
return built
load_build_environment()
BuildEnvironment.objects.get_or_create(
betype=BuildEnvironment.TYPE_LOCAL,
sourcedir=BuildTest.BUILDDIR,
builddir=BuildTest.BUILDDIR
)
release = Release.objects.get(name='local')
# Create a project for this build to run in
project = Project.objects.create_project(name=BuildTest.PROJECT_NAME,
release=release)
passthrough_variable_names = ["SSTATE_DIR", "DL_DIR", "SSTATE_MIRRORS", "BB_HASHSERVE", "BB_HASHSERVE_UPSTREAM"]
for variable_name in passthrough_variable_names:
current_variable = os.environ.get(variable_name)
if current_variable:
ProjectVariable.objects.get_or_create(
name=variable_name,
value=current_variable,
project=project)
if os.environ.get("TOASTER_TEST_USE_SSTATE_MIRROR"):
ProjectVariable.objects.get_or_create(
name="SSTATE_MIRRORS",
value="file://.* http://sstate.yoctoproject.org/all/PATH;downloadfilename=PATH",
project=project)
ProjectTarget.objects.create(project=project,
target=target,
task="")
build_request = project.schedule_build()
# run runbuilds command to dispatch the build
# e.g. manage.py runubilds
RunBuildsCommand().runbuild()
build_pk = build_request.build.pk
while Build.objects.get(pk=build_pk).outcome == Build.IN_PROGRESS:
sys.stdout.write("\rBuilding %s %d%%" %
(target,
build_request.build.completeper()))
sys.stdout.flush()
time.sleep(1)
self.assertEqual(Build.objects.get(pk=build_pk).outcome,
Build.SUCCEEDED,
"Build did not SUCCEEDED")
logger.info("\nBuild finished %s" % build_request.build.outcome)
return build_request.build
def target_already_built(self, target):
""" If the target is already built no need to build it again"""
for build in Build.objects.filter(
project__name=BuildTest.PROJECT_NAME):
targets = build.target_set.values_list('target', flat=True)
if target in targets:
return build
return None

View File

@@ -0,0 +1,363 @@
#! /usr/bin/env python3
#
# BitBake Toaster Implementation
#
# Copyright (C) 2016 Intel Corporation
#
# SPDX-License-Identifier: GPL-2.0-only
#
# Tests were part of openembedded-core oe selftest Authored by: Lucian Musat
# Ionut Chisanovici, Paul Eggleton and Cristian Iorga
import os
import pytest
from django.db.models import Q
from orm.models import Target_Image_File, Target_Installed_Package, Task
from orm.models import Package_Dependency, Recipe_Dependency, Build
from orm.models import Task_Dependency, Package, Target, Recipe
from orm.models import CustomImagePackage
from tests.builds.buildtest import BuildTest
@pytest.mark.order(4)
@pytest.mark.django_db(True)
class BuildCoreImageMinimal(BuildTest):
"""Build core-image-minimal and test the results"""
def setUp(self):
self.completed_build = self.target_already_built("core-image-minimal")
# Check if build name is unique - tc_id=795
def test_Build_Unique_Name(self):
all_builds = Build.objects.all().count()
distinct_builds = Build.objects.values('id').distinct().count()
self.assertEqual(distinct_builds,
all_builds,
msg='Build name is not unique')
# Check if build cooker log path is unique - tc_id=819
def test_Build_Unique_Cooker_Log_Path(self):
distinct_path = Build.objects.values(
'cooker_log_path').distinct().count()
total_builds = Build.objects.values('id').count()
self.assertEqual(distinct_path,
total_builds,
msg='Build cooker log path is not unique')
# Check task order sequence for one build - tc=825
def test_Task_Order_Sequence(self):
cnt_err = []
tasks = Task.objects.filter(
Q(build=self.completed_build),
~Q(order=None),
~Q(task_name__contains='_setscene')
).values('id', 'order').order_by("order")
cnt_tasks = 0
for task in tasks:
cnt_tasks += 1
if (task['order'] != cnt_tasks):
cnt_err.append(task['id'])
self.assertEqual(
len(cnt_err), 0, msg='Errors for task id: %s' % cnt_err)
# Check if disk_io matches the difference between EndTimeIO and
# StartTimeIO in build stats - tc=828
# def test_Task_Disk_IO_TC828(self):
# Check if outcome = 2 (SSTATE) then sstate_result must be 3 (RESTORED) -
# tc=832
def test_Task_If_Outcome_2_Sstate_Result_Must_Be_3(self):
tasks = Task.objects.filter(outcome=2).values('id', 'sstate_result')
cnt_err = []
for task in tasks:
if (task['sstate_result'] != 3):
cnt_err.append(task['id'])
self.assertEqual(len(cnt_err),
0,
msg='Errors for task id: %s' % cnt_err)
# Check if outcome = 1 (COVERED) or 3 (EXISTING) then sstate_result must
# be 0 (SSTATE_NA) - tc=833
def test_Task_If_Outcome_1_3_Sstate_Result_Must_Be_0(self):
tasks = Task.objects.filter(
outcome__in=(Task.OUTCOME_COVERED,
Task.OUTCOME_PREBUILT)).values('id',
'task_name',
'sstate_result')
cnt_err = []
for task in tasks:
if (task['sstate_result'] != Task.SSTATE_NA and
task['sstate_result'] != Task.SSTATE_MISS):
cnt_err.append({'id': task['id'],
'name': task['task_name'],
'sstate_result': task['sstate_result']})
self.assertEqual(len(cnt_err),
0,
msg='Errors for task id: %s' % cnt_err)
# Check if outcome is 0 (SUCCESS) or 4 (FAILED) then sstate_result must be
# 0 (NA), 1 (MISS) or 2 (FAILED) - tc=834
def test_Task_If_Outcome_0_4_Sstate_Result_Must_Be_0_1_2(self):
tasks = Task.objects.filter(
outcome__in=(0, 4)).values('id', 'sstate_result')
cnt_err = []
for task in tasks:
if (task['sstate_result'] not in [0, 1, 2]):
cnt_err.append(task['id'])
self.assertEqual(len(cnt_err),
0,
msg='Errors for task id: %s' % cnt_err)
# Check if task_executed = TRUE (1), script_type must be 0 (CODING_NA), 2
# (CODING_PYTHON), 3 (CODING_SHELL) - tc=891
def test_Task_If_Task_Executed_True_Script_Type_0_2_3(self):
tasks = Task.objects.filter(
task_executed=1).values('id', 'script_type')
cnt_err = []
for task in tasks:
if (task['script_type'] not in [0, 2, 3]):
cnt_err.append(task['id'])
self.assertEqual(len(cnt_err),
0,
msg='Errors for task id: %s' % cnt_err)
# Check if task_executed = TRUE (1), outcome must be 0 (SUCCESS) or 4
# (FAILED) - tc=836
def test_Task_If_Task_Executed_True_Outcome_0_4(self):
tasks = Task.objects.filter(task_executed=1).values('id', 'outcome')
cnt_err = []
for task in tasks:
if (task['outcome'] not in [0, 4]):
cnt_err.append(task['id'])
self.assertEqual(len(cnt_err),
0,
msg='Errors for task id: %s' % cnt_err)
# Check if task_executed = FALSE (0), script_type must be 0 - tc=890
def test_Task_If_Task_Executed_False_Script_Type_0(self):
tasks = Task.objects.filter(
task_executed=0).values('id', 'script_type')
cnt_err = []
for task in tasks:
if (task['script_type'] != 0):
cnt_err.append(task['id'])
self.assertEqual(len(cnt_err),
0,
msg='Errors for task id: %s' % cnt_err)
# Check if task_executed = FALSE (0) and build outcome = SUCCEEDED (0),
# task outcome must be 1 (COVERED), 2 (CACHED), 3 (PREBUILT), 5 (EMPTY) -
# tc=837
def test_Task_If_Task_Executed_False_Outcome_1_2_3_5(self):
builds = Build.objects.filter(outcome=0).values('id')
cnt_err = []
for build in builds:
tasks = Task.objects.filter(
build=build['id'], task_executed=0).values('id', 'outcome')
for task in tasks:
if (task['outcome'] not in [1, 2, 3, 5]):
cnt_err.append(task['id'])
self.assertEqual(len(cnt_err),
0,
msg='Errors for task id: %s' % cnt_err)
# Key verification - tc=888
def test_Target_Installed_Package(self):
rows = Target_Installed_Package.objects.values('id',
'target_id',
'package_id')
cnt_err = []
for row in rows:
target = Target.objects.filter(id=row['target_id']).values('id')
package = Package.objects.filter(id=row['package_id']).values('id')
if (not target or not package):
cnt_err.append(row['id'])
self.assertEqual(len(cnt_err),
0,
msg='Errors for target installed package id: %s' %
cnt_err)
# Key verification - tc=889
def test_Task_Dependency(self):
rows = Task_Dependency.objects.values('id',
'task_id',
'depends_on_id')
cnt_err = []
for row in rows:
task_id = Task.objects.filter(id=row['task_id']).values('id')
depends_on_id = Task.objects.filter(
id=row['depends_on_id']).values('id')
if (not task_id or not depends_on_id):
cnt_err.append(row['id'])
self.assertEqual(len(cnt_err),
0,
msg='Errors for task dependency id: %s' % cnt_err)
# Check if build target file_name is populated only if is_image=true AND
# orm_build.outcome=0 then if the file exists and its size matches
# the file_size value. Need to add the tc in the test run
def test_Target_File_Name_Populated(self):
cnt_err = []
builds = Build.objects.filter(outcome=0).values('id')
for build in builds:
targets = Target.objects.filter(
build_id=build['id'], is_image=1).values('id')
for target in targets:
target_files = Target_Image_File.objects.filter(
target_id=target['id']).values('id',
'file_name',
'file_size')
for file_info in target_files:
target_id = file_info['id']
target_file_name = file_info['file_name']
target_file_size = file_info['file_size']
if (not target_file_name or not target_file_size):
cnt_err.append(target_id)
else:
if (not os.path.exists(target_file_name)):
cnt_err.append(target_id)
else:
if (os.path.getsize(target_file_name) !=
target_file_size):
cnt_err.append(target_id)
self.assertEqual(len(cnt_err), 0,
msg='Errors for target image file id: %s' %
cnt_err)
# Key verification - tc=884
def test_Package_Dependency(self):
cnt_err = []
deps = Package_Dependency.objects.values(
'id', 'package_id', 'depends_on_id')
for dep in deps:
if (dep['package_id'] == dep['depends_on_id']):
cnt_err.append(dep['id'])
self.assertEqual(len(cnt_err), 0,
msg='Errors for package dependency id: %s' % cnt_err)
# Recipe key verification, recipe name does not depends on a recipe having
# the same name - tc=883
def test_Recipe_Dependency(self):
deps = Recipe_Dependency.objects.values(
'id', 'recipe_id', 'depends_on_id')
cnt_err = []
for dep in deps:
if (not dep['recipe_id'] or not dep['depends_on_id']):
cnt_err.append(dep['id'])
else:
name = Recipe.objects.filter(
id=dep['recipe_id']).values('name')
dep_name = Recipe.objects.filter(
id=dep['depends_on_id']).values('name')
if (name == dep_name):
cnt_err.append(dep['id'])
self.assertEqual(len(cnt_err), 0,
msg='Errors for recipe dependency id: %s' % cnt_err)
# Check if package name does not start with a number (0-9) - tc=846
def test_Package_Name_For_Number(self):
packages = Package.objects.filter(~Q(size=-1)).values('id', 'name')
cnt_err = []
for package in packages:
if (package['name'][0].isdigit() is True):
cnt_err.append(package['id'])
self.assertEqual(
len(cnt_err), 0, msg='Errors for package id: %s' % cnt_err)
# Check if package version starts with a number (0-9) - tc=847
def test_Package_Version_Starts_With_Number(self):
packages = Package.objects.filter(
~Q(size=-1)).values('id', 'version')
cnt_err = []
for package in packages:
if (package['version'][0].isdigit() is False):
cnt_err.append(package['id'])
self.assertEqual(
len(cnt_err), 0, msg='Errors for package id: %s' % cnt_err)
# Check if package revision starts with 'r' - tc=848
def test_Package_Revision_Starts_With_r(self):
packages = Package.objects.filter(
~Q(size=-1)).values('id', 'revision')
cnt_err = []
for package in packages:
if (package['revision'][0].startswith("r") is False):
cnt_err.append(package['id'])
self.assertEqual(
len(cnt_err), 0, msg='Errors for package id: %s' % cnt_err)
# Check the validity of the package build_id
# TC must be added in test run
def test_Package_Build_Id(self):
packages = Package.objects.filter(
~Q(size=-1)).values('id', 'build_id')
cnt_err = []
for package in packages:
build_id = Build.objects.filter(
id=package['build_id']).values('id')
if (not build_id):
# They have no build_id but if they are
# CustomImagePackage that's expected
try:
CustomImagePackage.objects.get(pk=package['id'])
except CustomImagePackage.DoesNotExist:
cnt_err.append(package['id'])
self.assertEqual(len(cnt_err),
0,
msg="Errors for package id: %s they have no build"
"associated with them" % cnt_err)
# Check the validity of package recipe_id
# TC must be added in test run
def test_Package_Recipe_Id(self):
packages = Package.objects.filter(
~Q(size=-1)).values('id', 'recipe_id')
cnt_err = []
for package in packages:
recipe_id = Recipe.objects.filter(
id=package['recipe_id']).values('id')
if (not recipe_id):
cnt_err.append(package['id'])
self.assertEqual(
len(cnt_err), 0, msg='Errors for package id: %s' % cnt_err)
# Check if package installed_size field is not null
# TC must be aded in test run
def test_Package_Installed_Size_Not_NULL(self):
packages = Package.objects.filter(
installed_size__isnull=True).values('id')
cnt_err = []
for package in packages:
cnt_err.append(package['id'])
self.assertEqual(
len(cnt_err), 0, msg='Errors for package id: %s' % cnt_err)
def test_custom_packages_generated(self):
"""Test if there is a corresponding generated CustomImagePackage"""
""" for each of the packages generated"""
missing_packages = []
for package in Package.objects.all():
try:
CustomImagePackage.objects.get(name=package.name)
except CustomImagePackage.DoesNotExist:
missing_packages.append(package.name)
self.assertEqual(len(missing_packages), 0,
"Some package were created from the build but their"
" corresponding CustomImagePackage was not found")