Project Number Date
test_Tails_ISO_feature-trixie 19 05 Jun 2025, 22:24

Failures Overview

The following summary displays scenarios that failed.

17.112
Scenario Erasure of memory freed by killed userspace processes
Before features/support/hooks.rb:266 0.000
Before features/support/hooks.rb:273 0.004
Steps
Given I have started Tails from DVD without network and logged in 9.025
And I prepare Tails for memory erasure tests 0.651
When I start a process allocating 128 MiB of memory with a known pattern 2.592
Then patterns cover at least 128 MiB in the guest's memory 4.843
99.610% of the expected size (128 MiB) has the pattern, but more than 100.000% was expected.
<false> is not true. (Test::Unit::AssertionFailedError)
./features/step_definitions/erase_memory.rb:186:in `/^patterns cover at least (\d+) MiB in the guest's memory$/'
features/erase_memory.feature:17:in `Then patterns cover at least 128 MiB in the guest's memory'
Output 1

Pattern coverage: 99.610% (127 MiB out of 128 MiB reference memory)

When I kill the allocating process 0.000
Then I find very few patterns in the guest's memory after 5 seconds 0.000
After features/support/hooks.rb:330 6.262

SCENARIO FAILED: 'Erasure of memory freed by killed userspace processes' (at time 00:41:20)

Boot log: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/00:41:20_Erasure_of_memory_freed_by_killed_userspace_processes.file_content_var_log_boot.log

Screenshot: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/00:41:20_Erasure_of_memory_freed_by_killed_userspace_processes.png

Video: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/00:41:20_Erasure_of_memory_freed_by_killed_userspace_processes.mkv

Systemd journal: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/00:41:20_Erasure_of_memory_freed_by_killed_userspace_processes.journal

After features/support/hooks.rb:100 0.000
12:14.272
Scenario I set up Additional Software when installing a package without persistent partition and the package is installed next time I start Tails
Before features/support/hooks.rb:266 0.000
Before features/support/hooks.rb:273 0.005
Before features/support/hooks.rb:513 0.006
Steps
Given I start Tails from a freshly installed USB drive with an administration password and the network is plugged and I login 1:39.959
And I update APT using apt 11.502
And I install "popularity-contest" using apt 27.375
Then I am proposed to add the "popularity-contest" package to my Additional Software 11.014
When I create a persistent storage and activate the Additional Software feature 23.569
Then Additional Software is correctly configured for package "popularity-contest" 2.905
When I shutdown Tails and wait for the computer to power off 15.330
And I start Tails from USB drive "__internal" with network unplugged 1:58.627
And I enable persistence 29.902
And I log in to a new session 40.520
And the Additional Software installation service has started 50.314
Then I am notified that the installation succeeded 5:3.249
try_for() timeout expired (Timeout::Error)
./features/support/helpers/misc_helpers.rb:145:in `rescue in try_for'
./features/support/helpers/misc_helpers.rb:51:in `try_for'
./features/step_definitions/common_steps.rb:595:in `/^I see the "(.+)" notification(?: after at most (\d+) seconds)?$/'
./features/step_definitions/additional_software_packages.rb:28:in `/^I am notified that the installation succeeded$/'
features/additional_software_packages.feature:39:in `Then I am notified that the installation succeeded'
And Additional Software is correctly configured for package "popularity-contest" 0.000
And the package "popularity-contest" is installed after Additional Software has been started 0.000
After features/support/hooks.rb:521 0.059
After features/support/hooks.rb:330 10.522

SCENARIO FAILED: 'I set up Additional Software when installing a package without persistent partition and the package is installed next time I start Tails' (at time 01:40:09)

Screenshot: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/01:40:09_I_set_up_Additional_Software_when_installing_a_package_without_persistent_partition_and_the_package_is_installed_next_time_I_start_Tails.png

Video: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/01:40:09_I_set_up_Additional_Software_when_installing_a_package_without_persistent_partition_and_the_package_is_installed_next_time_I_start_Tails.mkv

Systemd journal: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/01:40:09_I_set_up_Additional_Software_when_installing_a_package_without_persistent_partition_and_the_package_is_installed_next_time_I_start_Tails.journal

After features/support/hooks.rb:100 0.021
25:42.117
Scenario The Additional Software dpkg hook notices when persistence is locked down while installing a package
Before features/support/hooks.rb:266 0.000
Before features/support/hooks.rb:273 0.007
Before features/support/hooks.rb:513 0.009
Steps
Given a computer 0.293
And I start Tails from USB drive "__internal" and I login with an administration password 3:24.086
And I update APT using apt 11.555
When I install "makepp" using apt 22:6.182
Tor
Tor operation failed (despite retrying 10 times) with
Timeout::Error: try_for() timeout expired
Last ignored exception was: ExecutionFailedInVM: Command failed: dpkg -s 'makepp' 2>/dev/null | grep -qs '^Status:.*installed$'
error code: 1
stdout: 
stderr: .
<false> is not true.
./features/support/helpers/vm_helper.rb:496:in `rescue in execute_successfully'
./features/support/helpers/vm_helper.rb:493:in `execute_successfully'
./features/step_definitions/apt.rb:69:in `block in wait_for_package_installation'
./features/support/helpers/misc_helpers.rb:78:in `block (2 levels) in try_for'
./features/support/helpers/misc_helpers.rb:66:in `loop'
./features/support/helpers/misc_helpers.rb:66:in `block in try_for'
./features/support/helpers/misc_helpers.rb:65:in `try_for'
./features/step_definitions/apt.rb:68:in `wait_for_package_installation'
./features/step_definitions/apt.rb:86:in `block (3 levels) in <top (required)>'
./features/step_definitions/apt.rb:82:in `block (2 levels) in <top (required)>'
./features/support/helpers/misc_helpers.rb:211:in `block in retry_action'
./features/support/helpers/misc_helpers.rb:208:in `loop'
./features/support/helpers/misc_helpers.rb:208:in `retry_action'
./features/support/helpers/misc_helpers.rb:196:in `retry_tor'
./features/step_definitions/apt.rb:81:in `block in <top (required)>' (MaxRetriesFailure)
./features/support/helpers/misc_helpers.rb:226:in `rescue in block in retry_action'
./features/support/helpers/misc_helpers.rb:209:in `block in retry_action'
./features/support/helpers/misc_helpers.rb:208:in `loop'
./features/support/helpers/misc_helpers.rb:208:in `retry_action'
./features/support/helpers/misc_helpers.rb:196:in `retry_tor'
./features/step_definitions/apt.rb:81:in `/^I install "(.+)" using apt$/'
features/additional_software_packages.feature:48:in `When I install "makepp" using apt'
Then the Additional Software dpkg hook has been run for package "makepp" and notices the persistence is locked 0.000
And the package "makepp" is installed 0.000
After features/support/hooks.rb:521 0.033
After features/support/hooks.rb:330 5.267

SCENARIO FAILED: 'The Additional Software dpkg hook notices when persistence is locked down while installing a package' (at time 02:06:02)

Exception thrown while trying to save /tmp/TailsToaster/artifact.file_content_run_live-additional-software_log: RemoteShell::ServerFailure: FileNotFoundError: [Errno 2] No such file or directory: '/run/live-additional-software/log'

Boot log: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/02:06:02_The_Additional_Software_dpkg_hook_notices_when_persistence_is_locked_down_while_installing_a_package.file_content_var_log_boot.log

Output of ls -la --full-time /live/persistence/tailsdata_unlocked: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/02:06:02_The_Additional_Software_dpkg_hook_notices_when_persistence_is_locked_down_while_installing_a_package.cmd_output_persistent_volume

Output of ls -lar --full-time /var/cache/apt: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/02:06:02_The_Additional_Software_dpkg_hook_notices_when_persistence_is_locked_down_while_installing_a_package.cmd_output_var_cache_apt

Output of ls -lar --full-time /var/lib/apt: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/02:06:02_The_Additional_Software_dpkg_hook_notices_when_persistence_is_locked_down_while_installing_a_package.cmd_output_var_lib_apt

Output of mount: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/02:06:02_The_Additional_Software_dpkg_hook_notices_when_persistence_is_locked_down_while_installing_a_package.cmd_output_mount

Screenshot: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/02:06:02_The_Additional_Software_dpkg_hook_notices_when_persistence_is_locked_down_while_installing_a_package.png

Video: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/02:06:02_The_Additional_Software_dpkg_hook_notices_when_persistence_is_locked_down_while_installing_a_package.mkv

Systemd journal: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/02:06:02_The_Additional_Software_dpkg_hook_notices_when_persistence_is_locked_down_while_installing_a_package.journal

After features/support/hooks.rb:100 0.044
10:43.623
Scenario My Additional Software list is configurable through a GUI or through notifications when I install or remove packages with APT or Synaptic
Before features/support/hooks.rb:266 0.000
Before features/support/hooks.rb:273 0.005
Before features/support/hooks.rb:513 0.005
Steps
Given a computer 0.263
And I start Tails from USB drive "__internal" and I login with persistence enabled and an administration password 4:59.150
When I uninstall "popularity-contest" using apt 25.145
And I accept removing "popularity-contest" from Additional Software 2.872
Then "popularity-contest" is not in the list of Additional Software 0.826
When I start Synaptic 16.170
And I update APT using Synaptic 23.547
And I install "cowsay" using Synaptic 4:35.647
Tor
Tor operation failed (despite retrying 10 times) with
Dogtail::Failure: The Dogtail init script raised: SearchError: child of [application | synaptic]: 'Find' dialog
STDOUT:

STDERR:

 (MaxRetriesFailure)
./features/support/helpers/misc_helpers.rb:226:in `rescue in block in retry_action'
./features/support/helpers/misc_helpers.rb:209:in `block in retry_action'
./features/support/helpers/misc_helpers.rb:208:in `loop'
./features/support/helpers/misc_helpers.rb:208:in `retry_action'
./features/support/helpers/misc_helpers.rb:196:in `retry_tor'
./features/step_definitions/apt.rb:180:in `/^I install "(.+)" using Synaptic$/'
features/additional_software_packages.feature:61:in `And I install "cowsay" using Synaptic'
And I accept adding "cowsay" to Additional Software 0.000
Then Additional Software is correctly configured for package "cowsay" 0.000
When I uninstall "cowsay" using apt 0.000
And I refuse removing "cowsay" from Additional Software 0.000
Then Additional Software is correctly configured for package "cowsay" 0.000
When I start "Additional Software" via GNOME Activities Overview 0.000
And I remove "cowsay" from the list of Additional Software using Additional Software GUI 0.000
Then "cowsay" is not in the list of Additional Software 0.000
When I install "cowsay" using apt 0.000
And I refuse adding "cowsay" to Additional Software 0.000
Then "cowsay" is not in the list of Additional Software 0.000
After features/support/hooks.rb:521 0.043
After features/support/hooks.rb:330 5.842

SCENARIO FAILED: 'My Additional Software list is configurable through a GUI or through notifications when I install or remove packages with APT or Synaptic' (at time 02:16:51)

Boot log: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/02:16:51_My_Additional_Software_list_is_configurable_through_a_GUI_or_through_notifications_when_I_install_or_remove_packages_with_APT_or_Synaptic.file_content_var_log_boot.log

Content of /run/live-additional-software/log: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/02:16:51_My_Additional_Software_list_is_configurable_through_a_GUI_or_through_notifications_when_I_install_or_remove_packages_with_APT_or_Synaptic.file_content_run_live-additional-software_log

Output of ls -la --full-time /live/persistence/tailsdata_unlocked: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/02:16:51_My_Additional_Software_list_is_configurable_through_a_GUI_or_through_notifications_when_I_install_or_remove_packages_with_APT_or_Synaptic.cmd_output_persistent_volume

Output of ls -lar --full-time /var/cache/apt: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/02:16:51_My_Additional_Software_list_is_configurable_through_a_GUI_or_through_notifications_when_I_install_or_remove_packages_with_APT_or_Synaptic.cmd_output_var_cache_apt

Output of ls -lar --full-time /var/lib/apt: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/02:16:51_My_Additional_Software_list_is_configurable_through_a_GUI_or_through_notifications_when_I_install_or_remove_packages_with_APT_or_Synaptic.cmd_output_var_lib_apt

Output of mount: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/02:16:51_My_Additional_Software_list_is_configurable_through_a_GUI_or_through_notifications_when_I_install_or_remove_packages_with_APT_or_Synaptic.cmd_output_mount

Screenshot: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/02:16:51_My_Additional_Software_list_is_configurable_through_a_GUI_or_through_notifications_when_I_install_or_remove_packages_with_APT_or_Synaptic.png

Video: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/02:16:51_My_Additional_Software_list_is_configurable_through_a_GUI_or_through_notifications_when_I_install_or_remove_packages_with_APT_or_Synaptic.mkv

Systemd journal: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/02:16:51_My_Additional_Software_list_is_configurable_through_a_GUI_or_through_notifications_when_I_install_or_remove_packages_with_APT_or_Synaptic.journal

After features/support/hooks.rb:100 0.010
3:7.556
Scenario I am notified when Additional Software fails to install a package
Before features/support/hooks.rb:266 0.000
Before features/support/hooks.rb:273 0.006
Before features/support/hooks.rb:513 0.006
Steps
Given a computer 0.213
And I start Tails from USB drive "__internal" with network unplugged 1:51.078
And I enable persistence 14.737
And I remove the "cowsay" deb files from the APT cache 0.062
# Prevent the "Warning: virtual machine detected!" notification from racing
# with the one we'll be interacting with below.
And I disable the tails-virt-notify-user.service user unit 0.092
And I log in to a new session 39.205
Then I see the "The installation of your additional software failed" notification after at most 300 seconds 21.351
And I can open the Additional Software log file from the notification 0.814
The
The Dogtail init script raised: DBusException: org.freedesktop.DBus.Error.AccessDenied: GetWindows is not allowed
STDOUT:

STDERR:

 (Dogtail::Failure)
./features/support/helpers/dogtail.rb:93:in `run'
./features/support/helpers/dogtail.rb:339:in `initialize'
./features/support/helpers/dogtail.rb:268:in `new'
./features/support/helpers/dogtail.rb:268:in `call_tree_api_method'
./features/support/helpers/dogtail.rb:360:in `click'
./features/step_definitions/additional_software_packages.rb:40:in `click_gnome_shell_notification_button'
./features/step_definitions/additional_software_packages.rb:166:in `/^I can open the Additional Software log file from the notification$/'
features/additional_software_packages.feature:130:in `And I can open the Additional Software log file from the notification'
And the package "cowsay" is not installed 0.000
After features/support/hooks.rb:521 0.055
After features/support/hooks.rb:330 6.871

SCENARIO FAILED: 'I am notified when Additional Software fails to install a package' (at time 02:36:02)

Boot log: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/02:36:02_I_am_notified_when_Additional_Software_fails_to_install_a_package.file_content_var_log_boot.log

Content of /run/live-additional-software/log: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/02:36:02_I_am_notified_when_Additional_Software_fails_to_install_a_package.file_content_run_live-additional-software_log

Output of ls -la --full-time /live/persistence/tailsdata_unlocked: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/02:36:02_I_am_notified_when_Additional_Software_fails_to_install_a_package.cmd_output_persistent_volume

Output of ls -lar --full-time /var/cache/apt: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/02:36:02_I_am_notified_when_Additional_Software_fails_to_install_a_package.cmd_output_var_cache_apt

Output of ls -lar --full-time /var/lib/apt: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/02:36:02_I_am_notified_when_Additional_Software_fails_to_install_a_package.cmd_output_var_lib_apt

Output of mount: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/02:36:02_I_am_notified_when_Additional_Software_fails_to_install_a_package.cmd_output_mount

Screenshot: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/02:36:02_I_am_notified_when_Additional_Software_fails_to_install_a_package.png

Video: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/02:36:02_I_am_notified_when_Additional_Software_fails_to_install_a_package.mkv

Systemd journal: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/02:36:02_I_am_notified_when_Additional_Software_fails_to_install_a_package.journal

After features/support/hooks.rb:100 0.010
Tags: @product
2:28.812
Scenario Installing an old version of Tails to a pristine USB drive
Before features/support/hooks.rb:266 0.000
Before features/support/hooks.rb:273 0.008
Steps
Given a computer 0.213
And I create a 7200 MiB disk named "old" 0.050
And I plug USB drive "old" 1.060
And I write an old version of the Tails USB image to disk "old" 33.150
When I start Tails from USB drive "old" with network unplugged 1:54.337
The
The Dogtail init script raised: SearchError: descendent of [application | gnome-shell]: child with roleName='notification'
STDOUT:

STDERR:

 (Dogtail::Failure)
./features/support/helpers/dogtail.rb:93:in `run'
./features/support/helpers/dogtail.rb:339:in `initialize'
./features/support/helpers/dogtail.rb:268:in `new'
./features/support/helpers/dogtail.rb:268:in `call_tree_api_method'
./features/support/helpers/dogtail.rb:276:in `child'
./features/step_definitions/common_steps.rb:453:in `/^the computer (?:re)?boots Tails$/'
./features/step_definitions/common_steps.rb:276:in `/^I start Tails from (.+?) drive "(.+?)"( with network unplugged)?( and I login( with persistence enabled)?( with the changed persistence passphrase)?( (?:and|with) an administration password)?)?$/'
features/usb_upgrade.feature:25:in `When I start Tails from USB drive "old" with network unplugged'
Then the boot device has safe access rights 0.000
And Tails is running from USB drive "old" 0.000
And there is no persistence partition on USB drive "old" 0.000
And process "udev-watchdog" is running 0.000
And udev-watchdog is monitoring the correct device 0.000
And I unplug USB drive "old" 0.000
After features/support/hooks.rb:330 5.830

SCENARIO FAILED: 'Installing an old version of Tails to a pristine USB drive' (at time 02:38:38)

Boot log: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/02:38:38_Installing_an_old_version_of_Tails_to_a_pristine_USB_drive.file_content_var_log_boot.log

Screenshot: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/02:38:38_Installing_an_old_version_of_Tails_to_a_pristine_USB_drive.png

Video: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/02:38:38_Installing_an_old_version_of_Tails_to_a_pristine_USB_drive.mkv

Systemd journal: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/02:38:38_Installing_an_old_version_of_Tails_to_a_pristine_USB_drive.journal

After features/support/hooks.rb:100 0.057
Tags: @product
1:26.174
Scenario Creating a persistent partition with the old Tails USB installation
Before features/support/hooks.rb:266 0.000
Before features/support/hooks.rb:273 0.006
Steps
Given a computer 0.223
And I start Tails from USB drive "old" with network unplugged and I login 1:25.950
The
The Dogtail init script raised: SearchError: descendent of [application | gnome-shell]: child with roleName='notification'
STDOUT:

STDERR:

 (Dogtail::Failure)
./features/support/helpers/dogtail.rb:93:in `run'
./features/support/helpers/dogtail.rb:339:in `initialize'
./features/support/helpers/dogtail.rb:268:in `new'
./features/support/helpers/dogtail.rb:268:in `call_tree_api_method'
./features/support/helpers/dogtail.rb:276:in `child'
./features/step_definitions/common_steps.rb:453:in `/^the computer (?:re)?boots Tails$/'
./features/step_definitions/common_steps.rb:276:in `/^I start Tails from (.+?) drive "(.+?)"( with network unplugged)?( and I login( with persistence enabled)?( with the changed persistence passphrase)?( (?:and|with) an administration password)?)?$/'
features/usb_upgrade.feature:36:in `And I start Tails from USB drive "old" with network unplugged and I login'
Then Tails is running from USB drive "old" 0.000
And I create a persistent partition 0.000
And I take note of which tps features are available 0.000
Then a Tails persistence partition exists on USB drive "old" 0.000
And I shutdown Tails and wait for the computer to power off 0.000
After features/support/hooks.rb:330 5.836

SCENARIO FAILED: 'Creating a persistent partition with the old Tails USB installation' (at time 02:40:11)

Boot log: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/02:40:11_Creating_a_persistent_partition_with_the_old_Tails_USB_installation.file_content_var_log_boot.log

Screenshot: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/02:40:11_Creating_a_persistent_partition_with_the_old_Tails_USB_installation.png

Video: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/02:40:11_Creating_a_persistent_partition_with_the_old_Tails_USB_installation.mkv

Systemd journal: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/02:40:11_Creating_a_persistent_partition_with_the_old_Tails_USB_installation.journal

After features/support/hooks.rb:100 0.008
Tags: @product
1:22.953
Scenario Writing files to a read/write-enabled persistent partition with the old Tails USB installation
Before features/support/hooks.rb:266 0.000
Before features/support/hooks.rb:273 0.005
Steps
Given a computer 0.212
And I start Tails from USB drive "old" with network unplugged and I login with persistence enabled 1:22.740
The
The Dogtail init script raised: SearchError: descendent of [application | gnome-shell]: child with roleName='notification'
STDOUT:

STDERR:

 (Dogtail::Failure)
./features/support/helpers/dogtail.rb:93:in `run'
./features/support/helpers/dogtail.rb:339:in `initialize'
./features/support/helpers/dogtail.rb:268:in `new'
./features/support/helpers/dogtail.rb:268:in `call_tree_api_method'
./features/support/helpers/dogtail.rb:276:in `child'
./features/step_definitions/common_steps.rb:453:in `/^the computer (?:re)?boots Tails$/'
./features/step_definitions/common_steps.rb:276:in `/^I start Tails from (.+?) drive "(.+?)"( with network unplugged)?( and I login( with persistence enabled)?( with the changed persistence passphrase)?( (?:and|with) an administration password)?)?$/'
features/usb_upgrade.feature:46:in `And I start Tails from USB drive "old" with network unplugged and I login with persistence enabled'
Then Tails is running from USB drive "old" 0.000
And all tps features are active 0.000
When I write some files expected to persist 0.000
# Verify that our baseline for the next scenarios is sane
And all persistent filesystems have safe access rights 0.000
And all persistence configuration files have safe access rights 0.000
And all persistent directories from the old Tails version have safe access rights 0.000
And I take note of which tps features are available 0.000
And I shutdown Tails and wait for the computer to power off 0.000
# XXX: how does guestfs work vs snapshots?
Then only the expected files are present on the persistence partition on USB drive "old" 0.000
After features/support/hooks.rb:330 6.034

SCENARIO FAILED: 'Writing files to a read/write-enabled persistent partition with the old Tails USB installation' (at time 02:41:39)

Boot log: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/02:41:39_Writing_files_to_a_read_write-enabled_persistent_partition_with_the_old_Tails_USB_installation.file_content_var_log_boot.log

Screenshot: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/02:41:39_Writing_files_to_a_read_write-enabled_persistent_partition_with_the_old_Tails_USB_installation.png

Video: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/02:41:39_Writing_files_to_a_read_write-enabled_persistent_partition_with_the_old_Tails_USB_installation.mkv

Systemd journal: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/02:41:39_Writing_files_to_a_read_write-enabled_persistent_partition_with_the_old_Tails_USB_installation.journal

After features/support/hooks.rb:100 0.009
Tags: @product
2:14.239
Scenario Booting Tails from a USB drive upgraded from USB with persistence enabled
Before features/support/hooks.rb:266 0.000
Before features/support/hooks.rb:273 0.005
Steps
Given a computer 0.206
And I start Tails from USB drive "to_upgrade" with network unplugged and I login with persistence enabled 2:14.032
The
The Dogtail init script raised: SearchError: descendant of [application | Welcome to Tails!]: child with roleName='password text'
STDOUT:

STDERR:

 (Dogtail::Failure)
./features/support/helpers/dogtail.rb:93:in `run'
./features/support/helpers/dogtail.rb:339:in `initialize'
./features/support/helpers/dogtail.rb:268:in `new'
./features/support/helpers/dogtail.rb:268:in `call_tree_api_method'
./features/support/helpers/dogtail.rb:276:in `child'
./features/step_definitions/usb.rb:671:in `/^I try to enable persistence( with the changed passphrase)?$/'
./features/step_definitions/usb.rb:703:in `/^I enable persistence( with the changed passphrase)?$/'
./features/step_definitions/common_steps.rb:283:in `/^I start Tails from (.+?) drive "(.+?)"( with network unplugged)?( and I login( with persistence enabled)?( with the changed persistence passphrase)?( (?:and|with) an administration password)?)?$/'
features/usb_upgrade.feature:73:in `And I start Tails from USB drive "to_upgrade" with network unplugged and I login with persistence enabled'
Then all tps features from the old Tails version are active 0.000
And Tails is running from USB drive "to_upgrade" 0.000
And the boot device has safe access rights 0.000
And the expected persistent files created with the old Tails version are present in the filesystem 0.000
And all persistent directories from the old Tails version have safe access rights 0.000
After features/support/hooks.rb:330 6.085

SCENARIO FAILED: 'Booting Tails from a USB drive upgraded from USB with persistence enabled' (at time 02:46:55)

Boot log: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/02:46:55_Booting_Tails_from_a_USB_drive_upgraded_from_USB_with_persistence_enabled.file_content_var_log_boot.log

Screenshot: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/02:46:55_Booting_Tails_from_a_USB_drive_upgraded_from_USB_with_persistence_enabled.png

Video: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/02:46:55_Booting_Tails_from_a_USB_drive_upgraded_from_USB_with_persistence_enabled.mkv

Systemd journal: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/02:46:55_Booting_Tails_from_a_USB_drive_upgraded_from_USB_with_persistence_enabled.journal

After features/support/hooks.rb:100 0.011
23:2.996
Scenario Upgrading an initial Tails installation with an incremental upgrade
Before features/support/hooks.rb:266 0.000
Before features/support/hooks.rb:273 0.008
Steps
Given I have started Tails without network from a USB drive with a persistent partition enabled and logged in 15.110
And no SquashFS delta is installed 1.635
And Tails is fooled to think that version 6.0~testoverlayfs was initially installed 1.992
And Tails is fooled to think it is running version 6.0~testoverlayfs 0.271
And the file system changes introduced in version 6.2~testoverlayfs are not present 0.603
And the file system changes introduced in version 6.3~testoverlayfs are not present 0.505
When the network is plugged 0.090
And Tor is ready 29.798
Then I am proposed to install an incremental upgrade to version 6.2~testoverlayfs 22:12.988
Tor
Tor operation failed (despite retrying 10 times) with
FindFailed: can not find any of the patterns ["TailsUpgraderUpgradeTo6.2~testoverlayfs.png", "TailsUpgraderFailure.png"] on the screen (MaxRetriesFailure)
./features/support/helpers/misc_helpers.rb:226:in `rescue in block in retry_action'
./features/support/helpers/misc_helpers.rb:209:in `block in retry_action'
./features/support/helpers/misc_helpers.rb:208:in `loop'
./features/support/helpers/misc_helpers.rb:208:in `retry_action'
./features/support/helpers/misc_helpers.rb:196:in `retry_tor'
./features/step_definitions/usb.rb:1301:in `/^I am proposed to install an incremental upgrade to version (.+)$/'
features/usb_upgrade.feature:90:in `Then I am proposed to install an incremental upgrade to version 6.2~testoverlayfs'
And I can successfully install the incremental upgrade to version 6.2~testoverlayfs 0.000
Given I shutdown Tails and wait for the computer to power off 0.000
When I start Tails from USB drive "__internal" with network unplugged and I login with persistence enabled 0.000
Then Tails is running version 6.2~testoverlayfs 0.000
And all tps features are active 0.000
And the file system changes introduced in version 6.2~testoverlayfs are present 0.000
And only the 6.2~testoverlayfs SquashFS delta is installed 0.000
# Our IUK sets a release date that can make Tor bootstrapping impossible
Given Tails system time is magically synchronized 0.000
# We'll really install Tails_amd64_6.0~testoverlayfs_to_6.3~testoverlayfs.iuk
# but we need some way to force upgrading a second time in a row
# even if only the initially installed version is considered
And Tails is fooled to think that version 6.1~testoverlayfs was initially installed 0.000
When the network is plugged 0.000
And Tor is ready 0.000
Then I am proposed to install an incremental upgrade to version 6.3~testoverlayfs 0.000
And I can successfully install the incremental upgrade to version 6.3~testoverlayfs 0.000
Given I shutdown Tails and wait for the computer to power off 0.000
When I start Tails from USB drive "__internal" with network unplugged and I login with persistence enabled 0.000
Then Tails is running version 6.3~testoverlayfs 0.000
And all tps features are active 0.000
And the file system changes introduced in version 6.3~testoverlayfs are present 0.000
And only the 6.3~testoverlayfs SquashFS delta is installed 0.000
# Regression test for #17425 (i.e. the Upgrader would propose
# upgrading to the version that's already running)
Given Tails system time is magically synchronized 0.000
And Tails is fooled to think that version 6.1~testoverlayfs was initially installed 0.000
When the network is plugged 0.000
And Tor is ready 0.000
Then the Upgrader considers the system as up-to-date 0.000
# Regression test on #8158 (i.e. the IUK's filesystem is not part of the Unsafe Browser's chroot)
And I successfully start the Unsafe Browser 0.000
And the file system changes introduced in version 6.3~testoverlayfs are present in the Unsafe Browser's chroot 0.000
After features/support/hooks.rb:330 5.450

SCENARIO FAILED: 'Upgrading an initial Tails installation with an incremental upgrade' (at time 03:10:04)

Boot log: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/03:10:04_Upgrading_an_initial_Tails_installation_with_an_incremental_upgrade.file_content_var_log_boot.log

Screenshot: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/03:10:04_Upgrading_an_initial_Tails_installation_with_an_incremental_upgrade.png

Video: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/03:10:04_Upgrading_an_initial_Tails_installation_with_an_incremental_upgrade.mkv

Systemd journal: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/03:10:04_Upgrading_an_initial_Tails_installation_with_an_incremental_upgrade.journal

After features/support/hooks.rb:100 0.000
22:59.583
Scenario Upgrading a Tails whose signing key is outdated
Before features/support/hooks.rb:266 0.000
Before features/support/hooks.rb:273 0.005
Steps
Given I have started Tails without network from a USB drive with a persistent partition enabled and logged in 17.358
And Tails is fooled to think that version 6.0~testoverlayfs was initially installed 0.524
And Tails is fooled to think it is running version 6.0~testoverlayfs 0.076
And the signing key used by the Upgrader is outdated 0.099
But a current signing key is available on our website 0.000
When the network is plugged 0.073
And Tor is ready 28.012
Then I am proposed to install an incremental upgrade to version 6.2~testoverlayfs 22:13.439
Tor
Tor operation failed (despite retrying 10 times) with
FindFailed: can not find any of the patterns ["TailsUpgraderUpgradeTo6.2~testoverlayfs.png", "TailsUpgraderFailure.png"] on the screen (MaxRetriesFailure)
./features/support/helpers/misc_helpers.rb:226:in `rescue in block in retry_action'
./features/support/helpers/misc_helpers.rb:209:in `block in retry_action'
./features/support/helpers/misc_helpers.rb:208:in `loop'
./features/support/helpers/misc_helpers.rb:208:in `retry_action'
./features/support/helpers/misc_helpers.rb:196:in `retry_tor'
./features/step_definitions/usb.rb:1301:in `/^I am proposed to install an incremental upgrade to version (.+)$/'
features/usb_upgrade.feature:134:in `Then I am proposed to install an incremental upgrade to version 6.2~testoverlayfs'
After features/support/hooks.rb:330 6.064

SCENARIO FAILED: 'Upgrading a Tails whose signing key is outdated' (at time 03:33:09)

Boot log: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/03:33:09_Upgrading_a_Tails_whose_signing_key_is_outdated.file_content_var_log_boot.log

Screenshot: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/03:33:09_Upgrading_a_Tails_whose_signing_key_is_outdated.png

Video: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/03:33:09_Upgrading_a_Tails_whose_signing_key_is_outdated.mkv

Systemd journal: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/03:33:09_Upgrading_a_Tails_whose_signing_key_is_outdated.journal

After features/support/hooks.rb:100 0.000
Tags: @product
1:19.266
Scenario Using Tails' custom backup tool to backup a newly added file
Before features/support/hooks.rb:266 0.000
Before features/support/hooks.rb:273 0.014
Steps
Given I have started Tails without network from a USB drive with a persistent partition enabled and logged in 16.822
# The volume cannot be in use when GuestFs clones it, which we do below
And I shutdown Tails and wait for the computer to power off 12.404
And I clone USB drive "__internal" to a temporary USB drive "backup" 2.432
Given I have started Tails without network from a USB drive with a persistent partition enabled and logged in 11.791
And I write a file "/live/persistence/TailsData_unlocked/new" with contents "foo" 0.098
When I start Tails' custom backup tool 6.820
Then the backup tool displays "Plug in your backup Tails USB stick" 0.356
When I plug USB drive "backup" 2.783
And I give the Persistent Storage on drive "backup" its own UUID 0.740
And I click "Retry" in the backup tool 1.895
Then the backup tool displays "Do you want to back up your Persistent Storage now?" 0.878
When I click "Back Up" in the backup tool 1.201
And I enter my persistent storage passphrase into the polkit prompt 21.040
try_for() timeout expired (Timeout::Error)
./features/support/helpers/misc_helpers.rb:145:in `rescue in try_for'
./features/support/helpers/misc_helpers.rb:51:in `try_for'
./features/step_definitions/common_steps.rb:786:in `deal_with_polkit_prompt'
./features/step_definitions/backup.rb:18:in `/^I enter my persistent storage passphrase into the polkit prompt$/'
features/backup.feature:21:in `And I enter my persistent storage passphrase into the polkit prompt'
Then the backup tool displays "Your Persistent Storage was backed up successfully to your backup Tails USB stick!" 0.000
# The backup tool ejects the drive, so we have to replug it
When I unplug USB drive "backup" 0.000
And I plug USB drive "backup" 0.000
Then the USB drive "backup" contains the same files as my persistent storage 0.000
After features/support/hooks.rb:330 9.252

SCENARIO FAILED: 'Using Tails' custom backup tool to backup a newly added file' (at time 03:42:58)

Screenshot: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/03:42:58_Using_Tails__custom_backup_tool_to_backup_a_newly_added_file.png

Video: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/03:42:58_Using_Tails__custom_backup_tool_to_backup_a_newly_added_file.mkv

Systemd journal: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/03:42:58_Using_Tails__custom_backup_tool_to_backup_a_newly_added_file.journal

After features/support/hooks.rb:100 0.167
Tags: @product
0.109
Scenario Getting a DHCP lease with the default NetworkManager connection
Steps
Then the hostname should not have been leaked on the network 0.109
Hostname leak detected (DhcpLeakError)
./features/step_definitions/dhcp.rb:15:in `/^the hostname should not have been leaked on the network$/'
features/dhcp.feature:17:in `Then the hostname should not have been leaked on the network'
After features/support/hooks.rb:330 4.899

SCENARIO FAILED: 'Getting a DHCP lease with the default NetworkManager connection' (at time 03:43:55)

Boot log: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/03:43:55_Getting_a_DHCP_lease_with_the_default_NetworkManager_connection.file_content_var_log_boot.log

Network capture: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/03:43:55_Getting_a_DHCP_lease_with_the_default_NetworkManager_connection.pcap

Screenshot: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/03:43:55_Getting_a_DHCP_lease_with_the_default_NetworkManager_connection.png

Video: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/03:43:55_Getting_a_DHCP_lease_with_the_default_NetworkManager_connection.mkv

Systemd journal: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/03:43:55_Getting_a_DHCP_lease_with_the_default_NetworkManager_connection.journal

After features/support/hooks.rb:100 0.034
Tags: @product
0.897
Scenario Getting a DHCP lease with a manually configured NetworkManager connection
Steps
When I add a wired DHCP NetworkManager connection called "manually-added-con" 0.351
And I switch to the "manually-added-con" NetworkManager connection 0.401
Then the hostname should not have been leaked on the network 0.145
Hostname leak detected (DhcpLeakError)
./features/step_definitions/dhcp.rb:15:in `/^the hostname should not have been leaked on the network$/'
features/dhcp.feature:22:in `Then the hostname should not have been leaked on the network'
After features/support/hooks.rb:330 4.963

SCENARIO FAILED: 'Getting a DHCP lease with a manually configured NetworkManager connection' (at time 03:44:35)

Boot log: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/03:44:35_Getting_a_DHCP_lease_with_a_manually_configured_NetworkManager_connection.file_content_var_log_boot.log

Network capture: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/03:44:35_Getting_a_DHCP_lease_with_a_manually_configured_NetworkManager_connection.pcap

Screenshot: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/03:44:35_Getting_a_DHCP_lease_with_a_manually_configured_NetworkManager_connection.png

Video: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/03:44:35_Getting_a_DHCP_lease_with_a_manually_configured_NetworkManager_connection.mkv

Systemd journal: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/03:44:35_Getting_a_DHCP_lease_with_a_manually_configured_NetworkManager_connection.journal

After features/support/hooks.rb:100 0.049
Tags: @product
3:6.246
Scenario Outline GPT backup corruption with a persistent partition
Before features/support/hooks.rb:266 0.000
Before features/support/hooks.rb:273 0.019
Steps
Given I have started Tails without network from a USB drive with a persistent partition and stopped at Tails Greeter's login screen 13.806
And I corrupt the boot device's GPT backup header 0.789
And I power off the computer 0.339
When I start the computer 1.069
Then the computer boots Tails 2:7.997
When I log in to a new session 23.617
And all notifications have disappeared 18.210
Then I am recommended to migrate to a new USB stick due to partitioning errors 0.416
<nil> was expected to not be nil. (Test::Unit::AssertionFailedError)
./features/step_definitions/hardware.rb:84:in `/^I am recommended to migrate to a new USB stick due to partitioning errors$/'
features/hardware_failure.feature:45:in `Then I am recommended to migrate to a new USB stick due to partitioning errors'
features/hardware_failure.feature:41:in `Then I am recommended to migrate to a new USB stick due to partitioning errors'
And Tails detected partitioning error partitioning-corruption 0.000
After features/support/hooks.rb:330 6.787

SCENARIO FAILED: 'GPT backup corruption with a persistent partition, Examples (#1)' (at time 04:03:03)

Boot log: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/04:03:03_GPT_backup_corruption_with_a_persistent_partition,_Examples___1_.file_content_var_log_boot.log

Screenshot: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/04:03:03_GPT_backup_corruption_with_a_persistent_partition,_Examples___1_.png

Video: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/04:03:03_GPT_backup_corruption_with_a_persistent_partition,_Examples___1_.mkv

Systemd journal: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/04:03:03_GPT_backup_corruption_with_a_persistent_partition,_Examples___1_.journal

After features/support/hooks.rb:100 0.048
Tags: @product
2:57.851
Scenario Outline GPT backup corruption with a persistent partition
Before features/support/hooks.rb:266 0.000
Before features/support/hooks.rb:273 0.013
Steps
Given I have started Tails without network from a USB drive with a persistent partition and stopped at Tails Greeter's login screen 10.140
And I corrupt the boot device's GPT backup partition table 0.484
And I power off the computer 1.117
When I start the computer 1.579
Then the computer boots Tails 1:58.493
When I log in to a new session 31.019
And all notifications have disappeared 14.630
Then I am recommended to migrate to a new USB stick due to partitioning errors 0.386
<nil> was expected to not be nil. (Test::Unit::AssertionFailedError)
./features/step_definitions/hardware.rb:84:in `/^I am recommended to migrate to a new USB stick due to partitioning errors$/'
features/hardware_failure.feature:46:in `Then I am recommended to migrate to a new USB stick due to partitioning errors'
features/hardware_failure.feature:41:in `Then I am recommended to migrate to a new USB stick due to partitioning errors'
And Tails detected partitioning error partitioning-corruption 0.000
After features/support/hooks.rb:330 6.900

SCENARIO FAILED: 'GPT backup corruption with a persistent partition, Examples (#2)' (at time 04:06:08)

Boot log: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/04:06:08_GPT_backup_corruption_with_a_persistent_partition,_Examples___2_.file_content_var_log_boot.log

Screenshot: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/04:06:08_GPT_backup_corruption_with_a_persistent_partition,_Examples___2_.png

Video: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/04:06:08_GPT_backup_corruption_with_a_persistent_partition,_Examples___2_.mkv

Systemd journal: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/04:06:08_GPT_backup_corruption_with_a_persistent_partition,_Examples___2_.journal

After features/support/hooks.rb:100 0.009
Tags: @product
4:1.679
Scenario GPT backup corruption without a persistent partition
Before features/support/hooks.rb:266 0.000
Before features/support/hooks.rb:273 0.006
Steps
Given a computer 0.233
And I set Tails to boot with options "test_gpt_corruption=gpt_backup,gpt_backup_table" 0.000
And I temporarily create a 7200 MiB disk named "temp" 0.086
And I plug USB drive "temp" 1.071
And I write the Tails USB image to disk "temp" 39.117
When I start Tails from USB drive "temp" with network unplugged 2:39.648
Then Tails is running from USB drive "temp" 0.448
And the Greeter forbids creating a persistent partition 0.094
When I log in to a new session 24.991
And all notifications have disappeared 15.615
Then I am recommended to reinstall Tails due to partitioning errors 0.371
<"OK"> was expected to include
<"Creation of Persistent Storage has been disabled">. (Test::Unit::AssertionFailedError)
./features/step_definitions/hardware.rb:94:in `/^I am recommended to reinstall Tails due to partitioning errors$/'
features/hardware_failure.feature:59:in `Then I am recommended to reinstall Tails due to partitioning errors'
# We are gonna verify the dialog again so we need to clean up the
# first instance.
And I close the "zenity" window 0.000
And I am told that Persistent Storage cannot be created 0.000
And Tails detected partitioning error partitioning-corruption 0.000
After features/support/hooks.rb:330 8.923

SCENARIO FAILED: 'GPT backup corruption without a persistent partition' (at time 04:10:17)

Boot log: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/04:10:17_GPT_backup_corruption_without_a_persistent_partition.file_content_var_log_boot.log

Screenshot: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/04:10:17_GPT_backup_corruption_without_a_persistent_partition.png

Video: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/04:10:17_GPT_backup_corruption_without_a_persistent_partition.mkv

Systemd journal: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/04:10:17_GPT_backup_corruption_without_a_persistent_partition.journal

After features/support/hooks.rb:100 0.106
Tags: @product
2:55.939
Scenario Outline Disk partitioning errors without a persistent partition
Before features/support/hooks.rb:266 0.000
Before features/support/hooks.rb:273 0.006
Steps
Given a computer 0.217
And I set Tails to boot with options "test_partitioning_errors=part_resize" 0.000
And I temporarily create a 7200 MiB disk named "temp" 0.049
And I plug USB drive "temp" 1.049
And I write the Tails USB image to disk "temp" 29.734
When I start Tails from USB drive "temp" with network unplugged 2:24.888
cannot find TailsGreeter.png on the screen (FindFailed)
./features/support/helpers/screen.rb:181:in `rescue in wait'
./features/support/helpers/screen.rb:165:in `wait'
./features/step_definitions/common_steps.rb:449:in `/^the computer (?:re)?boots Tails$/'
./features/step_definitions/common_steps.rb:276:in `/^I start Tails from (.+?) drive "(.+?)"( with network unplugged)?( and I login( with persistence enabled)?( with the changed persistence passphrase)?( (?:and|with) an administration password)?)?$/'
features/hardware_failure.feature:81:in `When I start Tails from USB drive "temp" with network unplugged'
features/hardware_failure.feature:72:in `When I start Tails from USB drive "temp" with network unplugged'
Then Tails is running from USB drive "temp" 0.000
And the Greeter recommends reinstalling Tails due to partitioning errors 0.000
And the Greeter forbids starting Tails 0.000
And the Greeter forbids all settings but language 0.000
And Tails detected partitioning error system-partition-not-resized 0.000
After features/support/hooks.rb:330 7.689

SCENARIO FAILED: 'Disk partitioning errors without a persistent partition, Examples (#2)' (at time 04:17:06)

Boot log: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/04:17:06_Disk_partitioning_errors_without_a_persistent_partition,_Examples___2_.file_content_var_log_boot.log

Screenshot: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/04:17:06_Disk_partitioning_errors_without_a_persistent_partition,_Examples___2_.png

Video: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/04:17:06_Disk_partitioning_errors_without_a_persistent_partition,_Examples___2_.mkv

Systemd journal: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/04:17:06_Disk_partitioning_errors_without_a_persistent_partition,_Examples___2_.journal

After features/support/hooks.rb:100 0.137
Tags: @product
9.805
Scenario The included APT repository keys are up-to-date
Before features/support/hooks.rb:266 0.000
Before features/support/hooks.rb:273 0.006
Steps
Given I have started Tails from DVD without network and logged in 9.614
Then the keys trusted by APT are valid for the next 3 months 0.190
Command failed: apt-key adv --batch --with-colons --fingerprint --list-key
error code: 127
stdout: 
stderr: /bin/sh: 1: apt-key: not found
.
<false> is not true. (ExecutionFailedInVM)
./features/support/helpers/vm_helper.rb:496:in `rescue in execute_successfully'
./features/support/helpers/vm_helper.rb:493:in `execute_successfully'
./features/step_definitions/checks.rb:19:in `assert_all_keys_are_valid_for_n_months'
./features/step_definitions/checks.rb:8:in `/^the keys trusted by APT are valid for the next (\d+) months$/'
features/keys.feature:11:in `Then the keys trusted by APT are valid for the next 3 months'
After features/support/hooks.rb:330 6.161

SCENARIO FAILED: 'The included APT repository keys are up-to-date' (at time 04:21:41)

Boot log: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/04:21:41_The_included_APT_repository_keys_are_up-to-date.file_content_var_log_boot.log

Screenshot: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/04:21:41_The_included_APT_repository_keys_are_up-to-date.png

Video: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/04:21:41_The_included_APT_repository_keys_are_up-to-date.mkv

Systemd journal: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/04:21:41_The_included_APT_repository_keys_are_up-to-date.journal

After features/support/hooks.rb:100 0.000
Feature: Localization
Tags: @product
2:6.674
Scenario Outline Do not localize the XDG User Dirs to be able to use those dirs in Tor Browser (#19255)
Before features/support/hooks.rb:266 0.000
Before features/support/hooks.rb:273 0.005
Steps
Given I have started Tails from DVD without network and stopped at Tails Greeter's login screen 8.220
And I log in to a new session in German (de) 40.090
Then the live user's Pictures directory exists 0.289
And there is a GNOME bookmark for the Pictures directory 1:18.074
try_for() timeout expired
Last ignored exception was: Test::Unit::AssertionFailedError: <Dogtail::Failure> exception was expected but none was thrown.
./features/step_definitions/common_steps.rb:1198:in `block (2 levels) in <top (required)>'
./features/support/helpers/misc_helpers.rb:78:in `block (2 levels) in try_for'
./features/support/helpers/misc_helpers.rb:66:in `loop'
./features/support/helpers/misc_helpers.rb:66:in `block in try_for'
./features/support/helpers/misc_helpers.rb:65:in `try_for'
./features/step_definitions/common_steps.rb:1196:in `block in <top (required)>'
./features/step_definitions/common_steps.rb:1221:in `block in <top (required)>' (Timeout::Error)
./features/support/helpers/misc_helpers.rb:145:in `rescue in try_for'
./features/support/helpers/misc_helpers.rb:51:in `try_for'
./features/step_definitions/common_steps.rb:1196:in `/^I close the "([^"]+)" window via AltF4$/'
./features/step_definitions/common_steps.rb:1221:in `/^there is a GNOME bookmark for the (.*) directory$/'
features/localization.feature:17:in `And there is a GNOME bookmark for the Pictures directory'
features/localization.feature:11:in `And there is a GNOME bookmark for the <dir> directory'
After features/support/hooks.rb:330 6.062

SCENARIO FAILED: 'Do not localize the XDG User Dirs to be able to use those dirs in Tor Browser (#19255), Examples (#4)' (at time 04:28:19)

Boot log: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/04:28:19_Do_not_localize_the_XDG_User_Dirs_to_be_able_to_use_those_dirs_in_Tor_Browser___19255_,_Examples___4_.file_content_var_log_boot.log

Screenshot: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/04:28:19_Do_not_localize_the_XDG_User_Dirs_to_be_able_to_use_those_dirs_in_Tor_Browser___19255_,_Examples___4_.png

Video: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/04:28:19_Do_not_localize_the_XDG_User_Dirs_to_be_able_to_use_those_dirs_in_Tor_Browser___19255_,_Examples___4_.mkv

Systemd journal: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/04:28:19_Do_not_localize_the_XDG_User_Dirs_to_be_able_to_use_those_dirs_in_Tor_Browser___19255_,_Examples___4_.journal

After features/support/hooks.rb:100 0.000
Tags: @product
4:58.399
Scenario Changing the Persistent Storage passphrase
Before features/support/hooks.rb:266 0.000
Before features/support/hooks.rb:273 0.006
Steps
Given I have started Tails without network from a USB drive with a persistent partition enabled and logged in 10.328
# Note that if anything fails after the passphrase was changed and
# before it's changed back below, subsequent scenarios might fail
# because the Persistent Storage doesn't have the expected passphrase.
When I change the passphrase of the Persistent Storage 42.782
And I shutdown Tails and wait for the computer to power off 15.328
Then I start Tails from USB drive "__internal" with network unplugged and I login with the changed persistence passphrase 3:35.775
And I change the passphrase of the Persistent Storage back to the original 14.184
try_for() timeout expired (Timeout::Error)
./features/support/helpers/misc_helpers.rb:145:in `rescue in try_for'
./features/support/helpers/misc_helpers.rb:51:in `try_for'
./features/support/helpers/dogtail.rb:371:in `grabFocus'
./features/step_definitions/usb.rb:457:in `/^I change the passphrase of the Persistent Storage( back to the original)?$/'
features/persistence.feature:124:in `And I change the passphrase of the Persistent Storage back to the original'
After features/support/hooks.rb:330 5.977

SCENARIO FAILED: 'Changing the Persistent Storage passphrase' (at time 06:04:09)

Boot log: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/06:04:09_Changing_the_Persistent_Storage_passphrase.file_content_var_log_boot.log

Screenshot: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/06:04:09_Changing_the_Persistent_Storage_passphrase.png

Video: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/06:04:09_Changing_the_Persistent_Storage_passphrase.mkv

Systemd journal: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/06:04:09_Changing_the_Persistent_Storage_passphrase.journal

After features/support/hooks.rb:100 0.009
51.086
Background
Before features/support/hooks.rb:266 0.000
Before features/support/hooks.rb:273 0.010
Steps
Given I have started Tails without network from a USB drive without a persistent partition and logged in 11.649
And I create a persistent partition with the default settings 22.878
Then the "TorConfiguration" tps feature is not enabled 2.792
When the network is plugged 0.072
And the Tor Connection Assistant autostarts 8.280
And I configure some persistent obfs4 bridges in the Tor Connection Assistant 5.413
try_for() timeout expired (Timeout::Error)
./features/support/helpers/misc_helpers.rb:145:in `rescue in try_for'
./features/support/helpers/misc_helpers.rb:51:in `try_for'
./features/step_definitions/tor.rb:826:in `click_connect_to_tor'
./features/step_definitions/tor.rb:635:in `block (2 levels) in <top (required)>'
./features/step_definitions/tor.rb:499:in `tca_configure'
./features/step_definitions/tor.rb:626:in `/^I configure (?:some|the) (persistent )?(\w+) bridges (from a QR code )?in the Tor Connection Assistant(?: in (easy|hide) mode)?( without connecting|)$/'
features/persistent_tor_bridges.feature:13:in `And I configure some persistent obfs4 bridges in the Tor Connection Assistant'
And I wait until Tor is ready 0.000
Then tca.conf includes the configured bridges 0.000
And the "TorConfiguration" tps feature is enabled and active 0.000
When I cold reboot the computer 0.000
And the computer reboots Tails 0.000
And I enable persistence 0.000
And I capture all network traffic 0.000
And I log in to a new session 0.000
Then tca.conf includes the configured bridges 0.000
When the network is plugged 0.000
And the Tor Connection Assistant autostarts 0.000
Tags: @product
0.000
Scenario Disabling persistence of Tor bridges
Steps
When I choose to connect to Tor automatically 0.000
And I accept Tor Connection's offer to use my persistent bridges 0.000
And I disable saving bridges to Persistent Storage 0.000
And I configure a direct connection in the Tor Connection Assistant 0.000
Then I wait until Tor is ready 0.000
And tca.conf includes no bridge 0.000
And the "TorConfiguration" tps feature is not enabled and not active 0.000
And all Internet traffic has only flowed through Tor or connectivity check service 0.000
After features/support/hooks.rb:330 5.460

SCENARIO FAILED: 'Disabling persistence of Tor bridges' (at time 06:34:32)

Boot log: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/06:34:32_Disabling_persistence_of_Tor_bridges.file_content_var_log_boot.log

Screenshot: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/06:34:32_Disabling_persistence_of_Tor_bridges.png

Video: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/06:34:32_Disabling_persistence_of_Tor_bridges.mkv

Systemd journal: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/06:34:32_Disabling_persistence_of_Tor_bridges.journal

After features/support/hooks.rb:100 0.000
54.811
Scenario Chatting with some friend over XMPP
Before features/support/hooks.rb:266 0.000
Before features/support/hooks.rb:273 0.006
Before features/support/hooks.rb:513 0.004
Steps
Given I have started Tails from DVD and logged in and the network is connected 18.677
When I start "Pidgin Internet Messenger" via GNOME Activities Overview 10.863
Then I see Pidgin's account manager window 4.519
When I create my XMPP account 20.751
cannot find PidginAddAccountWindow.png on the screen (FindFailed)
./features/support/helpers/screen.rb:181:in `rescue in wait'
./features/support/helpers/screen.rb:165:in `wait'
./features/step_definitions/pidgin.rb:123:in `/^I create my XMPP account$/'
features/pidgin.feature:21:in `When I create my XMPP account'
And I close Pidgin's account manager window 0.000
Then Pidgin automatically enables my XMPP account 0.000
Given my XMPP friend goes online 0.000
When I start a conversation with my friend 0.000
And I say something to my friend 0.000
Then I receive a response from my friend 0.000
After features/support/hooks.rb:521 0.059
After features/support/hooks.rb:330 6.122

SCENARIO FAILED: 'Chatting with some friend over XMPP' (at time 06:36:05)

Boot log: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/06:36:05_Chatting_with_some_friend_over_XMPP.file_content_var_log_boot.log

Screenshot: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/06:36:05_Chatting_with_some_friend_over_XMPP.png

Video: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/06:36:05_Chatting_with_some_friend_over_XMPP.mkv

Systemd journal: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/06:36:05_Chatting_with_some_friend_over_XMPP.journal

After features/support/hooks.rb:100 0.000
51.634
Scenario Chatting with some friend over XMPP in a multi-user chat
Before features/support/hooks.rb:266 0.000
Before features/support/hooks.rb:273 0.006
Before features/support/hooks.rb:513 0.006
Steps
Given I have started Tails from DVD and logged in and the network is connected 15.034
When I start "Pidgin Internet Messenger" via GNOME Activities Overview 10.960
Then I see Pidgin's account manager window 4.813
When I create my XMPP account 20.825
cannot find PidginAddAccountWindow.png on the screen (FindFailed)
./features/support/helpers/screen.rb:181:in `rescue in wait'
./features/support/helpers/screen.rb:165:in `wait'
./features/step_definitions/pidgin.rb:123:in `/^I create my XMPP account$/'
features/pidgin.feature:34:in `When I create my XMPP account'
And I close Pidgin's account manager window 0.000
Then Pidgin automatically enables my XMPP account 0.000
When I join some empty multi-user chat 0.000
And I clear the multi-user chat's scrollback 0.000
And my XMPP friend goes online and joins the multi-user chat 0.000
Then I can see that my friend joined the multi-user chat 0.000
And I say something to my friend in the multi-user chat 0.000
Then I receive a response from my friend in the multi-user chat 0.000
When I say https://gitlab.tails.boum.org/tails to my friend in the multi-user chat 0.000
Then I see the Tails GitLab URL 0.000
When I wait 10 seconds 0.000
And I click on the Tails GitLab URL 0.000
Then the Tor Browser loads the Tails GitLab 0.000
After features/support/hooks.rb:521 0.043
After features/support/hooks.rb:330 6.371

SCENARIO FAILED: 'Chatting with some friend over XMPP in a multi-user chat' (at time 06:37:03)

Boot log: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/06:37:03_Chatting_with_some_friend_over_XMPP_in_a_multi-user_chat.file_content_var_log_boot.log

Screenshot: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/06:37:03_Chatting_with_some_friend_over_XMPP_in_a_multi-user_chat.png

Video: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/06:37:03_Chatting_with_some_friend_over_XMPP_in_a_multi-user_chat.mkv

Systemd journal: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/06:37:03_Chatting_with_some_friend_over_XMPP_in_a_multi-user_chat.journal

After features/support/hooks.rb:100 0.000
48.324
Scenario Connecting to the tails multi-user chat with my XMPP account
Before features/support/hooks.rb:266 0.000
Before features/support/hooks.rb:273 0.015
Before features/support/hooks.rb:513 0.009
Steps
Given I have started Tails from DVD and logged in and the network is connected 12.643
When I start "Pidgin Internet Messenger" via GNOME Activities Overview 10.962
Then I see Pidgin's account manager window 3.986
And I create my XMPP account 20.732
cannot find PidginAddAccountWindow.png on the screen (FindFailed)
./features/support/helpers/screen.rb:181:in `rescue in wait'
./features/support/helpers/screen.rb:165:in `wait'
./features/step_definitions/pidgin.rb:123:in `/^I create my XMPP account$/'
features/pidgin.feature:54:in `And I create my XMPP account'
And I close Pidgin's account manager window 0.000
Then Pidgin automatically enables my XMPP account 0.000
And I can join the "tails" channel on "chat.disroot.org" 0.000
After features/support/hooks.rb:521 0.041
After features/support/hooks.rb:330 6.069

SCENARIO FAILED: 'Connecting to the tails multi-user chat with my XMPP account' (at time 06:37:58)

Boot log: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/06:37:58_Connecting_to_the_tails_multi-user_chat_with_my_XMPP_account.file_content_var_log_boot.log

Screenshot: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/06:37:58_Connecting_to_the_tails_multi-user_chat_with_my_XMPP_account.png

Video: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/06:37:58_Connecting_to_the_tails_multi-user_chat_with_my_XMPP_account.mkv

Systemd journal: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/06:37:58_Connecting_to_the_tails_multi-user_chat_with_my_XMPP_account.journal

After features/support/hooks.rb:100 0.000
1:29.767
Scenario Using a persistent Pidgin configuration
Before features/support/hooks.rb:266 0.000
Before features/support/hooks.rb:273 0.005
Before features/support/hooks.rb:513 0.004
Steps
Given I have started Tails without network from a USB drive with a persistent partition enabled and logged in 12.994
And the network is plugged 0.299
And Tor is ready 32.132
And available upgrades have been checked 1.138
And all notifications have disappeared 5.539
When I start "Pidgin Internet Messenger" via GNOME Activities Overview 10.943
Then I see Pidgin's account manager window 5.897
When I create my XMPP account 20.821
cannot find PidginAddAccountWindow.png on the screen (FindFailed)
./features/support/helpers/screen.rb:181:in `rescue in wait'
./features/support/helpers/screen.rb:165:in `wait'
./features/step_definitions/pidgin.rb:123:in `/^I create my XMPP account$/'
features/pidgin.feature:68:in `When I create my XMPP account'
And I close Pidgin's account manager window 0.000
Then Pidgin automatically enables my XMPP account 0.000
When I close Pidgin 0.000
And I take note of the configured Pidgin accounts 0.000
And I shutdown Tails and wait for the computer to power off 0.000
Given a computer 0.000
And I start Tails from USB drive "__internal" and I login with persistence enabled 0.000
And Pidgin has the expected persistent accounts configured 0.000
When I start "Pidgin Internet Messenger" via GNOME Activities Overview 0.000
Then Pidgin automatically enables my XMPP account 0.000
And I join some empty multi-user chat 0.000
# Exercise Pidgin AppArmor profile with persistence enabled.
# This should really be in dedicated scenarios, but it would be
# too costly to set up the virtual USB drive with persistence more
# than once in this feature.
Given I start monitoring the AppArmor log of "/usr/bin/pidgin" 0.000
Then I cannot add a certificate from the "/home/amnesia/.gnupg" directory to Pidgin 0.000
And AppArmor has denied "/usr/bin/pidgin" from opening "/home/amnesia/.gnupg/test.crt" 0.000
When I close Pidgin's certificate import failure dialog 0.000
And I close Pidgin's certificate manager 0.000
Given I restart monitoring the AppArmor log of "/usr/bin/pidgin" 0.000
Then I cannot add a certificate from the "/live/persistence/TailsData_unlocked/gnupg" directory to Pidgin 0.000
And AppArmor has denied "/usr/bin/pidgin" from opening "/live/persistence/TailsData_unlocked/gnupg/test.crt" 0.000
When I close Pidgin's certificate import failure dialog 0.000
And I close Pidgin's certificate manager 0.000
Then I can add a certificate from the "/home/amnesia" directory to Pidgin 0.000
After features/support/hooks.rb:521 0.033
After features/support/hooks.rb:330 5.786

SCENARIO FAILED: 'Using a persistent Pidgin configuration' (at time 06:39:34)

Boot log: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/06:39:34_Using_a_persistent_Pidgin_configuration.file_content_var_log_boot.log

Screenshot: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/06:39:34_Using_a_persistent_Pidgin_configuration.png

Video: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/06:39:34_Using_a_persistent_Pidgin_configuration.mkv

Systemd journal: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/06:39:34_Using_a_persistent_Pidgin_configuration.journal

After features/support/hooks.rb:100 0.000
Tags: @product
8.845
Scenario No unexpected error messages in the journal after booting from DVD
Before features/support/hooks.rb:266 0.000
Before features/support/hooks.rb:273 0.008
Steps
Given I have started Tails from DVD without network and logged in 8.761
Then there are no unexpected messages of priority "err" or higher in the journal 0.084
Unexpected error messages in the journal: [
  {
    "MESSAGE": "/usr/lib/udev/rules.d/90-alsa-restore.rules:18 GOTO=\"alsa_restore_std\" has no matching label, ignoring.",
    "SYSLOG_IDENTIFIER": "systemd-udevd",
    "CODE_FILE": "src/udev/udev-rules.c",
    "CODE_FUNC": "rule_resolve_goto",
    "CODE_LINE": "1376",
    "PRIORITY": "3",
    "SYSLOG_FACILITY": "3",
    "TID": "575",
    "_BOOT_ID": "da05b10e66b5486fa4079a8b2e13c3a6",
    "_CAP_EFFECTIVE": "1f7fdffffff",
    "_CMDLINE": "/usr/lib/systemd/systemd-udevd",
    "_COMM": "systemd-udevd",
    "_EXE": "/usr/bin/udevadm",
    "_GID": "0",
    "_HOSTNAME": "amnesia",
    "_MACHINE_ID": "203552d5819c41f3800e2c8ef2545404",
    "_PID": "575",
    "_RUNTIME_SCOPE": "system",
    "_SELINUX_CONTEXT": "unconfined\n",
    "_SOURCE_REALTIME_TIMESTAMP": "1749133958199613",
    "_SYSTEMD_CGROUP": "/system.slice/systemd-udevd.service/udev",
    "_SYSTEMD_INVOCATION_ID": "997ef96379a041dc8e75334f3ec83eb6",
    "_SYSTEMD_SLICE": "system.slice",
    "_SYSTEMD_UNIT": "systemd-udevd.service",
    "_TRANSPORT": "journal",
    "_UID": "0",
    "__CURSOR": "s=c67c5105f1244c9eaaac7e53e5780977;i=37c;b=da05b10e66b5486fa4079a8b2e13c3a6;m=e18fe3;t=636d3fc3e581c;x=33204bd8aef3f456",
    "__MONOTONIC_TIMESTAMP": "14782435",
    "__REALTIME_TIMESTAMP": "1749133958207516",
    "__SEQNUM": "892",
    "__SEQNUM_ID": "c67c5105f1244c9eaaac7e53e5780977"
  },
  {
    "MESSAGE": "/usr/lib/udev/rules.d/90-alsa-restore.rules:22 GOTO=\"alsa_restore_std\" has no matching label, ignoring.",
    "SYSLOG_IDENTIFIER": "systemd-udevd",
    "CODE_FILE": "src/udev/udev-rules.c",
    "CODE_FUNC": "rule_resolve_goto",
    "CODE_LINE": "1376",
    "PRIORITY": "3",
    "SYSLOG_FACILITY": "3",
    "TID": "575",
    "_BOOT_ID": "da05b10e66b5486fa4079a8b2e13c3a6",
    "_CAP_EFFECTIVE": "1f7fdffffff",
    "_CMDLINE": "/usr/lib/systemd/systemd-udevd",
    "_COMM": "systemd-udevd",
    "_EXE": "/usr/bin/udevadm",
    "_GID": "0",
    "_HOSTNAME": "amnesia",
    "_MACHINE_ID": "203552d5819c41f3800e2c8ef2545404",
    "_PID": "575",
    "_RUNTIME_SCOPE": "system",
    "_SELINUX_CONTEXT": "unconfined\n",
    "_SOURCE_REALTIME_TIMESTAMP": "1749133958199638",
    "_SYSTEMD_CGROUP": "/system.slice/systemd-udevd.service/udev",
    "_SYSTEMD_INVOCATION_ID": "997ef96379a041dc8e75334f3ec83eb6",
    "_SYSTEMD_SLICE": "system.slice",
    "_SYSTEMD_UNIT": "systemd-udevd.service",
    "_TRANSPORT": "journal",
    "_UID": "0",
    "__CURSOR": "s=c67c5105f1244c9eaaac7e53e5780977;i=37e;b=da05b10e66b5486fa4079a8b2e13c3a6;m=e19160;t=636d3fc3e5999;x=46e471ff6ba788da",
    "__MONOTONIC_TIMESTAMP": "14782816",
    "__REALTIME_TIMESTAMP": "1749133958207897",
    "__SEQNUM": "894",
    "__SEQNUM_ID": "c67c5105f1244c9eaaac7e53e5780977"
  },
  {
    "MESSAGE": "The group `powerdev' does not exist.",
    "SYSLOG_IDENTIFIER": "adduser",
    "PRIORITY": "2",
    "SYSLOG_FACILITY": "1",
    "SYSLOG_PID": "1798",
    "SYSLOG_TIMESTAMP": "Jun  5 14:32:45 ",
    "_BOOT_ID": "da05b10e66b5486fa4079a8b2e13c3a6",
    "_CAP_EFFECTIVE": "1ffffffffff",
    "_CMDLINE": "adduser",
    "_COMM": "adduser",
    "_EXE": "/usr/bin/perl",
    "_GID": "0",
    "_HOSTNAME": "amnesia",
    "_MACHINE_ID": "203552d5819c41f3800e2c8ef2545404",
    "_PID": "1798",
    "_RUNTIME_SCOPE": "system",
    "_SELINUX_CONTEXT": "unconfined\n",
    "_SOURCE_REALTIME_TIMESTAMP": "1749133965467684",
    "_SYSTEMD_CGROUP": "/system.slice/live-config.service",
    "_SYSTEMD_INVOCATION_ID": "8e495494e64d4e388c73f99de968764d",
    "_SYSTEMD_SLICE": "system.slice",
    "_SYSTEMD_UNIT": "live-config.service",
    "_TRANSPORT": "syslog",
    "_UID": "0",
    "__CURSOR": "s=c67c5105f1244c9eaaac7e53e5780977;i=3fe;b=da05b10e66b5486fa4079a8b2e13c3a6;m=15057fc;t=636d3fcad2035;x=855a51537bf2ba55",
    "__MONOTONIC_TIMESTAMP": "22042620",
    "__REALTIME_TIMESTAMP": "1749133965467701",
    "__SEQNUM": "1022",
    "__SEQNUM_ID": "c67c5105f1244c9eaaac7e53e5780977"
  },
  {
    "MESSAGE": "/usr/lib/udev/rules.d/90-alsa-restore.rules:18 GOTO=\"alsa_restore_std\" has no matching label, ignoring.",
    "SYSLOG_IDENTIFIER": "systemd-udevd",
    "CODE_FILE": "src/udev/udev-rules.c",
    "CODE_FUNC": "rule_resolve_goto",
    "CODE_LINE": "1376",
    "PRIORITY": "3",
    "SYSLOG_FACILITY": "3",
    "TID": "575",
    "_BOOT_ID": "da05b10e66b5486fa4079a8b2e13c3a6",
    "_CAP_EFFECTIVE": "1f7fdffffff",
    "_CMDLINE": "/usr/lib/systemd/systemd-udevd",
    "_COMM": "systemd-udevd",
    "_EXE": "/usr/bin/udevadm",
    "_GID": "0",
    "_HOSTNAME": "amnesia",
    "_MACHINE_ID": "203552d5819c41f3800e2c8ef2545404",
    "_PID": "575",
    "_RUNTIME_SCOPE": "system",
    "_SELINUX_CONTEXT": "unconfined\n",
    "_SOURCE_REALTIME_TIMESTAMP": "1749134160778857",
    "_SYSTEMD_CGROUP": "/system.slice/systemd-udevd.service/udev",
    "_SYSTEMD_INVOCATION_ID": "997ef96379a041dc8e75334f3ec83eb6",
    "_SYSTEMD_SLICE": "system.slice",
    "_SYSTEMD_UNIT": "systemd-udevd.service",
    "_TRANSPORT": "journal",
    "_UID": "0",
    "__CURSOR": "s=c67c5105f1244c9eaaac7e53e5780977;i=8b3;b=da05b10e66b5486fa4079a8b2e13c3a6;m=6b06918;t=636d4085156d3;x=e1dadf19b4c4f956",
    "__MONOTONIC_TIMESTAMP": "112224536",
    "__REALTIME_TIMESTAMP": "1749134160778963",
    "__SEQNUM": "2227",
    "__SEQNUM_ID": "c67c5105f1244c9eaaac7e53e5780977"
  },
  {
    "MESSAGE": "/usr/lib/udev/rules.d/90-alsa-restore.rules:22 GOTO=\"alsa_restore_std\" has no matching label, ignoring.",
    "SYSLOG_IDENTIFIER": "systemd-udevd",
    "CODE_FILE": "src/udev/udev-rules.c",
    "CODE_FUNC": "rule_resolve_goto",
    "CODE_LINE": "1376",
    "PRIORITY": "3",
    "SYSLOG_FACILITY": "3",
    "TID": "575",
    "_BOOT_ID": "da05b10e66b5486fa4079a8b2e13c3a6",
    "_CAP_EFFECTIVE": "1f7fdffffff",
    "_CMDLINE": "/usr/lib/systemd/systemd-udevd",
    "_COMM": "systemd-udevd",
    "_EXE": "/usr/bin/udevadm",
    "_GID": "0",
    "_HOSTNAME": "amnesia",
    "_MACHINE_ID": "203552d5819c41f3800e2c8ef2545404",
    "_PID": "575",
    "_RUNTIME_SCOPE": "system",
    "_SELINUX_CONTEXT": "unconfined\n",
    "_SOURCE_REALTIME_TIMESTAMP": "1749134160778920",
    "_SYSTEMD_CGROUP": "/system.slice/systemd-udevd.service/udev",
    "_SYSTEMD_INVOCATION_ID": "997ef96379a041dc8e75334f3ec83eb6",
    "_SYSTEMD_SLICE": "system.slice",
    "_SYSTEMD_UNIT": "systemd-udevd.service",
    "_TRANSPORT": "journal",
    "_UID": "0",
    "__CURSOR": "s=c67c5105f1244c9eaaac7e53e5780977;i=8b5;b=da05b10e66b5486fa4079a8b2e13c3a6;m=6b06d52;t=636d408515b0e;x=88272f54479e3e95",
    "__MONOTONIC_TIMESTAMP": "112225618",
    "__REALTIME_TIMESTAMP": "1749134160780046",
    "__SEQNUM": "2229",
    "__SEQNUM_ID": "c67c5105f1244c9eaaac7e53e5780977"
  },
  {
    "MESSAGE": "The user `amnesia' is not a member of group `sudo'.",
    "SYSLOG_IDENTIFIER": "deluser",
    "PRIORITY": "2",
    "SYSLOG_FACILITY": "1",
    "SYSLOG_PID": "18779",
    "SYSLOG_TIMESTAMP": "Jun  5 14:36:08 ",
    "_AUDIT_LOGINUID": "1000",
    "_AUDIT_SESSION": "4",
    "_BOOT_ID": "da05b10e66b5486fa4079a8b2e13c3a6",
    "_CAP_EFFECTIVE": "1ffffffffff",
    "_CMDLINE": "/usr/bin/perl -T /usr/sbin/deluser amnesia sudo",
    "_COMM": "deluser",
    "_EXE": "/usr/bin/perl",
    "_GID": "0",
    "_HOSTNAME": "amnesia",
    "_MACHINE_ID": "203552d5819c41f3800e2c8ef2545404",
    "_PID": "18779",
    "_RUNTIME_SCOPE": "system",
    "_SELINUX_CONTEXT": "unconfined\n",
    "_SOURCE_REALTIME_TIMESTAMP": "1749134168521654",
    "_SYSTEMD_CGROUP": "/user.slice/user-1000.slice/session-4.scope",
    "_SYSTEMD_INVOCATION_ID": "fe83a84ab9b249d586097d40d642fdb0",
    "_SYSTEMD_OWNER_UID": "1000",
    "_SYSTEMD_SESSION": "4",
    "_SYSTEMD_SLICE": "user-1000.slice",
    "_SYSTEMD_UNIT": "session-4.scope",
    "_SYSTEMD_USER_SLICE": "-.slice",
    "_TRANSPORT": "syslog",
    "_UID": "0",
    "__CURSOR": "s=c67c5105f1244c9eaaac7e53e5780977;i=123c;b=da05b10e66b5486fa4079a8b2e13c3a6;m=7268e25;t=636d408c77be1;x=3978747d509c58",
    "__MONOTONIC_TIMESTAMP": "119967269",
    "__REALTIME_TIMESTAMP": "1749134168521697",
    "__SEQNUM": "4668",
    "__SEQNUM_ID": "c67c5105f1244c9eaaac7e53e5780977"
  }
].
<false> is not true. (Test::Unit::AssertionFailedError)
./features/step_definitions/checks.rb:237:in `/^there are no unexpected messages of priority "err" or higher in the journal$/'
features/sane_defaults.feature:21:in `Then there are no unexpected messages of priority "err" or higher in the journal'
After features/support/hooks.rb:330 6.180

SCENARIO FAILED: 'No unexpected error messages in the journal after booting from DVD' (at time 06:45:00)

Boot log: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/06:45:00_No_unexpected_error_messages_in_the_journal_after_booting_from_DVD.file_content_var_log_boot.log

Screenshot: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/06:45:00_No_unexpected_error_messages_in_the_journal_after_booting_from_DVD.png

Video: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/06:45:00_No_unexpected_error_messages_in_the_journal_after_booting_from_DVD.mkv

Systemd journal: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/06:45:00_No_unexpected_error_messages_in_the_journal_after_booting_from_DVD.journal

After features/support/hooks.rb:100 0.000
Tags: @product
31.717
Scenario No unexpected error messages in the journal after booting from USB drive
Before features/support/hooks.rb:266 0.000
Before features/support/hooks.rb:273 0.007
Steps
Given I have started Tails without network from a USB drive with a persistent partition enabled and logged in 31.592
Then there are no unexpected messages of priority "err" or higher in the journal 0.124
Unexpected error messages in the journal: [
  {
    "MESSAGE": "/usr/lib/udev/rules.d/90-alsa-restore.rules:18 GOTO=\"alsa_restore_std\" has no matching label, ignoring.",
    "SYSLOG_IDENTIFIER": "systemd-udevd",
    "CODE_FILE": "src/udev/udev-rules.c",
    "CODE_FUNC": "rule_resolve_goto",
    "CODE_LINE": "1376",
    "PRIORITY": "3",
    "SYSLOG_FACILITY": "3",
    "TID": "708",
    "_BOOT_ID": "b6dc815d40db4050b2a5273644c318bb",
    "_CAP_EFFECTIVE": "1f7fdffffff",
    "_CMDLINE": "/usr/lib/systemd/systemd-udevd",
    "_COMM": "systemd-udevd",
    "_EXE": "/usr/bin/udevadm",
    "_GID": "0",
    "_HOSTNAME": "amnesia",
    "_MACHINE_ID": "203552d5819c41f3800e2c8ef2545404",
    "_PID": "708",
    "_RUNTIME_SCOPE": "system",
    "_SELINUX_CONTEXT": "unconfined\n",
    "_SOURCE_REALTIME_TIMESTAMP": "1749136398114149",
    "_SYSTEMD_CGROUP": "/system.slice/systemd-udevd.service/udev",
    "_SYSTEMD_INVOCATION_ID": "5e31b6160e154725847461c4d8735709",
    "_SYSTEMD_SLICE": "system.slice",
    "_SYSTEMD_UNIT": "systemd-udevd.service",
    "_TRANSPORT": "journal",
    "_UID": "0",
    "__CURSOR": "s=cad0a8ec49254d6d865a01fae2d3c5ed;i=362;b=b6dc815d40db4050b2a5273644c318bb;m=f0e1df;t=636d48dac5d95;x=9d9c5cc71e728c9a",
    "__MONOTONIC_TIMESTAMP": "15786463",
    "__REALTIME_TIMESTAMP": "1749136398114197",
    "__SEQNUM": "866",
    "__SEQNUM_ID": "cad0a8ec49254d6d865a01fae2d3c5ed"
  },
  {
    "MESSAGE": "/usr/lib/udev/rules.d/90-alsa-restore.rules:22 GOTO=\"alsa_restore_std\" has no matching label, ignoring.",
    "SYSLOG_IDENTIFIER": "systemd-udevd",
    "CODE_FILE": "src/udev/udev-rules.c",
    "CODE_FUNC": "rule_resolve_goto",
    "CODE_LINE": "1376",
    "PRIORITY": "3",
    "SYSLOG_FACILITY": "3",
    "TID": "708",
    "_BOOT_ID": "b6dc815d40db4050b2a5273644c318bb",
    "_CAP_EFFECTIVE": "1f7fdffffff",
    "_CMDLINE": "/usr/lib/systemd/systemd-udevd",
    "_COMM": "systemd-udevd",
    "_EXE": "/usr/bin/udevadm",
    "_GID": "0",
    "_HOSTNAME": "amnesia",
    "_MACHINE_ID": "203552d5819c41f3800e2c8ef2545404",
    "_PID": "708",
    "_RUNTIME_SCOPE": "system",
    "_SELINUX_CONTEXT": "unconfined\n",
    "_SOURCE_REALTIME_TIMESTAMP": "1749136398114354",
    "_SYSTEMD_CGROUP": "/system.slice/systemd-udevd.service/udev",
    "_SYSTEMD_INVOCATION_ID": "5e31b6160e154725847461c4d8735709",
    "_SYSTEMD_SLICE": "system.slice",
    "_SYSTEMD_UNIT": "systemd-udevd.service",
    "_TRANSPORT": "journal",
    "_UID": "0",
    "__CURSOR": "s=cad0a8ec49254d6d865a01fae2d3c5ed;i=364;b=b6dc815d40db4050b2a5273644c318bb;m=f0e280;t=636d48dac5e36;x=ac871b9143142b47",
    "__MONOTONIC_TIMESTAMP": "15786624",
    "__REALTIME_TIMESTAMP": "1749136398114358",
    "__SEQNUM": "868",
    "__SEQNUM_ID": "cad0a8ec49254d6d865a01fae2d3c5ed"
  },
  {
    "MESSAGE": "The group `powerdev' does not exist.",
    "SYSLOG_IDENTIFIER": "adduser",
    "PRIORITY": "2",
    "SYSLOG_FACILITY": "1",
    "SYSLOG_PID": "2449",
    "SYSLOG_TIMESTAMP": "Jun  5 15:13:26 ",
    "_BOOT_ID": "b6dc815d40db4050b2a5273644c318bb",
    "_GID": "0",
    "_HOSTNAME": "amnesia",
    "_MACHINE_ID": "203552d5819c41f3800e2c8ef2545404",
    "_PID": "2449",
    "_RUNTIME_SCOPE": "system",
    "_SOURCE_REALTIME_TIMESTAMP": "1749136406882235",
    "_TRANSPORT": "syslog",
    "_UID": "0",
    "__CURSOR": "s=cad0a8ec49254d6d865a01fae2d3c5ed;i=410;b=b6dc815d40db4050b2a5273644c318bb;m=176f2aa;t=636d48e326e5f;x=e85b4c669fd1ac75",
    "__MONOTONIC_TIMESTAMP": "24572586",
    "__REALTIME_TIMESTAMP": "1749136406900319",
    "__SEQNUM": "1040",
    "__SEQNUM_ID": "cad0a8ec49254d6d865a01fae2d3c5ed"
  },
  {
    "MESSAGE": "/usr/lib/udev/rules.d/90-alsa-restore.rules:18 GOTO=\"alsa_restore_std\" has no matching label, ignoring.",
    "SYSLOG_IDENTIFIER": "systemd-udevd",
    "CODE_FILE": "src/udev/udev-rules.c",
    "CODE_FUNC": "rule_resolve_goto",
    "CODE_LINE": "1376",
    "PRIORITY": "3",
    "SYSLOG_FACILITY": "3",
    "TID": "708",
    "_BOOT_ID": "b6dc815d40db4050b2a5273644c318bb",
    "_CAP_EFFECTIVE": "1f7fdffffff",
    "_CMDLINE": "/usr/lib/systemd/systemd-udevd",
    "_COMM": "systemd-udevd",
    "_EXE": "/usr/bin/udevadm",
    "_GID": "0",
    "_HOSTNAME": "amnesia",
    "_MACHINE_ID": "203552d5819c41f3800e2c8ef2545404",
    "_PID": "708",
    "_RUNTIME_SCOPE": "system",
    "_SELINUX_CONTEXT": "unconfined\n",
    "_SOURCE_REALTIME_TIMESTAMP": "1749136559617626",
    "_SYSTEMD_CGROUP": "/system.slice/systemd-udevd.service/udev",
    "_SYSTEMD_INVOCATION_ID": "5e31b6160e154725847461c4d8735709",
    "_SYSTEMD_SLICE": "system.slice",
    "_SYSTEMD_UNIT": "systemd-udevd.service",
    "_TRANSPORT": "journal",
    "_UID": "0",
    "__CURSOR": "s=cad0a8ec49254d6d865a01fae2d3c5ed;i=e89;b=b6dc815d40db4050b2a5273644c318bb;m=a2403a8;t=636d4974cb684;x=f42ae0c86ccdb53e",
    "__MONOTONIC_TIMESTAMP": "170132392",
    "__REALTIME_TIMESTAMP": "1749136559617668",
    "__SEQNUM": "3721",
    "__SEQNUM_ID": "cad0a8ec49254d6d865a01fae2d3c5ed"
  },
  {
    "MESSAGE": "/usr/lib/udev/rules.d/90-alsa-restore.rules:22 GOTO=\"alsa_restore_std\" has no matching label, ignoring.",
    "SYSLOG_IDENTIFIER": "systemd-udevd",
    "CODE_FILE": "src/udev/udev-rules.c",
    "CODE_FUNC": "rule_resolve_goto",
    "CODE_LINE": "1376",
    "PRIORITY": "3",
    "SYSLOG_FACILITY": "3",
    "TID": "708",
    "_BOOT_ID": "b6dc815d40db4050b2a5273644c318bb",
    "_CAP_EFFECTIVE": "1f7fdffffff",
    "_CMDLINE": "/usr/lib/systemd/systemd-udevd",
    "_COMM": "systemd-udevd",
    "_EXE": "/usr/bin/udevadm",
    "_GID": "0",
    "_HOSTNAME": "amnesia",
    "_MACHINE_ID": "203552d5819c41f3800e2c8ef2545404",
    "_PID": "708",
    "_RUNTIME_SCOPE": "system",
    "_SELINUX_CONTEXT": "unconfined\n",
    "_SOURCE_REALTIME_TIMESTAMP": "1749136559617656",
    "_SYSTEMD_CGROUP": "/system.slice/systemd-udevd.service/udev",
    "_SYSTEMD_INVOCATION_ID": "5e31b6160e154725847461c4d8735709",
    "_SYSTEMD_SLICE": "system.slice",
    "_SYSTEMD_UNIT": "systemd-udevd.service",
    "_TRANSPORT": "journal",
    "_UID": "0",
    "__CURSOR": "s=cad0a8ec49254d6d865a01fae2d3c5ed;i=e8b;b=b6dc815d40db4050b2a5273644c318bb;m=a240544;t=636d4974cb820;x=9981a31bba565848",
    "__MONOTONIC_TIMESTAMP": "170132804",
    "__REALTIME_TIMESTAMP": "1749136559618080",
    "__SEQNUM": "3723",
    "__SEQNUM_ID": "cad0a8ec49254d6d865a01fae2d3c5ed"
  },
  {
    "MESSAGE": "The user `amnesia' is not a member of group `sudo'.",
    "SYSLOG_IDENTIFIER": "deluser",
    "PRIORITY": "2",
    "SYSLOG_FACILITY": "1",
    "SYSLOG_PID": "22882",
    "SYSLOG_TIMESTAMP": "Jun  5 15:16:09 ",
    "_AUDIT_LOGINUID": "1000",
    "_AUDIT_SESSION": "4",
    "_BOOT_ID": "b6dc815d40db4050b2a5273644c318bb",
    "_CAP_EFFECTIVE": "1ffffffffff",
    "_CMDLINE": "/usr/bin/perl -T /usr/sbin/deluser amnesia sudo",
    "_COMM": "deluser",
    "_EXE": "/usr/bin/perl",
    "_GID": "0",
    "_HOSTNAME": "amnesia",
    "_MACHINE_ID": "203552d5819c41f3800e2c8ef2545404",
    "_PID": "22882",
    "_RUNTIME_SCOPE": "system",
    "_SELINUX_CONTEXT": "unconfined\n",
    "_SOURCE_REALTIME_TIMESTAMP": "1749136569072372",
    "_SYSTEMD_CGROUP": "/user.slice/user-1000.slice/session-4.scope",
    "_SYSTEMD_INVOCATION_ID": "b29ab4c46e174f3dba9adc46211e9f45",
    "_SYSTEMD_OWNER_UID": "1000",
    "_SYSTEMD_SESSION": "4",
    "_SYSTEMD_SLICE": "user-1000.slice",
    "_SYSTEMD_UNIT": "session-4.scope",
    "_SYSTEMD_USER_SLICE": "-.slice",
    "_TRANSPORT": "syslog",
    "_UID": "0",
    "__CURSOR": "s=cad0a8ec49254d6d865a01fae2d3c5ed;i=187c;b=b6dc815d40db4050b2a5273644c318bb;m=ab4484c;t=636d497dcfb29;x=99be02ef95de835a",
    "__MONOTONIC_TIMESTAMP": "179587148",
    "__REALTIME_TIMESTAMP": "1749136569072425",
    "__SEQNUM": "6268",
    "__SEQNUM_ID": "cad0a8ec49254d6d865a01fae2d3c5ed"
  }
].
<false> is not true. (Test::Unit::AssertionFailedError)
./features/step_definitions/checks.rb:237:in `/^there are no unexpected messages of priority "err" or higher in the journal$/'
features/sane_defaults.feature:25:in `Then there are no unexpected messages of priority "err" or higher in the journal'
After features/support/hooks.rb:330 5.191

SCENARIO FAILED: 'No unexpected error messages in the journal after booting from USB drive' (at time 06:45:38)

Boot log: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/06:45:38_No_unexpected_error_messages_in_the_journal_after_booting_from_USB_drive.file_content_var_log_boot.log

Screenshot: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/06:45:38_No_unexpected_error_messages_in_the_journal_after_booting_from_USB_drive.png

Video: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/06:45:38_No_unexpected_error_messages_in_the_journal_after_booting_from_USB_drive.mkv

Systemd journal: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/06:45:38_No_unexpected_error_messages_in_the_journal_after_booting_from_USB_drive.journal

After features/support/hooks.rb:100 0.000
Tags: @product
28.144
Scenario I can shutdown Tails via the system menu
Before features/support/hooks.rb:266 0.000
Before features/support/hooks.rb:273 0.007
Steps
Given I have started Tails from DVD and logged in and the network is connected 14.154
When I request a shutdown using the system menu 13.990
try_for() timeout expired (Timeout::Error)
./features/support/helpers/misc_helpers.rb:145:in `rescue in try_for'
./features/support/helpers/misc_helpers.rb:51:in `try_for'
./features/support/helpers/dogtail.rb:371:in `grabFocus'
./features/step_definitions/common_steps.rb:887:in `/^I request a (shutdown|reboot) using the system menu$/'
features/shutdown_applet.feature:6:in `When I request a shutdown using the system menu'
Then Tails eventually shuts down 0.000
After features/support/hooks.rb:330 5.723

SCENARIO FAILED: 'I can shutdown Tails via the system menu' (at time 06:46:11)

Boot log: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/06:46:11_I_can_shutdown_Tails_via_the_system_menu.file_content_var_log_boot.log

Screenshot: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/06:46:11_I_can_shutdown_Tails_via_the_system_menu.png

Video: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/06:46:11_I_can_shutdown_Tails_via_the_system_menu.mkv

Systemd journal: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/06:46:11_I_can_shutdown_Tails_via_the_system_menu.journal

After features/support/hooks.rb:100 0.000
40.008
Scenario I can reboot Tails via the system menu
Before features/support/hooks.rb:266 0.000
Before features/support/hooks.rb:273 0.004
Steps
Given I have started Tails from DVD and logged in and the network is connected 26.012
When I request a reboot using the system menu 13.996
try_for() timeout expired (Timeout::Error)
./features/support/helpers/misc_helpers.rb:145:in `rescue in try_for'
./features/support/helpers/misc_helpers.rb:51:in `try_for'
./features/support/helpers/dogtail.rb:371:in `grabFocus'
./features/step_definitions/common_steps.rb:887:in `/^I request a (shutdown|reboot) using the system menu$/'
features/shutdown_applet.feature:12:in `When I request a reboot using the system menu'
Then Tails eventually restarts 0.000
After features/support/hooks.rb:330 5.750

SCENARIO FAILED: 'I can reboot Tails via the system menu' (at time 06:46:57)

Boot log: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/06:46:57_I_can_reboot_Tails_via_the_system_menu.file_content_var_log_boot.log

Screenshot: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/06:46:57_I_can_reboot_Tails_via_the_system_menu.png

Video: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/06:46:57_I_can_reboot_Tails_via_the_system_menu.mkv

Systemd journal: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/06:46:57_I_can_reboot_Tails_via_the_system_menu.journal

After features/support/hooks.rb:100 0.000
1:24.218
Scenario The Downloads directory is usable in Tor Browser
Before features/support/hooks.rb:266 0.000
Before features/support/hooks.rb:273 0.019
Before features/support/hooks.rb:513 0.011
Steps
Given I have started Tails from DVD and logged in and the network is connected 12.590
Then the live user's Downloads directory exists 0.057
And there is a GNOME bookmark for the Downloads directory 21.673
When I start the Tor Browser 24.906
And the Tor Browser loads the startup page 1.823
Then I can save the current page as "index.html" to the Downloads directory 23.167
The
The Dogtail init script raised: SearchError: child of [desktop frame | main]: 'org.gnome.Nautilus' application
STDOUT:

STDERR:

 (Dogtail::Failure)
./features/support/helpers/dogtail.rb:93:in `run'
./features/support/helpers/dogtail.rb:70:in `initialize'
./features/step_definitions/browser.rb:6:in `new'
./features/step_definitions/browser.rb:6:in `desktop_portal_save_as'
./features/step_definitions/browser.rb:38:in `browser_save_page_as'
./features/step_definitions/browser.rb:543:in `/^I (can|cannot) save the current page as "([^"]+[.]html)" to the (.*) (directory|GNOME bookmark)$/'
features/torified_browsing.feature:24:in `Then I can save the current page as "index.html" to the Downloads directory'
And I can print the current page as "output.pdf" to the Downloads directory 0.000
After features/support/hooks.rb:521 0.032
After features/support/hooks.rb:330 10.154

SCENARIO FAILED: 'The Downloads directory is usable in Tor Browser' (at time 07:14:28)

Screenshot: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/07:14:28_The_Downloads_directory_is_usable_in_Tor_Browser.png

Video: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/07:14:28_The_Downloads_directory_is_usable_in_Tor_Browser.mkv

Systemd journal: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/07:14:28_The_Downloads_directory_is_usable_in_Tor_Browser.journal

After features/support/hooks.rb:100 0.000
1:18.254
Scenario Downloading files with the Tor Browser
Before features/support/hooks.rb:266 0.000
Before features/support/hooks.rb:273 0.007
Before features/support/hooks.rb:513 0.008
Steps
Given I have started Tails from DVD and logged in and the network is connected 13.489
When I start the Tor Browser 25.604
Then the Tor Browser loads the startup page 1.725
When I download some file in the Tor Browser to the Downloads directory 37.435
The
The Dogtail init script raised: SearchError: child of [desktop frame | main]: 'org.gnome.Nautilus' application
STDOUT:

STDERR:

 (Dogtail::Failure)
./features/support/helpers/dogtail.rb:93:in `run'
./features/support/helpers/dogtail.rb:70:in `initialize'
./features/step_definitions/browser.rb:6:in `new'
./features/step_definitions/browser.rb:6:in `desktop_portal_save_as'
./features/step_definitions/browser.rb:310:in `/^I download some file in the Tor Browser to the (.*) directory$/'
features/torified_browsing.feature:32:in `When I download some file in the Tor Browser to the Downloads directory'
Then the file is saved to the Downloads directory 0.000
After features/support/hooks.rb:521 0.052
After features/support/hooks.rb:330 10.037

SCENARIO FAILED: 'Downloading files with the Tor Browser' (at time 07:15:56)

Screenshot: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/07:15:56_Downloading_files_with_the_Tor_Browser.png

Video: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/07:15:56_Downloading_files_with_the_Tor_Browser.mkv

Systemd journal: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/07:15:56_Downloading_files_with_the_Tor_Browser.journal

After features/support/hooks.rb:100 0.000
Tags: @product
43.969
Scenario The Tor Browser uses TBB's shared libraries
Before features/support/hooks.rb:266 0.000
Before features/support/hooks.rb:273 0.008
Steps
Given I have started Tails from DVD and logged in and the network is connected 13.323
When I start the Tor Browser 26.501
Then the Tor Browser uses all expected TBB shared libraries 4.143
The
The loaded shared libraries for the firefox process are not the way we expect them.
Expected TBB libs that are absent: ["libfreeblpriv3.so", "libgkcodecs.so", "liblgpllibs.so", "libmozgtk.so", "libmozsandbox.so", "libmozsqlite3.so", "libmozwayland.so", "libnspr4.so", "libnss3.so", "libnssckbi.so", "libnssutil3.so", "libplc4.so", "libplds4.so", "libsmime3.so", "libsoftokn3.so", "libssl3.so", "libxul.so"]
Native libs that we don't want: [].
<false> is not true. (Test::Unit::AssertionFailedError)
./features/step_definitions/browser.rb:259:in `xul_app_shared_lib_check'
./features/step_definitions/browser.rb:273:in `/^the (.*) uses all expected TBB shared libraries$/'
features/torified_browsing.feature:76:in `Then the Tor Browser uses all expected TBB shared libraries'
After features/support/hooks.rb:330 7.165

SCENARIO FAILED: 'The Tor Browser uses TBB's shared libraries' (at time 07:21:25)

Boot log: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/07:21:25_The_Tor_Browser_uses_TBB_s_shared_libraries.file_content_var_log_boot.log

Screenshot: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/07:21:25_The_Tor_Browser_uses_TBB_s_shared_libraries.png

Video: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/07:21:25_The_Tor_Browser_uses_TBB_s_shared_libraries.mkv

Systemd journal: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/07:21:25_The_Tor_Browser_uses_TBB_s_shared_libraries.journal

After features/support/hooks.rb:100 0.000
Tags: @product
2:12.845
Scenario The Persistent directory is usable in Tor Browser
Before features/support/hooks.rb:266 0.000
Before features/support/hooks.rb:273 0.006
Steps
Given I have started Tails without network from a USB drive with a persistent partition enabled and logged in 10.956
And the network is plugged 0.084
And I successfully configure Tor 28.044
And available upgrades have been checked 1.160
And all notifications have disappeared 5.883
And there is a GNOME bookmark for the Persistent directory 22.741
When I start the Tor Browser 25.601
And I download some file in the Tor Browser to the Persistent directory 38.374
The
The Dogtail init script raised: SearchError: child of [desktop frame | main]: 'org.gnome.Nautilus' application
STDOUT:

STDERR:

 (Dogtail::Failure)
./features/support/helpers/dogtail.rb:93:in `run'
./features/support/helpers/dogtail.rb:70:in `initialize'
./features/step_definitions/browser.rb:6:in `new'
./features/step_definitions/browser.rb:6:in `desktop_portal_save_as'
./features/step_definitions/browser.rb:310:in `/^I download some file in the Tor Browser to the (.*) directory$/'
features/torified_browsing.feature:113:in `And I download some file in the Tor Browser to the Persistent directory'
Then the file is saved to the Persistent directory 0.000
When I open the address "https://tails.net/about" in the Tor Browser 0.000
Then "Tails - How Tails works" has loaded in the Tor Browser 0.000
And I can print the current page as "output.pdf" to the Persistent directory 0.000
After features/support/hooks.rb:330 12.958

SCENARIO FAILED: 'The Persistent directory is usable in Tor Browser' (at time 07:27:11)

Screenshot: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/07:27:11_The_Persistent_directory_is_usable_in_Tor_Browser.png

Video: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/07:27:11_The_Persistent_directory_is_usable_in_Tor_Browser.mkv

Systemd journal: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/07:27:11_The_Persistent_directory_is_usable_in_Tor_Browser.journal

After features/support/hooks.rb:100 0.000
Tags: @product
1:23.327
Scenario Outline The default XDG directories are usable in Tor Browser
Before features/support/hooks.rb:266 0.000
Before features/support/hooks.rb:273 0.009
Steps
Given I have started Tails from DVD without network and logged in 10.253
Then the live user's Documents directory exists 0.093
And there is a GNOME bookmark for the Documents directory 21.007
Then I start the Tor Browser in offline mode 30.424
And I can save the current page as "index.html" to the Documents GNOME bookmark 21.548
The
The Dogtail init script raised: SearchError: child of [desktop frame | main]: 'org.gnome.Nautilus' application
STDOUT:

STDERR:

 (Dogtail::Failure)
./features/support/helpers/dogtail.rb:93:in `run'
./features/support/helpers/dogtail.rb:70:in `initialize'
./features/step_definitions/browser.rb:6:in `new'
./features/step_definitions/browser.rb:6:in `desktop_portal_save_as'
./features/step_definitions/browser.rb:38:in `browser_save_page_as'
./features/step_definitions/browser.rb:543:in `/^I (can|cannot) save the current page as "([^"]+[.]html)" to the (.*) (directory|GNOME bookmark)$/'
features/torified_browsing.feature:127:in `And I can save the current page as "index.html" to the Documents GNOME bookmark'
features/torified_browsing.feature:124:in `And I can save the current page as "index.html" to the <dir> GNOME bookmark'
After features/support/hooks.rb:330 9.722

SCENARIO FAILED: 'The default XDG directories are usable in Tor Browser, Examples (#1)' (at time 07:28:47)

Screenshot: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/07:28:47_The_default_XDG_directories_are_usable_in_Tor_Browser,_Examples___1_.png

Video: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/07:28:47_The_default_XDG_directories_are_usable_in_Tor_Browser,_Examples___1_.mkv

Systemd journal: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/07:28:47_The_default_XDG_directories_are_usable_in_Tor_Browser,_Examples___1_.journal

After features/support/hooks.rb:100 0.000
Tags: @product
1:19.976
Scenario Outline The default XDG directories are usable in Tor Browser
Before features/support/hooks.rb:266 0.000
Before features/support/hooks.rb:273 0.009
Steps
Given I have started Tails from DVD without network and logged in 9.998
Then the live user's Downloads directory exists 0.066
And there is a GNOME bookmark for the Downloads directory 20.389
Then I start the Tor Browser in offline mode 30.323
And I can save the current page as "index.html" to the Downloads GNOME bookmark 19.197
The
The Dogtail init script raised: SearchError: child of [desktop frame | main]: 'org.gnome.Nautilus' application
STDOUT:

STDERR:

 (Dogtail::Failure)
./features/support/helpers/dogtail.rb:93:in `run'
./features/support/helpers/dogtail.rb:70:in `initialize'
./features/step_definitions/browser.rb:6:in `new'
./features/step_definitions/browser.rb:6:in `desktop_portal_save_as'
./features/step_definitions/browser.rb:38:in `browser_save_page_as'
./features/step_definitions/browser.rb:543:in `/^I (can|cannot) save the current page as "([^"]+[.]html)" to the (.*) (directory|GNOME bookmark)$/'
features/torified_browsing.feature:128:in `And I can save the current page as "index.html" to the Downloads GNOME bookmark'
features/torified_browsing.feature:124:in `And I can save the current page as "index.html" to the <dir> GNOME bookmark'
After features/support/hooks.rb:330 9.293

SCENARIO FAILED: 'The default XDG directories are usable in Tor Browser, Examples (#2)' (at time 07:30:17)

Screenshot: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/07:30:17_The_default_XDG_directories_are_usable_in_Tor_Browser,_Examples___2_.png

Video: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/07:30:17_The_default_XDG_directories_are_usable_in_Tor_Browser,_Examples___2_.mkv

Systemd journal: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/07:30:17_The_default_XDG_directories_are_usable_in_Tor_Browser,_Examples___2_.journal

After features/support/hooks.rb:100 0.000
Tags: @product
1:18.757
Scenario Outline The default XDG directories are usable in Tor Browser
Before features/support/hooks.rb:266 0.000
Before features/support/hooks.rb:273 0.007
Steps
Given I have started Tails from DVD without network and logged in 9.765
Then the live user's Music directory exists 0.081
And there is a GNOME bookmark for the Music directory 20.337
Then I start the Tor Browser in offline mode 24.271
And I can save the current page as "index.html" to the Music GNOME bookmark 24.302
The
The Dogtail init script raised: SearchError: child of [desktop frame | main]: 'org.gnome.Nautilus' application
STDOUT:

STDERR:

 (Dogtail::Failure)
./features/support/helpers/dogtail.rb:93:in `run'
./features/support/helpers/dogtail.rb:70:in `initialize'
./features/step_definitions/browser.rb:6:in `new'
./features/step_definitions/browser.rb:6:in `desktop_portal_save_as'
./features/step_definitions/browser.rb:38:in `browser_save_page_as'
./features/step_definitions/browser.rb:543:in `/^I (can|cannot) save the current page as "([^"]+[.]html)" to the (.*) (directory|GNOME bookmark)$/'
features/torified_browsing.feature:129:in `And I can save the current page as "index.html" to the Music GNOME bookmark'
features/torified_browsing.feature:124:in `And I can save the current page as "index.html" to the <dir> GNOME bookmark'
After features/support/hooks.rb:330 9.835

SCENARIO FAILED: 'The default XDG directories are usable in Tor Browser, Examples (#3)' (at time 07:31:45)

Screenshot: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/07:31:45_The_default_XDG_directories_are_usable_in_Tor_Browser,_Examples___3_.png

Video: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/07:31:45_The_default_XDG_directories_are_usable_in_Tor_Browser,_Examples___3_.mkv

Systemd journal: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/07:31:45_The_default_XDG_directories_are_usable_in_Tor_Browser,_Examples___3_.journal

After features/support/hooks.rb:100 0.000
Tags: @product
1:3.020
Scenario Outline The default XDG directories are usable in Tor Browser
Before features/support/hooks.rb:266 0.000
Before features/support/hooks.rb:273 0.006
Steps
Given I have started Tails from DVD without network and logged in 8.864
Then the live user's Pictures directory exists 0.129
And there is a GNOME bookmark for the Pictures directory 20.316
Then I start the Tor Browser in offline mode 33.710
cannot find TorBrowserNewTabButton.png on the screen (FindFailed)
./features/support/helpers/screen.rb:181:in `rescue in wait'
./features/support/helpers/screen.rb:165:in `wait'
./features/step_definitions/browser.rb:479:in `/^the Tor Browser (?:has started|starts)$/'
./features/step_definitions/browser.rb:466:in `/^I start the Tor Browser( in offline mode)?$/'
features/torified_browsing.feature:130:in `Then I start the Tor Browser in offline mode'
features/torified_browsing.feature:123:in `Then I start the Tor Browser in offline mode'
And I can save the current page as "index.html" to the Pictures GNOME bookmark 0.000
After features/support/hooks.rb:330 8.448

SCENARIO FAILED: 'The default XDG directories are usable in Tor Browser, Examples (#4)' (at time 07:32:58)

Screenshot: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/07:32:58_The_default_XDG_directories_are_usable_in_Tor_Browser,_Examples___4_.png

Video: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/07:32:58_The_default_XDG_directories_are_usable_in_Tor_Browser,_Examples___4_.mkv

Systemd journal: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/07:32:58_The_default_XDG_directories_are_usable_in_Tor_Browser,_Examples___4_.journal

After features/support/hooks.rb:100 0.000
57.354
Scenario Starting a second instance of the Unsafe Browser results in an error message being shown
Before features/support/hooks.rb:266 0.000
Before features/support/hooks.rb:273 0.007
Steps
Given I have started Tails from DVD and logged in and the network is connected 12.503
When I successfully start the Unsafe Browser 23.988
# Wait for whatever facility the GNOME Activities Overview uses to
# learn about which applications are running to "settle". Without
# this sleep, it is confused and it's impossible to start a new
# instance (it will just switch to the one we already started).
And I wait 10 seconds 10.010

Slept for 10 seconds

And I try to start the Unsafe Browser 0.192
Then I see a warning about another instance already running 10.660
The
The Dogtail init script raised: SearchError: child of [desktop frame | main]: 'zenity' application
STDOUT:

STDERR:

 (Dogtail::Failure)
./features/support/helpers/dogtail.rb:93:in `run'
./features/support/helpers/dogtail.rb:70:in `initialize'
./features/step_definitions/unsafe_browser.rb:76:in `new'
./features/step_definitions/unsafe_browser.rb:76:in `/^I see a warning about another instance already running$/'
features/unsafe_browser.feature:59:in `Then I see a warning about another instance already running'
After features/support/hooks.rb:330 11.289

SCENARIO FAILED: 'Starting a second instance of the Unsafe Browser results in an error message being shown' (at time 07:49:35)

Screenshot: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/07:49:35_Starting_a_second_instance_of_the_Unsafe_Browser_results_in_an_error_message_being_shown.png

Video: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/07:49:35_Starting_a_second_instance_of_the_Unsafe_Browser_results_in_an_error_message_being_shown.mkv

Systemd journal: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/07:49:35_Starting_a_second_instance_of_the_Unsafe_Browser_results_in_an_error_message_being_shown.journal

After features/support/hooks.rb:100 0.000
24.433
Scenario Use Unlock VeraCrypt Volumes to unlock a USB drive that has a basic VeraCrypt volume
Steps
When I plug a USB drive containing a basic VeraCrypt volume 8.615
And I cancel the GNOME authentication prompt 2.306
And I unlock and mount this VeraCrypt volume with Unlock VeraCrypt Volumes 6.854
And I open this VeraCrypt volume in GNOME Files 6.054
Then I see the expected contents in this VeraCrypt volume 0.498
When I lock the currently opened VeraCrypt volume 0.103
Command failed: udisksctl unmount --block-device /dev/mapper/tcrypt-*
error code: 1
stdout: 
stderr: Error unmounting /dev/dm-0: GDBus.Error:org.freedesktop.UDisks2.Error.DeviceBusy: Error unmounting /dev/dm-0: target is busy
.
<false> is not true. (ExecutionFailedInVM)
./features/support/helpers/vm_helper.rb:496:in `rescue in execute_successfully'
./features/support/helpers/vm_helper.rb:493:in `execute_successfully'
./features/step_definitions/veracrypt.rb:330:in `/^I lock the currently opened VeraCrypt (volume|file container)$/'
features/veracrypt.feature:18:in `When I lock the currently opened VeraCrypt volume'
Then the VeraCrypt volume has been unmounted and locked 0.000
After features/support/hooks.rb:330 6.220

SCENARIO FAILED: 'Use Unlock VeraCrypt Volumes to unlock a USB drive that has a basic VeraCrypt volume' (at time 07:54:26)

Boot log: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/07:54:26_Use_Unlock_VeraCrypt_Volumes_to_unlock_a_USB_drive_that_has_a_basic_VeraCrypt_volume.file_content_var_log_boot.log

Screenshot: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/07:54:26_Use_Unlock_VeraCrypt_Volumes_to_unlock_a_USB_drive_that_has_a_basic_VeraCrypt_volume.png

Video: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/07:54:26_Use_Unlock_VeraCrypt_Volumes_to_unlock_a_USB_drive_that_has_a_basic_VeraCrypt_volume.mkv

Systemd journal: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/07:54:26_Use_Unlock_VeraCrypt_Volumes_to_unlock_a_USB_drive_that_has_a_basic_VeraCrypt_volume.journal

After features/support/hooks.rb:100 0.000
27.824
Scenario Use Unlock VeraCrypt Volumes to unlock a USB drive that has a hidden VeraCrypt volume
Steps
When I plug a USB drive containing a hidden VeraCrypt volume 11.476
And I cancel the GNOME authentication prompt 2.504
And I unlock and mount this VeraCrypt volume with Unlock VeraCrypt Volumes 8.568
And I open this VeraCrypt volume in GNOME Files 4.675
Then I see the expected contents in this VeraCrypt volume 0.479
When I lock the currently opened VeraCrypt volume 0.120
Command failed: udisksctl unmount --block-device /dev/mapper/tcrypt-*
error code: 1
stdout: 
stderr: Error unmounting /dev/dm-0: GDBus.Error:org.freedesktop.UDisks2.Error.DeviceBusy: Error unmounting /dev/dm-0: target is busy
.
<false> is not true. (ExecutionFailedInVM)
./features/support/helpers/vm_helper.rb:496:in `rescue in execute_successfully'
./features/support/helpers/vm_helper.rb:493:in `execute_successfully'
./features/step_definitions/veracrypt.rb:330:in `/^I lock the currently opened VeraCrypt (volume|file container)$/'
features/veracrypt.feature:27:in `When I lock the currently opened VeraCrypt volume'
Then the VeraCrypt volume has been unmounted and locked 0.000
After features/support/hooks.rb:330 5.968

SCENARIO FAILED: 'Use Unlock VeraCrypt Volumes to unlock a USB drive that has a hidden VeraCrypt volume' (at time 07:55:13)

Boot log: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/07:55:13_Use_Unlock_VeraCrypt_Volumes_to_unlock_a_USB_drive_that_has_a_hidden_VeraCrypt_volume.file_content_var_log_boot.log

Screenshot: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/07:55:13_Use_Unlock_VeraCrypt_Volumes_to_unlock_a_USB_drive_that_has_a_hidden_VeraCrypt_volume.png

Video: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/07:55:13_Use_Unlock_VeraCrypt_Volumes_to_unlock_a_USB_drive_that_has_a_hidden_VeraCrypt_volume.mkv

Systemd journal: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/07:55:13_Use_Unlock_VeraCrypt_Volumes_to_unlock_a_USB_drive_that_has_a_hidden_VeraCrypt_volume.journal

After features/support/hooks.rb:100 0.000
32.154
Scenario Use Unlock VeraCrypt Volumes to unlock a basic VeraCrypt file container with a PIM
Steps
When I plug and mount a USB drive containing a basic VeraCrypt file container with a PIM 16.796
And I unlock and mount this VeraCrypt file container with Unlock VeraCrypt Volumes 9.732
And I open this VeraCrypt volume in GNOME Files 5.051
Then I see the expected contents in this VeraCrypt volume 0.473
When I lock the currently opened VeraCrypt file container 0.100
Command failed: udisksctl unmount --block-device /dev/mapper/tcrypt-*
error code: 1
stdout: 
stderr: Error unmounting /dev/dm-0: GDBus.Error:org.freedesktop.UDisks2.Error.DeviceBusy: Error unmounting /dev/dm-0: target is busy
.
<false> is not true. (ExecutionFailedInVM)
./features/support/helpers/vm_helper.rb:496:in `rescue in execute_successfully'
./features/support/helpers/vm_helper.rb:493:in `execute_successfully'
./features/step_definitions/veracrypt.rb:330:in `/^I lock the currently opened VeraCrypt (volume|file container)$/'
features/veracrypt.feature:35:in `When I lock the currently opened VeraCrypt file container'
Then the VeraCrypt file container has been unmounted and locked 0.000
After features/support/hooks.rb:330 6.059

SCENARIO FAILED: 'Use Unlock VeraCrypt Volumes to unlock a basic VeraCrypt file container with a PIM' (at time 07:56:00)

Boot log: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/07:56:00_Use_Unlock_VeraCrypt_Volumes_to_unlock_a_basic_VeraCrypt_file_container_with_a_PIM.file_content_var_log_boot.log

Screenshot: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/07:56:00_Use_Unlock_VeraCrypt_Volumes_to_unlock_a_basic_VeraCrypt_file_container_with_a_PIM.png

Video: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/07:56:00_Use_Unlock_VeraCrypt_Volumes_to_unlock_a_basic_VeraCrypt_file_container_with_a_PIM.mkv

Systemd journal: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/07:56:00_Use_Unlock_VeraCrypt_Volumes_to_unlock_a_basic_VeraCrypt_file_container_with_a_PIM.journal

After features/support/hooks.rb:100 0.000
42.121
Scenario Use Unlock VeraCrypt Volumes to unlock a hidden VeraCrypt file container
Steps
When I plug and mount a USB drive containing a hidden VeraCrypt file container 27.210
And I unlock and mount this VeraCrypt file container with Unlock VeraCrypt Volumes 9.913
And I open this VeraCrypt volume in GNOME Files 4.378
Then I see the expected contents in this VeraCrypt volume 0.503
When I lock the currently opened VeraCrypt file container 0.115
Command failed: udisksctl unmount --block-device /dev/mapper/tcrypt-*
error code: 1
stdout: 
stderr: Error unmounting /dev/dm-0: GDBus.Error:org.freedesktop.UDisks2.Error.DeviceBusy: Error unmounting /dev/dm-0: target is busy
.
<false> is not true. (ExecutionFailedInVM)
./features/support/helpers/vm_helper.rb:496:in `rescue in execute_successfully'
./features/support/helpers/vm_helper.rb:493:in `execute_successfully'
./features/step_definitions/veracrypt.rb:330:in `/^I lock the currently opened VeraCrypt (volume|file container)$/'
features/veracrypt.feature:43:in `When I lock the currently opened VeraCrypt file container'
Then the VeraCrypt file container has been unmounted and locked 0.000
After features/support/hooks.rb:330 6.169

SCENARIO FAILED: 'Use Unlock VeraCrypt Volumes to unlock a hidden VeraCrypt file container' (at time 07:56:59)

Boot log: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/07:56:59_Use_Unlock_VeraCrypt_Volumes_to_unlock_a_hidden_VeraCrypt_file_container.file_content_var_log_boot.log

Screenshot: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/07:56:59_Use_Unlock_VeraCrypt_Volumes_to_unlock_a_hidden_VeraCrypt_file_container.png

Video: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/07:56:59_Use_Unlock_VeraCrypt_Volumes_to_unlock_a_hidden_VeraCrypt_file_container.mkv

Systemd journal: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/07:56:59_Use_Unlock_VeraCrypt_Volumes_to_unlock_a_hidden_VeraCrypt_file_container.journal

After features/support/hooks.rb:100 0.016
28.721
Scenario Use GNOME Disks to unlock a USB drive that has a basic VeraCrypt volume with a keyfile
Steps
When I plug a USB drive containing a basic VeraCrypt volume with a keyfile 9.439
And I cancel the GNOME authentication prompt 2.519
And I unlock and mount this VeraCrypt volume with GNOME Disks 11.037
And I open this VeraCrypt volume in GNOME Files 5.007
Then I see the expected contents in this VeraCrypt volume 0.568
When I lock the currently opened VeraCrypt volume 0.149
Command failed: udisksctl unmount --block-device /dev/mapper/tcrypt-*
error code: 1
stdout: 
stderr: Error unmounting /dev/dm-0: GDBus.Error:org.freedesktop.UDisks2.Error.DeviceBusy: Error unmounting /dev/dm-0: target is busy
.
<false> is not true. (ExecutionFailedInVM)
./features/support/helpers/vm_helper.rb:496:in `rescue in execute_successfully'
./features/support/helpers/vm_helper.rb:493:in `execute_successfully'
./features/step_definitions/veracrypt.rb:330:in `/^I lock the currently opened VeraCrypt (volume|file container)$/'
features/veracrypt.feature:54:in `When I lock the currently opened VeraCrypt volume'
Then the VeraCrypt volume has been unmounted and locked 0.000
After features/support/hooks.rb:330 6.288

SCENARIO FAILED: 'Use GNOME Disks to unlock a USB drive that has a basic VeraCrypt volume with a keyfile' (at time 07:57:43)

Boot log: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/07:57:43_Use_GNOME_Disks_to_unlock_a_USB_drive_that_has_a_basic_VeraCrypt_volume_with_a_keyfile.file_content_var_log_boot.log

Screenshot: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/07:57:43_Use_GNOME_Disks_to_unlock_a_USB_drive_that_has_a_basic_VeraCrypt_volume_with_a_keyfile.png

Video: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/07:57:43_Use_GNOME_Disks_to_unlock_a_USB_drive_that_has_a_basic_VeraCrypt_volume_with_a_keyfile.mkv

Systemd journal: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/07:57:43_Use_GNOME_Disks_to_unlock_a_USB_drive_that_has_a_basic_VeraCrypt_volume_with_a_keyfile.journal

After features/support/hooks.rb:100 0.001
30.282
Scenario Use GNOME Disks to unlock a USB drive that has a hidden VeraCrypt volume
Steps
When I plug a USB drive containing a hidden VeraCrypt volume 11.565
And I cancel the GNOME authentication prompt 2.280
And I unlock and mount this VeraCrypt volume with GNOME Disks 11.014
And I open this VeraCrypt volume in GNOME Files 4.768
Then I see the expected contents in this VeraCrypt volume 0.525
When I lock the currently opened VeraCrypt volume 0.127
Command failed: udisksctl unmount --block-device /dev/mapper/tcrypt-*
error code: 1
stdout: 
stderr: Error unmounting /dev/dm-0: GDBus.Error:org.freedesktop.UDisks2.Error.DeviceBusy: Error unmounting /dev/dm-0: target is busy
.
<false> is not true. (ExecutionFailedInVM)
./features/support/helpers/vm_helper.rb:496:in `rescue in execute_successfully'
./features/support/helpers/vm_helper.rb:493:in `execute_successfully'
./features/step_definitions/veracrypt.rb:330:in `/^I lock the currently opened VeraCrypt (volume|file container)$/'
features/veracrypt.feature:63:in `When I lock the currently opened VeraCrypt volume'
Then the VeraCrypt volume has been unmounted and locked 0.000
After features/support/hooks.rb:330 6.249

SCENARIO FAILED: 'Use GNOME Disks to unlock a USB drive that has a hidden VeraCrypt volume' (at time 07:58:29)

Boot log: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/07:58:29_Use_GNOME_Disks_to_unlock_a_USB_drive_that_has_a_hidden_VeraCrypt_volume.file_content_var_log_boot.log

Screenshot: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/07:58:29_Use_GNOME_Disks_to_unlock_a_USB_drive_that_has_a_hidden_VeraCrypt_volume.png

Video: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/07:58:29_Use_GNOME_Disks_to_unlock_a_USB_drive_that_has_a_hidden_VeraCrypt_volume.mkv

Systemd journal: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/07:58:29_Use_GNOME_Disks_to_unlock_a_USB_drive_that_has_a_hidden_VeraCrypt_volume.journal

After features/support/hooks.rb:100 0.001
52.880
Scenario Use GNOME Disks to unlock a basic VeraCrypt file container with a keyfile
Steps
When I plug and mount a USB drive containing a basic VeraCrypt file container with a keyfile 23.347
And I unlock and mount this VeraCrypt file container with GNOME Disks 23.638
And I open this VeraCrypt volume in GNOME Files 5.235
Then I see the expected contents in this VeraCrypt volume 0.547
When I lock the currently opened VeraCrypt file container 0.111
Command failed: udisksctl unmount --block-device /dev/mapper/tcrypt-*
error code: 1
stdout: 
stderr: Error unmounting /dev/dm-0: GDBus.Error:org.freedesktop.UDisks2.Error.DeviceBusy: Error unmounting /dev/dm-0: target is busy
.
<false> is not true. (ExecutionFailedInVM)
./features/support/helpers/vm_helper.rb:496:in `rescue in execute_successfully'
./features/support/helpers/vm_helper.rb:493:in `execute_successfully'
./features/step_definitions/veracrypt.rb:330:in `/^I lock the currently opened VeraCrypt (volume|file container)$/'
features/veracrypt.feature:71:in `When I lock the currently opened VeraCrypt file container'
Then the VeraCrypt file container has been unmounted and locked 0.000
After features/support/hooks.rb:330 6.360

SCENARIO FAILED: 'Use GNOME Disks to unlock a basic VeraCrypt file container with a keyfile' (at time 07:59:38)

Boot log: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/07:59:38_Use_GNOME_Disks_to_unlock_a_basic_VeraCrypt_file_container_with_a_keyfile.file_content_var_log_boot.log

Screenshot: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/07:59:38_Use_GNOME_Disks_to_unlock_a_basic_VeraCrypt_file_container_with_a_keyfile.png

Video: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/07:59:38_Use_GNOME_Disks_to_unlock_a_basic_VeraCrypt_file_container_with_a_keyfile.mkv

Systemd journal: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/07:59:38_Use_GNOME_Disks_to_unlock_a_basic_VeraCrypt_file_container_with_a_keyfile.journal

After features/support/hooks.rb:100 0.016
54.531
Scenario Use GNOME Disks to unlock a hidden VeraCrypt file container
Steps
When I plug and mount a USB drive containing a hidden VeraCrypt file container 29.021
And I unlock and mount this VeraCrypt file container with GNOME Disks 20.322
And I open this VeraCrypt volume in GNOME Files 4.720
Then I see the expected contents in this VeraCrypt volume 0.302
When I lock the currently opened VeraCrypt file container 0.164
Command failed: udisksctl unmount --block-device /dev/mapper/tcrypt-*
error code: 1
stdout: 
stderr: Error unmounting /dev/dm-0: GDBus.Error:org.freedesktop.UDisks2.Error.DeviceBusy: Error unmounting /dev/dm-0: target is busy
.
<false> is not true. (ExecutionFailedInVM)
./features/support/helpers/vm_helper.rb:496:in `rescue in execute_successfully'
./features/support/helpers/vm_helper.rb:493:in `execute_successfully'
./features/step_definitions/veracrypt.rb:330:in `/^I lock the currently opened VeraCrypt (volume|file container)$/'
features/veracrypt.feature:79:in `When I lock the currently opened VeraCrypt file container'
Then the VeraCrypt file container has been unmounted and locked 0.000
After features/support/hooks.rb:330 6.141

SCENARIO FAILED: 'Use GNOME Disks to unlock a hidden VeraCrypt file container' (at time 08:00:48)

Boot log: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/08:00:48_Use_GNOME_Disks_to_unlock_a_hidden_VeraCrypt_file_container.file_content_var_log_boot.log

Screenshot: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/08:00:48_Use_GNOME_Disks_to_unlock_a_hidden_VeraCrypt_file_container.png

Video: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/08:00:48_Use_GNOME_Disks_to_unlock_a_hidden_VeraCrypt_file_container.mkv

Systemd journal: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/08:00:48_Use_GNOME_Disks_to_unlock_a_hidden_VeraCrypt_file_container.journal

After features/support/hooks.rb:100 0.019
45.259
Scenario Use GNOME Disks to unlock a basic VeraCrypt file container with a PIM
Steps
When I plug and mount a USB drive containing a basic VeraCrypt file container with a PIM 17.271
And I unlock and mount this VeraCrypt file container with GNOME Disks 21.294
And I open this VeraCrypt volume in GNOME Files 5.915
Then I see the expected contents in this VeraCrypt volume 0.652
When I lock the currently opened VeraCrypt file container 0.126
Command failed: udisksctl unmount --block-device /dev/mapper/tcrypt-*
error code: 1
stdout: 
stderr: Error unmounting /dev/dm-0: GDBus.Error:org.freedesktop.UDisks2.Error.DeviceBusy: Error unmounting /dev/dm-0: target is busy
.
<false> is not true. (ExecutionFailedInVM)
./features/support/helpers/vm_helper.rb:496:in `rescue in execute_successfully'
./features/support/helpers/vm_helper.rb:493:in `execute_successfully'
./features/step_definitions/veracrypt.rb:330:in `/^I lock the currently opened VeraCrypt (volume|file container)$/'
features/veracrypt.feature:87:in `When I lock the currently opened VeraCrypt file container'
Then the VeraCrypt file container has been unmounted and locked 0.000
After features/support/hooks.rb:330 6.306

SCENARIO FAILED: 'Use GNOME Disks to unlock a basic VeraCrypt file container with a PIM' (at time 08:01:50)

Boot log: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/08:01:50_Use_GNOME_Disks_to_unlock_a_basic_VeraCrypt_file_container_with_a_PIM.file_content_var_log_boot.log

Screenshot: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/08:01:50_Use_GNOME_Disks_to_unlock_a_basic_VeraCrypt_file_container_with_a_PIM.png

Video: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/08:01:50_Use_GNOME_Disks_to_unlock_a_basic_VeraCrypt_file_container_with_a_PIM.mkv

Systemd journal: https://jenkins.tails.boum.org/job/test_Tails_ISO_feature-trixie/19/artifact/build-artifacts/08:01:50_Use_GNOME_Disks_to_unlock_a_basic_VeraCrypt_file_container_with_a_PIM.journal

After features/support/hooks.rb:100 0.000