All Projects β†’ Subito-it β†’ Mendoza

Subito-it / Mendoza

Licence: other
Dispatch Apple's UI Tests over multiple physical machines

Programming Languages

swift
15916 projects

🍷 Mendoza

Mendoza allows to parallelize Apple's UI Tests over multiple physical machines. While Xcode recently introduced parallelization it is limited to a single local machine requiring top-notch hardware to run tests on several instances at once. Both macOS and iOS projects are supported.

Mendoza is designed to parallelize tests on an unlimited number of different machines offering a scalable solution to reduce UI Testing's execution times. There are no particular contraints on the hardware that can be used, as an example we use rather oldish 2013 i5 MacBooks.

The tool is flexible thanks to plugins (that you can write in Swift 😎) allowing to heavily customize several steps in the dispatching pipeline.

The outcome of a test session will be a set of log files (.json, .html) and a single .xcresult bunble that will contain all results as if all tests were run on a single machine.

A snapshot of a session running on 8 concurrent nodes (each running 2 simulators at once) can be seen below.

Features
πŸƒβ€β™€οΈ Makes UI Test execution super fast!
πŸ‘¨πŸ»β€πŸ’» Written in Swift
πŸ”Œ Supports plugins (written in Swift!)
πŸ” Wide set of result formats
πŸ€– Supports both iOS and macOS projects

How does it work

The basic idea is simple: you compile a project on one machine, distribute the compiled package (test bundle) to a number of specified remote nodes, execute a subset of tests on each node and collect the results back together as if they were run on a single machine. On iOS projects, depending on the node hardware configuration, you can also run multiple simulators at once.

Mendoza hides all the complexity behind a single test command by leveraging built in command line tools to perform each of the aforementioned steps. To get an idea of what’s under the hood take a look here.

Installation

You'll need to install Mendoza on all nodes you'll use to distribute tests.

brew install Subito-it/made/mendoza

NOTE

You'll get a warning that sshpass was removed for security reasons. You can still install a copy by running brew install hudochenkov/sshpass/sshpass see repo. Please consider that sshpass is used only if you choose to connect to remote nodes via username/password authentication which is not the recommended way. When possible you should use ssh key based authentication.

Or you can build manually from sources

Building from sources

To build Mendoza make sure to install libssh2:

brew install libssh2
git clone https://github.com/Subito-it/Mendoza.git
cd Mendoza
swift package update
swift package generate-xcodeproj
xed .

From the target selection select the Mendoza project and add /usr/local/include to 'Header Search Paths' and /usr/local/lib to 'Library Search Paths'.

Quick start

Inside your project folder run

mendoza configuration init

this will prompt you with a series of (fairly self-explanatory) questions and produce a configuration file that you will feed to the test command as follows:

iOS project

mendoza test config.json --device_name="iPhone 8" --device_runtime="12.1"

macOS project

mendoza test configuration.json

This will compile your project, distribute the test bundles, execute the tests, collect the results together on the destination node that was specified during setup and generate a set of output files.

Commands

configuration init

Generates a new configuration file required to execute tests. you will be prompted with a series of (fairly self-explanatory) questions which will produce a json configuration file as an output.

Concepts

AppleID (iOS projects only)

During initialization you will be prompted for an Apple ID account that will be used to install new runtimes on testing nodes. Any low priviledge Apple ID account will work.

It is not mandatory to provide one, but test dispatching will fail if the requested simulator runtime isn't properly installed.

Nodes configuration

When setting up nodes you'll be asked:

  • label that identifies the node
  • address
  • authentication method
  • ram disk: you can optionally specify if the node should use a ram disk. This has a significant benefit in performances on older machines that have no SSD disk.

iOS projects only

  • concurrent simulators: manually enter the number of concurrent simulators to use at once. The rule of thumb is that you can run 1 simulator per physical CPU core. Specifying more that one simulator per core will work but this will result in slower total execution time because the node will be over-utilized.

Destination node

The destination node is the node that will be responsible to collect all the result logs.

configuration authentication

The credentials and passwords you will be asked during initialization are stored locally in your Keychain (see configuration file and security paragraph). This means that if the configuration was generated on a different machine you may be missing those credentials/passwords. The configuration authentication command will prompt and store missing credentials/password in your local Keychain.

plugin init

This command allows to create a plugin template script that will be used during the execution of the tests. Refer to the plugins paragraph.

test

Will launch tests as specified in the configuration files.

Required parameters

  • path to the configuration file

iOS projects only

  • --device_name=name: device name to use to run tests. e.g. 'iPhone 8'
  • --device_runtime=version: device runtime to use to run tests. e.g. '13.1'

Optional parameters

  • --timeout=[minutes]: maximum allowed idle time (in seconds) in test standard output before dispatch process is automatically terminated. (default: 60s)
  • --failure_retry=[count]: the number of times a failing tests should be repeated
  • --device_language=[language]: the device's language. e.g. 'en-EN'
  • --device_locale=[locale]: the device's locale. e.g. 'en_US'
  • --include_files=[files]: specify from which files UI tests should be extracted. Accepts wildcards and comma separated. e.g SBTA*.swift,SBTF*.swift. (default: '*.swift')
  • --exclude_files=[files]: specify which files should be skipped when extracting UI tests. Accepts wildcards and comma separated. e.g SBTA*.swift,SBTF*.swift. (default: '')
  • --plugin_data=[data]: a custom string that can be used to inject data to plugins
  • --plugin_debug: write log files for plugin development. Refer to the plugins paragraph
  • --nonHeadlessSimulatorsFlag: run simulators in non headless mode. This will make testing slightly slower because devices need to be booted an arranged properly
  • --use_localhost: πŸ”₯ when passing these flag tests will be dispatched on the localhost as well even if it is not specified in the configuration file. This is useful when launching tests locally leveraging additional simulators of your own development machine
  • --xcode_buildnumber=[number]: when multiple xcode versions are installed on the dispatching nodes automatically switch to the specified Xcode build number version (e.g. 12E507 as shown in the About window)
  • --xcresult_blob_threshold_kb=[size]: to reduce the size of the final xcresult it is possible to specify the maximum blob size of the xcresult Data objects. Big blobs are usually trace files that are collected during test execution which might not be needed in all cases. A reasonable value is 1024
  • --llvm_cov_equivalence_path=[path]: the path equivalence that will be passed to 'llvm-cov show' when generating code coverage (,)
  • --clear_derived_data_on_failure: on compilation failure derived data will be cleared and compilation will be retried once
  • --skip_result_merge: skip xcresult merge (keep one xcresult per test in the result folder)

Test output

Mendoza will write a set of log files containing information about the test session:

  • test_details.json: provides a detailed insight of the test session
  • test_result.json: the list of tests that passed/failed in json format
  • test_result.html: the list of tests that passed/failed in html format
  • repeated_test_result.json: the list of tests that had to be repeated in json format
  • repeated_test_result.html: the list of tests that had to be repeated in html format
  • merged.xcresult: the result bundle containing all test data. Can be opened with Xcode
  • coverage.json: the coverage summary file generated by running `xcrun llvm-cov export --summary-only -instr-profile [path] [executable_path]
  • coverage.html: the coverage html file generated by running `xcrun llvm-cov show --format=html -instr-profile [path] [executable_path]

If you're interested in seeing the specific actions that made a test fail can manually inspect the merged.xcresult using Xcode. Alternatively you may also consider using Cachi which is able to parse and present results in a web browser.

Plugins

Plugins allow to customize various steps of Mendoza's pipeline. This opens up to several optimizations that are stricly related to your own specific workflows.

A plugin is initialized with the plugin init command. The plugin file that is generated contains a struct with a single handle() method with a signature that depends on the plugin. The types in the signature are showed as comments above the struct definition.

The following plugins are available:

  • extract: allows to specify the test methods that should be performed in every test file
  • sorting: plugin to add estimated execution time to test cases
  • event: plugin to perform actions (e.g. notifications) based on dispatching events
  • precompilation: plugin to perform actions before compilation starts
  • postcompilation: plugin to perform actions after compilation completes
  • teardown: plugin to perform actions at the end of the dispatch process

You should definitely consider using swift sh if your plugins requires additional dependencies.

extract

By default test methods will be extracted from all files in the UI testing target. This should work most of the times however in some advanced cases this could not be the desired behaviour, for example if there is some custom tagging to run tests on specific devices (e.g. only iPhone/iPad). This plugin allows to override the default behaviour and put in place a custom implementation

See an example TestExtractionPlugin.swift.

sorting

By default test cases will be executed randomly because Mendoza has no information about the execution time of test cases. Mendoza can significantly improve total execution time of test if you provide an estimate of the execution time of tests.

Using tools like Cachi you can automatically store and retrieve tests statistics.

See an example TestSortingPlugin.swift.

event

This plugin will be invoked during the different steps of Mendoza's pipeline. You'll be notified when compilation starts/ends, when tests bundles start/end being distributed and so on. Based on these event you could for example send notifications.

precompilation

Your project might be so heavily customized that you might need to perform some changes to the project before the compilation of the UI testing target begins.

postcompilation

If you're using a precompilation plugin you might also need a post compilation plugin to restore any change previously made.

teardown

This plugin allows to perform custom actions once the test session ends. You'll get some result information as input in order to perform action accoring to the test session outcome

Debugging plugins

The files used by Mendoza internally to execute plugins won't be deleted when you run tests passing the --plugin_debug flag. After a test session, in the same folder of your plugins, you'll find 2 additional files: one with the same filename of the plugin but prefixed with an underscore and a file with a .debug extension. The .debug file will invoke the plugin file and pass arguments to it.

An easy way to debug a plugins is to create a new macOS command line tool project in Xcode, copy paste the content of the _Pluginfile.swft and add the arguments to the scheme settings. From there you can use Xcode to run an debug the plugin.

Configuration file and security

The configure init command will generate a configuration file containing all the information needed to compile, execute and distribute tests to a set of specified remote nodes. You can create different configuration files to test different test targets or use different remote nodes.

All the access credentials and passwords that are requested during initialization are stored locally in the current user’s Keychain. This means that you may be asked to update them by running configure authentication if something doesn’t match with what is specified in the configuration file (e.g. access credentials to a remote node).

Node configuration

It is suggested that nodes are configured as follows:

SSH MaxSessions

Set MaxSessions to 100 in /etc/ssh/sshd_config

Increase maxfiles

launchctl limit maxfiles 64000 524288

Increase max processes

/usr/sbin/sysctl -w kern.maxprocperuid=4096 kern.maxproc=2500

Increase pseudo terminals

/usr/sbin/sysctl -w kern.tty.ptmx_max=999

Contributions

Contributions are welcome! If you have a bug to report, feel free to help out by opening a new issue or sending a pull request.

Authors

Tomas Camin (@tomascamin)

License

Mendoza is available under the Apache License, Version 2.0. See the LICENSE file for more info.

Note that the project description data, including the texts, logos, images, and/or trademarks, for each open source project belongs to its rightful owner. If you wish to add or remove any projects, please contact us at [email protected].