Skip to content

Comparing AVA, Jest, Mocha, and mocha-parallel-tests testing frameworks

License

Notifications You must be signed in to change notification settings

scraggo/comparing-javascript-test-runners

Repository files navigation

Comparing JavaScript Test Runners

comparing-javascript-test-runners.png

https://github.com/scraggo/comparing-javascript-test-runners

This article is a comparison of the AVA, Jest, and Mocha JavaScript test runners. Permalink

To generate the speed metrics in the article, I created a node application (part of this repo) that runs tests in all the frameworks listed above. See the documentation here.

Overview

The Problem

Technology is always changing. As a result, the popular frameworks of today may become unpopular tomorrow. Despite this, we can assume that code should be tested and there are general principles we can adhere to.

In recent years, JavaScript has become a more robust language thanks to the steady cadence of enhancements starting with ES6/2015. As a result, many front-end frameworks can accomplish the task of creating serious web applications. In general, these frameworks are much easier to test than frameworks of years past. Three test runners have risen far above the rest as the most popular choices:

Goals

The immediate practical goal of this article is to help you choose a JavaScript testing framework. (Note: I'll be using "test runner" and "testing framework" interchangeably.) My aim is to address the following questions:

How does one choose the right testing framework for their use case? What criteria should one base their decision on?

In order to do this, we'll explore some general principles regarding both testing frameworks and testing in general. Then, after outlining the criteria for evaluating the frameworks, we can explore them in detail.

Testing in general

There are differing opinions on what the "best practices" are for testing code. I have a pragmatic opinion - we should test what is necessary to gain confidence that our application is running correctly and we should do it in a way that leverages existing tools that have been refined and are trusted by the developer community.

The general principles I think are important are:

Documentation: writing tests that outline the functionality of the application

  • A test description should be simple to understand. Written in this way, the tests are another form of documentation for the source code. A product owner should be able to read through the tests and understand their relation to the application as a whole.
  • In the case of testing abstractions, we should take time to be even clearer about the utility of such a test.

Philosophy: "What" should we test? What level of "granularity" are we aiming for?

  • A unit test has the smallest scope of the test types. What may be under test is a function or a class.
  • An integration test has a wider scope than a unit test. This sort of test combines individual units and tests them as a group with the aim of assuring that they run correctly when interacting with each other. (We want to avoid is the scenario where units work perfectly in isolation, but not together.)
  • An acceptance test (a.k.a. end to end test) has the widest scope. It's goal is to test as an end user would use the application without directly calling source code.
  • When in doubt:
    • It should be clear where unit, integration, and acceptance tests are.
    • Think of the unit under test as a "black box" whenever possible. The goal should be to test the behavior of the unit, not how it's implemented.
    • Don't test external library code. We should be using high-quality libraries that have their own test suites.

State: the pros and cons of sharing state between tests

  • Ideally, one doesn't share state between tests, period. Test suites (a file that contains test blocks) should definitely aim to steer clear of shared state. Individual test blocks should have a clean version of the unit under test.
  • In a few cases, shared setup between test blocks may be necessary. An example of this is when we're making assertions during an end-to-end test that has many steps.

Coverage: the extent to which one should measure test coverage

  • Using test coverage tools gives us a metric for how many lines, statements, or blocks of your code are tested. Your team may decide that 100% test coverage is necessary. My feeling is that you should cover all essential functions, the goal being that after changing code, your tests will fail if your changes affected anything that it could have affected.
  • Another "metric" that I like to use while reviewing code is analyzing the changes for the big picture: Do the tests cover what was in the "acceptance criteria" of the task?

Tips

  • Learn more about testing by reading the test specs of a well-written library.
  • Is an external library under-documented? Read through the test-suites to get a quick view into how it works. (Reading the open and closed issues may be helpful too.)

Requirements

Now that we've outlined some testing concepts, we can dive into what we'll require from a test framework.

Popularity and Community

A testing framework should have community support. An unpopular framework may be out of date or may not have the kinks ironed out. It might have incomplete documentation (including stack overflow questions.) It might not have enough developers working on it to fix its issues.

Speed

A testing framework should not be slow. We can only define this relatively - for one person's slow may be another person's acceptable. We will be benchmarking speed for all the frameworks.

Ease of Use

A testing framework should be un-difficult to use. The setup, configuration, command-line options, and test writing itself should be relatively straightforward. If it's too confusing to a developer to use, it will be less likely that tests will get written.

Failure Reporting and Debugging

A testing framework should give ample information when a particular test fails. We should know exactly which test failed and the stack trace of the unit being called. We should also be able to put breakpoints in with our favorite debugging tools.

Works with your framework and environment of choice (React, Redux, Electron, etc)

A testing framework must be compatible with what you're trying to test. It should be flexible enough to be able to adapt to the changing needs of those frameworks.

  • Ability to run in multiple environments: node, in-browser, CI/CD
  • Versatility to run unit, integration, and end-to-end tests.

Nice to Have

Depending on what you need to test, a framework should support:

  • Organization strategies: describe blocks, it blocks
  • "watch" and "inspect" modes
  • A variety of assertion capabilities
  • Ability to add tools like coverage (nyc), snapshot testing, etc.
  • Mocking / Injecting modules (intercepting require statements)
  • Webpack compilation (injecting of webpack-defined global variables)
    • simulating a CICD build step
  • Babel transpilation of ESNext code
    • Use of alias module import statements (removing the need for ../../)

Comparing the Test Runners

Now, onto comparing the test runners themselves. Here's an overview of each one:

AVA

Ava Logo

https://github.com/avajs/ava

The magic of AVA is in its simplicity. It's minimal, fast, concurrent, and has a simple syntax that entirely removes the use of globals (like describe, it, etc.) It supports asynchronous behavior out of the box. AVA has a small team of three developers, one being open-source heavyweight @sindresorhus. Some other selling points directly from their readme:

AVA is a test runner for Node.js with a concise API, detailed error output, embrace of new language features and process isolation that let you write tests more effectively.

AVA adds code excerpts and clean diffs for actual and expected values. If values in the assertion are objects or arrays, only a diff is displayed, to remove the noise and focus on the problem.

AVA automatically removes unrelated lines in stack traces, allowing you to find the source of an error much faster, as seen above.

AVA automatically detects whether your CI environment supports parallel builds. Each build will run a subset of all test files, while still making sure all tests get executed. See the ci-parallel-vars package for a list of supported CI environments.

Jest

Jest Logo

https://jestjs.io/

Jest is feature-packed, aiming to solve everything in one package, with a focus on making the experience delightful for the test author. It's written and maintained by Facebook and is extremely popular and community supported due to the ubiquity of React and create-react-app. The CLI output is colorful and interactive with detailed exception reporting and diffing. Snapshot testing, mocking, and coverage reporting are all built-in. Also included are globals like it and describe as well as a custom assertion library (similar to chai). It also touts:

zero config - Jest aims to work out of the box, config free, on most JavaScript projects.

isolated - Tests are parallelized by running them in their own processes to maximize performance.

great api - From it to expect - Jest has the entire toolkit in one place. Well documented, well maintained, well good.

Mocha

Mocha Logo

https://mochajs.org/

Being the most established of the testing frameworks, Mocha enjoys a solid place in the JavaScript community. It's been around since 2011 and is maintained by the OpenJS Foundation (growth stage) and contributors. Mocha supports numerous command-line options and configurations. It's generally used in tandem with external libraries - assert or chai could take care of your assertion needs and sinon could take care of your mocking needs. The it and describe blocks mentioned by Jest were pioneered by Mocha (along with the beforeEach, afterEach, and other pre/post hooks). In addition to being able to run in node, you can also run tests in the browser giving you full access to the DOM. There's also a dizzying array of test reporting styles (one being Nyan cat.) In its own words:

Mocha is a feature-rich JavaScript test framework running on Node.js and in the browser, making asynchronous testing simple and fun. Mocha tests run serially, allowing for flexible and accurate reporting, while mapping uncaught exceptions to the correct test cases.

Mocha is the most-depended-upon module on npm (source: libraries.io)

The SuperAgent request library test documentation was generated with Mocha's "doc" reporter

Update 8.0.0 / 2020-06-10: Mocha 8 has built-in support for running tests in parallel!

mocha-parallel-tests

https://github.com/mocha-parallel/mocha-parallel-tests

⚠️ Important note: As of Mocha 8x, there's built-in support for running tests in parallel. In case you can't currently upgrade to mocha 8x, mocha-parallel-tests is a viable choice!

mocha-parallel-tests is not a testing framework. It's a wrapper over Mocha designed to significantly speed it up. It's new in 2019 and has a small team. I'll go into detail on why I'm including it here (and what "parallel" means) in the "speed" portion of this article. From the readme:

mocha-parallel-tests is a test runner for tests written with mocha testing framework which allows you to run them in parallel.

mocha-parallel-tests executes each of your test files in a separate process while maintaining the output structure of mocha.

Compared to the other tools which try to parallelize mocha tests execution, mocha-parallel-tests doesn't require you to write the code in a different way or use some specific APIs - just run your tests with mocha-parallel-tests instead of mocha and you will see the difference. Or if you prefer to use mocha programmatic API replace it with mocha-parallel-tests default export and you're done!

If you're using Node.JS >= 12 your tests execution will be even faster because mocha-parallel-tests supports running tests with Node.JS worker threads API.

Popularity and Community Comparison

Now that we know a bit about each framework, lets look at some of their popularity, publish frequency, and other community metrics.

chart-popularity-slide chart-downloads-slide

Charts made with https://npm-stat.com/charts.html?package=ava&package=jest&package=mocha&from=2015-01-01&to=2020-05-27

Overall, we can see that all the frameworks are rising in popularity. To me, this indicates that more people are writing JavaScript applications and testing them - which is quite exciting. The fact that none of them are on a downward trend makes all of them viable in this category.

Weekly Downloads * Last Publish Publishes in 1 Year Contributors
Jest 7.2 million 2020-05-05 27 1083
Mocha 4.3 million 2020-04-24 11 439
AVA 227,179 2020-05-08 20 243
mocha-parallel-tests 18,097 2020-02-08 4 14

* Weekly Downloads as of May 15, 2020

🥇Jest is clearly the most popular framework with 7.2 million weekly downloads. It was published most recently and is updated very frequently. Its popularity can be partially attributed to the popularity of the React library. Jest is shipped with create-react-app and is recommended for use in React's documentation.

🥈Mocha comes in second place with 4.3 million weekly downloads. It was the de facto standard long before Jest hit the scene and is the test runner of many, many applications. It isn't published as frequently as the other two which I believe is a testament to it being tried, true, and more stable.

🥉AVA has 227,179 weekly downloads, an order of magnitude fewer than the most popular frameworks. It is published frequently, which positively signals a focus on improvement and iteration. This may be due to its (arguably niche) focus on minimalism or it having a small team that doesn't have the resources to promote the library.

mocha-parallel-tests has 18,097 weekly downloads and doesn't enjoy as frequent updates as the major three. It's extremely new and not a framework.

In general, more popularity brings more community involvement. The number of open and closed issues tends to increase as a result. To create a loose maintenance ratio *metric , we divide the open issues by the total number of issues (open + closed issues):

Open Issues Closed Issues Total Ratio
Mocha 254 2225 2479 10.2%
AVA 154 1169 1323 11.6%
Jest 844 4343 5187 16.2%
mocha-parallel-tests 37 111 148 25.0%

🥇Mocha has the lowest ratio of open to closed issues, making it the most successfully maintained library. It's stability surely correlates with its longevity (and vice versa.)

🥈AVA is 2nd place. This is quite impressive given its small team.

🥉Jest is 3rd place. This comes as no surprise given that it has the most issues to deal with.

mocha-parallel-tests has the fewest number of total issues by far but the highest ratio. It doesn't have a significant financial backing like the other frameworks do.

* Caveat: I'm assuming the open issues in these libraries aren't crippling to the core functionality of the library.

Speed Comparison

Before we get to the comparison, I'd like to discuss a few concepts. All of the frameworks run the tests in "parallel", with the exception of Mocha, which runs its tests in "serial."

What do "serial" and "parallel" mean?

"Serial" - one at a time, ie - the first must complete before the second, the second must complete before the third, etc. Code may run asynchronously but doesn't start a separate process. This type of processing is also known as sequential processing.

"Parallel" - happening simultaneously, ie - the first, second, third, etc can happen at the same time. Multiple tasks are completed at a time by different processes (which may be different threads or literally different processors).

For all of the frameworks with parallel capabilities, only separate test files are run in parallel. describe and it blocks in a given file/suite are run serially. Given this, writing more test files and putting slow tests can be put into their own files may increase the speed of running the complete test suite.

Benchmarks

To generate the speed metrics in the article, I created a node application that runs tests in all the frameworks listed above. The documentation for it explains how I wrote and ran the tests. The aim was to simulate a "true" test run in a significantly sized enterprise codebase. Here are the results (node version 12):

Speed Type
mocha 8.x --parallel 5.3s parallel
mocha-parallel-tests 7.3s parallel
AVA 9.6s parallel
Jest 12.5s parallel
Mocha 16.2s serial

A caveat with all benchmarking tests: the hardware environment (the make, model, RAM, processes running, etc) will affect measured results. For this reason, we'll only be considering the speeds relative to each other.

🥇Mocha 8.x with --parallel enabled is the fastest of the frameworks in this run (and most runs.) (Mocha versions prior to 8.x can't enable this option.)

🥈mocha-parallel-tests and AVA are close behind (AVA actually ran faster than mocha-parallel-tests in a few of the runs.)

🥉Jest is also fast, but seems to have a bit more overhead than the other two.

Mocha in serial mode lags far behind the parallel runners - which is to be expected because it tests must take "real" time to execute, one after the other. If speed is your most important criteria (and its drawbacks are not an issue), you'll see a 200-1000% increase in test speed using mocha-parallel-tests instead (depending on your machine, node version, and the tests themselves).

Ease of Use Comparison

I'll split "ease of use" into a few categories:

  • Amount of necessary configuration/dependencies
  • Writing the tests
  • Running the tests

Amount of necessary configuration/dependencies

Configuration Dependencies
Jest close-to-zero-config: lots of defaults All dependencies included: snapshot testing, mocking, coverage reporting, assertions
AVA Sensible defaults some externals necessary. Included: snapshot testing, assertions
Mocha & mocha-parallel-tests Many, many options most externals necessary (all if in-browser)

🥇Jest takes the cake in this department. Using its defaults wherever possible, you could have close to zero configuration.

Jest's configuration can be defined in the package.json file of your project, or through a jest.config.js file or through the --config <path/to/file.js|cjs|mjs|json> option. If you'd like to use your package.json to store Jest's config, the "jest" key should be used on the top level so Jest will know how to find your settings

  • Sensible defaults for finding tests (default: [ "**/__tests__/**/*.[jt]s?(x)", "**/?(*.)+(spec|test).[jt]s?(x)" ])
  • A huge number of options
  • Built-in snapshot tests, coverage reporting, mocking modules and libraries
  • Excellent documentation, lots of tutorials and examples

🥈AVA comes in 2nd place.

  • Sensible defaults for finding tests (see Jest's defaults)
  • Configure in package.json, an ava.config.* file, or another override file in the directory root
  • Many CLI options
  • Built-in snapshot tests
  • @ava/babel for Babel compilation
  • @ava/typescript for TypeScript support
  • Good documentation, few tutorials and examples
  • Coverage reporting, mocking modules and libraries must be imported from elsewhere

🥉Mocha comes in 3rd place.

By default, mocha looks for the glob "./test/*.js", so you may want to put your tests in test/ folder. If you want to include subdirectories, pass the --recursive option.

  • One default for finding tests (above)
  • Configure in package.json or an override file
  • Many, many CLI options
  • Good documentation (slightly opaque and a lot to read through), lots of tutorials and examples (in and out of Mocha's docs)
  • Assertions*, coverage reporting, snapshot tests, mocking modules and libraries (everything) must be imported from elsewhere

* node's built-in assert is commonly used with Mocha for assertions. While it's not built into Mocha, it can be easily imported: const assert = require('assert'). If testing in-browser, you wouldn't have access to assert and would have to use a library like chai.

For mocha-parallel-tests, run tests as you would with Mocha. There is a caveat:

Most of mocha CLI options are supported. If you're missing some of the options support you're welcome to submit a PR: all options are applied in a same simple way.

Writing the tests

Summary
Mocha & mocha-parallel-tests describe and it blocks
Jest like Mocha, but everything built in
AVA import test context, customizable

Mocha's influence on test-writing is undeniable. From Mocha's getting started section, we can see how tests are organized in nested describe blocks that can contain any number of it blocks which make test assertions.

const assert = require('assert'); // only works in node
describe('Array', function() {
  describe('#indexOf()', function() {
    it('should return -1 when the value is not present', function() {
      assert.equal([1, 2, 3].indexOf(4), -1);
    });
  });
});

Chai's expect is commonly used instead of assert:

const { expect } = require('chai'); // works in both node and browser

it('should return -1 when the value is not present', function() {
  expect([1, 2, 3].indexOf(4)).to.equal(-1);
});

More Mocha test examples can be found here: https://mochajs.org/#examples

Jest follows Mocha's example, but everything is included (no need to use external assertion, etc libraries). A lot of the syntax is compatible with Mocha, for example - it can be used instead of test:

describe('Array', function() {
  describe('#indexOf()', function() {
    test('should return -1 when the value is not present', function() {
      expect([1, 2, 3].indexOf(4)).toEqual(-1);
    });
  });
});

(Notice that Jest's toEqual is very much like chai's to.equal. Many of Jest's assertions are camel-cased analogues of chai's assertions.)

AVA takes a different approach to writing tests. It prides itself on not injecting globals into your tests. Everything comes from test and the t "execution context" variable in the callback:

// Array-indexOf.spec.js
const test = require('ava');

test('should return -1 when the value is not present', t => {
  t.is([1, 2, 3].indexOf(4), -1);
});

Built-in assertions are available on the t object: true, false, truthy, falsy, is, etc. You can also create custom assertions.

Like Mocha and Jest, AVA has before, beforeEach, after, and afterEach hooks documented here. These work on a per-file basis.

AVA does not have a way to nest test blocks (an ability describe affords in Mocha and Jest.) This issue goes into detail on why the maintainers haven't adopted this functionality. This article has some alternatives.

An excellent example of a single AVA test suite shows a wide range of AVA's capabilities.

Since the frameworks have drastically different styles and similar capabilities, this was difficult to rank from easiest to most difficult. AVA has a noticeable lack of organizing tests with nested describe blocks, but its contained API is extremely flexible. Mocha forces the user to make decisions on which libraries to use, but this makes it almost limitless in capability. Jest includes everything, but some of the built-in magic makes it difficult or confusing to get certain things done ( see this issue on mocks ).

Running the tests

Summary
Jest interactive CLI or GUI
Mocha & mocha-parallel-tests non-interactive CLI or browser
AVA non-interactive CLI

🥇Jest has an incredible interactive command line interface. (Using Majestic adds a web-based GUI to the experience.) There are numerous options for choosing which tests run and updating snapshots - all keyboard-driven. It watches for test file changes in watch mode and only runs the tests that have been updated. There isn't as much of a need to use .only because filtering terms is a breeze.

Jest CLI1 Jest CLI2 Jest CLI3

It's text output is not only a pleasure to look at, it also includes all the information you could need:

...
PASS  test/file.spec.js
...

Test Suites: 50 passed, 50 total
Tests:       1250 passed, 1250 total
Snapshots:   0 total
Time:        10.753s
Ran all test suites.

Jest Output

Above: Jest output on successful test run

🥈Mocha, being highly configurable, sometimes necessitates long and difficult to read commands to run test suites. Once this command is set, it may need to be altered to filter for files. I wind up making a lot of slightly varied package.json script commands to capture the variations I'm looking for. Once these are set, tests run smoothly and the output and diffs are informative and legible. (Also remember the wide range of output styles available.)

Mocha Output

Mocha Parallel Tests Output

Above: Mocha and mocha-parallel-tests output

🥉AVA is highly configurable, but there are a few things not included by default that I miss. One is seeing the output of all the tests. Enable verbose to do that. Another is seeing the time it takes to execute the complete test suite. Adding time (a bash function) before the command, but it's not as immediately comprehendible - see this issue. AVA's defaults are extremely minimal, so like Mocha, you may need a complex configuration to serve your needs. Filtering for tests is very similar to the other frameworks (there's a --match, -m command that can be repeated.)

AVA Output

Above: AVA output. Refer to "real" for time taken to run all the tests

Failure Reporting and Debugging Comparison

All of the frameworks have pleasant to read output, detailed informative diffs for failures, and stack traces that show exactly where errors occurred. They all also offer the ability to debug with node or debugger of your choice. Mocha has an option to enable full stack traces, which may make it slightly more optimal for test failure troubleshooting.

Mocha

Jest

AVA

Works with your framework and environment of choice (React, Redux, Electron, etc) Comparison

🥇Mocha works for everything. I've had success running React applications and end-to-end tests with Spectron (for Electron applications), amongst other things. It can also run in browser, which is extremely helpful for testing applications that use libraries which are tied directly to browser functionality.

🥈Jest is an all-purpose test-runner and is recommended for testing React applications. It doesn't have support for browser testing (as far as I can tell.) It supports Puppeteer for acceptance testing (and may support others).

🥉AVA is an all-purpose test-runner, though, it has yet to support browser testing. It supports Puppeteer for acceptance testing (and may support others).

Full Comparison (with "Nice to Haves")

Let's recap our findings and fill in some gaps with our "nice to haves." (M8 = Mocha 8.x with --parallel enabled. MPT = mocha-parallel-tests)

Feature Notes
Popularity Jest - 1st place. Mocha - 2nd place. AVA - 3rd place. MPT - 4th place.
Maintenance Mocha - 1st place. AVA - 2nd place. Jest - 3rd place. MPT - 4th place.
Speed M8/MPT - 1st place. AVA - 2nd place. Jest - 3rd place. Mocha - 4th place.
Amount of necessary configuration/dependencies Jest - 1st place. AVA - 2nd place. Mocha - 3rd place. MPT - 4th place.
Organization strategies: describe and it blocks Mocha and Jest use these, AVA does not (omits all test globals)
Running the tests Jest - 1st place. Mocha - 2nd place. MPT - 3rd place. AVA - 4th place.
Failure Reporting and Debugging Mocha slightly above others
Works with your framework and environment of choice Mocha wins for supporting browser testing
"watch" and "inspect" modes Included in all the frameworks.
A variety of assertion capabilities All the frameworks have these as well as the ability to add custom matchers
Ability to add tools like coverage (nyc), snapshot testing, etc Mocha and AVA allow you to add what you need, Jest has these built-in
Mocking / Injecting modules (intercepting require statements) Jest has its own module/library mocking system. Mocha and AVA require external libraries (examples: sinon, proxyquire, inject-loader)
Webpack compilation (injecting of webpack-defined global variables) For Mocha, mochapack is a library that allows this. Jest's docs on webpack. There's no AVA documentation for webpack specifically, see below on Babel.
Babel transpilation of ESNext code and use of alias module import statements (removing the need for ../../) All the frameworks allow for babel configuration

Recommendations

As you can see, all the frameworks are incredibly robust for most testing needs. However, if you picked one at random, it might not work for a specific use case. It's not an easy choice, but here's how I'd break it down:

  • 🏅Mocha is recommended if you want your tests to run in any environment. It's incredibly community-supported and is extend-able with your favorite 3rd-party packages. Using mocha-parallel-tests (or 8.x with --parallel enabled) would give you a speed advantage.
  • 🏅Jest is recommended if you want to get tests up and running quickly. It has everything built in and requires very little configuration. The command line and GUI experience is unmatched. Finally, it's the most popular and makes an excellent pair with React.
  • 🏅AVA is recommended if you want a minimalist framework with no globals. AVA is fast, easy to configure, and you get ES-Next transpilation out of the box. You don't want hierarchical describe blocks and you want to support a smaller project.

Conclusion

I hope this article was helpful - whether it be in furthering your own investigation, giving you a place to start with learning, or helping your team choose the right test runner for your applications. Don't hesitate to contact me if you have questions or feedback.

Want to contribute?

Found a typo? Want to add details or make a correction? This repo is open-source and your contributions are 100% welcome 💥.

External Resources

Articles:

More on Mocha:

More on Jest:

More on AVA:

Packages:

Package URL Category Works With Notes
nyc - istanbul coverage AVA, Mocha
sinon mocking AVA, Mocha
proxyquire mock-require AVA, Mocha
inject-loader mock-require AVA, Mocha
chai assertions Mocha see chai plugins including chai-immutable, chai-as-promised, chai-sinon
mocha-snapshots snapshot Mocha
mochapack webpack Mocha
Majestic GUI Jest

About

Comparing AVA, Jest, Mocha, and mocha-parallel-tests testing frameworks

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Contributors 4

  •  
  •  
  •  
  •