Contact Free trial Login

Create a Monitor

You can create a monitor by filling in fields if you do not know how to write tests in the Behavior Driven Development (BDD) test-writing language or how to use the BAT CLI. This method of creating a monitor creates a test that tests one or more endpoints, and schedules the test suite to run at intervals that you specify.

After every interval, you can view the results of the test and statistics about API performance.

Before You Begin

  • Ensure that your user ID has the Anypoint Monitoring User permission and is assigned the Exchange Administrators role.

    An administrator for your organization can grant you this permission in Anypoint Access Management.

  • Determine which endpoints you want to test and the REST methods that you want to test them with.

  • Decide upon the return codes that you expect the methods to return.

  • Figure out the frequency at which you would like the monitor to test the endpoints.

    The shortest interval is every fifteen minutes. You can also choose to run the monitor hourly, daily, weekly, or monthly. The schedule begins immediately after you create the monitor. You can schedule another monitor later for the same test suite that is more detailed, if you prefer. You can also modify the schedule after you create the monitor, if you want to specify any of these options: Set an hourly schedule to start at a certain number of minutes after the hour begins. Set a daily, weekly, or monthly schedule to start at a certain time of day.

  • Determine which execution location you would like to run the monitor in.

  • Decide whether you would like the request messages to include additional header information.

  • Obtain one of the following pieces of information, if you would like to send reports to one of these reporting tools:

    Reporting Tool Required Information


    License Key








    Email address


To create a monitor:

  1. In the Functional Monitoring section of Anypoint Monitoring, click Create.

  2. Specify a name and a short description for the monitor.

  3. Fill in the fields with the information that you have gathered.

  4. Click Create.


The dashboard opens. The monitor that you created is listed on the left.

Beneath the name of the monitor is an indicator that explains when the current run of the monitor began. A line at the bottom of the list item shows how much time remains before the next run of the monitor. Also, the Monitored Endpoints, Tests, and Schedules sections will all be empty.

After the monitor has run, the list item for the test suite will indicate whether the monitored endpoints passed or failed.

Tests silently fail when they take more than 120 seconds to run. As a workaround, you can split a long-running test into multiple tests. Download the monitor, split the test across multiple monitors in the BAT CLI, run and time each test to be sure that it is under 120 seconds, and then upload the monitors.

If you do not want to split a long-running test, download the monitor and run it from the BAT CLI.

What to do next

Click the Scheduling tab to see the schedule for the monitor. You can add additional schedules and delete schedules that you no longer want to run. You can also disable schedules, if you want to pause them, and then enable them again.

After the monitor runs, click the Monitored Endpoints tab to see:

  • Whether all endpoints returned the expected codes, as indicated by PASSED in the first column of the table

  • Statistics about the execution of the test of the endpoint

As your monitor continues to run at the scheduled intervals, the statistics in this table are updated. They show the minimum, average, and maximum runtime for up to five executions of the test.

Click Tests to see statistics and other information about the monitor. Because you have only one test in the monitor that you just created, the statistics and information in this section apply to the test, too.

If you want to delete the monitor, delete all versions of it in the Tests section.

Was this article helpful?

💙 Thanks for your feedback!

Edit on GitHub