Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
The CSV provider collects metrics from CSV files and makes them available to Akamas. It offers a very versatile way to integrate custom data sources.
This section provides the minimum requirements that you should match before using the CSV File telemetry provider.
The following requirements should be met to enable the provider to gather CSV files from remote hosts:
Port 22 (or a custom one) should be open from Akamas installation to the host where the files reside.
The host where the files reside should support SCP or SFTP protocols.
Read access to the CSV files target of the integration
Versions < 2.0.0 are compatibile with Akamas until version 1.8.0
Versions >= 2.0.0 are compatible with Akamas from version 1.9.0
The CSV File provider is generic and allows integration with any data source, therefore it does not come with support for a specific component type.
To operate properly, the CSV file provider expects the presence of four fields in each processed CSV file:
A timestamp field used to identify the point in time a certain sample refers to.
A component field used to identify the Akamas entity.
A metric field used to identify the name of the metric.
A value field used to store the actual value of the metric.
These fields can have custom names in the CSV file, you can specify them in the provider configuration.
The Install CSV provider page describes how to get this Telemetry Provider installed. Once installed, this provider is shared with all users of your Akamas installation and can be used to monitor many different systems, by configuring appropriate telemetry provider instances as described in the Create a CSV provider instance page.
The installed provider is shared with all users of your Akamas installation and can monitor many different systems, by configuring appropriate telemetry provider instances.
To create an instance of the Dynatrace provider, build a YAML file (instance.yml
in this example) with the definition of the instance:
Then you can create the instance for the system
using the Akamas CLI:
When you create an instance of the Dynatrace provider, you should specify some configuration information to allow the provider to correctly extract and process metrics from Dynatrace.
You can specify configuration information within the config
part of the YAML of the instance definition.
url
- URL of the Dynatrace installation API (see https://www.dynatrace.com/support/help/extend-dynatrace/dynatrace-api/ to retrieve the URL of your installation)
token
- A Dynatrace API Token with the proper permissions
You can collect additional metrics with the Dynatrace provider by using the metrics
field:
In the case in which Akamas cannot reach directly your Dynatrace installation, you can configure an HTTP proxy by using the proxy
field:
This section reports the complete reference for the definition of a telemetry instance.
This table shows the reference for the config
section within the definition of the Dynatrace provider instance:
This table reports the reference for the config
→ proxy
section within the definition of the Dynatrace provider instance:
This table reports the reference for the metrics
section within the definition of the Dynatrace provider instance. The section contains a collection of objects with the following properties:
This section reports common use cases addressed by this provider.
Check the Linux optimization pack for a list of all the system metrics available in Akamas.
As a first step to start extracting metrics from Dyntrace, generate your API token and make sure it has the right permissions.
As a second step, choose a strategy to map your Linux component (MyLinuxComponent) with the corresponding Dyntrace entity.
Let’s assume you want to map by id your Dynatrace entity, you can find the id in the URL bar of a Dyntrace dashboard of the entity:
Grab the id and add it to the Linux component definition:
You can leverage the name of the entity as well:
As a third and final step, once the component is all set, you can create an instance of the Dynatrace provider and then build your first studies:
Field | Type | Value restrictions | Required | Default Value | Description |
---|---|---|---|---|---|
Field | Type | Value restrictions | Required | Default value | Description |
---|---|---|---|---|---|
Field | Type | Value Restrictions | Required | Default value | Description |
---|---|---|---|---|---|
url
String
It should be a valid URL
Yes
The URL of the Dynatrace installation API (see the official reference)
token
String
Yes
The Dynatrace API Token the provider should use to interact with Dynatrace. The token should have sufficient permissions.
proxy
Object
See Proxy options reference
No
The specification of the HTTP proxy to use to communicate with Dynatrace.
pushEvents
String
true, false
No
true
If set to true the provider will inform dynatrace of the configuration change event which will be visible in the Dynatrace UI.
tags
Object
No
A set of global tags to match Dynatrace entities. The provider uses these tags to apply a default filtering of Dynatrace entities for every component.
address
String
It should be a valid URL
Yes
The URL of the HTTP proxy to use to communicate with the Dynatrace installation API
port
Number (integer)
1 <port
<65535
Yes
The port at which the HTTP proxy listens for connections
username
String
No
The username to use when authenticating against the HTTP proxy, if necessary
password
String
No
The username to use when authenticating against the HTTP proxy, if necessary
metric
String
It must be an Akamas metric
Yes
The name of an Akamas metric that should map to the new metric you want to gather
datasourceMetric
String
A valid Dynatrace metric
Yes
The Dynatrace query to use to extract metric
labels
Array of strings
-
No
The list of Dynatrace labels that should be retained when gathering the metric
staticLabels
Key-Value
-
No
Static labels that will be attached to metric samples
aggregation
String
No
avg
The aggregation to perform if the mergeEntities property under the extras section is set to true
extras
Object
Only the parameter mergeEntities
can be defined to either true
or false
No
Section for additional properties
To install the CSV File provider, create a YAML file (called provider.yml
in this example) with the specification of the provider:
Then, you can then install the provider with the Akamas CLI:
This page describes how to set up an OracleDB exporter in order to gather metrics regarding an Oracle Database instance through the Prometheus provider.
The OracleDB exporter repository is available on the official project page. The suggested deploy mode is through a Docker image, since the Prometheus instance can easily access the running container through the Akamas network.
Use the following command line to run the container, where cust-metrics.toml
is your configuration file defining the queries for additional custom metrics (see paragraph below) and DATA_SOURCE_NAME
an environment variable containing the Oracle EasyConnect string:
You can refer to the official guide for more details or alternative deployment modes.
It is possible to define additional queries to expose custom metrics using any data in the database instance that is readable by the monitoring user (see the guide for more details about the syntax).
The following is an example of exporting system metrics from the Dynamic Performance (V$) Views used by the Prometheus provider default queries for the Oracle Database optimization pack:
The Spark History Server provider collects metrics from a Spark History Server instance and makes them available to Akamas.
Prerequisites
This section provides the minimum requirements that you should match before using the Spark History Server telemetry provider.
Apache Spark 2.3
Spark History Server API must be reachable at the provided address and port (the default port is 18080
).
spark-application
You can check Spark History Server provider metrics mapping to see how component-types metrics are extracted by this provider.
Versions < 2.0.0 are compatible with Akamas until version 1.8.0
Versions >= 2.0.0 are compatible with Akamas from version 1.9.0
This section lists the workflow operators this provider depends on:
Akamas uses components to identify specific elements of the system to be monitored and optimized. Your system might contain multiple components to model, for example, a Spark application and each host of the cluster. To point Akamas to the right component when extracting metrics you need to add a property called sparkApplication
to your Spark Application component. The provider will only extract metrics for components for which this property has been specified.
To install the Spark History Server provider, create a YAML file (called provider.yml
in this example) with the definition of the provider:
Then you can install the provider using the Akamas CLI:
The installed provider is shared with all users of your Akamas installation and can monitor many different systems, by configuring appropriate telemetry provider instances.
The Dynatrace provider collects metrics from Dynatrace and makes them available to Akamas.
This provider includes support for several technologies. In any case, custom queries can be defined to gather the desired metrics.
Dynatrace SaaS/Managed version 1.187 or later
Kubernetes and Docker
Web Application
Ubuntu-16.04, Rhel-7.6
java-openjdk-8, java-openjdk-11, java-openjdk-17
java-ibm-j9vm-6, java-ibm-j9vm-8, java-eclipse-openj9-11
Refer to Dynatrace provider metrics mapping to see how component-types metrics are extracted by this provider.
This section provides the minimum requirements that you should match before using the Prometheus provider.
Dynatrace SaaS/Managed version 1.187 or later
A valid Dynatrace license
Dynatrace OneAgent installed on the servers where the Dynatrace entities to be monitored are running
Connectivity between Akamas and the Dynatrace server on port 443
A Dynatrace API token with the privileges described here.
The Dynatrace provider needs a Dynatrace API token with the following privileges:
metrics.read (Read metrics)
entities.read (Read entities and tags)
DataExport (Access problem and event feed, metrics, and topology)
ReadSyntheticData (Read synthetic monitors, locations, and nodes)
DataImport (Data ingest, e.g.: metrics and events). This permission is used to inform Dynatrace about configuration changes.
To generate an API Token for your Dynatrace installation you can follow these steps.
To instruct Akamas from which Dynatrace entities (e.g. Workloads, Services, Process Groups) metrics should be collected you can some specific properties on components.
Different strategies can be used to map Dynatrace entities to Akamas components:
By id
By name
By tags
By Kubernetes properties
You can map a component to a Dynatrace entity by leveraging the unique id of the entity, which you should put under the id
property in the component. This strategy is best used for long-lived instances whose ID does not change during the optimization such as Hosts, Process Groups, or Services.
Here is an example of how to setup host monitoring via id:
You can find the id of a Dynatrace entity by looking at the URL of a Dynatrace dashboard relative to the entity. Watch out that the "host" key is valid only for Linux components, other components (e.g. the JVM) must drill down into the host entities to get the PROCESS_GROUP_INSTANCE or PROCESS_GROUP id.
You can map a component to a Dynatrace entity by leveraging the entity’s display name. This strategy is similar to the map by id but provides a more friendly way to identify the mapped entity. Beware that id multiple entities in your Dynatrace installation share the same name they will all be mapped to the same component. The Dynatrace display name should be put under the name
property in the component definition:
You can map a component to a Dynatrace entity by leveraging Dynatrace tags that match the entity, tags which you should put under the tags
property in the component definition.
If multiple tags are specified, instances matching any of the specified tags will be selected.
This sample configuration maps to the component all Dynatrace entities with tag environment: test
or [AWS]dynatrace-monitored: true
Dynatrace supports both key-value and key-only tags. Key-only tags can be specified as Key-value tags with an empty value as in the following example
By leveraging dedicated properties, you can map a component to a Dynatrace entity referring to a Kubernetes cluster (e.g., a Pod or a Container).
To properly identify the set of containers to be mapped, you can specify the following properties. Any container matching all the properties will be mapped to the component.
You can retrieve all the information to setup the properties on the top of the Dynatrace container dashboard.
The following example shows how to map a component to a container running in Kubernetes:
To properly identify the set of pods to be mapped, you can specify the following properties. Any pod matching all the properties will be mapped to the component.
If you need to narrow your pod selection further you can also specify a set of tags as described in the by tags. Note that tags for Kubernetes resources are called Labels in the Dynatrace dashboard.
Labels are specified as key-value in the Akamas configuration. In Dynatrace’s dashboard key and value are separated by a column (:
)
The following example shows how to map a component to a pod running in Kubernetes:
Please note, that when you are mapping components to Kubernetes entities the property type
is required to instruct Akamas on which type of entity you want to map.
Dynatrace maps Kubernetes entities to the following types:
You can improve the matching of components with Dynatrace by adding a type
property in the component definition, this property will help the provider match only those Dynatrace entities of the given type.
The type of an entity can be retrieved from the URL of the entity’s dashboard
Available entity types can be retrieved, from your Dynatrace instance, with the following command:
In some circumstances, you might want to map multiple Dyantrace entities (e.g. a set of hosts) to the same Akamas component and import aggregated metrics.
This can be easily done by using tags. If Akamas detects that multiple entities have been mapped to the same component it will try to aggregate metrics; some metrics, however, can not be automatically aggregated.
To force aggregation on all available metrics you can add the mergeable: true
property to the component under the Dynatrace element.
To create an instance of the CSV provider, build a YAML file (instance.yml
in this example) with the definition of the instance:
Then you can create the instance for the system
using the Akamas CLI:
timestampFormat
formatNotice that the week-year format YYYY
is compliant with the ISO-8601 specification, but you should replace it with the year-of-era format yyyy
if you are specifying a timestampFormat
different from the ISO one. For example:
Correct: yyyy-MM-dd HH:mm:ss
Wrong: YYYY-MM-dd HH:mm:ss
You can find detailed information on timestamp patterns in the Patterns for Formatting and Parsing section on the DateTimeFormatter (Java Platform SE 8) page.
When you create an instance of the CSV provider, you should specify some configuration information to allow the provider to correctly extract and process metrics from your CSV files.
You can specify configuration information within the config
part of the YAML of the instance definition.
address
- a URL or IP identifying the address of the host where CSV files reside
username
- the username used when connecting to the host
authType
- the type of authentication to use when connecting to the file host; either password
or key
auth
- the authentication credential; either a password or a key according to authType
. When using keys, the value can either be the value of the key or the path of the file to import from
remoteFilePattern
- a list of remote files to be imported
protocol
- the protocol to use to retrieve files; either scp
or sftp
. Default is scp
fieldSeparator
- the character used as a field separator in the CSV files. Default is ,
componentColumn
- the header of the column containing the name of the component. Default is COMPONENT
timestampColumn
- the header of the column containing the timestamp. Default is TS
timestampFormat
- the format of the timestamp (e.g. yyyy-MM-dd HH:mm:ss zzz
). Default is YYYY-MM-ddTHH:mm:ss
You should also specify the mapping between the metrics available in your CSV files and those provided by Akamas. This can be done in the metrics
section of the telemetry instance configuration. To map a custom metric you should specify at least the following properties:
metric
- the name of a metric in Akamas
datasourceMetric
- the header of a column that contains the metric in the CSV file
The provider ignores any column not present as datasourceMetric
in this section.
The sample configuration reported in this section would import the metric cpu_util
from CSV files formatted as in the example below:
The following represents the complete configuration reference for the telemetry provider instance.
The following table reports the configuration reference for the config
section
The following table reports the configuration reference for the metrics
section
Here you can find common use cases addressed by this provider.
In this use case, you are going to import some metrics coming from SAR, a popular UNIX tool to monitor system resources. SAR can export CSV files in the following format.
Note that the metrics are percentages (between 1 and 100), while Akamas accepts percentages as values between 0 and 1, therefore each metric in this configuration has a scale factor of 0.001.
You can import the two CPU metrics and the memory metric from a SAR log using the following telemetry instance configuration.
Using the configured instance, the CSV File provider will perform the following operations to import the metrics:
Retrieve the file "/csv/sar.csv" from the server "127.0.0.1" using the SCP protocol authenticating with the provided password.
Use the column hostname
to lookup components by name.
Use the column timestamp
to find the timestamps of the samples (that are expected to be in the format specified by timestampFormat
).
Collect the metrics (two with the same name, but different labels, and one with a different name):
cpu_util
: in the CSV file is in the column %user and attach to its samples the label "mode" with value "user".
cpu_util
: in the CSV file is in the column %system and attach to its samples the label "mode" with value "system".
mem_util
: in the CSV file is in the column %memory.
Akamas supports the integration with virtually any telemetry and observability tool.
The following table describes the supported Telemetry Providers, which are created automatically at installation time.
Telemetry Provider | Description |
---|---|
Notice that Telemetry Providers are shared across all the workspaces within the same Akamas installation, and only users with administrative privileges can manage them.
By default, only requests at the service level are imported by the telemetry provider.
To import specific key requests you can follow these steps.
Currently only average response time, throughput, and error rate metrics are available for key requests.
Create a new component of type Web Application for each key request you want to import. This allows tracking response time, throughput, and error rates separately.
You can use the following yaml file as an example and customize it to suit your needs.
In order to instruct Akamas to import a specific key request you just need to change the id
field of the yaml above to the one that matches your key request on Dynatarce.
To obtain that ID open the analysis page for the request as in the example below, take note of the URL of the page, and look for the SERVICE_METHOD
keyword. The id is the one starting with SERVICE_METHOD
and ending before the character %14
Considering the example below the id is SERVICE_METHOD-D4BCC949D5DD656A
Create a telemetry instance for your system using the yaml specified below as an example and modify it to provide your Dynatrace account and credentials. This will instruct Akamas to use key request metrics instead of service metrics.
The NeoLoad Web provider collects metrics from a NeoLoad Web instance and makes them available to Akamas.
This section provides the minimum requirements that you should match before using the NeoLoad Web telemetry provider.
NeoLoad Web SaaS or managed version 7.1 or later.
The NeoLoad Web API must be reachable at a provided address and port (by default ).
NeoLoad Web API access token.
Versions < 2.0.0 are compatibile with Akamas untill version 1.8.0
Versions >= 2.0.0 are compatible with Akamas from version 1.9.0
Web Application
This section lists the workflow operators this provider depends on.
Akamas reasons in terms of a system to be optimized and in terms of parameters and metrics of components of that system. To understand which metrics collected from NeoloadWeb should refer to which component, the NeoloadWeb provider looks up the property neoloadweb
in the components of a system:
The Prometheus provider collects metrics from a Prometheus instance and makes them available to Akamas.
This provider includes support for several technologies (Prometheus exporters). In any case, custom queries can be defined to gather the desired metrics.
This section provides the minimum requirements that you should match before using the Prometheus provider.
Akamas supports Prometheus starting from version2.26.
Using also theprometheus-operator
requires Prometheus 0.47 or greater. This version is bundled with the kube-prometheus-stack
since version 15.
Connectivity between the Akamas server and the Prometheus server is also required. By default, Prometheus is run on port 9090.
(Linux system metrics)
(Java metrics)
(Docker container metrics)
exporter (AWS resources metrics)
(Web application metrics)
The Prometheus provider includes queries for most of the monitoring use cases these exporters cover. If you need to specify custom queries or make use of exporters not currently supported you can specify them as described in creating .
Kubernetes (Pod, Container, Workload, Namespace)
Web Application
Java (java-ibm-j9vm-6, java-ibm-j9vm-8, java-eclipse-openj9-11, java-openjdk-8, java-openjdk-11, java-openjdk-17)
Linux (Ubuntu-16.04, Rhel-7.6)
Akamas reasons in terms of a system to be optimized and in terms of parameters and metrics of components of that system. To understand which metrics collected from Prometheus should be mapped to a component, the Prometheus provider looks up some properties in the components of a system grouped under prometheus
property. These properties depend on the exporter and the component type.
Nested under this property you can also include any additional field your use case may require to filter the imported metrics further. These fields will be appended in queries to the list of label matches in the form field_name=~'field_value'
, and can specify either exact values or patterns.
It is important that you add instance
and, optionally, the job
properties to the components of a system so that the Prometheus provider can gather metrics from them:
The Prometheus provider does not usually require a specific configuration of the Prometheus instance it uses.
When gathering metrics for hosts it's usually convenient to set the value of the instance
label so that it matches the value of the instance
property in a component; in this way, the Prometheus provider knows which system component each data point refers to.
Here’s an example configuration for Prometheus that sets the instance
label:
To create an instance of the Spark History Server provider, build a YAML file (instance.yml
in this example) with the definition of the instance:
Then you can create the instance for the system spark-system
using the Akamas CLI:
When you create an instance of the Spark History Server provider, you should specify some configuration information to allow the provider to correctly extract and process metrics from the Spark History server.
You can specify configuration information within the config
part of the YAML of the instance definition.
address
- hostname of the Spark History Server instance
The following YAML file describes the definition of a telemetry instance.
The following table reports the reference for the config
section within the definition of the Spark History Server provider instance:
This section reports common use cases addressed by this provider.
As a first step, you need to create a YAML file (spark_instance.yml
) containing the configuration the provider needs to connect to the Spark History Server, plus the filter on the desired level of granularity for the imported metrics:
and then create the telemetry instance using the Akamas CLI:
This section reports common best practices you can adopt to ease the use of this telemetry provider.
configure metrics granularity: in order to reduce the collection time, configure the importLevel
to import metrics with a granularity no finer than the study requires.
wait for metrics publication: make sure in the workflow there is a few-minute interval between the end of the Spark application and the execution of the Spark telemetry instance, since the Spark History Server may take some time to complete the publication of the metrics.
The LoadRunner provider collects metrics generated by a LoadRunner instance (converted to their JSON format and placed in CIFS network share) and makes them available to Akamas.
Prerequisites
This section provides the minimum requirements that you should match before using the LoadRunner telemetry provider.
Micro Focus LoadRunner 12.60 or 2020
The network share is reachable at port 445/TCP
The network share is reachable at port 139/UDP
Username, domain (if required), and password of the network share.
Read permission on the network share.
Web Application
You can check to see how component-types metrics are extracted by this provider.
To share a folder on Windows, please follow these steps:
Right-click on the folder, then select Properties
Go to Sharing tab, then select Advanced Sharing
In the opened window, enable Share this folder
In the "Share name" textbox type the name of the share. This is the name of the share over the network\
Then click on Permissions, then Add
In the textbox type the name of the user or the group (with the domain if required) that you want to grant access to the share, then click "OK"\
Select the added user (or group) and grant the required permissions\
Click OK, OK, and then Close
Open "This PC" from the Start menu, then click on "Map network drive"\
In the "Map Network Drive" window, select a suitable drive letter name and enter the remote folder path of the network share that has been given to you by your storage admin (it should be something with the format \\mycompanyshareserver.mycompany\foldername)\
Make sure to check "Reconnect at sign-in" and "Connect using different credentials".
Click Finish
In the "Windows Security" window enter the username (with the domain, if required) and the password for the network share and check "Remember my credentials".
Click "OK"
Akamas reasons in terms of a system to be optimized and in terms of parameters and metrics of components of that system. To understand the link between metrics collected from LoadRunner and a specific component, the LoadRunner provider looks up the property loadrunner
in the components of a system:
To install the LoadRunner provider, create a YAML file (called provider.yml
in this example) with the definition of the provider:
Then you can install the provider using the Akamas CLI:
The installed provider is shared with all users of your Akamas installation and can monitor many different systems, by configuring appropriate telemetry provider instances.
To create an instance of the LoadRunner provider, build a YAML file (instance.yml
in this example) with the definition of the instance:
Then you can create the instance for the system
using the Akamas CLI:
When you create an instance of the LoadRunner provider, you should specify some configuration information to allow the provider to correctly extract and process metrics from LoadRunner.
You can specify configuration information within the config
part of the YAML of the instance definition.
hostname
: The hostname or IP of the server that hosts the CIFS share where the LoadRunner results have been exported. See .
username
: The username and the domain required to access the network share. Supported formats are:
username
domain\username
username@domain
password
: The password required to access the network share
shareName
: The name of the network share as it is exposed by the server
The following YAML file provides the reference for the definition of a telemetry instance.
The following table describes the reference for the config
section within the definition of the LoadRunner provider instance:
To create an instance of the Prometheus provider, edit a YAML file (instance.yml
in this example) with the definition of the instance:
Then you can create the instance for the system
using the Akamas CLI:
When you create an instance of the Prometheus provider, you should specify some configuration information to allow the provider to extract and process metrics from Prometheus correctly.
You can specify configuration information within the config
part of the YAML of the instance definition.
address
, a URL or IP identifying the address of the host where Prometheus is installed
port
, the port exposed by Prometheus
user
, the username for the Prometheus service
password
, the user password for the Prometheus service
job
, a string to specify the scraping job name. The default is ".*" for all scraping jobs
logLevel
, set this to "DETAILED" for some extra logs when searching for metrics (default value is "INFO")
headers
, to specify additional custom headers
e.g: headers:
"custom_key": "custom_value"
namespace
, a string to specify the namespace
duration
, integer to determine the duration in seconds for data collection (use a number between 1 and 3600)
enableHttps
, boolean to enable HTTPS in Prometheus (since 3.2.6)
ignoreCertificates
, boolean to ignore SSL certificates
disableConnectionCheck
, boolean to disable initial connection check to Prometheus
The Prometheus provider allows defining additional queries to populate custom metrics or redefine the default ones according to your use case. You can configure additional metrics using the metrics
field as shown in the configuration below:
In this example, the telemetry instance will populate cust_metric
with the results of the query specified in datasource
, maintaining the value of the labels listed under labels
.
Akamas pre-processes the queries before running them, replacing special-purpose placeholders with the fields provided in the components. For example, given the following component definition:
the query sum(jvm_memory_used_bytes{instance=~"$INSTANCE$", job=~"$JOB$"})
will be expanded for this component into sum(jvm_memory_used_bytes{instance=~"service01", job=~"jmx"})
. This provides greater flexibility through the templatization of the queries, allowing the same query to select the correct data sources for different components.
The following is the list of available placeholders:
This section reports common use cases addressed by this provider.
To gather kubernetes metrics, the following exporters are required:
kube-state-metrics
cadvisor
As an example, you can define a component with type Kubernetes Container
in this way:
The command will expose on localhost on port 9100 Java metrics of youJar.jar
__ which can be scraped by Prometheus.
config.yaml
is a configuration file useful for the activity of this exporter. It is suggested to use this configuration for an optimal experience with the Prometheus provider:
As a next step, add a new scraping target in the configuration of the Prometheus used by the provider:
You can then create a YAML file with the definition of a telemetry instance (prom_instance.yml
) of the Prometheus provider:
And you can create the telemetry instance using the Akamas CLI:
Finally, to bind the extracted metrics to the related component, you should add the following field to the properties
of the component’s definition:
Here’s the manifest of the node_exporter
service:
The service will expose on localhost on port 9100 system metrics __ which can be scraped by Prometheus.
As a final step, add a new scraping target in the configuration of the Prometheus used by the provider:
You can then create a YAML file with the definition of a telemetry instance (prom_instance.yml
) of the Prometheus provider:
And you can create the telemetry instance using the Akamas CLI:
Finally, to bind the extracted metrics to the related component, you should add the following field to the properties
of the component’s definition:
When you create an instance of the NeoLoad Web provider, you should specify some configuration information to allow the provider to correctly extract and process metrics from NeoLoad Web.
You can specify configuration information within the config
part of the YAML of the instance definition.
accountToken
- NeoLoad Web API access token.
The following YAML file describes the definition of a telemetry instance.
The following table provides the reference for the config
section within the definition of the NeoLoad Web provider instance:
Notice: the NeoLoadWeb provider imports data points matching at least one of the configured values for both metrics
and actions
.
This section reports common use cases addressed by this provider.
This example shows how to configure the NeoLoad Web provider in order to collect performance metrics published on the SaaS web API.
You must create a YAML file with the definition of a telemetry instance (neoload_instance.yml
) of the NeoLoad Web provider:
and then create the telemetry instance using the Akamas CLI:
You can then configure the workflow in order to trigger the execution of a NeoLoad test using the NeoLoadWeb provider, as in the following example:
This section reports common best practices you can adopt to ease the use of this telemetry provider.
filter the imported metrics: import only the required metrics using the metrics
and actions
filters, in order to avoid throttling on the NeoLoad Web instance.
To install the NeoLoad Web provider, create a YAML file (called provider.yml
in this example) with the definition of the provider:
Then you can install the provider using the Akamas CLI:
The installed provider is shared with all users of your Akamas installation and can monitor many different systems, by configuring appropriate telemetry provider instances.
This page describes how to set up a CloudWatch exporter in order to gather AWS metrics through the Prometheus provider. This is especially useful to monitor system metrics when you don’t have direct SSH access to AWS resources like EC2 Instances or if you want to gather AWS-specific metrics not available in the guest OS.
In order to fetch metrics fromCloudWatch, the exporter requires an IAM user or role with the following privileges:
cloudwatch:GetMetricData
cloudwatch:GetMetricStatistics
cloudwatch:ListMetrics
tag:GetResources
You can assign AWS-managed policies CloudWatchReadOnlyAccess and ResourceGroupsandTagEditorReadOnlyAccess to the desired user to enable these permissions.
The CloudWatch exporter repository is available on the . It requires a minimal configuration to fetch metrics from the desired AWS instances. Below is a short list of the parameters needed for a minimal configuration:
region: AWS region of the monitored resource
metrics: a list of objects containing filters for the exported metrics
aws_namespace: the namespace of the monitored resource
aws_metric_name: the name of the AWS metric to fetch
aws_dimensions: the dimension to expose as labels
aws_dimension_select: the dimension to filter over
aws_statistics: the list of metric statistics to expose
aws_tag_select: optional tags to filter on
tag_selections: map containing the list of values to select for each tag
resource_type_selection: resource type to fetch the tags from (see: )
resource_id_dimension: dimension to use for the resource id (see: )
Notice: AWS bills CloudWatch usage in batches of 1 million requests, where every metric counts as a single request. To avoid unnecessary expenses configure only the metrics you need.
Notice: AWS bills CloudWatch usage in batches of 1 million requests, where every metric counts as a single request. To avoid unnecessary expenses configure an appropriate scraping interval.
Once you configured the exporter in the Prometheus configuration you can start to fetch metrics using the Prometheus provider. The following sections describe some scripts you can add as tasks in your workflow.
Since Amazon bills your CloudWatch queries is wise to run the exporter only when needed. The following script allows you to manage the exporter from the workflow by adding the following tasks:
start the container right before the beginning of the load test (command: bash script.sh start
)
The example below is the Akamas-supported configuration, fetching metrics of EC2 instances named server1 and server2.
To install the Prometheus provider, create a YAML file (provider.yml
in this example) with the definition of the provider:
Then you can install the provider using the Akamas CLI:
The installed provider is shared with all users of your Akamas installation and can monitor many different systems, by configuring appropriate telemetry provider instances.
Akamas property | Dynatrace property | Location |
---|---|---|
Akamas property | Dynatrace property | Location |
---|---|---|
Kubernetes type | Dynatrace type |
---|---|
Field | Type | Description | Default Value | Restrictions | Required |
---|---|---|---|---|---|
Field | Type | Description | Restrictions | Required |
---|---|---|---|---|
You can check to see how component-types metrics are extracted by this provider.
Refer to to see how component-type metrics are extracted by this provider.
Notice: you should configure your Prometheus instances so that the Prometheus provider can leverage the instance
property of components, as described in the section here above.
Field | Type | Description | Default value | Restriction | Required |
---|
Check for a list of all Spark application metrics available in Akamas
This example shows how to configure a Spark History Server provider in order to collect performance metrics about a Spark application submitted to the cluster using the operator.
Finally, you will need to define for your study a workflow that includes the submission of the Spark application to the cluster, in this case using the :
The provider expects that its required data are on a CIFS share. If you are using the , please follow the instruction on here below on how to export LoadRunner results in a network share by setup a share on the LoadRunner Controller.
Field | Type | Description | Default Value | Restrictions | Required |
---|
Please refer to for a complete reference of PromQL
Placeholder | Usage example | Component definition example | Expanded query | Description |
---|
Check for a list of all the Java metrics available in Akamas
You can leverage the Prometheus provider to collect Java metrics by using the . The JMX Exporter is a collector of Java metrics for Prometheus that can be run as an agent for any Java application. Once downloaded, you execute it alongside a Java application with this command:
Check the for a list of all the system metrics available in Akamas
You can leverage the Prometheus provider to collect system metrics (Linux) by using the . The Node exporter is a collector of system metrics for Prometheus that can be run as a standalone executable or a service within a Linux machine to be monitored. Once downloaded, schedule it as a service using, for example, systemd:
Field | Type | Description | Default value | Restrictions | Required |
---|
Check the page for a list of all web application metrics available in Akamas
For a complete list of possible values for namespaces, metrics, and dimensions please refer to the official .
The suggested deployment mode for the exporter is through a . The following snippet provides a command line example to run the container (remember to provide your AWS credentials if needed and the path of the configuration file):
You can refer to the for more details or alternative deployment modes.
In order to scrape the newly created exporter add a new job to the configuration file. You will also need to define some in order to add the instance
label required by Akamas to properly filter the incoming metrics.
In the example below the instance
label is copied from the instance’s Name
tag:
It’s worth noting that CloudWatch may require some minutes to aggregate the stats according to the configured granularity, causing the telemetry provider to fail while trying to fetch data points not available yet. To avoid such issues you can add at the end of your workflow a task using an to wait for the CloudWatch metrics to be ready. The following script is an example of implementation:
stop the container after the metrics publication, as described in the (command: bash script.sh stop
).
collects metrics from CSV files
collects metrics from Dynatrace
collects metrics from Prometheus
collects metrics from Spark History Server
collects metrics from Tricentis Neoload Web
collects metrics from MicroFocus Load Runner Professional
collects metrics from MicroFocus Load Runner Enterprise
collects price metrics for Amazon Elastic Compute Cloud (ec2) from Amazon’s own APIs
namespace
Kubernetes namespace
Container dashboard
containerName
Kubernetes container name
Container dashboard
basePodName
Kubernetes base pod name
Container dashboard
state
State
Pod dashboard
namespace
Namespace
Pod dashboard
workload
Workload
Pod dashboard
Docker container
CONTAINER_GROUP_INSTANCE
Pod
CLOUD_APPLICATION_INSTANCE
Workload
CLOUD_APPLICATION
Namespace
CLOUD_APPLICATION_NAMESPACE
Cluster
KUBERNETES_CLUSTER
address
String
The address of the machine where the CSV file resides
A valid URL or IP
Yes
port
Number (integer)
The port to connect to, in order to retrieve the file
22
1≤port
≤65536
No
username
String
The username to use in order to connect to the remote machine
Yes
protocol
String
scp
scp
sftp
No
authType
String
Specify which method is used to authenticate against the remote machine:
password: use the value of the parameter auth
as a password
key: use the value of the parameter auth
as a private key. Supported formats are RSA and DSA
password
key
Yes
auth
String
A password or an RSA/DSA key (as YAML multi-line string, keeping new lines)
Yes
remoteFilePattern
String
The path of the remote file(s) to be analyzed. The path can contains GLOB expressio
A list of valid path for linux
Yes
componentColumn
String
The CSV column containing the name of the component.
The column's values must match (case sensitive) the name of a component specified in the System
COMPONENT
The column must exists in the CSV file
Yes
timestampColumn
String
The CSV column containing the timestamps of the samples
TS
The column must exists in the CSV file
No
timestampFormat
String
Timestamps' format
YYYY-mm-ddTHH:MM:ss
Must be specified using Java syntax.
No
fieldSeparator
String
Specify the field separator of the CSV
,
,
;
No
metric
String
The name of the metric in Akamas
An existing Akamas metric
Yes
datasourceMetric
String
The name (header) of the column that contains the specific metric
An existing column in the CSV file
Yes
scale
Decimal number
The scale factor to apply when importing the metric
staticLabels
List of key-value pairs
A list of key-value pairs that will be attached to the specific metric sample
No
| URL | Spark History Server address | Yes |
| String | Granularity of the imported metrics |
| Allowed values: | No |
| Integer | Spark History Server listening port |
| No |
| String | A valid access token | Yes |
| URL | Hostname of the NeoLoad Web API |
| No |
| List of strings | List of component metrics to import |
| No |
| List of strings | List of "User Paths" to import | No |
To install the LoadRunnerEnterprise provider, create a YAML file (called provider.yml
in this example) with the definition of the provider:
Then you can install the provider using the Akamas CLI:
The installed provider is shared with all users of your Akamas installation and can monitor many different systems, by configuring appropriate telemetry provider instances.
The LoadRunner Enterprise provider collects metrics from a LoadRunner Enterprise instance and makes them available to Akamas.
This section provides the minimum requirements that you should match before using the LoadRunnerEnterprise telemetry provider.
LoadRunner Enterprise 12.60, 12.63 and 2020 SP3, 2022 and 2022 R1
InfluxDB 1.7 and 1.8
Port 8086 between Akamas VM to InfluxDB host, opened in both directions. This port is used to gather metrics.
The provider requires a user that can access InfluxDB.
The user must have read permission on the database containing the LoadRunner metrics.
Web Application
You can check LoadRunner provider metrics mapping to see how component-types metrics are extracted by this provider.
This section lists the workflow operators this provider depends on.
To set up the integration between Loadrunner Enterprise and InfluxDB please follow the official Microfocus documentation. Akamas does not require any additional setup on the data source.
Akamas reasons in terms of a system to be optimized and in terms of parameters and metrics of components of that system. To understand the link between metrics collected from LoadRunnerEnterprise through InfluxDB and a specific component, the LoadRunnerEnterprise provider looks up some properties in the components of a system:
loadrunnerenterprise
You can use this example to start building your component specification:
To install the AWS provider, create a YAML file (called provider.yml
in this example) with the specification of the provider:
Then you can install the provider with the Akamas CLI:
The installed provider is shared with all users of your Akamas installation and can monitor many different systems, by configuring appropriate telemetry provider instances.
hostname | String | - | IP address or FQDN | Yes |
username | String | The username and the domain required to access the share | - | Supported formats:
| Yes |
password | String | The password required to access the share | - | - | Yes |
shareName | String | The name of the share as it is exposed by the server | - | - | Yes |
pathPrefix | String | A prefix for the default path where the provider looks for the data. The default path is {studyName}{experimentId}{trialId} | - | A valid Windows path | No |
|
|
| These placeholders are replaced respectively with the |
|
|
| This placeholder is replaced with a list containing any additional filter in the component’s definition (other than |
|
|
|
|
|
| These placeholders are used within kubernetes environments |
The AWS provider collects price metrics for Amazon Elastic Compute Cloud (EC2) from Amazon’s own APIs.
This provider imports just one metricaws_ec2_price
which is available in the EC2 component type of the AWS Optimization pack.
This section provides the minimum requirements that you should match before using the AWS telemetry provider.
This telemetry provider makes use of parameters aws_ec2_instance_type
andaws_ec2_instance_size
to identify the price. When using this provider make sure that your system has a component of type EC2 and that those parameters are defined in the baseline and, if optimized, in the parameters selection.
An IAM user who has been granted the AWSPriceListServiceFullAccess
policy, that is the following permissions:
DescribeServices
GetAttributeValues
GetProducts
You may find more information on AWS cost permissions here.
Versions >= 2.0.0 are compatible with Akamas from version 1.9.0
EC2
In order to gather price information about a component you’re required to input an extra field in its definition:
region
, which tells the provider the AWS region of the modeled instance
Please note this field is mandatory and must be specified as follows:
Here is a complete list of AWS region names, together with their Akamas-compatible codes:
To create an instance of the LoadRunnerEnterprise provider, build a YAML file (instance.yml
in this example) with the definition of the instance:
Then you can create the instance using the Akamas CLI:
When you create an instance of the LoadRunnerEnterprise provider, you should specify some configuration information to allow the provider to correctly extract and process metrics from Prometheus.
You can specify configuration information within the config
part of the YAML of the instance definition.
address
: The address of the InfluxDB instance, in the form of schema://address (i.e https://influxdb.mycompay.com)
port
: The InfluxDB port
username
: The username required to connect to InfluxDB.
password
: The password for the username
database
: The database name where LoadRunner metrics are stored
This table reports the reference for the config
section within the definition of the LoadRunnerEnterprise provider instance:
To create an instance of the CSV provider, build a YAML file (instance.yml
in this example) with the definition of the instance:
Then you can create the instance for the aws-system
using the Akamas CLI:
When you create an instance of the AWS provider, you should specify some configuration information to allow the provider to correctly extract and process metrics from your CSV files.
You can specify configuration information within the config
part of the YAML of the instance definition.
accessKeyId
- the access key id of your chosen IAM use
secretAccessKey
- the secret access key of your chosen IAM user
The following represents the complete configuration reference for the telemetry provider instance.
Then you can create the instance for the aws-system
using the Akamas CLI:
When you create an instance of the AWS provider, you should specify some configuration information to allow the provider to correctly extract and process metrics from your CSV files.
You can specify configuration information within the config
part of the YAML of the instance definition.
accessKeyId
- the access key id of your chosen IAM use
secretAccessKey
- the secret access key of your chosen IAM user
The following YAML file represents a template to define the telemetry provider instance.
The following table describes the configuration reference for the config
section
the hostname or IP of the server that hosts the CIFS share where the LoadRunner results have been exported. See
See
See below
See below
If not set in the component properties, this placeholder is replaced with the duration field configured in the telemety-instance. You should use it with instead of hardcoding a fixed value.
See
Region Name | Region |
---|---|
Field | Type | Description | Default value | Restrictions | Required |
---|---|---|---|---|---|
Field | Type | Description | Default Value | Restrictions | Required |
---|---|---|---|---|---|
US East (Ohio)
us-east-2
US East (N. Virginia)
us-east-1
US West (N. California)
us-west-1
US West (Oregon)
us-west-2
Africa (Cape Town)
af-south-1
Asia Pacific (Hong Kong)
ap-east-1
Asia Pacific (Mumbai)
ap-south-1
Asia Pacific (Osaka-Local)
ap-northeast-3
Asia Pacific (Seoul)
ap-northeast-2
Asia Pacific (Singapore)
ap-southeast-1
Asia Pacific (Sydney)
ap-southeast-2
Asia Pacific (Tokyo)
ap-northeast-1
Canada (Central)
ca-central-1
China (Beijing)
cn-north-1
China (Ningxia)
cn-northwest-1
Europe (Frankfurt)
eu-central-1
Europe (Ireland)
eu-west-1
Europe (London)
eu-west-2
Europe (Milan)
eu-south-1
Europe (Paris)
eu-west-3
Europe (Stockholm)
eu-north-1
Middle East (Bahrain)
me-south-1
South America (São Paulo)
sa-east-1
AWS GovCloud (US-East)
us-gov-east-1
AWS GovCloud (US)
us-gov-west-1
address
String
The address of the InfluxDB
-
A valid URL
Yes
port
Integer
The port of the InfluxDB
-
A valid port
yes
username
String
The username to connect to InfluxDB
-
-
Yes
password
String
The password of the specified user
-
-
Yes
database
String
The database with the metrics
-
-
Yes
verifySSL
Boolean
Wheter to check the certificate, of InfluxDB API is exposed in HTTPS
false
true/false
No
transactions
List of string
Restrict metrics collection to the listed transactions names. If not specified the provider will collect metrics of all transactions
Empty
accessKeyId
String
The access key id of your chosen IAM user
Valid IAM credentials
Yes
secretAccessKey
String
The secret access key of your chosen IAM user
Valid IAM credentials
Yes