Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
The CSV provider collects metrics from CSV files and makes them available to Akamas. It offers a very versatile way to integrate custom data sources.
This section provides the minimum requirements that you should match before using the CSV File telemetry provider.
The following requirements should be met to enable the provider to gather CSV files from remote hosts:
Port 22 (or a custom one) should be open from Akamas installation to the host where the files reside.
The host where the files reside should support SCP or SFTP protocols.
Read access to the CSV files target of the integration
Versions < 2.0.0 are compatibile with Akamas until version 1.8.0
Versions >= 2.0.0 are compatible with Akamas from version 1.9.0
The CSV File provider is generic and allows integration with any data source, therefore it does not come with support for a specific component type.
To operate properly, the CSV file provider expects the presence of four fields in each processed CSV file:
A timestamp field used to identify the point in time a certain sample refers to.
A component field used to identify the Akamas entity.
A metric field used to identify the name of the metric.
A value field used to store the actual value of the metric.
These fields can have custom names in the CSV file, you can specify them in the provider configuration.
The Install CSV provider page describes how to get this Telemetry Provider installed. Once installed, this provider is shared with all users of your Akamas installation and can be used to monitor many different systems, by configuring appropriate telemetry provider instances as described in the Create a CSV provider instance page.
To install the CSV File provider, create a YAML file (called provider.yml
in this example) with the specification of the provider:
Then, you can then install the provider with the Akamas CLI:
To install the Prometheus provider, create a YAML file (provider.yml
in this example) with the definition of the provider:
Then you can install the provider using the Akamas CLI:
The installed provider is shared with all users of your Akamas installation and can monitor many different systems, by configuring appropriate telemetry provider instances.
To install the Spark History Server provider, create a YAML file (called provider.yml
in this example) with the definition of the provider:
Then you can install the provider using the Akamas CLI:
The installed provider is shared with all users of your Akamas installation and can monitor many different systems, by configuring appropriate telemetry provider instances.
To create an instance of the Spark History Server provider, build a YAML file (instance.yml
in this example) with the definition of the instance:
Then you can create the instance for the system spark-system
using the Akamas CLI:
When you create an instance of the Spark History Server provider, you should specify some configuration information to allow the provider to correctly extract and process metrics from the Spark History server.
You can specify configuration information within the config
part of the YAML of the instance definition.
address
- hostname of the Spark History Server instance
The following YAML file describes the definition of a telemetry instance.
The following table reports the reference for the config
section within the definition of the Spark History Server provider instance:
address
URL
Spark History Server address
Yes
importLevel
String
Granularity of the imported metrics
job
Allowed values: job
, stage
, task
No
port
Integer
Spark History Server listening port
18080
No
This section reports common use cases addressed by this provider.
Check Spark Application page for a list of all Spark application metrics available in Akamas
This example shows how to configure a Spark History Server provider in order to collect performance metrics about a Spark application submitted to the cluster using the Spark SSH Submit operator.
As a first step, you need to create a YAML file (spark_instance.yml
) containing the configuration the provider needs to connect to the Spark History Server, plus the filter on the desired level of granularity for the imported metrics:
and then create the telemetry instance using the Akamas CLI:
Finally, you will need to define for your study a workflow that includes the submission of the Spark application to the cluster, in this case using the Spark SSH Submit operator:
This section reports common best practices you can adopt to ease the use of this telemetry provider.
configure metrics granularity: in order to reduce the collection time, configure the importLevel
to import metrics with a granularity no finer than the study requires.
wait for metrics publication: make sure in the workflow there is a few-minute interval between the end of the Spark application and the execution of the Spark telemetry instance, since the Spark History Server may take some time to complete the publication of the metrics.
To install the NeoLoad Web provider, create a YAML file (called provider.yml
in this example) with the definition of the provider:
Then you can install the provider using the Akamas CLI:
The installed provider is shared with all users of your Akamas installation and can monitor many different systems, by configuring appropriate telemetry provider instances.
When you create an instance of the NeoLoad Web provider, you should specify some configuration information to allow the provider to correctly extract and process metrics from NeoLoad Web.
You can specify configuration information within the config
part of the YAML of the instance definition.
accountToken
- NeoLoad Web API access token.
The following YAML file describes the definition of a telemetry instance.
The following table provides the reference for the config
section within the definition of the NeoLoad Web provider instance:
account
String
A valid access token
Yes
neoloadApi
URL
Hostname of the NeoLoad Web API
https://neoload-api.saas.neotys.com
No
metrics
List of strings
List of component metrics to import
['<all transactions>']
No
actions
List of strings
List of "User Paths" to import
No
Notice: the NeoLoadWeb provider imports data points matching at least one of the configured values for both metrics
and actions
.
This section reports common use cases addressed by this provider.
Check the Web Application page for a list of all web application metrics available in Akamas
This example shows how to configure the NeoLoad Web provider in order to collect performance metrics published on the SaaS web API.
You must create a YAML file with the definition of a telemetry instance (neoload_instance.yml
) of the NeoLoad Web provider:
and then create the telemetry instance using the Akamas CLI:
You can then configure the workflow in order to trigger the execution of a NeoLoad test using the NeoLoadWeb provider, as in the following example:
This section reports common best practices you can adopt to ease the use of this telemetry provider.
filter the imported metrics: import only the required metrics using the metrics
and actions
filters, in order to avoid throttling on the NeoLoad Web instance.
The LoadRunner Enterprise provider collects metrics from a LoadRunner Enterprise instance and makes them available to Akamas.
This section provides the minimum requirements that you should match before using the LoadRunnerEnterprise telemetry provider.
LoadRunner Enterprise 12.60, 12.63 and 2020 SP3
InfluxDB 1.7 and 1.8
Port 8086 between Akamas VM to InfluxDB host, opened in both directions. This port is used to gather metrics.
The provider requires a user that can access InfluxDB.
The user must have read permission on the database containing the LoadRunner metrics.
Web Application
You can check LoadRunner provider metrics mapping to see how component-types metrics are extracted by this provider.
This section lists the workflow operators this provider depends on.
To set up the integration between Loadrunner Enterprise and InfluxDB please follow the official Microfocus documentation. Akamas does not require any additional setup on the data source.
Akamas reasons in terms of a system to be optimized and in terms of parameters and metrics of components of that system. To understand the link between metrics collected from LoadRunnerEnterprise through InfluxDB and a specific component, the LoadRunnerEnterprise provider looks up some properties in the components of a system:
loadrunnerenterprise
You can use this example to start building your component specification:
To install the LoadRunner provider, create a YAML file (called provider.yml
in this example) with the definition of the provider:
Then you can install the provider using the Akamas CLI:
The installed provider is shared with all users of your Akamas installation and can monitor many different systems, by configuring appropriate telemetry provider instances.
To create an instance of the CSV provider, build a YAML file (instance.yml
in this example) with the definition of the instance:
Then you can create the instance for the aws-system
using the Akamas CLI:
When you create an instance of the AWS provider, you should specify some configuration information to allow the provider to correctly extract and process metrics from your CSV files.
You can specify configuration information within the config
part of the YAML of the instance definition.
accessKeyId
- the access key id of your chosen IAM use
secretAccessKey
- the secret access key of your chosen IAM user
The following represents the complete configuration reference for the telemetry provider instance.
Then you can create the instance for the aws-system
using the Akamas CLI:
When you create an instance of the AWS provider, you should specify some configuration information to allow the provider to correctly extract and process metrics from your CSV files.
You can specify configuration information within the config
part of the YAML of the instance definition.
accessKeyId
- the access key id of your chosen IAM use
secretAccessKey
- the secret access key of your chosen IAM user
The following YAML file represents a template to define the telemetry provider instance.
The following table describes the configuration reference for the config
section
accessKeyId
String
The access key id of your chosen IAM user
Valid IAM credentials
Yes
secretAccessKey
String
The secret access key of your chosen IAM user
Valid IAM credentials
Yes
The Prometheus provider collects metrics from a Prometheus instance and makes them available to Akamas.
This provider includes support for several technologies. In any case, you can specify your own queries to gather any metric you want.
This section provides the minimum requirements that you should match before using the Prometheus provider.
Prometheus 1.x and 2.x.
Prometheus is reachable at a provided URL or IP address and port (the default Prometheus port is 9090)
Node exporter (Linux system metrics)
JMX exporter (Java metrics)
cAdvisor (Docker container metrics)
CloudWatch exporter (AWS resources metrics)
Jmeter (Web application metrics)
Versions < 2.0.0 are compatibile with Akamas untill version 1.8.0
Versions >= 2.0.0 are compatible with Akamas from version 1.9.0
Ubuntu-16.04, Rhel-7.6
java-openjdk-8, java-openjdk-11
java-ibm-j9vm-6, java-ibm-j9vm-8, java-eclipse-openj9-11
Container
ec2
Web Application
You can check Prometheus provider metrics mapping to see how component-types metrics are extracted by this provider
The Prometheus provider does not require a specific configuration of the Prometheus instance it uses, except for the need for setting the value of the instance
label of each data point so that it matches the value of the instance
property in a component; in this way, the Prometheus provider can know which data point refers to which component in a system.
Here’s an example configuration for Prometheus that sets the instance
label:
Akamas reasons in terms of a system to be optimized and in terms of parameters and metrics of components of that system. To understand which metrics collected from Prometheus should refer to which component, the Prometheus provider looks up some properties in the components of a system grouped under prometheus
property. These properties are:
instance
, which tells the provider "where" a component is
job
(optional), which filters the imported metrics by exporter identifier, useful when different components run on the same instance. If not provided, this value defaults to the one configured globally in the telemetry-instance definition.
Nested under this property you can also include any additional field your use case may require to further filter the imported metrics. These fields will be appended in queries to the list of label matches in the form field_name=~'field_value'
, and can specify either exact values or patterns.
Notice: you should configure your Prometheus instances so that the Prometheus provider can leverage the instance
property of components, as described in the Setup datasource section here above.
It is important that you add instance
and, optionally, the job
properties to the components of a system so that the Prometheus provider can gather metrics from them:
The AWS provider collects price metrics for Amazon Elastic Compute Cloud (EC2) from Amazon’s own APIs.
This section provides the minimum requirements that you should match before using the AWS telemetry provider.
A IAM user that has been granted the AWSPriceListServiceFullAccess
policy, that is the following permissions:
DescribeServices
GetAttributeValues
GetProducts
You may find more information on AWS cost permissions here.
Versions >= 2.0.0 are compatible with Akamas from version 1.9.0
EC2
In order to gather price information about a component you’re required to input an extra field in its definition:
region
, which tells the provider the AWS region of the modeled instance
Please note this field is mandatory and must be specified as follows:
Here is a complete list of AWS region names, together with their Akamas-compatible codes:
US East (Ohio)
us-east-2
US East (N. Virginia)
us-east-1
US West (N. California)
us-west-1
US West (Oregon)
us-west-2
Africa (Cape Town)
af-south-1
Asia Pacific (Hong Kong)
ap-east-1
Asia Pacific (Mumbai)
ap-south-1
Asia Pacific (Osaka-Local)
ap-northeast-3
Asia Pacific (Seoul)
ap-northeast-2
Asia Pacific (Singapore)
ap-southeast-1
Asia Pacific (Sydney)
ap-southeast-2
Asia Pacific (Tokyo)
ap-northeast-1
Canada (Central)
ca-central-1
China (Beijing)
cn-north-1
China (Ningxia)
cn-northwest-1
Europe (Frankfurt)
eu-central-1
Europe (Ireland)
eu-west-1
Europe (London)
eu-west-2
Europe (Milan)
eu-south-1
Europe (Paris)
eu-west-3
Europe (Stockholm)
eu-north-1
Middle East (Bahrain)
me-south-1
South America (São Paulo)
sa-east-1
AWS GovCloud (US-East)
us-gov-east-1
AWS GovCloud (US)
us-gov-west-1
This page describes how to set up an OracleDB exporter in order to gather metrics regarding an Oracle Database instance through the Prometheus provider.
The OracleDB exporter repository is available on the official project page. The suggested deploy mode is through a Docker image, since the Prometheus instance can easily access the running container through the Akamas network.
Use the following command line to run the container, where cust-metrics.toml
is your configuration file defining the queries for additional custom metrics (see paragraph below) and DATA_SOURCE_NAME
an environment variable containing the Oracle EasyConnect string:
You can refer to the official guide for more details or alternative deployment modes.
It is possible to define additional queries to expose custom metrics using any data in the database instance that is readable by the monitoring user (see the guide for more details about the syntax).
The following is an example of exporting system metrics from the Dynamic Performance (V$) Views used by the Prometheus provider default queries for the Oracle Database optimization pack:
This page describes how to set up a CloudWatch exporter in order to gather AWS metrics through the Prometheus provider. This is especially useful to monitor system metrics when you don’t have direct SSH access to AWS resources like EC2 Instances or if you want to gather AWS-specific metrics not available in the guest OS.
In order to fetch metrics fromCloudWatch, the exporter requires an IAM user or role with the following privileges:
cloudwatch:GetMetricData
cloudwatch:GetMetricStatistics
cloudwatch:ListMetrics
tag:GetResources
You can assign AWS-managed policies CloudWatchReadOnlyAccess and ResourceGroupsandTagEditorReadOnlyAccess to the desired user to enable these permissions.
The CloudWatch exporter repository is available on the official project page. It requires a minimal configuration to fetch metrics from the desired AWS instances. Below is a short list of the parameters needed for a minimal configuration:
region: AWS region of the monitored resource
metrics: a list of objects containing filters for the exported metrics
aws_namespace: the namespace of the monitored resource
aws_metric_name: the name of the AWS metric to fetch
aws_dimensions: the dimension to expose as labels
aws_dimension_select: the dimension to filter over
aws_statistics: the list of metric statistics to expose
aws_tag_select: optional tags to filter on
tag_selections: map containing the list of values to select for each tag
resource_type_selection: resource type to fetch the tags from (see: Resource Types)
resource_id_dimension: dimension to use for the resource id (see: Resource Types)
For a complete list of possible values for namespaces, metrics, and dimensions please refer to the official AWS CloudWatch User Guide.
Notice: AWS bills CloudWatch usage in batches of 1 million requests, where every metric counts as a single request. To avoid unnecessary expenses configure only the metrics you need.
The suggested deployment mode for the exporter is through a Docker image. The following snippet provides a command line example to run the container (remember to provide your AWS credentials if needed and the path of the configuration file):
You can refer to the official guide for more details or alternative deployment modes.
In order to scrape the newly created exporter add a new job to the configuration file. You will also need to define some relabeling rules in order to add the instance
label required by Akamas to properly filter the incoming metrics.
In the example below the instance
label is copied from the instance’s Name
tag:
Notice: AWS bills CloudWatch usage in batches of 1 million requests, where every metric counts as a single request. To avoid unnecessary expenses configure an appropriate scraping interval.
Once you configured the exporter in the Prometheus configuration you can start to fetch metrics using the Prometheus provider. The following sections describe some scripts you can add as tasks in your workflow.
It’s worth noting that CloudWatch may require some minutes to aggregate the stats according to the configured granularity, causing the telemetry provider to fail while trying to fetch data points not available yet. To avoid such issues you can add at the end of your workflow a task using an Executor operator to wait for the CloudWatch metrics to be ready. The following script is an example of implementation:
Since Amazon bills your CloudWatch queries is wise to run the exporter only when needed. The following script allows you to manage the exporter from the workflow by adding the following tasks:
start the container right before the beginning of the load test (command: bash script.sh start
)
stop the container after the metrics publication, as described in the previous section (command: bash script.sh stop
).
The example below is the Akamas-supported configuration, fetching metrics of EC2 instances named server1 and server2.
To create an instance of the CSV provider, build a YAML file (instance.yml
in this example) with the definition of the instance:
Then you can create the instance for the system
using the Akamas CLI:
timestampFormat
formatRegarding the timestamp format, please notice that while the week-year format YYYY
is compliant with the ISO-8601 specification, but you should replace it with the year-of-era format yyyy
if you are specifying a timestampFormat
different from the ISO one. For example:
Correct: yyyy-MM-dd HH:mm:ss
Wrong: YYYY-MM-dd HH:mm:ss
You can find detailed information on timestamp patterns in the Patterns for Formatting and Parsing section on the DateTimeFormatter (Java Platform SE 8) page.
When you create an instance of the CSV provider, you should specify some configuration information to allow the provider to correctly extract and process metrics from your CSV files.
You can specify configuration information within the config
part of the YAML of the instance definition.
address
- a URL or IP identifying the address of the host where CSV files reside
username
- the username used when connecting to the host
authType
- the type of authentication to use when connecting to the file host; either password
or key
auth
- the authentication credential; either a password or a key according to authType
. When using keys, the value can either be the value of the key or the path of the file to import from
remoteFilePattern
- a list of remote files to be imported
protocol
- the protocol to use to retrieve files; either scp
or sftp
. Default is scp
fieldSeparator
- the character used as a field separator in the csv files. Default is ,
componentColumn
- the header of the column containing the name of the component. Default is COMPONENT
timestampColumn
- the header of the column containing the timestamp. Default is TS
timestampFormat
- the format of the timestamp (e.g. yyyy-MM-dd HH:mm:ss zzz
). Default is YYYY-MM-ddTHH:mm:ss
You should also specify the mapping between the metrics available in your CSV files and those provided by Akamas. This can be done in the metrics
section of the telemetry instance configuration. To map a custom metric you should specify at least the following properties:
metric
- the name of a metric in Akamas
datasourceMetric
- the header of a column that contains the metric in the CSV file
The provider ignores any column not present as datasourceMetric
in this section.
The sample configuration reported in this section would import the metric cpu_util
from CSV files formatted as in the example below:
The following represents the complete configuration reference for the telemetry provider instance.
The following table reports the configuration reference for the config
section
address
String
The address of the machine where the CSV file resides
A valid URL or IP
Yes
port
Number (integer)
The port to connect to, in order to retrieve the file
22
1≤port
≤65536
No
username
String
The username to use in order to connect to the remote machine
Yes
protocol
String
scp
scp
sftp
No
authType
String
Specify which method is used to authenticate against the remote machine:
password: use the value of the parameter auth
as a password
key: use the value of the parameter auth
as a private key. Supported formats are RSA and DSA
password
key
Yes
auth
String
A password or an RSA/DSA key (as YAML multi-line string, keeping new lines)
Yes
remoteFilePattern
String
A list of valid path for linux
Yes
componentColumn
String
The CSV column containing the name of the component.
The column's values must match (case sensitive) the name of a component specified in the System
COMPONENT
The column must exists in the CSV file
Yes
timestampColumn
String
The CSV column containing the timestamps of the samples
TS
The column must exists in the CSV file
No
timestampFormat
String
Timestamps' format
YYYY-mm-ddTHH:MM:ss
No
fieldSeparator
String
Specify the field separator of the CSV
,
,
;
No
The following table reports the configuration reference for the metrics
section
metric
String
The name of the metric in Akamas
An existing Akamas metric
Yes
datasourceMetric
String
The name (header) of the column that contains the specific metric
An existing column in the CSV file
Yes
scale
Decimal number
The scale factor to apply when importing the metric
staticLabels
List of key-value pairs
A list of key-value pairs that will be attached to the specific metric sample
No
Here you can find common use cases addressed by this provider.
In this use case, you are going to import some metrics coming from SAR, a popular UNIX tool to monitor system resources. SAR can export CSV files in the following format.
Note that the metrics are percentages (between 1 and 100), while Akamas accepts percentages as values between 0 and 1, therefore each metric in this configuration has a scale factor of 0.001.
You can import the two CPU metrics and the memory metric from a SAR log using the following telemetry instance configuration.
Using the configured instance, the CSV File provider will perform the following operations to import the metrics:
Retrieve the file "/csv/sar.csv" from the server "127.0.0.1" using the SCP protocol authenticating with the provided password.
Use the column hostname
to lookup components by name.
Use the column timestamp
to find the timestamps of the samples (that is expected to be in the format specified by timestampFormat
).
Collect the metrics (two with the same name, but different labels, and one with a different name):
cpu_util
: in the CSV file is in the column %user and attach to its samples the label "mode" with value "user".
cpu_util
: in the CSV file is in the column %system and attach to its samples the label "mode" with value "system".
mem_util
: in the CSV file is in the column %memory.
Dynatrace SaaS or Managed version 1.187 or later
Versions < 2.0.0 are compatible with Akamas until version 1.8.0
Versions >= 2.0.0 are compatible with Akamas from version 1.9.0
Linux
Ubuntu-16.04, Rhel-7.6
JVM
java-openjdk-8, java-openjdk-11
java-ibm-j9vm-6, java-ibm-j9vm-8, java-eclipse-openj9-11
Web Application
Kubernetes and Docker
Refer to Dynatrace provider metrics mapping to see how component-types metrics are extracted by this provider.
a valid Dynatrace license
Dynatrace OneAgent installed on the servers where the Dynatrace entities to be monitored are running
Connectivity between:
Akamas server and Dynatrace server over port 443 (if Dynatrace is deployed on-premises)
Akamas server and internet (over https) if Dynatrace is managed as a SaaS platform
URL (+ port if it is not the default one) of your Dynatrace server (SaaS or Managed)
API token from your Dynatrace server (Saas or Managed) with the rights to "Access problems and event feed, metrics, and topology" - see more details here
A Dynatrace user belonging to the group "Monitoring viewer", which allows access to the environment in read-only mode, without being able to change settings, download or install OneAgent
The Dynatrace provider needs a Dynatrace API token with the following privileges:
metrics.read (Read entities)
entities.read (Read metrics)
DataExport (Access problem and event feed, metrics, and topology)
DataImport (Data ingest, e.g.: metrics and events)
ReadSyntheticData (Read synthetic monitors, locations, and nodes)
To generate an API Token for your Dynatrace installation you can follow these steps.
To map metrics collected from Dynatrace to Akamas component, the Dynatrace provider looks up some properties in the components of a system grouped under dynatrace
.
These properties give you access to different strategies to map Akamas components onto Dynatrace entities:
Map by id
Map by name
Map by tags
You can map a component to a Dynatrace entity by leveraging the unique id of the entity, which you should put under the id
property in the component:
You can find the id of a Dynatrace entity by looking at the URL of a Dynatrace dashboard relative to the entity. Watch out that the "host" key is valid only for Linux components, other components (e.g. the JVM) require to drill down into the host entities to get the PROCESS_GROUP_INSTANCE or PROCESS_GROUP id
You can map a component to a Dynatrace entity by leveraging the entity’s display name which you should put under the name
property in the component definition:
You can map a component to a Dynatrace entity by leveraging Dynatrace tags that match the entity, tags which you should put under the tags
property in the component definition. If multiple tags are specified all instances matching any of the specified tags will be selected.
In the case of a key-only tag, the example above would be:
You can improve the matching of components with Dynatrace by adding a type
property in the component definition, this property will help the Provider match only those Dynatrace entities of the given type.
The type of an entity can be retrieved from the URL of the entity’s dashboard
Available entities types can be retrieved, from your Dynatrace instance, with the following command:
A set of Kubernetes-specific properties can be used to allow users to retrieve data related to the monitored Kubernetes clusters.
Please note, that the property type
is required to retrieve Kubernetes entities. Currently, the supported types in Dynatrace are the following
CONTAINER_GROUP_INSTANCE
Docker container
CLOUD_APPLICATION_INSTANCE
Pod
CLOUD_APPLICATION
Workload
CLOUD_APPLICATION_NAMESPACE
Namespace
KUBERNETES_CLUSTER
Cluster
The following Kubernetes object can be retrieved from Dynatrace.
The following properties are supported for containers:
namespace
Kubernetes namespace
Container dashboard
-
containerName
Kubernetes container name
Container dashboard
-
basePodName
Kubernetes base pod name
Container dashboard
-
The following properties are supported for pods
state
State
Pod dashboard
-
labels
Labels
Pod dashboard
Labels are specified as key-value in Akamas configuration.
In Dynatrace’s dashboard key and value are separated with a column (:
)
The LoadRunner provider collects metrics generated by a LoadRunner instance (converted to their JSON format and placed in CIFS network share) and makes them available to Akamas.
Prerequisites
This section provides the minimum requirements that you should match before using the LoadRunner telemetry provider.
Micro Focus LoadRunner 12.60 or 2020
The network share is reachable at port 445/TCP
The network share is reachable at port 139/UDP
Username, domain (if required), and password of the network share.
Read permission on the network share.
Web Application
You can check LoadRunner provider metrics mapping to see how component-types metrics are extracted by this provider.
The provider expects that its required data are on a CIFS share. If you are using the LoadRunner operator, please follow the instruction on here below on how to export LoadRunner results in a network share by setup a share on the LoadRunner Controller.
To share a folder on Windows, please follow these steps:
Right-click on the folder, then select Properties
Go to Sharing tab, then select Advanced Sharing
In the opened window, enable Share this folder
In the "Share name" textbox type the name of the share. This is the name of the share over the network
Then click on Permissions, then Add
In the textbox type the name of the user or the group (with the domain if required) that you want to grant access to the share, then click OK
Select the added user (or group) and grant the required permissions
Click OK, OK, and then Close
Open "This PC" from the Start menu, then click on "Map network drive"
In the "Map Network Drive" window, select a suitable drive letter name and enter the remote folder path of the network share that has been given to you by your storage admin (it should be something with the format \\mycompanyshareserver.mycompany\foldername)
Make sure to check "Reconnect at sign-in" and "Connect using different credentials".
Click Finish
In the "Windows Security" window enter the username (with the domain, if required) and the password for the network share and check "Remember my credentials".
Click Ok
Akamas reasons in terms of a system to be optimized and in terms of parameters and metrics of components of that system. To understand the link between metrics collected from LoadRunner and a specific component, the LoadRunner provider looks up the property loadrunner
in the components of a system:
The installed provider is shared with all users of your Akamas installation and can monitor many different systems, by configuring appropriate telemetry provider instances.
To create an instance of the Dynatrace provider, build a YAML file (instance.yml
in this example) with the definition of the instance:
Then you can create the instance for the system
using the Akamas CLI:
When you create an instance of the Dynatrace provider, you should specify some configuration information to allow the provider to correctly extract and process metrics from Dynatrace.
You can specify configuration information within the config
part of the YAML of the instance definition.
url
- URL of the Dynatrace installation API (see https://www.dynatrace.com/support/help/extend-dynatrace/dynatrace-api/ to retrieve the URL of your installation)
token
- A Dynatrace API Token with the proper permissions
You can collect additional metrics with the Dynatrace provider by using the metrics
field:
In the case in which Akamas cannot reach directly your Dynatrace installation, you can configure an HTTP proxy by using the proxy
field:
This section reports the complete reference for the definition of a telemetry instance.
This table shows the reference for the config
section within the definition of the Dynatrace provider instance:
url
String
It should be a valid URL
Yes
token
String
Yes
proxy
Object
See Proxy options reference
Yes
The specification of the HTTP proxy to use to communicate with Dynatrace.
pushEvents
String
true, false
No
true
If set to true the provider will inform dynatrace of the configuration change event which will be visible in the Dynatrace UI.
tags
Object
No
A set of global tags to match Dynatrace entities. The provider uses these tags to apply a default filtering of Dynatrace entities for every component.
This table reports the reference for the config
→ proxy
section within the definition of the Dynatrace provider instance:
address
String
It should be a valid URL
Yes
The URL of the HTTP proxy to use to communicate with the Dynatrace installation API
port
Number (integer)
1 <port
<65535
Yes
The port at which the HTTP proxy listens for connections
username
String
No
The username to use when authenticating against the HTTP proxy, if necessary
password
String
No
The username to use when authenticating against the HTTP proxy, if necessary
This table reports the reference for the metrics
section within the definition of the Dynatrace provider instance. The section contains a collection of objects with the following properties:
metric
String
It must be an Akamas metric
Yes
The name of an Akamas metric that should map to the new metric you want to gather
datasourceMetric
String
A valid Dynatrace metric
Yes
The Dynatrace query to use to extract metric
labels
Array of strings
-
No
The list of Dynatrace labels that should be retained when gathering the metric
staticLabels
Key-Value
-
No
Static labels that will be attached to metric samples
This section reports common use cases addressed by this provider.
Check the Linux optimization pack for a list of all the system metrics available in Akamas.
As a first step to start extracting metrics from Dyntrace, generate your API token and make sure it has the right permissions.
As a second step, choose a strategy to map your Linux component (MyLinuxComponent) with the corresponding Dyntrace entity.
Let’s assume you want to map by id your Dynatrace entity, you can find the id in the URL bar of a Dyntrace dashboard of the entity:
Grab the id and add it to the Linux component definition:
You can leverage the name of the entity as well:
As a third and final step, once the component is all set, you can create an instance of the Dynatrace provider and then build your first studies:
Akamas supports the integration with virtually any telemetry and observability tool.
The following table describes the supported Telemetry Providers, which are created automatically at installation time.
collects metrics from CSV files
collects metrics from Dynatrace
collects metrics from Prometheus
collects metrics from Spark History Server
collects metrics from Tricentis Neoload Web
collects metrics from MicroFocus Load Runner Professional
collects metrics from MicroFocus Load Runner Enterprise
collects price metrics for Amazon Elastic Compute Cloud (ec2) from Amazon’s own APIs
Notice that Telemetry Providers are shared across all the workspaces within the same Akamas installation, and only users with administrative privileges can manage them.
The protocol used to connect to the remote machine: or
The path of the remote file(s) to be analyzed. The path can contains expressio
Must be specified using .
The URL of the Dynatrace installation API (see )
The Dynatrace API Token the provider should use to interact with Dynatrace. The token should have .
The NeoLoad Web provider collects metrics from a NeoLoad Web instance and makes them available to Akamas.
This section provides the minimum requirements that you should match before using the NeoLoad Web telemetry provider.
NeoLoad Web SaaS or managed version 7.1 or later.
The NeoLoad Web API must be reachable at a provided address and port (by default https://neoload-api.saas.neotys.com).
NeoLoad Web API access token.
Versions < 2.0.0 are compatibile with Akamas untill version 1.8.0
Versions >= 2.0.0 are compatible with Akamas from version 1.9.0
Web Application
You can check NeoLoadWeb provider metrics mapping to see how component-types metrics are extracted by this provider.
This section lists the workflow operators this provider depends on.
Akamas reasons in terms of a system to be optimized and in terms of parameters and metrics of components of that system. To understand which metrics collected from NeoloadWeb should refer to which component, the NeoloadWeb provider looks up the property neoloadweb
in the components of a system:
To create an instance of the Prometheus provider, edit a YAML file (instance.yml
in this example) with the definition of the instance:
Then you can create the instance for the system
using the Akamas CLI:
When you create an instance of the Prometheus provider, you should specify some configuration information to allow the provider to extract and process metrics from Prometheus correctly.
You can specify configuration information within the config
part of the YAML of the instance definition.
address
, a URL or IP identifying the address of the host where Prometheus is installed
port
, the port exposed by Prometheus
user
, the username for the Prometheus service
password
, the user password for the Prometheus service
job
, a string to specify the scraping job name. The default is ".*" for all scraping jobs
logLevel
, set this to "DETAILED" for some extra logs when searching for metrics (default value is "INFO")
headers
, to specify additional custom headers
e.g: headers:
"custom_key": "custom_value"
namespace
, a string to specify the namespace
duration
, integer to determine the duration in seconds for data collection (use a number between 1 and 3600)
enableHttps
, boolean to enable HTTPS in Prometheus (since 3.2.6)
ignoreCertificates
, boolean to ignore SSL certificates
disableConnectionCheck
, boolean to disable initial connection check to Prometheus
The Prometheus provider allows defining additional queries to populate custom metrics or redefine the default ones according to your use case. You can configure additional metrics using the metrics
field as shown in the configuration below:
In this example, the telemetry instance will populate cust_metric
with the results of the query specified in datasource
, maintaining the value of the labels listed under labels
.
Please refer to Querying basics | Prometheus for a complete reference of PromQL
Akamas pre-processes the queries before running them, replacing special-purpose placeholders with the fields provided in the components. For example, given the following component definition:
the query sum(jvm_memory_used_bytes{instance=~"$INSTANCE$", job=~"$JOB$"})
will be expanded for this component into sum(jvm_memory_used_bytes{instance=~"service01", job=~"jmx"})
. This provides greater flexibility through the templatization of the queries, allowing the same query to select the correct datasources for different components.
The following is the list of available placeholders:
Placeholder
Usage example
Component definition example
Expanded query
Description
$INSTANCE$
$JOB$
node_load1{instance=~"$INSTANCE$", job=~"$JOB$"}
node_load1{instance=~"frontend", job=~"node"}
These placeholders are replaced respectively with the instance
and job
fields configured in the component’s prometheus
configuration. These are the main fields used to filter the right datasources for this component/metric pair, and should always be defined in the label matchers of your queries.
%FILTERS%
container_memory_usage_bytes{job=~"$JOB$" %FILTERS%}
container_memory_usage_bytes{job=~"advisor", name=~"db-.*"}
This placeholder is replaced with a list containing any additional filter in the component’s definition (other than instance
and job
), where each field is expanded as field_name=~"field_value"
. This is useful to define additional label matches in the query without the need to hardcode them.
$DURATION$
rate(http_client_requests_seconds_count[$DURATION$])
rate(http_client_requests_seconds_count[30s])
This section reports common use cases addressed by this provider.
Check Java OpenJDK page for a list of all the Java metrics available in Akamas
You can leverage the Prometheus provider to collect Java metrics by using the JMX Exporter. The JMX Exporter is a collector of Java metrics for Prometheus that can be run as an agent for any Java application. Once downloaded, you execute it alongside a Java application with this command:
The command will expose on localhost on port 9100 Java metrics of youJar.jar
__ which can be scraped by Prometheus.
config.yaml
is a configuration file useful for the activity of this exporter. It is suggested to use this configuration for an optimal experience with the Prometheus provider:
As a next step, add a new scraping target in the configuration of the Prometheus used by the provider:
You can then create a YAML file with the definition of a telemetry instance (prom_instance.yml
) of the Prometheus provider:
And you can create the telemetry instance using the Akamas CLI:
Finally, to bind the extracted metrics to the related component, you should add the following field to the properties
of the component’s definition:
Check the Linux page for a list of all the system metrics available in Akamas
You can leverage the Prometheus provider to collect system metrics (Linux) by using the Node exporter. The Node exporter is a collector of system metrics for Prometheus that can be run as a standalone executable or a service within a Linux machine to be monitored. Once downloaded, schedule it as a service using, for example, systemd:
Here’s the manifest of the node_exporter
service:
The service will expose on localhost on port 9100 system metrics __ which can be scraped by Prometheus.
As a final step, add a new scraping target in the configuration of the Prometheus used by the provider:
You can then create a YAML file with the definition of a telemetry instance (prom_instance.yml
) of the Prometheus provider:
And you can create the telemetry instance using the Akamas CLI:
Finally, to bind the extracted metrics to the related component, you should add the following field to the properties
of the component’s definition:
To install the AWS provider, create a YAML file (called provider.yml
in this example) with the specification of the provider:
Then you can install the provider with the Akamas CLI:
The installed provider is shared with all users of your Akamas installation and can monitor many different systems, by configuring appropriate telemetry provider instances.
To install the LoadRunnerEnterprise provider, create a YAML file (called provider.yml
in this example) with the definition of the provider:
Then you can install the provider using the Akamas CLI:
The installed provider is shared with all users of your Akamas installation and can monitor many different systems, by configuring appropriate telemetry provider instances.
The Spark History Server provider collects metrics from a Spark History Server instance and makes them available to Akamas.
Prerequisites
This section provides the minimum requirements that you should match before using the Spark History Server telemetry provider.
Apache Spark 2.3
Spark History Server API must be reachable at the provided address and port (the default port is 18080
).
spark-application
You can check to see how component-type metrics are extracted by this provider.
Versions < 2.0.0 are compatible with Akamas until version 1.8.0
Versions >= 2.0.0 are compatible with Akamas from version 1.9.0
This section lists the workflow operators this provider depends on:
Akamas uses components to identify specific elements of the system to be monitored and optimized. Your system might contain multiple components to model, for example, a Spark application and each host of the cluster. To point Akamas to the right component when extracting metrics you need to add a property called sparkApplication
to your Spark Application component. The provider will only extract metrics for components for which this property has been specified.
To create an instance of the LoadRunner provider, build a YAML file (instance.yml
in this example) with the definition of the instance:
Then you can create the instance for the system
using the Akamas CLI:
When you create an instance of the LoadRunner provider, you should specify some configuration information to allow the provider to correctly extract and process metrics from LoadRunner.
You can specify configuration information within the config
part of the YAML of the instance definition.
hostname
: The hostname or IP of the server that hosts the CIFS share where the LoadRunner results have been exported. See .
username
: The username and the domain required to access the network share. Supported formats are:
username
domain\username
username@domain
password
: The password required to access the network share
shareName
: The name of the network share as it is exposed by the server
The following YAML file provides the reference for the definition of a telemetry instance.
The following table describes the reference for the config
section within the definition of the LoadRunner provider instance:
To create an instance of the LoadRunnerEnterprise provider, build a YAML file (instance.yml
in this example) with the definition of the instance:
Then you can create the instance using the Akamas CLI:
When you create an instance of the LoadRunnerEnterprise provider, you should specify some configuration information to allow the provider to correctly extract and process metrics from Prometheus.
You can specify configuration information within the config
part of the YAML of the instance definition.
address
: The address of the InfluxDB instance, in the form of schema://address (i.e https://influxdb.mycompay.com)
port
: The InfluxDB port
username
: The username required to connect to InfluxDB.
password
: The password for the username
database
: The database name where LoadRunner metrics are stored
This table reports the reference for the config
section within the definition of the LoadRunnerEnterprise provider instance:
See below
See below
If not set in the component properties, this placeholder is replaced with the duration field configured in the telemety-instance. You should use it with instead of hardcoding a fixed value.
hostname
String
the hostname or IP of the server that hosts the CIFS share where the LoadRunner results have been exported. See LoadRunner operator
-
IP address or FQDN
Yes
username
String
The username and the domain required to access the share
-
Supported formats:
username
domain\username
username@domain
Yes
password
String
The password required to access the share
-
-
Yes
shareName
String
The name of the share as it is exposed by the server
-
-
Yes
pathPrefix
String
A prefix for the default path where the provider looks for the data.
The default path is {studyName}{experimentId}{trialId}
-
A valid Windows path
No
address
String
The address of the InfluxDB
-
A valid URL
Yes
port
Integer
The port of the InfluxDB
-
A valid port
yes
username
String
The username to connect to InfluxDB
-
-
Yes
password
String
The password of the specified user
-
-
Yes
database
String
The database with the metrics
-
-
Yes
verifySSL
Boolean
Wheter to check the certificate, of InfluxDB API is exposed in HTTPS
false
true/false
No
transactions
List of string
Restrict metrics collection to the listed transactions names. If not specified the provider will collect metrics of all transactions
Empty