Skip to content

Commit

Permalink
→ AWS, +Prometheus +Grafana, +doc, versions++ (#340)
Browse files Browse the repository at this point in the history
Install NATS metrics exporters

NATS exporter, format data disks

disks, monitoring

Mount NATS dashboards

streaming vs JetStream

Typo

Start exporter after NATS

Correct docker usage

Correct working directory

Unambiguous name

Include NATs, reference STAN

Fix scrape port

Formatting

License
  • Loading branch information
teabot authored Oct 18, 2022
1 parent 07895af commit eb7d284
Show file tree
Hide file tree
Showing 20 changed files with 879 additions and 50 deletions.
3 changes: 2 additions & 1 deletion README.md
Original file line number Diff line number Diff line change
Expand Up @@ -14,7 +14,8 @@ This repository houses user-friendly, cloud-ready benchmarking suites for the fo
* [Apache RocketMQ](https://rocketmq.apache.org)
* Generic [JMS](https://javaee.github.io/jms-spec/)
* [KoP (Kafka-on-Pulsar)](https://github.com/streamnative/kop)
* [NATS Streaming](https://nats.io/)
* [NATS](https://docs.nats.io/)
* [NATS Streaming (STAN)](https://docs.nats.io/legacy/stan)
* [NSQ](https://nsq.io)
* [Pravega](https://pravega.io/)
* [RabbitMQ](https://www.rabbitmq.com/)
Expand Down
151 changes: 151 additions & 0 deletions driver-nats-streaming/README.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,151 @@
# NATS Streaming benchmarks

> **Warning**
> This module relates to [NATS Streaming / STAN](https://docs.nats.io/legacy/stan) which has been deprecated
> and replaced by [NATS JetStream](https://docs.nats.io/nats-concepts/jetstream). We do not yet have a benchmark
> driver for JetStream.
This folder houses the assets necessary to run benchmarks for
[NATS Streaming](https://docs.nats.io/legacy/stan). In order to run these benchmarks, you'll need to:

* [Create the necessary local artifacts](#creating-local-artifacts)
* [Stand up a NATS cluster](#creating-a-nats-cluster-on-amazon-web-services-aws-using-terraform-and-ansible)
on Amazon Web Services (which includes a client host for running the benchmarks)
* [SSH into the client host](#sshing-into-the-client-host)
* [Run the benchmarks from the client host](#running-the-benchmarks-from-the-client-host)

## Creating local artifacts

In order to create the local artifacts necessary to run the NATS benchmarks in AWS, you'll need to have
[Maven](https://maven.apache.org/install.html) installed. Once Maven's installed, you can create the necessary
artifacts with a single Maven command:

```bash
$ git clone https://github.com/openmessaging/benchmark.git
% cd messaging-benchmark
$ mvn install
```

## Creating a NATS cluster on Amazon Web Services (AWS) using Terraform and Ansible

In order to create an NATS cluster on AWS, you'll need to have the following installed:

* [Terraform](https://terraform.io)
* [The `terraform-inventory` plugin for Terraform](https://github.com/adammck/terraform-inventory)
* [Ansible](http://docs.ansible.com/ansible/latest/intro_installation.html)

In addition, you will need to:

* [Create an AWS account](https://aws.amazon.com/account/) (or use an existing account)
* [Install the `aws` CLI tool](https://aws.amazon.com/cli/)
* [Configure the `aws` CLI tool](http://docs.aws.amazon.com/cli/latest/userguide/cli-chap-getting-started.html)

Once those conditions are in place, you'll need to create an SSH public and private key at `~/.ssh/nats_streaming_aws`
(private) and `~/.ssh/nats_streaming_aws.pub` (public), respectively.

```bash
$ ssh-keygen -f ~/.ssh/nats_streaming_aws
```

When prompted to enter a passphrase, simply hit **Enter** twice. Then, make sure that the keys have been created:

```bash
$ ls ~/.ssh/nats_streaming_aws*
```

With SSH keys in place, you can create the necessary AWS resources using a single Terraform command:

```bash
$ cd driver-nats-streaming/deploy
$ terraform init
$ terraform apply
```

That will install the following [EC2](https://aws.amazon.com/ec2) instances (plus some other resources, such as a
[Virtual Private Cloud](https://aws.amazon.com/vpc/) (VPC)):

| Resource | Description | Count |
|:--------------------|:------------------------------------------------------------|:------|
| NATS instances | The VMs on which NATS brokers will run | 3 |
| Client instance | The VM from which the benchmarking suite itself will be run | 4 |
| Prometheus instance | The VM on which metrics services will be run | 1 |

When you run `terraform apply`, you will be prompted to type `yes`. Type `yes` to continue with the installation or
anything else to quit.

Once the installation is complete, you will see a confirmation message listing the resources that have been installed.

### Variables

There's a handful of configurable parameters related to the Terraform deployment that you can alter by modifying the
defaults in the `terraform.tfvars` file.

| Variable | Description | Default |
|:------------------|:------------------------------------------------------------------------------------------------------------------------------------|:-----------------------------------------------------------------|
| `region` | The AWS region in which the NATS cluster will be deployed | `us-west-2` |
| `az` | The availability zone in which the NATS cluster will be deployed | `us-west-2a` |
| `public_key_path` | The path to the SSH public key that you've generated | `~/.ssh/rabbitmq_aws.pub` |
| `ami` | The [Amazon Machine Image](http://docs.aws.amazon.com/AWSEC2/latest/UserGuide/AMIs.html) (AWI) to be used by the cluster's machines | [`ami-9fa343e7`](https://access.redhat.com/articles/3135091) |
| `instance_types` | The EC2 instance types used by the various components | `i3.4xlarge` (NATS brokers), `c4.8xlarge` (benchmarking clients) |

> If you modify the `public_key_path`, make sure that you point to the appropriate SSH key path when running the
> [Ansible playbook](#running-the-ansible-playbook).
### Running the Ansible playbook

With the appropriate infrastructure in place, you can install and start the RabbitMQ cluster using Ansible with just
one command. Note that a `TFSTATE` environment must point to the folder in which the `tf.state` file is located.

```bash
$ TF_STATE=. ansible-playbook \
--user ec2-user \
--inventory `which terraform-inventory` \
deploy.yaml
```

> If you're using an SSH private key path different from `~/.ssh/nats_streaming_aws`, you can specify that path using
> the `--private-key` flag, for example `--private-key=~/.ssh/my_key`.
## SSHing into the client host

In the [output](https://www.terraform.io/intro/getting-started/outputs.html) produced by Terraform, there's a
`client_ssh_host` variable that provides the IP address for the client EC2 host from which benchmarks can be run.
You can SSH into that host using this command:

```bash
$ ssh -i ~/.ssh/nats_streaming_aws ec2-user@$(terraform output client_ssh_host)
```

## Running the benchmarks from the client host

Once you've successfully SSHed into the client host, you can run all
[available benchmark workloads](../#benchmarking-workloads) like this:

```bash
$ cd /opt/benchmark
$ sudo bin/benchmark --drivers driver-nats-streaming/nats-streaming.yaml workloads/*.yaml
```

You can also run specific workloads in the `workloads` folder. Here's an example:

```bash
$ sudo bin/benchmark --drivers driver-nats-streaming/nats-streaming.yaml workloads/1-topic-1-partitions-1kb.yaml
```

## Monitoring

### Prometheus

The [`prometheus-nats-exporter`](https://github.com/nats-io/prometheus-nats-exporter)
service is installed and Prometheus is installed on a standalone instance, along with
[Node Exporter](https://github.com/prometheus/node_exporter) on all brokers to allow the collection of system metrics.
Prometheus exposes a public endpoint `http://${prometheus_host}:9090`.

### Grafana

Grafana and a set of standard dashboards are installed alongside Prometheus. These are exposed on a public endpoint
`http://${prometheus_host}:3000`. Credentials are `admin`/`admin`. Dashboards included:
* [NATS dashboards](https://github.com/nats-io/prometheus-nats-exporter/blob/main/walkthrough/README.md)
* [NATS streaming dashboard](https://github.com/nats-io/prometheus-nats-exporter/blob/main/walkthrough/streaming.md)
* [Node Exporter dashboard](https://grafana.com/grafana/dashboards/1860-node-exporter-full/)

3 changes: 3 additions & 0 deletions driver-nats-streaming/deploy/ali/ansible.cfg
Original file line number Diff line number Diff line change
@@ -0,0 +1,3 @@
[defaults]
host_key_checking=false
private_key_file=alicloud.pem
15 changes: 15 additions & 0 deletions driver-nats-streaming/deploy/ali/terraform.tfvars
Original file line number Diff line number Diff line change
@@ -0,0 +1,15 @@
region = "cn-shenzhen"
availability_zone = "cn-shenzhen-c"
private_key_file = "alicloud.pem"
key_name = "key-pair-from-terraform-nats-streaming"
image_id = "centos_7_04_64_20G_alibase_201701015.vhd"

instance_types = {
"nats-streaming-server" = "ecs.se1.4xlarge" #4c16g
"client" = "ecs.n4.4xlarge"
}

num_instances = {
"nats-streaming-server" = 3
"client" = 2
}
7 changes: 6 additions & 1 deletion driver-nats-streaming/deploy/ansible.cfg
Original file line number Diff line number Diff line change
@@ -1,3 +1,8 @@
[defaults]
host_key_checking=false
private_key_file=alicloud.pem
private_key_file=~/.ssh/nats_streaming_aws

[privilege_escalation]
become=true
become_method='sudo'
become_user='root'
Loading

0 comments on commit eb7d284

Please sign in to comment.