If you clone a Cassandra VM with the goal of creating a cassandra cluster – you may find that every Cassandra node has the same hostID.Continue reading
Creating a mixed read/write workload with fio can be a bit confusing. Assume we want to create a fixed rate workload of 100 IOPS split 70:30 between reads and writes.
Specify the rate directly with rate_iops=<read-rate>,<write-rate> do not try to use rwmixread with rate_iops. For the example above use.
Additionally older versions of fio exhibit problems when using rate_poisson with rate_iops . fio version 3.7 that I was using did not exhibit the problem.Continue reading
The parameters norandommap and randrepeat significantly change the way that repeated random IO workloads will be executed, and also can meaningfully change the results of an experiment due to the way that caching works on most storage system.Continue reading
From the SQL Window of SQL*Server. Issue these commands to drop the tables and procedures created by HammerDB. This will allow you (for instance) to re-create the database, or create a new database with more warehouses (larger size) while retaining the same name/DB layout.Continue reading
How to use the “jobs” and “clients” parameters in pgbench without going crazy.Continue reading
How to speed up your X-ray benchmark development cycle by re-using/re-cycling benchmark VMs and more importantly data-sets.Continue reading
I have VMs running on bare-metal instances. Each bare-metal instance is in a separate rack by design (for fault tolerance). The bandwidth is 25GbE however, the response time between the hosts is so high that I need multiple streams to consume that bandwidth.
Compared to my local on-prem lab I need many more streams to get the observed throughput close to the theoretical bandwidth of 25GbE
|# iperf Streams||AWS Throughput||On-Prem Throughput|
|1||4.8 Gbit||21.4 Gbit|
|2||9 Gbit||22 Gbit|
|8||23 Gbit||23 Gbit|
End to End Creation of a Nutanix Cluster on AWS and Running X-RayContinue reading
Scale factor to workingset size lookup for tiny databasesContinue reading
A series of videos showing how to install, run, modify and analyze HCI clusters with the Nutanix X-ray toolContinue reading
How to identify optane drives in linux OS using lspci.Continue reading
Use the following SQL to drop the tables and indexes in the HammerDB TPC-H schema, so that you can re-load it.Continue reading
Tips and tricks for using diskspd especially useful for those familar with tools like fioContinue reading
How to ensure performance testing with diskspd is stressing the underlying storage devices, not the OS filesystem.Continue reading
How to install and setup diskspd before starting your first performance tests and avoiding wrong results due to null byte issues.Continue reading
How can database density be measured?
- How does database performance behave as more DBs are consolidated?
- What impact does running the CVM have on available host resources?
- The cluster was able to achieve ~90% of the theoretical maximum.
- CVM overhead was 5% for this workload.
The goal was to establish how database performance is affected as additional database workloads are added into the cluster. As a secondary metric, measure the overhead from running the virtual storage controller on the same host as the database servers themselves. We use the Postgres database with pgbench workload and measure the total transactions per second.
- 4 Node Nutanix cluster, with 2x Xeon CPU’s per host with 20 cores per socket.
Each database is identically configured with
- Postgres 9.3
- Ubuntu Linux
- 4 vCPU
- 8GB of memory
- pgbench benchmark, running the “simple” query set.
The database is sized so that it fits entirely in memory. This is a test of CPU/Memory not IO.
The experiment starts with a single Database on a single host. We add more databases into the cluster until we reach 40 databases in total. At 40 databases with 4 vCPU each and a CPU bound workload we use all 160 CPU cores on the cluster.
The database is configured to fit into the host DRAM memory, and the benchmark runs as fast as it can – the benchmark is CPU bound.
Below are the measured results from running 1-40 databases on the 4 node cluster.
Performance scales almost linearly from 4 to 160 CPU with no obvious bottlenecks before all of the CPU cores are saturated in the host at 40 databases.Continue reading
Many storage performance testers are familiar with vdbench, and wish to use it to test Hyper-Converged (HCI) performance. To accurately performance test HCI you need to deploy workloads on all HCI nodes. However, deploying multiple VMs and coordinating vdbench can be tricky, so with X-ray we provide an easy way to run vdbench at scale. Here’s how to do it.Continue reading