In this guide, you will learn how to deploy an ExtraHop Explore virtual appliance on a Linux kernel-based virtual machine (KVM) and to join multiple Explore appliances to create an Explore cluster. You should be familiar with basic KVM administration before proceeding.
|Important:||If you want to deploy more than one ExtraHop virtual appliance, create the new instance with the original deployment package or clone an existing instance that has never been started.|
Your environment must meet the following requirements to deploy a virtual Explore appliance:
|Important:||ExtraHop tests virtual Explore clusters on local storage for optimal performance. ExtraHop strongly recommends deploying virtual Explore clusters on continuously available, low latency storage, such as a local disk, direct-attached storage (DAS), network-attached storage (NAS), or storage area network (SAN).|
- A KVM hypervisor environment capable of hosting the Explore virtual appliance. The
Explore virtual appliance is available in the following configurations:
EXA Manager Node EXA-XS EXA-S EXA-M EXA-L 4 CPUs 4 CPUs 8 CPUs 16 CPUs 32 CPUs 8 GB RAM 8 GB RAM 16 GB RAM 32 GB RAM 64 GB RAM 4 GB boot disk 4 GB boot disk 4 GB boot disk 4 GB boot disk 4 GB boot disk 12 GB 250 GB or smaller datastore disk 500 GB or smaller datastore disk 1 TB or smaller datastore disk 2 TB or smaller datastore disk
The hypervisor CPU should provide Supplemental Streaming SIMD Extensions 3 (SSSE3) support.
Note: The Explore manager node is preconfigured with a 12 GB datastore disk. You must manually configure a second virtual disk to the other EXA configurations to store record data.
Consult with your ExtraHop sales representative or Technical Support to determine the datastore disk size that is best for your needs.
Note: For KVM deployments, virtio-scsi interface is recommended for the boot and datastore disks.
- An Explore virtual appliance license key.
- The following TCP ports must be open:
- TCP port 443: Enables browser access to Administration settings. Requests sent to port 80 are automatically redirected to HTTPS port 443.
- TCP port 9443: Enables Explore nodes to communicate with other Explore nodes in the same cluster.
The installation package for KVM systems is a tar.gz file that contains the following items:
- The domain XML configuration file
- The domain XML checksum file
- The boot disk
- The boot disk checksum file
To deploy the Explore virtual appliance, complete the following procedures:
Identify the bridge through which you will access the management interface of your Explore appliance.
- Make sure the management bridge is accessible to the Explore virtual appliance and to all users who must access the management interface.
- If you need to access the management interface from an external computer, configure a physical interface on the management bridge.
After you identify the management bridge, edit the configuration file, and create the Explore virtual appliance.
- Contact ExtraHop Support (firstname.lastname@example.org) to obtain and download the Explore KVM package.
- Extract the tar.gz file that contains the installation package.
- Copy the extrahop-boot.qcow2 file to your KVM system.
Open the domain XML configuration file in a text editor and edit the following
Change the VM name to a name for your ExtraHop virtual appliance.
Change the source file path ([PATH_TO_STORAGE]) to the
location where you stored the virtual disk file in step 3.
Change the source bridge for the management network
(ovsbr0) to match the name of your management
<interface type='bridge'> <source bridge='ovsbr0'/> <model type='virtio'/> <alias name='net0'/> <address type='pci' domain='0x0000' bus='0x00' slot='0x03' function='0x0'/> </interface>
If your virtual bridge is configured through Open vSwitch virtual
switch software, add the following virtualport type setting to the
interface (after the source bridge setting):
<virtualport type='openvswitch'> </virtualport>
- Change the VM name to a name for your ExtraHop virtual appliance.
- Save the XML file.
Create the new Explore virtual appliance with your revised domain XML
configuration file by running the following command:
virsh define <EXA-5100v-<x>.xml>Where <EXA-5100v-<x>.xml> is the name of your domain XML configuration file.
Create the datastore disk so that the allotted space is large enough to store the type of records you want to store for the amount of lookback desired.
Start the VM by running the following command:
virsh start <vm_name>Where <vm_name> is the name of your ExtraHop virtual appliance you configured in step 4 of the Edit the domain XML file section.
Log in to the KVM console and view the IP address for your new ExtraHop virtual
appliance by running the following command:
virsh console <vm_name>
By default, ExtraHop systems ship with DHCP enabled. If your network does not support DHCP, you must configure a static address manually.
- Log in to the KVM host.
Run the following command to connect to the ExtraHop system through the
virtual serial console:
virsh console <vm_name>
Where <vm_name> is the name of your virtual machine.
Press ENTER twice to get to the system login prompt.
ExtraHop Discover Appliance Version 220.127.116.116 IP: 192.0.2.81 exampleium login:
- At the login prompt, type shell, and then press ENTER.
- At the password prompt, type default, and then press ENTER.
To configure the static IP address, run the following commands:
Enable privileged commands:
- At the password prompt, type default, and then press ENTER.
Enter configuration mode:
Enter the interface configuration mode:
Run the ip command and specify the IP address and
DNS settings in the following format: ip ipaddr <ip_address>
<netmask> <gateway> <dns_server>
ip ipaddr 10.10.2.14 255.255.0.0 10.10.1.253 10.10.1.254
Leave the interface configuration section:
Save the running config file:
- Type y and then press ENTER.
- Enable privileged commands:
After you obtain the IP address for the Explore appliance, log in to the appliance through https://<explore_ip_address>/admin and complete the following recommended procedures.
|Note:||The default login username is setup and the password is default.|
- Register your ExtraHop system
- Connect the Discover and Command appliances to Explore appliances
- Send record data to the Explore appliance
- Review the Explore Post-deployment Checklist and configure additional Explore appliance settings.
For the best performance, data redundancy, and stability, you must configure at least three Explore appliances in an Explore cluster.
|Important:||If you are creating an Explore cluster with six or more nodes, you must configure the cluster with manager nodes. For manager node instructions, see Deploying manager nodes.|
In the following example, the Explore appliances have the following IP addresses:
- Node 1: 10.20.227.177
- Node 2: 10.20.227.178
- Node 3: 10.20.227.179
You will join nodes 2 and 3 to node 1 to create the Explore cluster. All three nodes are data nodes. You cannot join a data node to a manager node or join a manager node to a data node to create a cluster.
|Important:||Each node that you join must have the same configuration (physical or virtual) and the same ExtraHop firmware version. EXA 5100 and EXA 5200 physical appliances can be in the same cluster.|
Before you beginYou must have already installed or provisioned the Explore appliances in your environment before proceeding.
- Log in to the Administration settings on all three Explore appliances with the setup user account in three separate browser windows or tabs.
- Select the browser window of node 1.
- In the Status and Diagnostics section, click Fingerprint and note the fingerprint value. You will later confirm that the fingerprint for node 1 matches when you join the remaining two nodes.
- Select the browser window of node 2.
- In the Explore Cluster Settings section, click Join Cluster.
In the Host field, type the hostname or IP address of data
node 1 and then click Continue.
Note: For cloud-based deployments, be sure to type the IP address listed in the Interfaces table on the Connectivity page.
Confirm that the fingerprint on this page matches the fingerprint you noted in
In the Setup Password field, type the password for the
node 1 setup user account and then click
When the join is complete, the Explore Cluster Settings section has two new entries: Cluster Members and Cluster Data Management.
Click Cluster Members. You should see node 1 and node 2 in
- In the Status and Diagnostics section, click Explore Cluster Status. Wait for the Status field to change to Green before adding the next node.
Repeat steps 5 - 10 to join each additional node to the new cluster.
Note: To avoid creating multiple clusters, always join a new node to an existing cluster and not to another single appliance.
When you have added all of your Explore appliances to the cluster, click
Cluster Members in the Explore
Cluster Settings section. You should see all of the joined nodes
in the list, similar to the following figure.
- In the Explore Cluster Settings section, click Cluster Data Management and make sure that Replication Level is set to 1 and Shard Reallocation is ON.
After you deploy the Explore appliance, you must establish a connection from all ExtraHop Discover and Command appliances to the Explore appliance before you can query records.
|Important:||Connect the Discover appliance to each Explore node so that the Discover appliance can distribute the workload across the entire Explore cluster.|
|Note:||If you manage all of your Discover appliances from a Command appliance, you only need to perform this procedure from the Command appliance.|
- Log in to the Administration settings on the ExtraHop system through https://<extrahop-hostname-or-IP-address>/admin.
- In the ExtraHop Explore Settings section, click Connect Explore Appliances.
- Click Add New.
- In the Explore node field, type the hostname or IP address of any Explore appliance in the Explore cluster.
For each additional Explore appliance in the
cluster, click Add New and enter the individual hostname
or IP address in the corresponding Explore node
- Click Save.
- Confirm that the fingerprint on this page matches the fingerprint of node 1 of the Explore cluster.
- In the Explore Setup Password field, type the password for the Explore node 1 setup user account and then click Connect.
- When the Explore Cluster settings are saved, click Done.
After your Explore appliance is connected to all of your Discover and Command appliances, you must configure the type of records you want to store.