Create Remote Archive Volume

You need a remote archive volume to store your backups. During the installation if you did not create the remote archive volume, you can create it from EXAoperation later. Exasol recommends using S3 bucket to store your backups for Exasol on AWS.

This section explains how you can use an S3 bucket to create a remote archive volume and store your backup.

Remote Archive Volume Options

You can use the following volume options (depending on the protocol):

  • cleanvolume: Delete remote backups on expiry
  • noverifypeer: Do not check server certificate
  • nocompression: Write plain data
  • forcessl: Use STARTTLS in FTP connection
  • webdav: Use WebDAV for http-URL
  • webhdfs: For WebHDFS URLs
  • s3: For remote archive volumes on Amazon S3
  • timeout=seconds: Allows higher client/server response time

Prerequisites

  • All nodes must be able to reach the remote target.
  • If DNS names will be used, ensure the cluster has DNS servers configured.
  • An existing S3 bucket. The URL must be in the following format:

    • http://bucketname.s3.amazonaws.com/optional-directory/

    • https://bucketname.s3.amazonaws.com/optional-directory/

    If you do not have a bucket, see How Do I Create an S3 Bucket?.

  • Read-write access to the S3 bucket. In case the nodes are on a private network, make sure that an S3 endpoint is configured for your VPC and that the route table for your subnet is updated accordingly to store backups in the bucket. For more information, see Endpoints for Amazon S3.
  • A secret access key for the S3 bucket. If you do not have a key, see Managing Access Keys for IAM Users.

An Amazon S3 bucket URL using the legacy global endpoint format (<bucket-name>.s3.amazonaws.com) may need up to 24 hours after bucket creation before it becomes available.

A fully qualified Amazon S3 bucket URL that includes the AWS region (<bucket-name>.s3.<region-code>.amazonaws.com) will become available immediately.

Configure Remote Archive Volume in EXAoperation

Do the following to configure the remote archive volume:

  1. Log in into EXAoperation.
  2. Go to Services > EXAStorage, click Add Remote Volume.
  3. Enter the following details:
    • Archive URL: Use the URL from the S3 bucket that you created in AWS Console.
    • User: If you have configured EC2 instance role, leave this field empty. If not, enter the aws_access_key_id here.
    • Password: If you have configured EC2 instance role, leave this field empty. If not, enter the aws_secret_access_key here.
    • Allowed Users: Select the EXAoperation users that can use the remote archive volume.
    • Read-only Users: Select the EXAoperation users that can read from the remote archive volume.

      Any user that you add as Read-only User, is limited to read-only access even if it's added to the Allowed Users list. If you have only one user that's added into list of allowed user, make sure you don't add the user in the list of Read-only Users.

    • Labels: Add label to the remote archive volume for an easy identification.
    • Options:You can use one of the options mentioned in section Remote Archive Volume Options. For example, cleanvolume option to clean expired backups.
  4. Click Add.

 

Other Remote Archive Volume Options

The port numbers used in this table are examples, most protocols allow you to set the port.

Protocol Archive URL Example
FTP

ftp://<ftpserver>:2021/optional-directory/

FTPS ftps://<ftpserver>:2021/optional-directory/
SMB smb:////<smbserver>:2139/optional-directory/
WebHDFS

http://<hadoop-server>:2080/optional-directory/

https://<hadoop-server>:20443/optional-directory/

Azure Blob Storage http://storage_container_name.blob.corewindows.net/container_name

https://storage_container_name.blob.corewindows.net/container_name

For more information, see Create Remote Archive Volume with Azure Blob Storage.

Google Cloud Storage

http://<bucketname>.storage.googleapis.com

https://<bucketname>.storage.googleapis.com

For more information, see Create Remote Archive Volume with Google Cloud Storage .