1
0
Fork 0
You cannot select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.

86 lines
2.6 KiB
Python

#!/usr/bin/env python
#
# End-to-end testing using actual rqlited binary.
#
# To run a specific test, execute
#
# python system_test/full_system_test.py Class.test
import os
import json
import unittest
import sqlite3
import time
from helpers import Node, deprovision_node, write_random_file, random_string, env_present
from s3 import download_s3_object, delete_s3_object
S3_BUCKET = 'rqlite-testing-circleci'
S3_BUCKET_REGION = 'us-west-2'
RQLITED_PATH = os.environ['RQLITED_PATH']
class TestAutoBackupS3(unittest.TestCase):
@unittest.skipUnless(env_present('RQLITE_S3_ACCESS_KEY'), "S3 credentials not available")
def test(self):
'''Test that automatic backups to AWS S3 work'''
self.node = None
self.cfg = None
self.path = None
self.backup_file = None
self.access_key_id = os.environ['RQLITE_S3_ACCESS_KEY']
self.secret_access_key_id = os.environ['RQLITE_S3_SECRET_ACCESS_KEY']
# Create the auto-backup config file
self.path = random_string(32)
auto_backup_cfg = {
"version": 1,
"type": "s3",
"interval": "1s",
"sub" : {
"access_key_id": self.access_key_id,
"secret_access_key": self.secret_access_key_id,
"region": S3_BUCKET_REGION,
"bucket": S3_BUCKET,
"path": self.path
}
}
self.cfg = write_random_file(json.dumps(auto_backup_cfg))
# Create a node, enable automatic backups, and start it. Then
# create a table and insert a row. Wait for a backup to happen.
self.node = Node(RQLITED_PATH, '0', auto_backup=self.cfg)
self.node.start()
self.node.wait_for_leader()
self.node.execute('CREATE TABLE foo (id INTEGER NOT NULL PRIMARY KEY, name TEXT)')
self.node.execute('INSERT INTO foo(name) VALUES("fiona")')
self.node.wait_for_all_fsm()
time.sleep(5)
# Download the backup file from S3 and check it.
backup_data = download_s3_object(self.access_key_id, self.secret_access_key_id,
S3_BUCKET, self.path)
self.backup_file = write_random_file(backup_data, mode='wb')
conn = sqlite3.connect(self.backup_file)
c = conn.cursor()
c.execute('SELECT * FROM foo')
rows = c.fetchall()
self.assertEqual(len(rows), 1)
self.assertEqual(rows[0][1], 'fiona')
conn.close()
def tearDown(self):
if self.node is not None:
deprovision_node(self.node)
if self.cfg is not None:
os.remove(self.cfg)
if self.backup_file is not None:
os.remove(self.backup_file)
delete_s3_object(self.access_key_id, self.secret_access_key_id,
S3_BUCKET, self.path)
if __name__ == "__main__":
unittest.main(verbosity=2)