diff --git a/.github/workflows/PBM-FULL.yml b/.github/workflows/PBM-FULL.yml index f07f4510..c8364e89 100644 --- a/.github/workflows/PBM-FULL.yml +++ b/.github/workflows/PBM-FULL.yml @@ -19,7 +19,7 @@ on: jobs: test: runs-on: ubuntu-latest - timeout-minutes: 120 + timeout-minutes: 180 strategy: fail-fast: false matrix: diff --git a/pbm-functional/pytest/cluster.py b/pbm-functional/pytest/cluster.py index f359dc0c..e5503b5c 100644 --- a/pbm-functional/pytest/cluster.py +++ b/pbm-functional/pytest/cluster.py @@ -490,10 +490,19 @@ def destroy(self,**kwargs): print("\n") cleanup=kwargs.get('cleanup_backups', False) if cleanup: + timeout = time.time() + 30 + self.disable_pitr() result=self.exec_pbm_cli("delete-pitr --all --force --yes ") Cluster.log(result.stdout + result.stderr) + while True: + if not self.get_status()['running'] or time.time() > timeout: + break result=self.exec_pbm_cli("delete-backup --older-than=0d --force --yes") Cluster.log(result.stdout + result.stderr) + while True: + if not self.get_status()['running'] or time.time() > timeout: + break + for host in self.all_hosts: try: container = docker.from_env().containers.get(host) diff --git a/pbm-functional/pytest/test_PBM-1252.py b/pbm-functional/pytest/test_PBM-1252.py index 3b4975c9..21fdb5ba 100644 --- a/pbm-functional/pytest/test_PBM-1252.py +++ b/pbm-functional/pytest/test_PBM-1252.py @@ -43,7 +43,7 @@ def start_cluster(cluster,request): cluster.destroy(cleanup_backups=True) @pytest.mark.timeout(3600,func_only=True) -def test_load(start_cluster,cluster): +def test_load_PBM_T250(start_cluster,cluster): cluster.check_pbm_status() indexes = [] diff --git a/pbm-functional/pytest/test_PBM-1344.py b/pbm-functional/pytest/test_PBM-1344.py new file mode 100644 index 00000000..1ade8276 --- /dev/null +++ b/pbm-functional/pytest/test_PBM-1344.py @@ -0,0 +1,116 @@ +import pytest +import pymongo +import bson +import testinfra +import time +import os +import docker +import threading + +from datetime import datetime +from cluster import Cluster + +documents = [{"a": 1}, {"b": 2}, {"c": 3}, {"d": 4}] + + +@pytest.fixture(scope="package") +def docker_client(): + return docker.from_env() + + +@pytest.fixture(scope="package") +def config(): + return { + "_id": "rs1", + "members": [{"host": "rs101"}, {"host": "rs102"}, {"host": "rs103"}], + } + + +@pytest.fixture(scope="package") +def cluster(config): + return Cluster(config) + + +@pytest.fixture(scope="function") +def start_cluster(cluster, request): + try: + cluster.destroy() + cluster.create() + cluster.setup_pbm() + os.chmod("/backups", 0o777) + os.system("rm -rf /backups/*") + yield True + finally: + if request.config.getoption("--verbose"): + cluster.get_logs() + cluster.destroy(cleanup_backups=True) + + +@pytest.mark.timeout(600, func_only=True) +def test_physical_PBM_T279(start_cluster, cluster): + cluster.check_pbm_status() + client = pymongo.MongoClient(cluster.connection) + backup = cluster.make_backup("physical") + cluster.enable_pitr(pitr_extra_args="--set pitr.oplogSpanMin=0.5") + for i in range(10): + pymongo.MongoClient(cluster.connection)["test"]["test"].insert_one({"doc": i}) + cluster.disable_pitr() + time.sleep(10) + cluster.delete_backup(backup) + cluster.destroy() + + cluster.create() + cluster.setup_pbm() + time.sleep(10) + cluster.check_pbm_status() + backup = cluster.make_backup("physical") + cluster.enable_pitr(pitr_extra_args="--set pitr.oplogSpanMin=0.5") + for i in range(10): + pymongo.MongoClient(cluster.connection)["test"]["test"].insert_one({"doc": i}) + time.sleep(5) + pitr = datetime.utcnow().strftime("%Y-%m-%dT%H:%M:%S") + backup = " --time=" + pitr + Cluster.log("Time for PITR is: " + pitr) + cluster.disable_pitr() + time.sleep(10) + cluster.make_restore(backup, restart_cluster=True, check_pbm_status=True) + assert ( + pymongo.MongoClient(cluster.connection)["test"]["test"].count_documents({}) + == 10 + ) + Cluster.log("Finished successfully") + + +@pytest.mark.timeout(300, func_only=True) +def test_logical_PBM_T280(start_cluster, cluster): + cluster.check_pbm_status() + client = pymongo.MongoClient(cluster.connection) + backup = cluster.make_backup("logical") + cluster.enable_pitr(pitr_extra_args="--set pitr.oplogSpanMin=0.5") + for i in range(10): + pymongo.MongoClient(cluster.connection)["test"]["test"].insert_one({"doc": i}) + cluster.disable_pitr() + time.sleep(10) + cluster.delete_backup(backup) + cluster.destroy() + + cluster.create() + cluster.setup_pbm() + time.sleep(10) + cluster.check_pbm_status() + backup = cluster.make_backup("logical") + cluster.enable_pitr(pitr_extra_args="--set pitr.oplogSpanMin=0.5") + for i in range(10): + pymongo.MongoClient(cluster.connection)["test"]["test"].insert_one({"doc": i}) + time.sleep(5) + pitr = datetime.utcnow().strftime("%Y-%m-%dT%H:%M:%S") + backup = " --time=" + pitr + Cluster.log("Time for PITR is: " + pitr) + cluster.disable_pitr() + time.sleep(10) + cluster.make_restore(backup, check_pbm_status=True) + assert ( + pymongo.MongoClient(cluster.connection)["test"]["test"].count_documents({}) + == 10 + ) + Cluster.log("Finished successfully") diff --git a/pbm-functional/pytest/test_PBM-1391.py b/pbm-functional/pytest/test_PBM-1391.py new file mode 100644 index 00000000..b82c2f12 --- /dev/null +++ b/pbm-functional/pytest/test_PBM-1391.py @@ -0,0 +1,87 @@ +import pytest +import pymongo +import bson +import testinfra +import time +import os +import docker +import threading + +from datetime import datetime +from cluster import Cluster + +documents = [{"a": 1}, {"b": 2}, {"c": 3}, {"d": 4}] + + +@pytest.fixture(scope="package") +def docker_client(): + return docker.from_env() + + +@pytest.fixture(scope="package") +def config(): + return { + "mongos": "mongos", + "configserver": { + "_id": "rscfg", + "members": [{"host": "rscfg01"}, {"host": "rscfg02"}, {"host": "rscfg03"}], + }, + "shards": [ + { + "_id": "rs1", + "members": [{"host": "rs101"}, {"host": "rs102"}, {"host": "rs103"}], + }, + { + "_id": "rs2", + "members": [{"host": "rs201"}, {"host": "rs202"}, {"host": "rs203"}], + }, + ], + } + + +@pytest.fixture(scope="package") +def cluster(config): + return Cluster(config) + + +@pytest.fixture(scope="function") +def start_cluster(cluster, request): + try: + cluster.destroy() + os.chmod("/backups", 0o777) + os.system("rm -rf /backups/*") + cluster.create() + cluster.setup_pbm() + client = pymongo.MongoClient(cluster.connection) + yield True + + finally: + if request.config.getoption("--verbose"): + cluster.get_logs() + cluster.destroy(cleanup_backups=True) + + +@pytest.mark.timeout(600, func_only=True) +def test_physical_PBM_T278(start_cluster, cluster): + cluster.check_pbm_status() + client = pymongo.MongoClient(cluster.connection) + client.admin.command("enableSharding", "test") + client.admin.command("shardCollection", "test.test", key={"_id": "hashed"}) + cluster.make_backup("physical") + cluster.enable_pitr(pitr_extra_args="--set pitr.oplogSpanMin=0.1") + time.sleep(5) + client["test"]["test"].insert_many(documents) + time.sleep(5) + pitr = datetime.utcnow().strftime("%Y-%m-%dT%H:%M:%S") + Cluster.log("Time for PITR is: " + pitr) + cluster.disable_pitr() + time.sleep(5) + client.drop_database("test") + backup = " --time=" + pitr + cluster.make_restore(backup, restart_cluster=True, check_pbm_status=True) + cluster.make_backup("physical") + cluster.enable_pitr(pitr_extra_args="--set pitr.oplogSpanMin=0.1") + time.sleep(10) + assert client["test"]["test"].count_documents({}) == len(documents) + assert client["test"].command("collstats", "test").get("sharded", False) + Cluster.log("Finished successfully") diff --git a/pbm-functional/pytest/test_directoryperdb.py b/pbm-functional/pytest/test_directoryperdb.py index 821ce2cd..55e44812 100644 --- a/pbm-functional/pytest/test_directoryperdb.py +++ b/pbm-functional/pytest/test_directoryperdb.py @@ -36,7 +36,7 @@ def start_cluster(cluster,request): cluster.destroy(cleanup_backups=True) @pytest.mark.timeout(300,func_only=True) -def test_physical(start_cluster,cluster): +def test_physical_PBM_T220(start_cluster,cluster): cluster.check_pbm_status() pymongo.MongoClient(cluster.connection)["test"]["test"].insert_many(documents) backup=cluster.make_backup("physical") @@ -47,7 +47,7 @@ def test_physical(start_cluster,cluster): Cluster.log("Finished successfully") @pytest.mark.timeout(300,func_only=True) -def test_incremental(start_cluster,cluster): +def test_incremental_PBM_T219(start_cluster,cluster): cluster.check_pbm_status() cluster.make_backup("incremental --base") pymongo.MongoClient(cluster.connection)["test"]["test"].insert_many(documents) diff --git a/pbm-functional/pytest/test_encryption.py b/pbm-functional/pytest/test_encryption.py index fe07f26a..5821eb1d 100644 --- a/pbm-functional/pytest/test_encryption.py +++ b/pbm-functional/pytest/test_encryption.py @@ -36,7 +36,7 @@ def start_cluster(cluster,request): cluster.destroy(cleanup_backups=True) @pytest.mark.timeout(300,func_only=True) -def test_physical(start_cluster,cluster): +def test_physical_PBM_T223(start_cluster,cluster): cluster.check_pbm_status() pymongo.MongoClient(cluster.connection)["test"]["test"].insert_many(documents) backup=cluster.make_backup("physical") @@ -47,7 +47,7 @@ def test_physical(start_cluster,cluster): print("Finished successfully") @pytest.mark.timeout(300,func_only=True) -def test_incremental(start_cluster,cluster): +def test_incremental_PBM_T222(start_cluster,cluster): cluster.check_pbm_status() cluster.make_backup("incremental --base") pymongo.MongoClient(cluster.connection)["test"]["test"].insert_many(documents) diff --git a/pbm-functional/pytest/test_replicaset.py b/pbm-functional/pytest/test_replicaset.py index 26b88c83..8f3e3447 100644 --- a/pbm-functional/pytest/test_replicaset.py +++ b/pbm-functional/pytest/test_replicaset.py @@ -46,23 +46,80 @@ def test_logical(start_cluster,cluster): cluster.check_pbm_status() pymongo.MongoClient(cluster.connection)["test"]["test"].insert_many(documents) pymongo.MongoClient(cluster.connection)["test"]["test1"].insert_many(documents) - pymongo.MongoClient(cluster.connection)["test2"]["test21"].insert_many(documents) - pymongo.MongoClient(cluster.connection)["test2"]["test22"].insert_many(documents) - backup_partial=cluster.make_backup("logical --ns=test.test,test2.*") backup_full=cluster.make_backup("logical") pymongo.MongoClient(cluster.connection).drop_database('test') - pymongo.MongoClient(cluster.connection).drop_database('test2') - cluster.make_restore(backup_partial,check_pbm_status=True) - assert pymongo.MongoClient(cluster.connection)["test"]["test"].count_documents({}) == len(documents) - assert pymongo.MongoClient(cluster.connection)["test"]["test1"].count_documents({}) == 0 - assert pymongo.MongoClient(cluster.connection)["test2"]["test21"].count_documents({}) == len(documents) - assert pymongo.MongoClient(cluster.connection)["test2"]["test22"].count_documents({}) == len(documents) - pymongo.MongoClient(cluster.connection).drop_database('test') cluster.make_restore(backup_full,check_pbm_status=True) assert pymongo.MongoClient(cluster.connection)["test"]["test"].count_documents({}) == len(documents) assert pymongo.MongoClient(cluster.connection)["test"]["test1"].count_documents({}) == len(documents) Cluster.log("Finished successfully") + +@pytest.mark.timeout(300, func_only=True) +def test_logical_selective_PBM_T274(start_cluster, cluster): + cluster.check_pbm_status() + client = pymongo.MongoClient(cluster.connection) + for i in range(10): + client["test1"]["test_coll11"].insert_one({"key": i, "data": i}) + client["test2"]["test_coll21"].insert_one({"key": i, "data": i}) + client["test2"]["test_coll22"].insert_one({"key": i, "data": i}) + client["test1"]["test_coll11"].create_index(["key"], name="test_coll11_index_old") + client["test2"]["test_coll21"].create_index(["key"], name="test_coll21_index_old") + backup_full = cluster.make_backup("logical") + backup_partial = cluster.make_backup("logical --ns=test1.test_coll11,test2.*") + cluster.enable_pitr(pitr_extra_args="--set pitr.oplogSpanMin=0.1") + time.sleep(5) + client.drop_database("test1") + for i in range(10): + client["test1"]["test_coll11"].insert_one({"key": i + 10, "data": i + 10}) + client["test1"]["test_coll11"].create_index("data", name="test_coll11_index_new") + client["test2"]["test_coll22"].create_index("data", name="test_coll22_index_new") + time.sleep(10) + pitr = datetime.utcnow().strftime("%Y-%m-%dT%H:%M:%S") + pitr = " --time=" + pitr + Cluster.log("Time for PITR is: " + pitr) + cluster.disable_pitr() + time.sleep(10) + client.drop_database("test1") + client.drop_database("test2") + backup_partial = " --base-snapshot=" + backup_partial + pitr + backup_full = ( + " --base-snapshot=" + backup_full + pitr + " --ns=test1.test_coll11,test2.*" + ) + cluster.make_restore(backup_partial, check_pbm_status=True) + assert client["test1"]["test_coll11"].count_documents({}) == 10 + assert client["test2"]["test_coll21"].count_documents({}) == 10 + assert client["test2"]["test_coll22"].count_documents({}) == 10 + for i in range(10): + assert client["test1"]["test_coll11"].find_one({"key": i + 10, "data": i + 10}) + assert client["test2"]["test_coll21"].find_one({"key": i, "data": i}) + assert client["test2"]["test_coll22"].find_one({"key": i, "data": i}) + assert ( + "test_coll11_index_old" + not in client["test1"]["test_coll11"].index_information() + ) + assert "test_coll11_index_new" in client["test1"]["test_coll11"].index_information() + assert "test_coll21_index_old" in client["test2"]["test_coll21"].index_information() + assert "test_coll22_index_new" in client["test2"]["test_coll22"].index_information() + client.drop_database("test1") + client.drop_database("test2") + cluster.make_restore(backup_full, check_pbm_status=True) + assert client["test1"]["test_coll11"].count_documents({}) == 10 + assert client["test2"]["test_coll21"].count_documents({}) == 10 + assert client["test2"]["test_coll22"].count_documents({}) == 10 + for i in range(10): + assert client["test1"]["test_coll11"].find_one({"key": i + 10, "data": i + 10}) + assert client["test2"]["test_coll21"].find_one({"key": i, "data": i}) + assert client["test2"]["test_coll22"].find_one({"key": i, "data": i}) + assert ( + "test_coll11_index_old" + not in client["test1"]["test_coll11"].index_information() + ) + assert "test_coll11_index_new" in client["test1"]["test_coll11"].index_information() + assert "test_coll21_index_old" in client["test2"]["test_coll21"].index_information() + assert "test_coll22_index_new" in client["test2"]["test_coll22"].index_information() + Cluster.log("Finished successfully") + + @pytest.mark.timeout(300,func_only=True) def test_physical(start_cluster,cluster): cluster.check_pbm_status() diff --git a/pbm-functional/pytest/test_sharded.py b/pbm-functional/pytest/test_sharded.py index e601a2fb..1d8e2669 100644 --- a/pbm-functional/pytest/test_sharded.py +++ b/pbm-functional/pytest/test_sharded.py @@ -40,11 +40,9 @@ def start_cluster(cluster,request): cluster.setup_pbm() client=pymongo.MongoClient(cluster.connection) client.admin.command("enableSharding", "test") - client.admin.command("enableSharding", "test2") client.admin.command("shardCollection", "test.test", key={"_id": "hashed"}) client.admin.command("shardCollection", "test.test2", key={"_id": "hashed"}) client.admin.command("shardCollection", "test.test3", key={"_id": "hashed"}) - client.admin.command("shardCollection", "test2.test21", key={"_id": "hashed"}) yield True finally: @@ -53,25 +51,12 @@ def start_cluster(cluster,request): cluster.destroy(cleanup_backups=True) @pytest.mark.timeout(600,func_only=True) -def test_logical_PBM_T218(start_cluster,cluster): +def test_logical(start_cluster,cluster): cluster.check_pbm_status() pymongo.MongoClient(cluster.connection)["test"]["test"].insert_many(documents) pymongo.MongoClient(cluster.connection)["test"]["test1"].insert_many(documents) - pymongo.MongoClient(cluster.connection)["test2"]["test21"].insert_many(documents) - pymongo.MongoClient(cluster.connection)["test2"]["test22"].insert_many(documents) - backup_partial=cluster.make_backup("logical --ns=test.test,test2.*") backup_full=cluster.make_backup("logical") pymongo.MongoClient(cluster.connection).drop_database('test') - cluster.make_restore(backup_partial,check_pbm_status=True) - assert pymongo.MongoClient(cluster.connection)["test"]["test"].count_documents({}) == len(documents) - assert pymongo.MongoClient(cluster.connection)["test"].command("collstats", "test").get("sharded", False) - assert pymongo.MongoClient(cluster.connection)["test"]["test1"].count_documents({}) == 0 - assert pymongo.MongoClient(cluster.connection)["test"].command("collstats", "test1").get("sharded", True) is False - assert pymongo.MongoClient(cluster.connection)["test2"]["test21"].count_documents({}) == len(documents) - assert pymongo.MongoClient(cluster.connection)["test2"].command("collstats", "test21").get("sharded", False) - assert pymongo.MongoClient(cluster.connection)["test2"]["test22"].count_documents({}) == len(documents) - assert pymongo.MongoClient(cluster.connection)["test2"].command("collstats", "test22").get("sharded", True) is False - pymongo.MongoClient(cluster.connection).drop_database('test') cluster.make_restore(backup_full,check_pbm_status=True) assert pymongo.MongoClient(cluster.connection)["test"]["test1"].count_documents({}) == len(documents) assert pymongo.MongoClient(cluster.connection)["test"].command("collstats", "test1").get("sharded", True) is False @@ -79,6 +64,93 @@ def test_logical_PBM_T218(start_cluster,cluster): assert pymongo.MongoClient(cluster.connection)["test"].command("collstats", "test").get("sharded", False) Cluster.log("Finished successfully") + +@pytest.mark.timeout(300, func_only=True) +def test_logical_selective_PBM_T218(start_cluster, cluster): + cluster.check_pbm_status() + client = pymongo.MongoClient(cluster.connection) + client.admin.command("enableSharding", "test2") + client.admin.command("shardCollection", "test2.test_coll21", key={"_id": "hashed"}) + for i in range(10): + client["test1"]["test_coll11"].insert_one({"key": i, "data": i}) + client["test2"]["test_coll21"].insert_one({"key": i, "data": i}) + client["test2"]["test_coll22"].insert_one({"key": i, "data": i}) + client["test1"]["test_coll11"].create_index(["key"], name="test_coll11_index_old") + client["test2"]["test_coll21"].create_index(["key"], name="test_coll21_index_old") + backup_full = cluster.make_backup("logical") + backup_partial = cluster.make_backup("logical --ns=test1.test_coll11,test2.*") + cluster.enable_pitr(pitr_extra_args="--set pitr.oplogSpanMin=0.1") + time.sleep(5) + client.drop_database("test1") + for i in range(10): + client["test1"]["test_coll11"].insert_one({"key": i + 10, "data": i + 10}) + client["test1"]["test_coll11"].create_index("data", name="test_coll11_index_new") + client["test2"]["test_coll22"].create_index("data", name="test_coll22_index_new") + time.sleep(10) + pitr = datetime.utcnow().strftime("%Y-%m-%dT%H:%M:%S") + pitr = " --time=" + pitr + Cluster.log("Time for PITR is: " + pitr) + cluster.disable_pitr() + time.sleep(10) + client.drop_database("test1") + client.drop_database("test2") + backup_partial = " --base-snapshot=" + backup_partial + pitr + backup_full = ( + " --base-snapshot=" + backup_full + pitr + " --ns=test1.test_coll11,test2.*" + ) + cluster.make_restore(backup_partial, check_pbm_status=True) + assert client["test1"]["test_coll11"].count_documents({}) == 10 + assert ( + client["test1"].command("collstats", "test_coll11").get("sharded", True) + is False + ) + assert client["test2"]["test_coll21"].count_documents({}) == 10 + assert client["test2"].command("collstats", "test_coll21").get("sharded", False) + assert client["test2"]["test_coll22"].count_documents({}) == 10 + assert ( + client["test2"].command("collstats", "test_coll22").get("sharded", True) + is False + ) + for i in range(10): + assert client["test1"]["test_coll11"].find_one({"key": i + 10, "data": i + 10}) + assert client["test2"]["test_coll21"].find_one({"key": i, "data": i}) + assert client["test2"]["test_coll22"].find_one({"key": i, "data": i}) + assert ( + "test_coll11_index_old" + not in client["test1"]["test_coll11"].index_information() + ) + assert "test_coll11_index_new" in client["test1"]["test_coll11"].index_information() + assert "test_coll21_index_old" in client["test2"]["test_coll21"].index_information() + assert "test_coll22_index_new" in client["test2"]["test_coll22"].index_information() + client.drop_database("test1") + client.drop_database("test2") + cluster.make_restore(backup_full, check_pbm_status=True) + assert client["test1"]["test_coll11"].count_documents({}) == 10 + assert ( + client["test1"].command("collstats", "test_coll11").get("sharded", True) + is False + ) + assert client["test2"]["test_coll21"].count_documents({}) == 10 + assert client["test2"].command("collstats", "test_coll21").get("sharded", False) + assert client["test2"]["test_coll22"].count_documents({}) == 10 + assert ( + client["test2"].command("collstats", "test_coll22").get("sharded", True) + is False + ) + for i in range(10): + assert client["test1"]["test_coll11"].find_one({"key": i + 10, "data": i + 10}) + assert client["test2"]["test_coll21"].find_one({"key": i, "data": i}) + assert client["test2"]["test_coll22"].find_one({"key": i, "data": i}) + assert ( + "test_coll11_index_old" + not in client["test1"]["test_coll11"].index_information() + ) + assert "test_coll11_index_new" in client["test1"]["test_coll11"].index_information() + assert "test_coll21_index_old" in client["test2"]["test_coll21"].index_information() + assert "test_coll22_index_new" in client["test2"]["test_coll22"].index_information() + Cluster.log("Finished successfully") + + @pytest.mark.timeout(600, func_only=True) def test_logical_pitr_PBM_T194(start_cluster,cluster): cluster.check_pbm_status() diff --git a/pbm-functional/pytest/test_upgrade.py b/pbm-functional/pytest/test_upgrade.py index 3193b0c6..05b3d4b2 100644 --- a/pbm-functional/pytest/test_upgrade.py +++ b/pbm-functional/pytest/test_upgrade.py @@ -47,7 +47,7 @@ def start_cluster(cluster,request): cluster.destroy(cleanup_backups=True) @pytest.mark.timeout(600,func_only=True) -def test_logical(start_cluster,cluster): +def test_logical_PBM_T275(start_cluster,cluster): cluster.check_pbm_status() pymongo.MongoClient(cluster.connection)["test"]["test"].insert_many(documents) backup=cluster.make_backup("logical") @@ -61,7 +61,7 @@ def test_logical(start_cluster,cluster): Cluster.log("Finished successfully") @pytest.mark.timeout(600,func_only=True) -def test_physical(start_cluster,cluster): +def test_physical_PBM_T276(start_cluster,cluster): cluster.check_pbm_status() pymongo.MongoClient(cluster.connection)["test"]["test"].insert_many(documents) backup=cluster.make_backup("physical") @@ -75,7 +75,7 @@ def test_physical(start_cluster,cluster): Cluster.log("Finished successfully") @pytest.mark.timeout(600,func_only=True) -def test_incremental(start_cluster,cluster): +def test_incremental_PBM_T277(start_cluster,cluster): cluster.check_pbm_status() cluster.make_backup("incremental --base") pymongo.MongoClient(cluster.connection)["test"]["test"].insert_many(documents) diff --git a/pbm-functional/pytest/test_vault.py b/pbm-functional/pytest/test_vault.py index d4df52f0..2ee19645 100644 --- a/pbm-functional/pytest/test_vault.py +++ b/pbm-functional/pytest/test_vault.py @@ -38,7 +38,7 @@ def start_cluster(cluster,request): cluster.destroy(cleanup_backups=True) @pytest.mark.timeout(300,func_only=True) -def test_physical_PBM_T195(start_cluster,cluster): +def test_physical_PBM_T196(start_cluster,cluster): cluster.check_pbm_status() pymongo.MongoClient(cluster.connection)["test"]["test"].insert_many(documents) backup=cluster.make_backup("physical")