-
Notifications
You must be signed in to change notification settings - Fork 34
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Schema restore failed with: "Unrecognized strategy option {us-east} passed to org.apache.cassandra.locator.NetworkTopologyStrategy" #4041
Schema restore failed with: "Unrecognized strategy option {us-east} passed to org.apache.cassandra.locator.NetworkTopologyStrategy" #4041
Comments
Well, this is expected. A workaround is to take the schema file from backup location, modify it to fit your needs, and apply it manually. |
@yarongilor Is there anything you suggest to change in Scylla Manager ? As per #4041 (comment) this is expected behavior of the manager. It looks that there is no datacenter of the "us-east" name in the destination cluster.
|
It's not a new issue, mostly a usability issue. |
Issue about restoring schema into a differenct DC setting: #4049. |
So currently the user is supposed to do the schema restore manually @mikliapko so I'll say we should at least skip the nemesis if the region of the snapshots doesn't match. At least until it would be implemented on the test end or manager end. |
I don't remember we have an issue for that. |
@mikliapko - IMO we can plan for a workaround, depends when this issue will be fixed from Manager side. |
there was an issue about this, long long ago: I don't know if anything was done to try to apply any workaround. |
The issue reproduced at https://jenkins.scylladb.com/job/scylla-master/job/tier1/job/longevity-mv-si-4days-streaming-test/7/ |
@mikliapko we want to have at least a workaround for this issue until it's being fixed in manager. |
Sorry, lost track of this issue for a while. I’ll try to come up with a solution no later than next week. |
If there is no fix planned in manager or an easy workaround on the manager side, you can have a workaround in SCT by uploading backups to several regions (each with correct region in schema). |
@Michal-Leszczynski |
It's in the backup location under Uncompressed file can look like that:
|
reproduced there: PackagesScylla version: Kernel Version: Installation detailsCluster size: 5 nodes (i4i.8xlarge) Scylla Nodes used in this run:
OS / Image: Test: Logs and commands
Logs:
|
Was trying to workaround it via manual restore of the schema (with putting the right region into the CQL statements for keyspace creation) - failed because of Looks like backup uploading to several regions is the only way left so far. |
those backup were create with KMS keys which are long gone, regardless of region. |
I recreated those backups a few months ago. |
I take it back, we clear the aliases and not the keys. |
Hm, we have an issue related to In test_defaults.yaml this parameter is empty string:
Then it gets rewritten by aws_config.yaml:
This parameter is used to configure manager agent: node.update_manager_agent_backup_config(
region=self.params.get("backup_bucket_region"),
general_config=agent_backup_general_config,
) As a result, if the I'm thinking about adding a validation rule for if self.get("backup_bucket_region") != self.get("region_name"):
self["backup_bucket_region"] = self.get("region_name") @fruch May this change lead to any unexpected consequences? How do you think? |
as long as you have buckets in all regions for doing the backup nemesis (including of in gce / azure) I would recommend removing |
Yes, as for now we can't set different |
After duplicating all snapshots and fixing backup location issues,
The test alters keyspace with RF=3 to 5 what is prohibited. |
I suppose to fix it, we need to implement the procedure described here: |
I wonder how this test was introduces at first time? |
reproduced there: |
Packages
Scylla version:
2024.2.0~rc2-20240904.4c26004e5311
with build-ida8549197de3c826053f88ddfd045b365b9cd8692
Kernel Version:
5.15.0-1068-aws
Issue description
The backup restore failed with error:
The restore task was started like:
then failed:
Describe your issue in detail and steps it took to produce it.
Impact
Describe the impact this issue causes to the user.
How frequently does it reproduce?
Describe the frequency with how this issue can be reproduced.
Installation details
Cluster size: 4 nodes (i4i.4xlarge)
Scylla Nodes used in this run:
OS / Image:
ami-0555cb82c50d0d5f1
(aws: undefined_region)Test:
longevity-alternator-1h-scan-12h-ttl-no-lwt-2h-grace-4loaders-sisyphus-test
Test id:
4afc0c3a-7457-4d8b-a69a-8ee387d26369
Test name:
enterprise-2024.2/alternator_tablets/longevity-alternator-1h-scan-12h-ttl-no-lwt-2h-grace-4loaders-sisyphus-test
Test method:
longevity_test.LongevityTest.test_custom_time
Test config file(s):
Logs and commands
$ hydra investigate show-monitor 4afc0c3a-7457-4d8b-a69a-8ee387d26369
$ hydra investigate show-logs 4afc0c3a-7457-4d8b-a69a-8ee387d26369
Logs:
Jenkins job URL
Argus
The text was updated successfully, but these errors were encountered: