Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[bug] Signadot CRDs left in the cluster after sandbox TTL has passed or was deleted #48

Open
gamino-brex opened this issue Feb 14, 2024 · 1 comment
Labels
bug Something isn't working

Comments

@gamino-brex
Copy link

Describe the bug
Sandbox and RouteConfig resources with months old were not cleaned up from deleted sandboxes (either deleted or expired post-TTL). Resources can be found with kubectl but the sandboxes do not show up in the Dashboard or CLI.

To Reproduce

  • Currently on operator v0.15.0 but can be tracked to older versions since this is happening for months already
  • No explicit known steps to reproduce

Expected behavior
If a sandbox has expired or was deleted, we should not have any related orphan resources hanging around our K8s cluster

Observed behavior
Stale resources in our cluster

Additional context
More details in the shared Slack channel

@gamino-brex gamino-brex added the bug Something isn't working label Feb 14, 2024
@scott-cotton
Copy link
Member

scott-cotton commented Feb 16, 2024

I agree we should not be leaving around Signadot k8s resources.

The only known circumstances in which we can leave k8s resources behind are

  1. Running cluster delete from the dashboard (this gives a warning about leaving resources around); and
  2. changing the cluster agent token to a token for a different cluster as identified in our dashboard/cli

As in general we have in place guards which prevent sandbox deletion on our SaaS unless the corresponding cluster has confirmed deletion.

So, it would be good to get more information here. Do you think that these orphaned k8s resources could be the result of one of the 2 circumstances above? If not, could you provide some more info, like how often this occurs, perhaps the metadata of some orphaned Signadot k8s resources?

If, on the other hand the 2 circumstances above do seem to explain your orphaned Signadot k8s resources, would you prefer that we delete them anyway?

(To clean up anything manually, it would involve removing finalizers with kubectl prior to deletion, or alternatively, creating a sandbox with name .Spec.Name of the orphaned SDS resources and deleting it from our saas)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working
Projects
None yet
Development

No branches or pull requests

2 participants