Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[epic] v1.0.0 Performance and Scale #920

Open
joelanford opened this issue Jun 11, 2024 · 9 comments
Open

[epic] v1.0.0 Performance and Scale #920

joelanford opened this issue Jun 11, 2024 · 9 comments
Assignees
Labels
epic triage/needs-information Indicates an issue needs more information in order to work on it. v1.x Issues related to OLMv1 features that come after 1.0

Comments

@joelanford
Copy link
Member

Epic Goal

  • Measure and implement any necessary improvements to ensure OLM v1.0.0 meets or exceeds OCP guidelines around performance and scalability.

Why is this important?

  • OLM v1.0.0 will be a payload component that always runs in OCP clusters. In order to reduce SD and customer costs, we need to minimize this overhead.
  • OLM v1.0.0 is intended to be used on a wide variety of clusters, ranging from single node clusters with just a few namespaces to clusters 2-3 orders of magnitude larger. We need make sure that it runs just as well on a small cluster as it does a large cluster.
  • In order to reduce user frustration, we need to provide a responsive user experience. Reconciliation needs to be fast and non-blocking to ensure users receive the experience they have come to expect from OCP. To the extent possible, long-running tasks (e.g. catalog fetching/caching and image pulling) should be performed asynchronously.

Scenarios

  1. Collect pprof profiles for CPU and memory when running standard user flows around installing, upgrading, and removing operators from public catalogs (e.g. operatorhub)
  2. Find the most resource intensive code paths. Provide documentation and recommendations related to making improvements in those areas.
  3. Coordinate with OLM maintainers to make improvements in areas deemed to provide the most significant performance and scale gain.
  4. Implement automated performance and scale regression tests in the existing upstream CI test suite.

Examples of known areas for improvement include:

  • When reconciling a ClusterExtension to resolve a bundle from the criteria provided by a user, the reconciler should return a desired bundle within 100ms and allocate no more memory than the size of the catalog metadata for the named spec.packageName.
  • When the ClusterExtension reconciler does not have the contents of a resolved image bundle available, it does not block waiting for the image to be pulled and processed. Rather, it starts an asychronous job, reports the pending image pull via the ClusterExtension status, and returns from reconcile.
@joelanford joelanford added v1.0 Issues related to the initial stable release of OLMv1 epic labels Jun 11, 2024
@OchiengEd
Copy link

/assign

@joelanford
Copy link
Member Author

joelanford commented Jul 14, 2024

I think I've found one unexpected slowdown: the bundle handler that converts a registry+v1 bundle to plain and then to helm. It takes 5s on my machine in the "Force upgrade" e2e test.

@kevinrizza kevinrizza moved this to Implementing in OLM v1 Jul 16, 2024
@joelanford
Copy link
Member Author

@OchiengEd just wanted to check to make sure you didn't find any critical (as in "must fix for 1.0.0") issues in your performance and scale research?

If not, can we move the remaining scope of this epic to v1.x?

@everettraven
Copy link
Contributor

everettraven commented Aug 20, 2024

Item to include in performance and scale (although maybe more of a release blocker based on other discussions):

EDIT: After discussion in the community meeting, this issue is not in scope for this epic.

@everettraven everettraven added v1.x Issues related to OLMv1 features that come after 1.0 and removed v1.0 Issues related to the initial stable release of OLMv1 labels Aug 20, 2024
@OchiengEd
Copy link

No critical issues were identified. This epic was slated to be moved to 1.x

@OchiengEd OchiengEd removed their assignment Oct 24, 2024
@LalatenduMohanty LalatenduMohanty added the triage/needs-information Indicates an issue needs more information in order to work on it. label Oct 29, 2024
@LalatenduMohanty
Copy link
Member

Lets re-access this and identify the acceptance criteria for this epic.

@joelanford joelanford added v1.1 and removed v1.x Issues related to OLMv1 features that come after 1.0 labels Nov 5, 2024
@LalatenduMohanty LalatenduMohanty moved this from Implementing to Accepted in OLM v1 Dec 3, 2024
@LalatenduMohanty LalatenduMohanty removed the status in OLM v1 Dec 3, 2024
@LalatenduMohanty LalatenduMohanty added the v1.x Issues related to OLMv1 features that come after 1.0 label Dec 10, 2024
@LalatenduMohanty LalatenduMohanty self-assigned this Jan 21, 2025
@grokspawn
Copy link
Contributor

We should remodel this issue to design the assessment/reporting infrastructure, with MVP implementation. However, it should include the scope of the new query catalogd web API discussed in #1607.
We can create subsequent epics to give us measurable progress and continue to refine the implementation.

@LalatenduMohanty
Copy link
Member

Next step is to call a meeting and agree on the things we want to achieve with this epic. cc @dtfranz

@grokspawn
Copy link
Contributor

grokspawn commented Jan 28, 2025

From the committee meeting, this is analogous to work that we're doing w.r.t. feature-gates, where we have

  1. a general framework which provides for measurement/assessment/detection capabilities in a standardized approach; and
  2. features which hook into the framework to inform their measurement/assessment/detection functionality, starting with catalogd web api benchmarking #1607

@LalatenduMohanty LalatenduMohanty moved this to Designing in OLM v1 Feb 4, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
epic triage/needs-information Indicates an issue needs more information in order to work on it. v1.x Issues related to OLMv1 features that come after 1.0
Projects
Status: Designing
Development

No branches or pull requests

6 participants