Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Stress test on NiLab mongodb #162

Open
FaroutYLq opened this issue Jun 18, 2024 · 1 comment
Open

Stress test on NiLab mongodb #162

FaroutYLq opened this issue Jun 18, 2024 · 1 comment
Assignees
Labels

Comments

@FaroutYLq
Copy link
Collaborator

Not really an outsource "issue", but just a to-do test. We hope to use that mirror exclusively for reprocessing or MC.

@FaroutYLq
Copy link
Collaborator Author

FaroutYLq commented Oct 11, 2024

Right now given the LazyInit PR in admix, we are prepared to test the similar processing scenario for SR2 reprocessing, in terms of DB connection per job. That means, from a stress test on NiLab, we will be able to project how much connection we will have in SR2 reprocessing, and see if it is OK to make NiLab mirror exclusively reserved for reprocessing pipeline and nothing else.

Here is a draft of test proposal, which should be a project recorded here once we figure out exact runlist.

  • Find a list of SR1 calibration runs whose peaks are not available (to coordinate with @dantonmartin. It should be a subset of those runs whose raw_records not removed from disk, because of missing intermediate data)
  • Merge LazyInit PR in admix and make sure outsource is using it. (whether through overriding admix version, or just in a new environment with this updated admix)
  • Make sure in xenon config that the only mirror involved is NiLab.
  • Process v15 for this decided run list. Hopefully we get something around 400 runs (that is a typical batch size we submit per day for SR1)
  • Make sure in xenon config that - Ideally handled by @minzhong98 or other people who have admin log of NiLab mirror: Closely monitor RunDB load, and compare the connection amount trend with running job number you learned from condor_q. Understand the amount phenomenologically:
    • ~x connections per y job
    • connection decay time if any
  • Deliver those result peaks to DaLi.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
Projects
None yet
Development

No branches or pull requests

2 participants