History
- Bug fixes for corrections versioning
- Switch to new common account
- Rucio updates
- Tape updates
- Corrections versioning (#95)
- Rucio updates
- Processed file purger (#85)
- hax TailCut minitree (#89)
- Purge updates for Midway (#88, #91)
- More verbose debug info
- Add tsm-server and login as master checksums
- Do not use "--append" for processed files (to recopy failed local processing)
- Add mem-per-cpu and walltime options for massive-cax * Default 2000 MB and 48 hours, respectively
- Allow no specified partition if using reservation * Defaults to sandyb on Midway
- Link ROOT libraries into CWD to avoid rebuilding for each job
- Add "Extended" minitree production
- Propagate custom ncpu into pax process task
- Revert back to deleting error'ed processed data
- Specify partition when counting jobs
- Remove wall time limit
- Disable buffer purging
- Switch everything to /project2
- hax only if processed is "transferred"
- Implement command line arguments for: * Run tag (e.g. "_sciencerun0") * Cluster partition (e.g. "xenon1t", "kicp") * Number of CPUs per job
- Fix bug in run range in massive cax
- Switch to rsync for processed data
- Fix drift velocity function bug (#79)
- Allow for massive cax to have stop range
- Make proximity trees
- First Ruciax merge
- Revert to checksumming "verifying" only
- Fix: only modify RunsDB if actually deleting data
- Add check of pax version when deleting processed data
- Switch back to xenon1t partition
- Switch to kicp partition
- Copy only processed files, not raw, and fix version check
- Disable iterative mode to always work on old runs
- Gains for acquisition monitor
- Tweaks for getting the slow control interface to be used for gains.
- Gains can depend on time (#77)
- Drift velicity (#75)
- Tweaks for _sciencerun0 data
- Decrease purge threshold to 5 days (from 25)
- Allow purge of Rn220, disallow AmBe
- Do not purge until processed
- Switch processed dir back to /project (from /project2)
- Request 9 cores per job (instead of 4)
- Specify 8 cores for pax (instead of 4), 1 extra for I/O worker
- Switch to /project2 space on Midway
- Switch to rsync for data transfers (instead of scp)
- Compute gains at Midway to speed it up (Closes #50)
- Added tape backup upload (TSM) to master branch
- Fix hax logging bug (was preventing all minitree creation)
- Delay error'ed purging on xe1t-datamanager
- Reduce massive-cax job submission delay back to 1 second
- Fix pax input type for MV
- Increase pax max_blocks_on_heap
- Adjust setpermission for PDC and Midway
- Add other treemakers to process_hax
- Do not table scan run database
- Increase delay between job submission to 10 s
- Increase timeout for pax processing
- Purge Rn220+Kr83m only on Midway
- Disable iterative mode for massive-cax (for now)
- Fix batch queue checking for MV jobs
- Do not purge Kr83m and Rn220
- Fix bug in queue checking
- hax minitree production
- Reduce pax queue block size in batch processing #51
- Fix pax queue size configuration option
- Remove gfal environment setup (may be clashing with pax)
- Remove "once" functionality from massive-cax (strange error with "watch", and hanging without "watch")
- Revert to file:// instead of SRM address for Midway gfal-copy upload
- Remove extraenous AddChecksum's and put ProcessBatchQueue at the end
- Reduce max_queue_blocks from 100 to 50 (otherwise AmBe hits memory limit)
- Do not purge processed files
- Consider only same file type when counting copies
- Remove hardcoded midway-srm treatment
- Hardcode Midway SRM address for gfal-copy uploads.
- Switch back to Stash round-robin address.
- Fix missing "dir_processed" key error.
- Change from Stash to Nikhef for Midway GRID upload
- Specify ADLER32 checksum for gfal-copy
- Skip "verifying" stage for GRID transfers (assume gfal-copy checksum is sufficient)
- Change Stash GSIFTP site to round-robin address
- Change raw directory on nikhef-srm
- Fix variable name for GSIFTP source server
- Extend gfal-copy time even more (to 9 hours)
- Should cover typical 40 GB file and slow 3 MB/s speed
- Use GSIFTP address of Stash (login) as source with gfal-copy
- Upload to nikhef-srm from Stash (login)
- Switch from lcg-cp to gfal-copy in cax.json
- Purge using gfal-rm on Stash
- Pass GRID certificate for worker nodes.
- Change raw directory for Stash GRID copy.
- Load GRID tools within job on Midway
- Switch Stash SRM address to gsiftp address
- Allow specification in cax.json for number of streams in GRID copy
- Increase gfal-copy timeout to 3 hours (-t 10800)
- Disable LFC registration (Rucio should pick up the bookkeeping later)
- Gains less than 1 are set exactly to zero.
- Another bug fix (forgot a :)
- Bug fix (commented wrong line in job script in previous commit)
- Command-line option to specify one run or starting run with massive-cax
- Hardcoded (commented out) option to switch to Midway sandyb (public) partition
- Check queue in all partitions on Midway instead of just xenon1t
- Add "login" (Stash/ci-connect) to Midway upload option (remove Midway from Stash download)
- Do not recheck checksums on Stash ("login"), too slow since no batch queue for massive-cax
- Verify that data is still there (#42)
- MV processing support
- Add command line options --once and --config for massive-cax
- Bug fix: Job name should have pax version, not cax version
- Temporarily disable 'sacct' call (seems to be broken on Midway after restart)
- Bug fix: check for actual version number in queue list instead of "head"
- Reduce allowed number of jobs in queue to 500 (since we only have 28*16=448 cores)
- Change path to Anaconda Installation at PDC
- Process after copy.
- Have LED mode have gains equal to 1, which is the same as XENON1T_LED.ini.
- Fix permissions at PDC when new folder for new processed versions of data (#35).
- Forgot to remove line about batch queue statistics that only works at Midway.
- Forgot to update HISTORY.rst
- Gains derived from HV measurements using HTTP interface (#34).
- Statistics at end of job on usage.
- Slow control variables defined in hax added to rundoc.
- Allow more jobs at Midway
- Error if default gains
- Create output directory prior to pax processing
- Gains support (#32).
- Create and use run sub-directory for logs
- Fix bug where "[]" in cax.json is not treated as "None"
- Yet more PDC changes
- Changes for Stockholm PDC
- Forgot to switch environment outside of test environment
- cax now operates by sending jobs to the batch queue for every run (See PR #30).
- Only send email notifications for failed jobs
- Remove Nikhef ability to pull new data
- Switch queue check command for public nodes on Midway
- Switch to public nodes on Midway for next mass reprocessing
- Increase 1 CPU processing threshold to 1000 events (XENON1T/pax#390)
- Change all cax.json entries from xenon1t-daq to xe1t-datamanager
- Checksum comes from xe1t-datamanager
- Grid copy functionality
- Use datamanager machine at LNGS.
- Remove race condition check since didn't work
- Raise timeout for deleting to 24 hours
- Tune race condition logic (make stricter)
- Log exceptions then reraise
- Fix bug/typo in logic checking that data location doesn't already exist before transferring processed data.
- Do not retransfer processed data now also checks pax_version because otherwise would stop after any version
- Execute one candidate transfer instead of all candidate transfers so it can recheck next time task is executed what candidates are
- Avoid race condition if two cax running with copies.
- Purity is float and not sympy float type. Otherwise, MongoDB doesn't understand it.
- Generalized purification evolution function in run database
- Handle reconnect signal from Mongo if LNGS connection unstable.
- Handle modified times even if file does not exist
- Check modified times before deleting data for timeout
- Catch FileNotFoundError when deleting files, then warn.
- Process with pax 5.0
- Revert PROCESSING_DIR to separate directories for each job
- Stockholm grabs data from LNGSx
- Fix bug in parameter manipulation for lifetime fit.
- Add electron lifetime support
- When task looks for runs, have it only return the _id then fetch that id later. Helps with timeouts.
- If task timeout of mongo find, have it skip that task.
- Remove mv command for logs from job, doesn't work with new {processing_dir}. Keep them all in same location for now.
- Uncomment submit command for automatic processing
- Use different folder for raw and root data
- Add cax-mv, cax-rm, cax-stray
- Don't need to clear DAQ buffer anymore in cax.
- Cleanup and fixes related to processing.
- Transfer bug that made bad element in data location list
- Specify the partition in qsub.py
- filesystem.py: Add a class to ask for the status of a file or folder
- Retry if errored instead of waiting two days.
- Specify log level on command line.
- Version number only in file log, not screen
- Add release support
- Add version number to log output
- Initial stable release
- SCP support for transfer
- Checksumming
- Retry failed transfers if checksum fails or timeout
- Processing on batch queue
- Initial release