Date: Sat, 30 Mar 2024 07:01:05 +0000 (UTC) Message-ID: <1410378539.22451.1711782070378@confluence> Subject: Exported From Confluence MIME-Version: 1.0 Content-Type: multipart/related; boundary="----=_Part_22450_1814833156.1711782065368" ------=_Part_22450_1814833156.1711782065368 Content-Type: text/html; charset=UTF-8 Content-Transfer-Encoding: quoted-printable Content-Location: file:///C:/exported.html
May 15, 2014
* Created issues, will space out across sprints and add estimates<= /p>
* One of them is to follow up with HTCondor team on dynamic slot C= lassAds
* Adding visit id to worker job for monitoring
* Working on pre-script to write DAG dynamically
+ Will keep simple; probably not so necessary for Nightly C= omputing
* Will estimate story points
* Have replicator jobs talking to replicator nodes and distributor= nodes
+ Just metadata for now, not data yet
+ How much data? Will be 500+MB; maybe use 50MB scaled to n= etwork
+ How to transfer data?
- Job to node: shm, socket
- Replicator to distributor: memory to memory via so= cket, file transfer
- Doesn't matter much; whatever is easiest
* Using regular lsst-dev HTCondor pool
* Right now, some scripts point to SRP's home directory
+ Will work on making relocatable
OCS will have DDS interface available sooner than expected, maybe June/J= uly
* Log messages gathered to one location? Yes
+ Also critical messages can go via real-time messages
+ Also status updated directly in databases for monitoring<= /p>
* Need to make sure operators have sufficient insight into system<= /p>
* Is Tucson headquarters an overall control center?
+ Anticipate HQ is mostly overall management, not day-to-da= y
+ Four operational centers (Base, Archive, Chile DAC, US DA= C)
+ Operational centers run mostly independently
+ Base runs Nightly Computing in conjunction with NCSA
+ Archive runs DRP
* DES is trying out large solid-state disk behind GPFS
+ K-T will see if there are any potential uses for LSST tes= ting
+ Maybe DB group?