14
AGLT2 Site Report Shawn McKee University of Michigan HEPiX Fall 2014 / UNL

AGLT2 Site Report Shawn McKee University of Michigan HEPiX Fall 2014 / UNL

Embed Size (px)

Citation preview

Page 1: AGLT2 Site Report Shawn McKee University of Michigan HEPiX Fall 2014 / UNL

AGLT2 Site Report

Shawn McKeeUniversity of MichiganHEPiX Fall 2014 / UNL

Page 2: AGLT2 Site Report Shawn McKee University of Michigan HEPiX Fall 2014 / UNL

Outline

Site Summary and StatusMonitoringProvisioning with CobblerHTCondorMCORE detailsVirtualization StatusNetworking UpgradeUpdates on projectsPlans for the future

AGLT2-HEPiX 14-Oct-14

Outline

Page 3: AGLT2 Site Report Shawn McKee University of Michigan HEPiX Fall 2014 / UNL

AGLT2-HEPiX 14-Oct-14

Site SummaryThe ATLAS Great Lake Tier-2 (AGLT2) is a distributed LHC Tier-2 for ATLAS spanning between UM/Ann Arbor and MSU/East Lansing. Roughly 50% of storage and compute at each site

5722 single core job slots (added 480 cores)MCORE slots increased from 240 to 420 (dynamic)269 Tier-3 job slots usable by Tier-2 Average 9.26 HS06/slot3.5 Petabytes of storage (adding 192 TB, retiring 36 TB)Total of 54.4 kHS06, up from 49.0 kHS06 in spring

Most Tier-2 services virtualized in VMware2x40 Gb inter-site connectivity, UM has 100G to WAN, MSU has 10G to WAN, lots of 10Gb internal ports and 16 x 40Gb portsHigh capacity storage systems have 2 x 10Gb bonded links40Gb link between Tier-2 and Tier-3 physical locations

Page 4: AGLT2 Site Report Shawn McKee University of Michigan HEPiX Fall 2014 / UNL

AGLT2-HEPiX 14-Oct-14

AGLT2 MonitoringAGLT2 has a number of monitoring components in useAs shown in Annecy we have:• Customized “summary” page->• OMD (Open Monitoring

Distribution) at both UM/MSU• Ganglia• Central syslog’ing via ELK:

Elasticsearch, Logstash, Kibana• SRMwatch to track dCache SRM

status• GLPI to track tickets (with

FusionInventory)

Page 5: AGLT2 Site Report Shawn McKee University of Michigan HEPiX Fall 2014 / UNL

Provisioning with Cobbler

AGLT2-HEPiX 14-Oct-14

AGLT2 Provisioning/Config MgmtAGLT2 uses a Cobbler server configuration managed by CFEngine and duplicated at both sites for building service nodes (excepting site-specific network/host info)

Created flexible default kickstart template with Cobbler’s template language (Cheetah) to install a variety of “profiles” as selected when adding system to Cobbler (server, cluster-compute, desktop, etc).Simple PXE based installation from networkCobbler handles (with included post-install scripts) creating bonded NIC configurations – used to deal with those manuallyCobbler manages mirroring of OS and extra repositories

Kickstart setup is kept minimal and most configuration done by CFEngine on first bootDell machines get BIOS and Firmware updates in post-install using utils/packages from Dell yum repositories See Ben Meekhof’s talk Thursday for details (https://indico.cern.ch/event/320819/session/7/contribution/23)

Page 6: AGLT2 Site Report Shawn McKee University of Michigan HEPiX Fall 2014 / UNL

AGLT2-HEPiX 14-Oct-14

HTCondor CE at AGLT2• Bob Ball worked for ~1 month at AGLT2 setup

– Steep learning curve for newbies– Lots of non-apparent niceties in preparing job-router configuration– RSL no longer available for routing decisions

• Cannot change content of job route except during condor-ce restart

• However, CAN modify variables and place them in ClassAd variables set in the router– Used at AGLT2 to control MCORE slot access

• Currently in place on test gatekeeper only• Will extend to the primary GK ~10/22/14• See full details of our experience and setup at

https://www.aglt2.org/wiki/bin/view/AGLT2/CondorCE

Page 7: AGLT2 Site Report Shawn McKee University of Michigan HEPiX Fall 2014 / UNL

AGLT2-HEPiX 14-Oct-14

MCORE at AGLT2• AGLT2 has supported MCORE jobs

for many months now• Condor configured for two MCORE

job types– Static slots (10 total, 8 cores each)– Dynamic slots (420 of 8 cores each)

• Requirements statements added by the “condor_submit” script– Depends on count of queued MP8 jobs

• Result is instant access for a small number with gradual release of cores for more with time.

Full details at https://www.aglt2.org/wiki/bin/view/AGLT2/MCoreSetup

QUEUED

RUNNING

Page 8: AGLT2 Site Report Shawn McKee University of Michigan HEPiX Fall 2014 / UNL

Virtualization Status

AGLT2-HEPiX 14-Oct-14

Virtualization at AGLT2Most Tier-2 services run on VMware (vSphere 5.5)UM uses iSCSI storage backends

Dell MD3600i, MD3000i and SUN NAS 7410vSphere manages virtual disk allocation between units and RAID volumes based on various volume performance capabilities and VM demand

MSU runs on DAS – Dell MD3200Working on site resiliency details

Multisite SSO operational between sites (SSO at either site manages both sites) MSU is operating site-specific Tier-2 VMs (dcache doors, xrootd, cobbler) on vSphereVMware Replication Appliance is used to perform daily replications of critical UM VMs to MSU’s site. This is working wellOur goal is to have MSU capable of bringing up Tier-2 service VMs within 1 day of loss of UM site. Queued: a real test of this process

Page 9: AGLT2 Site Report Shawn McKee University of Michigan HEPiX Fall 2014 / UNL

AGLT2-HEPiX 14-Oct-14

AGLT2 100G Network Details

Link down problematic optics

Page 10: AGLT2 Site Report Shawn McKee University of Michigan HEPiX Fall 2014 / UNL

AGLT2-HEPiX 14-Oct-14

Software-Defined Storage ResearchNSF proposal submitted involving campus and our Tier-2

Exploring Ceph for future software-defined storage

Goal is centralized storage that supports in place access from CPUs across campus

Intends to leverage Dell “dense” storage MD3xxx (12 Gbps SAS) in JBOD modeStill waiting for news…

Page 11: AGLT2 Site Report Shawn McKee University of Michigan HEPiX Fall 2014 / UNL

AGLT2-HEPiX 14-Oct-14

Update on DIIRT• At Ann Arbor Gabriele Carcassi presented on “

Using Control Systems for Operation and Debugging”• This effort has continued and is now called DIIRT (Data

Integration In Real Time)

Control System Studio

UI for operators

NFS

CSV or JSON

diirt server

Websockets+

JSON

Web pages

HTML + Javascript

scripts

dependency

data flow

Currently implemented• Scripts populate NFS directory from condor/ganglia• Files are served by diirt server through web sockets• Control System Studio can create “drag’n’drop” UI

Page 12: AGLT2 Site Report Shawn McKee University of Michigan HEPiX Fall 2014 / UNL

AGLT2-HEPiX 14-Oct-14

DIIRT UI

Canvas allows drag-n-drop of elements to assemble views, no programming requiredServer can feed remote clients in real-time. Project info at http://diirt.org/

Page 13: AGLT2 Site Report Shawn McKee University of Michigan HEPiX Fall 2014 / UNL

AGLT2-HEPiX 14-Oct-14

Future Plans• Participating in SC14 (simple WAN data-pump system)• Our Tier-3 uses Lustre 2.1 and has ~500TB

– Approximately 35M files averaging 12MB/file– We will purchase new hardware providing another 500TB.– Intend to go to Lustre 2.5+ and VERY interested in using Lustre on ZFS

for this– Plan: install new Lustre instance, then migrate existing Lustre data over,

then rebuild older hardware into the new instance, retiring some components for spare parts.

• Still exploring OpenStack as an option for our site. Would like to use Ceph for a back-end.

• New network components support Software Defined Networking (OpenFlow). Once v1.3 is supported we intend to experiment with SDN in our Tier-2 and as part of LHCONE point-to-point testbed.

• Working on IPv6 dual-stack for all nodes in our Tier-2

Page 14: AGLT2 Site Report Shawn McKee University of Michigan HEPiX Fall 2014 / UNL

Conclusion

AGLT2-HEPiX 14-Oct-14

Summary

• Monitoring is helping us easily find/fix issues• Virtualization tools working well and we are close to

meeting our site resiliency goals• Network upgrade in place 2x40G inter-site, 100G

WAN• DIIRT is a new project allowing us to customize how

we manage and correlated diverse data.• FUTURE: OpenStack, IPv6, Lustre on ZFS for Tier-3,

SDN

Questions ?