Upload
domenic-stephens
View
213
Download
0
Embed Size (px)
Citation preview
EGEE-II INFSO-RI-031688
Enabling Grids for E-sciencE
www.eu-egee.org
EGEE Site Architecture
Resource Center Deployment Considerations
MIMOS EGEE Tutorial
2
Enabling Grids for E-sciencE
EGEE-II INFSO-RI-031688
Basic Site Architecture
• User Interface (UI)– User login environment– gLite clients
• MON– R-GMA Server for accounting
• SE (Disk Pool Manager)– Storage resource services
• Computing Element (CE)– Gateway to computing resources– Small sites will also install:
Site-BDII Batch system manager NFS file system for VO software
• Worker Node (WN)– Job execution machine– gLite clients used by jobs
3
Enabling Grids for E-sciencE
EGEE-II INFSO-RI-031688
Central Services
• BDII– Top level information system
service– Available regionally
• Resource Broker– RB or new WMS– Job management– Only RB available regionally
(Dec 2007)
• VO Services– One service for each VO– LCG File Catalogue
Maps VO’s logical file names to physical file names
– VO Management Service Manages list of VO members
4
Enabling Grids for E-sciencE
EGEE-II INFSO-RI-031688
Information System
• Top Level BDII on BDII
• Site BDII running on CE
• GRIS on services– RB/WMS– LFC– SE– CE
5
Enabling Grids for E-sciencE
EGEE-II INFSO-RI-031688
Additional Services
• R-GMA– Central Registry and Schema server
• MyProxy– Dedicated node for better security
• File Transfer Service– Designed to manage reliable file transfer– Create “channels” to with associated transfer parameters to
optimize and tune transfers Concurrent files and number of parallel streams
• AMGA
6
Enabling Grids for E-sciencE
EGEE-II INFSO-RI-031688
Network Considerations
• Grid Services– Public IP Required by each Grid service– Forward and reverse DNS configuration
• Worker Node– Public IP for parallel stream file transfer– Private IP is possible
Single stream transfer for WNs to SE Possible bottleneck via NAT/Firewall device
• Storage Elements– Bandwidth to and from Worker Nodes– Bandwidth to WAN Network
• Firewall requirements– https://twiki.cern.ch/twiki/bin/view/LCG/LCGPortTable
7
Enabling Grids for E-sciencE
EGEE-II INFSO-RI-031688
Hardware Requirements
• Minimum: only for very small sites
• Resource Broker– At least 20 GB Disk
• Worker Node– Depends on applications– X GB scratch space for each job– X MB Memory per job
• Large sites: +100 WNs– SMP or multi-core servers for CE and BDII– Install Site-BDII, Batch server and NFS server on dedicated node
Spec Minimum Recommend
CPU P3 500Mhz P4 2Ghz
RAM 256 MB 1GB
Disk 10 GB 30 GB
Network 100 Mbps 1 Gbps
8
Enabling Grids for E-sciencE
EGEE-II INFSO-RI-031688
OS and Middleware Installation
• gLite is certified on Scientific Linux CERN– But should work on RHEL binary compatible distributions
Include SLC yum/apt repository Mirror SLC and gLite repository for faster Installation
• Currently porting from SLC3/i386 to SLC4/i386– SL3/gLite 3.0: MON, DPM, LFC, VOMS, WMS, RB– SL4/gLite 3.1: BDII, lcgCE, WN, UI
• X86_64 is the next port being worked on
9
Enabling Grids for E-sciencE
EGEE-II INFSO-RI-031688
Additional Servers Requirements
• Installation of Java SDK– Installed separately due to licensing restrictions– RPMs packages required to resolve dependencies of Middleware– Java SDK 1.4.2 for glite 3.0– Java SDK 1.5 for glite 3.1
• Synchronize server time– Configure Network Time Protocol (NTP) for every server– Required by GSI security– Configure time zone and hardware clock to UTC
Troubleshooting and comparing log files across time zones
• Host certificates are required on all services– Except for UI, WN and BDII– Mutual authentication by clients
10
Enabling Grids for E-sciencE
EGEE-II INFSO-RI-031688
Thanks for you attention
• Questions