1 / 17

Building Cyberinfrastructure into a University Culture

Building Cyberinfrastructure into a University Culture. EDUCAUSE Live! March 30, 2010 Curt Hillegas Director, TIGRESS HPC Center Princeton University. Context.

oistin
Download Presentation

Building Cyberinfrastructure into a University Culture

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Building Cyberinfrastructure into a University Culture EDUCAUSE Live! March 30, 2010 Curt Hillegas Director, TIGRESS HPC Center Princeton University

  2. Context As a world-renowned research university, Princeton seeks to achieve the highest levels of distinction in the discovery and transmission of knowledge and understanding. At the same time, Princeton is distinctive among research universities in its commitment to undergraduate teaching. www.Princeton.EDU

  3. Cyberinfrastructure Cyberinfrastructure consists of computational systems, data and information management, advanced instruments, visualization environments, and people, all linked together by software and advanced networks to improve scholarly productivity and enable knowledge breakthroughs and discoveries not otherwise possible. Developing a Coherent Cyberinfrastructure from Local Campus to National Facilities: Challenges and Strategies A Workshop Report and Recommendations

  4. Contents • Yesterday • Today • Tomorrow • Lessons

  5. 2002-The Beginning • OIT, Academic Services – Computational Science and Engineering Support group (CSES) • The Princeton Institute for Computational Science and Engineering Support (PICSciE) • Research Computing Advisory Group (RCAG)

  6. Trust • AdrOIT – small Beowulf cluster • Princeton Software Repository • Maintenance of old systems • Elders image – rebuilt RHEL distribution

  7. Relationships • Hire new faculty • New PICSciE Director, Prof. Jerry Ostriker • 1st annual RCAG Presentation • Wow!!! • Let’s buy something big together • Collaborative selection process • Cobble together funding

  8. Partnerships • 64 processor SGI Altix BX2, Hecate • 1024 node IBM BlueGene/L, Orangena • Faculty • OIT • Development • Facilities • Collaborative administration (without fees) • Housed in central Data Center (without fees) • 256 node Dell Beowulf cluster, Della • HPC Steering Committee

  9. Storage • Hire a pair of junior faculty • 192 node Beowulf cluster including startup • 35 TB • Fees to recover 50% of capital cost • 10% utilization within the first 6 months • No fees!!! • 95% utilization within 4 months

  10. Hierarchical Storage Management • 96 node SGI Altix ICE, Artemis • RCAG – Need a scalable storage system that provides appropriate performance and availability for aging data • OIT proposal to Provost’s office • 1 PB total • IBM DS4800, GPFS, GPFS HSM, TSM HSM • Added benefit – free backups!!! • All systems have high performance access

  11. Success Brings New Challenges • Senior faculty hire • 448 node Dell cluster • New scheduling policies • (Supercomputing|Scientific Computing) Administrators Meeting – SCAM • DataSpace • New PICSciE Director – Prof. Jeroen Tromp

  12. Visualization • Visualization Expert • Sony SRX-S110 Projector • 8,847,360 pixels (4096x2160) • Rear projection ultra wide angle fabric screen • 9’3” (H) X 16’6” (W) • Open Source and Proprietary software

  13. Cyberinfrastructure at Princeton Programming Support Research Computing Base HPC Hardware Software Infrastructure Visualization Collaboration Storage Infrastructure

  14. HPC Hardware

  15. Collaboration • PICSciE • RCAG • TIGRESS Steering Committee • TIGRESS Users • SCAM • PLUG

  16. Future • New Data Center • Coordinated supervision of departmental scientific/Linux system administrators • Collaboration with the Library • Lifecycle management • New technologies • GPGPU • Power7 • X86_64 based single image • Participation in Virtual Organizations

  17. Lessons Learned • Research is driven by the faculty • Trust, relationships, and partnerships are essential to success • Research computing relies on the complete cyberinfrastructure of the University • Avoid fees • Change is a constant • Start small, do things well, and growth will follow • It’s all about the people

More Related