FutureGrid and US Cyberinfrastructure Collaboration with EU Symposium on transatlantic EU-U.S. cooperation in the field of large scale research infrastructures.

Similar presentations

Presentation on theme: "FutureGrid and US Cyberinfrastructure Collaboration with EU Symposium on transatlantic EU-U.S. cooperation in the field of large scale research infrastructures."— Presentation transcript:

1
FutureGrid and US Cyberinfrastructure Collaboration with EU Symposium on transatlantic EU-U.S. cooperation in the field of large scale research infrastructures 1st October 2010 CNR, Rome, Italy Geoffrey Fox gcf@indiana.edu http://www.infomall.org http://www.futuregrid.orghttp://www.infomall.orghttp://www.futuregrid.org Director, Digital Science Center, Pervasive Technology Institute Associate Dean for Research and Graduate Studies, School of Informatics and Computing Indiana University Bloomington

4
Dynamic Provisioning Results Time elapsed between requesting a job and the jobs reported start time on the provisioned node. The numbers here are an average of 2 sets of experiments. Time minutes Number of nodes

6
FutureGrid Partners Indiana University (Architecture, core software, Support) – Collaboration between research and infrastructure groups Purdue University (HTC Hardware) San Diego Supercomputer Center at University of California San Diego (INCA, Monitoring) University of Chicago/Argonne National Labs (Nimbus) University of Florida (ViNE, Education and Outreach) University of Southern California Information Sciences (Pegasus to manage experiments) University of Tennessee Knoxville (Benchmarking) University of Texas at Austin/Texas Advanced Computing Center (Portal) University of Virginia (OGF, Advisory Board and allocation) Center for Information Services and GWT-TUD from Technische Universtität Dresden. (VAMPIR) EUROPE! Red institutions have FutureGrid hardware

8
Grid Interoperability from Andrew Grimshaw Colleagues, FutureGrid has as two of its many goals the creation of a Grid middleware testing and interoperability testbed as well as the maintenance of standards compliant endpoints against which experiments can be executed. We at the University of Virginia are tasked with bringing up three stacks as well as maintaining standard- endpoints against which these experiments can be run. We currently have UNICORE 6 and Genesis II endpoints functioning on X-Ray (a Cray). Over the next few weeks we expect to bring two additional resources, India and Sierra (essentially Linux clusters), on-line in a similar manner (Genesis II is already up on Sierra). As called for in the FutureGrid program execution plan, once those two stacks are operational we will begin to work on g-lite (with help we may be able to accelerate that). Other standards-compliant endpoints are welcome in the future, but not part of the current funding plan. I’m writing the PGI and GIN working groups to see if there is interest in using these resources (endpoints) as a part of either the GIN or PGI work, in particular in demonstrations or projects for OGF in October or SC in November. One of the key differences between these endpoints and others is that they can be expected to persist. These resources will not go away when a demo is done. They will be there as a testbed for future application and middleware development (e.g., a metascheduler that works across g-lite and Unicore 6). http://futuregrid.org8 We RENKEI/NAREGI project are interested in the participation for interoperation demonstrations or projects for OGF and SC. We have a prototype middleware which can submit and receive jobs using the HPCBP specification. Can we have more detailed information of your endpoints(authentication, data staging method, and so on) and the participation conditions of the demonstrations/projects.

13
Interactions with Europe II Joint Summer Schools in areas of Computational and Data-enabled Science and Engineering – GPU’s, Petascale, Big Data, Clouds – Define “appliances” to support laboratories Study link US Cyberinfrastructure to EGI starting with setting up an “EGI node” on FutureGrid – Build on work of GIN group in OGF – Suite of Interoperability experiments between Grids and Clouds – US Europe Asia South America Studies of importance of GreenIT using Grid5000, FutureGrid and ? – Generalize to other distributed computing research http://futuregrid.org13

14
194 papers submitted to main track; 48 accepted; 4 days of tutorials including OpenNebula