SCinet
Bandwidth
Challenge Schedule
Bandwidth
Challenge Winners Announced
High-Performance
Bandwidth Challenge
Continuing the tradition started at SC2000, SCinet and Qwest Communications
are sponsoring the Third Annual High-performance Bandwidth Challenge.
For the Bandwidth Challenge, applicants from science and engineering
research communities across the globe will use the unique SCinet infrastructure
to demonstrate emerging techniques or applications, many of which
consume enormous amounts of network resources. At SC2001, a single
application consumed more than 3 billion bits per second (Gbps). Another
application creatively consumed (squandered?) significant aggregate
amounts of bandwidth by connecting to a large number of remote sites.
For SC2002, applicants
are challenged to significantly stress the SCinet network infrastructure
while delivering innovative application value on an OC-48 or higher
(OC-192!) interconnect. In turn, SCinet facilitates access to the
networks, provides technical support to applicants, and makes arrangements
for equipment and floor and rack space to applicants with demonstrable
needs.
Qwest Communications
is sponsoring the award of one or more monetary prizes for the applications
that make the most effective and/or courageous use of SCinet resources.
The primary measure of performance will be the verifiable network
throughput as measured from the contestant's booth through the SCinet
switches and routers to external connections.
Seven full proposals
were received prior to the initial application deadline. The abstracts
received with these proposals are provided here.
Bandwidth Gluttony
- Distributed Grid-Enabled Particle Physics Event Analysis over enhanced
TCP/IP
Submitted by Julian Bunn, California Institute of Technology
"Using distributed databases at Caltech, CERN, possibly UCSD
and other HEP institutes, we will show a particle physics analysis
application that issues remote database selection queries, prepares
virtual data object collections, moves those collections across the
WAN using specially enhanced TCP/IP stacks, and renders the results
in real time on the analysis client workstation in Baltimore. This
scheme is a preview of a general "Grid Enabled Analysis Environment"
that is being developed for CERN's LHC experiments. It makes use of
modified TCP with Adaptive Queue Management to achieve excellent throughput
for large data transfers in the WAN."
Bandwidth to
the World
Submitted by Les Cottrell, Stanford Linear Accelerator Center
"The unprecedented avalanche of data already being generated
by and for new and future High Energy and Nuclear Physics (HENP) experiments
at Labs such as SLAC, FNAL, KEK and CERN is demanding new strategies
for how the data is collected, shared, analyzed and presented. For
example, the SLAC BaBar experiment and JLab are each already collecting
over a TByte/day, and BaBar expects to increase by a factor of 2 in
the coming year. The Fermilab CDF and D0 experiments are ramping up
to collect similar amounts of data, and the CERN LHC experiment expects
to collect over ten million TBytes. The strategies being adopted to
analyze and store this unprecedented amount of data is the coordinated
deployment of Grid technologies such as those being developed for
the Particle Physics Data Grid and the Grid Physics Network. It is
anticipated that these technologies will be deployed at hundreds of
institutes that will be able to search out and analyze information
from an interconnected worldwide grid of tens of thousands of computers
and storage devices. This in turn will require the ability to sustain
over long periods the transfer of large amounts of data between collaborating
sites with relatively low latency. The Bandwidth to the World project
is designed to demonstrate the current data transfer capabilities
to several sites with high performance links, worldwide. In a sense
the site at SC2002 is acting like a HENP tier 0 or tier 1 site (an
accelerator or major computation site) in distributing copies of the
raw data to multiple replica sites. The demonstration will be over
real live production networks with no efforts to manually limit other
traffic. Since, by turning off the 10Gbps link, we will be able to
saturate our 1 Gbps link to SCinet and control the router in our booth,
which will be at one end of the congested link, we also hope to be
able to investigate/demonstrate the effectiveness of QBone Scavenger
Service (QBSS) in managing competing traffic flows and on the response
time of lower volume interactive traffic on high performance links."
Data Services
and Visualization
Submitted by Helen Chen, Sandia National Laboratories
"The ability to do effective problem setup and analysis of simulation
results is critical to a complete, balanced problem-solving environment
for ASCI. Research scientists and analysts must be able to efficiently
design complex computational experiments, as well as be able to see
and understand the results of those experiments in a manner that enables
unprecedented confidence in simulation results. We demonstrate two
visualization techniques that utilize a 10Gbps network infrastructure
to deliver robust data services for the management and comprehension
of very large and complex data sets. The LBL Visapult application
visualizes a large scientific data set that is stored in an InfinARRAY
File System (IAFS). IAFS aggregates, over TCP/IP, the storage of a
number of physical file systems into a single virtual file system.
Its performance is determined by the interconnect bandwidth, and the
number of parallel storage processors in the system, thereby offering
a very scalable solution for servicing scientific supercomputing.
A possible alternative to IAFS is to employ remote storage, using
the emerging IP storage technology such as iFCP and iSCSI as the data
source to Visapult. The Sandia V2001 is an Interactive Video System
that transports high-resolution video over local- or wide-area IP
networks. A V2001 Encoder connects to the DVI or RGB video from a
computer as if it were a flat panel monitor. It delivers the video
signal presented at its video port over Gigabit Ethernet. At the visualization
end, a V2001 Decoder displays the received signal to a panel monitor
connected via its DVI or RGB port. The V2001 Decoder provides low-latency
interactivity on remote images using an USBbased keyboard and mouse."
Global Telescience
featuring IPv6
Submitted by Mark Ellisman, National Center for Microscopy and Imaging
Research (NCMIR)
"The NCMIR Lab at the University of California (UCSD) and the
San Diego Supercomputer Center (SDSC) intends to demonstrate another
"real" scientific application utilizing native IPv6 and
a mixture of high bandwidth and low latency. In our demonstration
we will feature a network enabled end-to-end system for 3D tomography
utilizing network resources to remotely control an Intermediate Voltage
Electron Microscope, transfer data to remote storage resources and
complete compute jobs on distributed heterogeneous computer systems.
The process of finding features using the microscope is a visually
guided task in which users must distinguish features in a low contrast
high noise medium. The nature of this task calls for the highest video
quality possible when navigating the specimen in the microscope. To
address these challenges, the Telescience system is actively integrating
digital video over native IPv6 networks providing high quality low
latency video for use in navigation of the specimen in the microscope.
Similar to our presentation last year, we will continue to improve
upon past achievements. This year's demo will feature higher bandwidth
usage as well as other technological improvements."
Grid Datafarm
for a HEP Application
Submitted by Osamu Tatebe, National Institute of Advanced Industrial
Science and Technology (AIST)
"This is a high-energy physics application that simulates the
ATLAS detector, which will be operational by 2007 at CERN. Currently,
six clusters, three in US and three in Japan, comprise a cluster-of-cluster
filesystem (Gfarm filesystem). The FADS/Goofy simulation code based
on the Geant4 toolkit simulates the ATLAS detector and generates hits
collection (raw data) in this Gfarm filesystem. Physically, each cluster
generates the corresponding part of hits collection, and stores it
to its cluster filesystem, which will be replicated to all other cluster
filesystems, although every operation is uniformly performed on the
Gfarm cluster-of-cluster filesystem. Replicated files will be transparently
accessed, and used for fault tolerance and load balancing."
Real-Time Terascale
Computation for 3D Tele-Immersion
Submitted by Herman Towles, University of North Carolina at Chapel
Hill
"Tele-Immersion, an ability to share presence with distant individuals,
situations and environments, may provide vast new possibilities for
human experience and interaction in the near future. Combining correlation-based
3D reconstruction and view-dependent stereo display technologies,
the University of Pennsylvania and the University of North Carolina
have previously demonstrated a prototype 3D tele-immersion system
running over Abilene between Philadelphia and Chapel Hill. Working
with the Pittsburgh Supercomputing Center, we are now working to harness
the massive computational power of PSC s new Terascale Computing System
(lemieux) to improve our reconstruction volume, reconstruction quality
and frame rate performance. We propose to demonstrate 3D tele-immersion
using the remote computational resources of the Pittsburgh Supercomputing
Center to do high-quality, 3D reconstruction of a dynamically changing,
room-size environment at SC2002 and to display it with view-dependent
stereo at a nearby office-like booth. We will set up both these environments
at SC2002 an office for acquisition and an office with display. In
the latter office, there will be a large projective stereo display
providing a live, 3D portal into the acquisition office. The display
will use passive stereo technology with position tracking of the user
for view-dependent rendering. The first office and occupant(s) will
be acquired with an array of up to 45 cameras. Raw 2D images from
the acquisition office (at SC2002 in Baltimore) will be transported
to PSC (in Pittsburgh) and processed into a 3D scene description in
real-time. This 3D scene description will then be transported back
to the display office (at SC2002 in Baltimore) for 3D rendering. In
addition to this remote interactive compute mode, we will provide
a local reconstruction mode (that runs without PSC) and a local playback
mode of pre-computed reconstructions for those times when remote computing
or networking resources are not available."
Wide Area Distributed
Simulations using Cactus, Globus and Visapult
Submitted by John Shalf, Lawrence Berkeley National Laboratory
"We will perform a "hero calculation" of unprecedented
scale that will consume unprecedented amounts of network bandwidth.
The calculation will model gravitational waves generated during the
collision of black holes. A single simulation will be distributed
amongst several MPP supercomputers at several sites. The distributed
simulation will send simulation results over multiple high-capacity
network links to the SC02 show floor for visualization and analysis.
We expect this year's entry to set new records for extreme bandwidth
consumption. This year's entry builds on the effort of our winning
SC2001 entry in several important aspects. We will demonstrate a system
that provides tracking, remote monitoring, management interfaces,
and high performance parallel visualization of supercomputing simulations
that span across multiple parallel supercomputers on the Grid. We
will also demonstrate both fault-resilient unreliable protocols and
even some custom reliable protocols are capable of using all available
networking resources with a high degree of efficiency. This endeavor
will lay the groundwork for the kind of tools necessary for making
efficient use of high performance research networks to support Grid
and metacomputing activities. The entry will consist of the Cactus
code framework, the Visapult parallel visualization system and Globus
Grid infrastructure. The Cactus code will run across multiple parallel
supercomputers at different sites including NERSC, LBNL, Sandia, Poland
and other similar resources using MPICH-G2 for MPI and Globus GRAM
for job launching/management. The Cactus steering infrastructure will
be used to remotely control the running code and the Visapult system
will provide extremely high performance parallel visualization of
the current physics being evolved by Cactus using cluster computers
located on the SC2002 show floor."
Data Reservoir
Submitted by Kei Hiraki, University of Tokyo
Data Reservoir is a data sharing system that uses very high-speed
internet connections (up to 100Gbps) between distant locations. The
Data Reservoir utilizes the low-level iSCSI protocol and has filesystem
transparency.
Parallel data transfer with hierarchical data striping is a key factor
in achieving full bandwidth of the highspeed network. Our system has
two nodes that are connected by a 10Gbps link, and each node consists
of dozens of 1U IA-servers and a 10Gbps-capable switch. Our software
environment is RedHat Linux and uses NFS and the iSCSI driver.
Below is
a schedule for presentation of the High-Performance Bandwidth Challenge
Entries. All demonstrations will take place on the exhibit hall show
floor.
| Tuesday,
November 19 |
| 2:00pm |
Entry 1 |
Kei Hiraki,
"Data Reservoir" |
| 3:00pm |
Entry 2 |
Mark Ellisman,
"Global Telescience Featuring IPv6" |
| 4:00pm |
Entry 3 |
currently
unused |
| Wednesday,
November 20 |
| 11:00am |
Entry 4 |
Osamu Tatebe,
"Grid Datafarm for a HEP Application " |
| 12:00pm |
Entry 5 |
Les Cottrell,
"Bandwidth To The World" |
| 1:00pm |
Entry 6 |
Herman Towles,
"Terascale Computation for 3D Tele-Immersion" |
| Wednesday,
November 20 |
| 2:00pm |
Entry 7 |
Helen Chen,
"Data Services and Visualization" |
| 3:00pm |
Entry 8 |
Julian Bunn,
"Bandwidth Gluttony" |
| 4:00pm |
Entry
9 |
John
Shalf "Simulations using Cactus, Globus and Visapult" |
| SCinet
will attempt to take all measurements during the above time slots.
In case network or operational problems are encountered, the judges
will set aside additional time slots on Tuesday and Wednesday
evening. |
| Thursday,
November 21 |
| 8:30am -
11:30am |
Judging |
| 11:30am -
1:00pm |
SC2002
Netcasting |
|
Please
keep bandwidth utilization to a MINIMUM between 11:30am and
1:00pm due to SC2002 Netcasting. SCinet will be watching!
|
| 1:30pm |
Awards
Ceremony |
| 3:00pm |
All
entries are encouraged to run simultaneously. |