Super Computing

Advanced research in 3-D reflection seismology requires the solution of large computational and data intensive problems. SEP has a long tradition of early adoption and effective utilization of the leading edge scientific computers available (PDP-11 - Vax 780 - Convex C1 - CM-5 - Power Challenge).

The late 1990s saw a shift from large shared memory machines to inexpensive PC hardware formed into clusters running Linux (often referred to as "beowulf" clusters). These new machines have the advantage of providing inexpensive computational power, but normally require considerably more programming knowledge for full utilization.

In 1999 SEP bought five four-processor 1400L boxes from SGI. In 2001, SEP bought a 16 node dual processor machine, omu from Linux Networks. In 2001-2002 expanded its computational and storage capacity further by building its own Linux machines out of components. We built a 2TB disk server, koko and a 32 node, 64 processor cluster, okok.

In 2003 SEP started a Linux Cluster Initiative with some of its industrial sponsors. This initiative proved funds for three additional Beowulf clusters from California Digital . The first cluster, bought in 2003, consisted of 40 dual 2.4 xeon processors with 2 GBs of ram per node. For applications that are memory intensive in 2003 we also bought an 8 node, dual 2.4 xeon cluster with 6 gigabytes of RAM (the maximum practical memory for a 32 bit chip). Our memory needs continue to grow so in January 2005 we bought a 32 node, dual Xeon64 cluster. Half of the nodes with 4 GB of ram, the other half with 8 GB.

SEP researchers use these computer facilities to develop parallel applications to solve computationally intensive problems, such as:

  • 2-D and 3-D prestack and poststack imaging and inversion
  • 2-D and 3-D seismic tomography.
  • 2-D and 3-D acoustic and elastic modeling
  • 3-D seismic visualization
  • 2-D and 3-D shear wave processing
  • 2-D and 3-D multiple elimination
  • 4-D seismic processing

To find out more about the research performed on these topics you can look at SEP student and faculty research pages.

Research projects are usually developed using SEP's seismic processing system (a.k.a. SEPlib). SEP provides a unique opportunity to learn first-hand about all aspects of scientific computing, including system building and design, system administration, and software development.

If you are interested in scientific computing, and parallel computing in particular, you may find it interesting to look to the WWW pages of these other research programs at Stanford: the Scientific Computing/Computational Mathematics program and the FLASH project.

sep/about/computing.txt · Last modified: 2015/05/27 02:06 (external edit)
CC Attribution-Share Alike 4.0 International
Driven by DokuWiki Recent changes RSS feed Valid CSS Valid XHTML 1.0