GeoWulf

UCDavis
Current status:


History:

1998

February: My first parallel cluster computer (@Clarkson):
1999

September: My second parallel machine (first @UCDavis) (First 4 nodes have been installed. By the end of the school year we'll have 24 nodes. Moreover, we are working in close collaboration with Professor Kleeman and will be able to connect GeoWulf with his Beowulf Machine and have total of more than 128 computational nodes by fall of 2000.

October: 8 nodes installed and the front-end computer just arrived. MPI 1.1.2. installed.
November: Portland Group compilers installed; Parasoft insure and codewizard installed;


2000

January: New MPI 1.2 installed and tested. Minor problems with front-end computer fixed. Mysterious networking problem fixed (cron job pings outside world every 15 minutes).

February: Single pile and pile groups jobs runing in parallel on GeoWulf.



March: Memory upgrade for the controller machine. 0.5GB added. Controller machine now has 1GB of RAM memory.

October: BIG1 disk (IBM Ultrastar 34.6GB SCSI) is failing on main machine. Still holding on, but it is quite noisy and reporting problems frequently.

November: Eight more nodes (AMD Duron 750MHz, 128MB RAM, 20GB disk) + service node (AMD Duron 750MHz, 128MB RAM, 2 x 46GB disks) ordered.

December: Nodes arrived.


2001

January: Putting together upgrade nodes. Upgrade team: Mark Olton, Frank McKenna, Zhaohui Yang and Boris Jeremic.



February: System installed on upgrade computers. GeoWulf now consists of 16 heterogenous node computers based on the Intel Pentium II (8 nodes) and AMD Duron (8 nodes) processors, one controller computer (dual Pentium III) and one service computer (AMD Duron). It features: over 3GB of distributed memory and over 420 GB of distributed disk space.

February: BIG1 disk almost unusable, need to get replacement.

May: BIG1 disk replaced.

June: Service machine upgraded to a full workstation. To be used for testing firewalls, backup systems and various other things.

October: Design of new head machine Koyaanisquatsi (Dual AMD 1.2MHz, 2GB RAM, 36GB SCSI disk...)

December: Koyaanisquatsi complete, problems with RedHat (unsuported hardware?)




2002

Februar: SuSE installed instead of RedHat, still problem with the last 0.5GB RAM chip.

May: Installing globus on the cluster.

July: Reinstalling nodes with SuSE.

August: reinstalling MPI for globus device, testing of OpenSeesGrid in progress.




2003

April: Geomechanics added to the head section of the cluster.

May: Reinstalling nodes with RedHat.

June: GeoWulf and Geomechanics upgraded to dual AMD 2400 with 2GB of RAM each. Current setup is as follows: 16 heterogenous node computers based on the Intel Pentium II (8 nodes) and AMD Duron (8 nodes) processors, two controller computer (dual AMD 24000) and one service computer (AMD Duron). The system features: over 6GB of distributed memory and over 420 GB of distributed disk space.

September: 35GB SCSI disk drive replaced BIG1 on GeoWulf (disk failure?).

July - September: Koyaanisqatsi machine becomes RAIDS server (over 0.5TB).


2004

February: Fried CPU and motherboard on Koyaanisqatsi.

March: Disk failure on Koyaanisqatsi.

June: GeoWulf moved next door to a airconditioned room (temperature is kept below 18 degrees centigrade).

November: Backup USB disk (115GB) added for backup.


2005

May: Shunya SMP machine (4xIntel, Dual CPU, double core) machine added to the GeoWulf Cluster(Prof. Kunnath)

September: Initial upgrade to GigB network.


2006

March: Construction of 4 new nodes, dual core Pentium D with 2GB each... This was in part class project (Computational Geomechanics, ECI285)



June: System hacked (due to a little stupid omission). Operating system reinstall from scratch (fedora core 5). Recompilation of all numerical and message passing libraries to optimize performance.

July: New 4x2 core "TeachingMachine" connected to the system.

October: Deployment of 4 new nodes (dual core, Intel core-duo technology), each one with 4GB of memory.

December: Current GeoWulf system (note much smaller "footprint", yet the most powerfull machine so far).


2007

July: Addition of another head node (dual CPU, 4GB RAM, 400MB disc

August: problem with cooling in the computer room...


2008

April: Cooling problems, computer nodes survived 12 hours of over 60C...

May: implementation of alternative (additional) air flow (cooling)


2009

September: Adding/replacing nodes, going for many multi core CPUs as well as GPGUs

November: TESLA1060 is here (together with new Koyaanisqatsi machine, Xeon E55520 Quad core, 2.26GHz, 12GB RAM, 858GB disk RAID)



Boris Jeremić