SC08: Day 0 part 3

This show is for the birds …. the pidgeons that is … the ones wandering near the booth. I hope that … er … ah … nothing gets into the machines … Not exactly bugs, but it is possible that someone could tell you that your machine is full of pidgeon droppings … and mean … Read moreSC08: Day 0 part 3

SC08: Day 0 part 2

Ok, we have the machines set up in the Pervasive Software booth, #203.
Had a power hiccup (e.g. Joe knocked the power out while moving the rack) so we used this as an excuse to reconfigure the RAID to its RAID10 state. There really was no advantage to the RAID0 version, and the risk of problems was higher. RAID should be resynched in another 53 minutes.
We have a single 10 GbE handling all the traffic … looks like Windows didn’t like sending traffic simultaneously on both links.

Read moreSC08: Day 0 part 2

SC08: Day 0 … monday morning …

Here we are, sitting in that most important spot, by the coffee, waiting for our cohorts and colleagues … Austin is a nice place… the SC08 map is kinda … I dunno … a little small-ish? Some of us [old : Dougs suggestion, talking about me] folks can’t quite read it …

SC08: Come see ΔV3 in Pervasive Software's booth

Make sure you look at their data mining demo. DataRush (as indicated in the previous post) is a cool technology, and we are happy to be helping out.
Pervasive Software has a vision for data intensive HPC that aligns well with what we have been saying. Personal supercomputing has been something we have been talking about for about 8 years, since I developed CT-BLAST. That was a tool to completely hide the pain of dealing with clusters for running one application, NCBI BLAST. Later I developed a more refined methodology to accelerate multiple applications, by dividing up the data set, distributing it to a parallel resource with a job scheduler, and recombining on the back end. The point here was to make this a general framework to accelerate informatics applications … apps without significant data dependencies per iteration or instance, and little/no communication between iterations.

Read moreSC08: Come see ΔV3 in Pervasive Software's booth

Another side of HPC: data intensive HPC

We have called this other names in the past, but basically data intensive HPC is pretty much anything that involves streaming huge amounts of data past processing elements to effect the calculation or analysis at hand. This type of HPC is not usually typified by large linear algebra solvers, so things like HPCC and LINPACK are less meaningful characterizations for data intensive performance, as this often relies upon significant IO firepower, as well as many cores. You can’t start with a slow disk or network and assume you will be able to handle a data intensive calculation.
This is in part why we developed JackRabbit and now ΔV. We wanted to provide very large pipes to disk and network with JackRabbit, and very affordable performance with ΔV. As it turns out, ΔV is directly relevant to what I write here, and I will tell you why at the end of this post.

Read moreAnother side of HPC: data intensive HPC

Looks like the updated security measures are holding …

Quite early in the process, but we see probes have dropped off. Some of this may be due to the IP level (draconian) restrictions.
The outline of the new security measures are as follows:
1) The user has to have a valid VPN certificate to ssh to the system.
2) Users cannot share a certificate. This isn’t simply policy, it is enforced on a technological level.
3) outgoing traffic is pretty much restricted to VPN, and set specific ports. If there is abuse of those ports, they will lose them.

Read moreLooks like the updated security measures are holding …