Monthly Archives: February 2011
I do mention in my tagline that this blog may contain some information about computers, so you’ve been warned. The following article is quite technical, and may not be of interest to the photography contingent that might otherwise be interested in my blog articles. I posted this because I spent several days attempting a particular task, and managed to get it down to a few simple instructions. It may be useful for other attempting the same task.
I’ve recently set up an Infiniband fabric at home, and had a lot of trouble getting it working on Ubuntu. Windows 7 was a breeze, I just instlled the OFED drivers from openfbrics.org.
Here’s the steps to get it working on Ubuntu 10.10. It might not seem a lot, but there was a lot of messing to get it down to these few steps.
Install Ubuntu. Everything below is done as root – “sudo bash”.
Add a file to /etc/udev/rules.d called, say, 99-udev-umad.rules. This will cause the correct entries to be created in /sys, which otherwise do not get created for Ubuntu 10.10.
Insert the following:
KERNEL==”umad*”, NAME=”infiniband/%k”, MODE=”0666″
KERNEL==”issm*”, NAME=”infiniband/%k”, MODE=”0666″
Edit /etc/modules and add the following modules:
Next, “apt-get install opensm”. This will install the subnet manager and all the relevant dependencies, libibverbs, etc.
Then add the relevant entries for the interface into /etc/network/interfaces file:
iface ib0 inet static
Then reboot. This will create the relevant infiniband entries in /sys, load the ipoib modules, and bring up the infiniband port with an ip address.
You should now have a functioning infiniband port on your Ubuntu machine.
However, there is still some more investigation to be done. Initially, when I was mixing custom kernels with OFED drivers, and stock linux kernel drivers, netperf was showing 7 gbps throughput. With the configurtation above, it’s down at about 25 mbps. Dreadfully slow. I’ll have to find out what optimisations are needed (or other drivers) in order to get the speed back up to 7 gbps.
Note: iperf maxed out at 1.2 gbps, and on the current linux install, I couldnt get netperf client working at all. netserver would work, but only showed a throughput of 25mbps from a Win7 client. HOWEVER, when I set up the raid with 6 old 160G drives, the “hdparm -t /dev/md0p1″ showed 250MB/sec reads, and I got the same from the Win7 machine using samba across the infiniband fabric. This seems to indicate that iperf and netperf are completely unreliable for testing this type of connection. Bear in mind though that I did have netperf running on the previous ubuntu install, but that installation was so messy I don’t know what drivers and user-space software was running. I reckon it’s the kind of think that may be fixed in the stock Ubuntu install in the near future. For the moment, just go with real-world testing, i.e. copying large files from ramdisk to ramdisk, for example.
I just got the following great news. Mark H. McCormick-Goodhart, Director of the ink longevity research programme at http://www.aardenburg-imaging.com/news.html has just announced that the subscription fees to the website are being dropped, making all the research available to everyone for free. I’ve been a member since the very early days, and found in invaluable as a printer in selecting (or ignoring) third party inks for use in my printers.
I’m actually in the proud position of having the WORST EVER performing sample in that programme. An Ink I played around with in the early days was a dye ink off eBay called “Signal Inkjet”. I got one batch of 600ml (6x100m bottles), which cost me $10. It seemed ridiculously cheap, and sure enough, the results matched the price. The prints were only good for a few months before visible fading and colour shifting kicked in. Once I saw the results coming out of Aardenburg-Imaging, I quickly switched back to Claria OEM inks (I was on a dye printer at the time) which was showing much better results.
I eventually got and Epson R2880, and as a result of this research, I settled on Inkjetfly inks, as they have longevity and colour gamut approaching that of OEM inks for a fraction of the price. I’m printing all of my on photos now (and also for some friends, club competitions, etc.) and we’re very happy with the results.
Also, if you’re interested in the research, you can also submit your own samples. It’s very interesting to see your own samples put through the light-fade tests, and you can be confident that the results are accurate and measured in a way that can be compared consistently with the other samples under test.
Have a look at Mark’s website, and take a look at the longevity test results. It lists a huge amount of printer/paper/ink combinations, invaluable for someone who’s looking to use third party inks at either an amateur or professional level.
I was asked to take some shots for Scene Hairdressing in Shannon recently. The were having a bridal theme day where models would come in to the salon, a make-up artist would take care of the make up, the Salon staff looked after the hair, all I had to do was press the shutter release. Well, sort of.
So early Sunday morning we arrived at the salon with multiple strobes, tripods, backdrops, triggers, etc. By the time we were ready with two setups (one white backdrop, one grey backdrop), the first model was almost ready. So we started shooting while the other model was getting prepared.
Because the hair and makeup took about an hour each time, we got to shoot the two models in two separate hair styles over about 5 hours. Below is a small selection of shots from the day. Most have had a background texture added in post.
More shots to be ‘scene’ here. 😉
Would you like to have over 700MB/sec throughput between your PCs at home for under €110? That’s like a full CD’s worth of data every second! If you do, then read on….
Since this article was originally written, I’ve found the real-world throughput of infiniband from a windows machine and an ubuntu machine gives me a max of 135MB/sec, just under twice my 1gbps ethernet (75MB.sec). Thats with a raid array capable of 350MB/sec on the linux side, feeding a samba link to the windows machine at 95% CPU. So, it falls a lot short of the desired 700MB/sec that I thought may be possible. It’s not possible with IP over Infininband. And iSER isnt available on windows, so no SRP targets could be used, which uses RDMA. So a whole lotta research leading to block walls and 135MB/sec max.
With the increasing amout of data that I have to manage on my computers at home, I started looking into a faster way of moving data around the place. I started with a RAID array in my PC, which gives me read-write speeds of 250MB/sec. Not being happy with that, I looked a creating a bigger external array, with more disks, for faster throughput. I happened to have a decent linux box sitting there doing very little. It had a relatively recent motherboard , and 8 SATA connectors. But no matter how fast I got the drives in that linux box to go, I’d always be limited by the throughput of the 1Gb ethernet network between the machines, so I researched several different ways of inter-PC communication that might break the 1gbps barrier. The 1GB ethernet was giving me about 75MB/sec throughput.
The first I looked at was USB 3.0 (5 gbit/s). While that’s very good for external hard drives, there didnt seem to be a decent solution out there for allowing multiple drives to be added together to increase throughput. We are now starting to see raid boxes appear with USB3.0 interfaces, but they are still quite expensive. To connect my existing linux box to my windows desktop, I’d need a card with a USB 3.0 slave port so that the external array would look like one big drive, and max out the 5Gbps bandwidth of a USB 3.0 link . However, these do not seem to exist, so I moved onto the next option.
Then I moved on to 10G Ethernet (10 gbit/s). One look at the prices here and I immediately ruled it out. Several hundred Euro for a single adapter.
Fibre channel (2-8 gbit/s). Again the pricing was prohibitive, especially for the higher throughput cards. Even the 2Gbps cards were expensive, and would not give me much of a boost over 1Gbps ethernet.
Then came Infiniband (10-40 gbit/s). I came across this while looking through the List of Device Bit Rates page on Wikipedia. I had heard of it as an interconnect in cluster environments and high-end data-centres. I also assumed that the price would be prohibitive. A 10G adapter would theoretically give up to a Gigabyte per second throughput between the machines. However, I wasn’t ruling it out until I had a look on eBay at a few prices. To my surprise, there was a whole host of adapters available ranging from several hundred dollars down to about fifty dollars. $50? for a 10Gig adapter? Surely this couldn’t be right. I looked again, and I spotted some dual port Mellanox MHEA28-XTC cards at $35.99. This worked out at about €27 per adapter, plus €25 shipping. Incredible, if I could get it to work. I’d also read that it is possible to use a standard infiniband cable to directly connect two machines together without a switch, saving me about €700 in switch costs. If I wanted to bring another machine into the Infiniband fabric, though, I’d have to bear that cost. For the moment, two machines directly connected was all I needed.
With a bit more research, I found that drivers for the card were available for Windows 7 and Linux from OpenFabrics.org, so I ordered 2 cards from the U.S. and a cable from Hong Kong.
About 10 days later the adapters arrived. I installed one adapter in the Windows 7 machine. Windows initially failed to find a driver, so I then went on the OpenFabrics.org website and downloaded OFED_2-3_win7_x64.zip. After installation I had two new network connections available in windows (the adapter was dual-port), ready for me to connect to the other machine.
Next I moved onto the Linux box. I won’t even start with the hassle I had to install the card in my linux box. After days of research, driver installation, kernel re-compilation, driver re-compilation, etc. etc., etc., etc., I eventually tried swapping the slot that I had the card plugged into. Low and below, the f&*cking thing worked. So, my mother board has two PCI-Ex16 slots, and the infiniband adapter would work in one, but not in the other. Who would have thought. All I had to do then was assign an IP address to it. –EDIT– here’s a quick HOWTO on getting the fabric up on Ubuntu 10.10. About 10 minutes should get it working – http://davidhunt.ie/wp/?p=375 –EDIT–
Without a cable (it still had not arrived from Hong Kong), all I could do was sit there and wait until it arrived to test the setup. Would the machines be able to feed the cards fast enough to get a decent throughput? On some forums I’d seen throughput tests of 700MB/sec. Would I get anywhere close to that with a 3GHz dual core athlon to a 3GHz i7 950?
A few days later, the cable arrived. I connected the cable into each machine, and could immediately send pings between the machines. I’d previously assigned static IP addresses to the infiniband ports on each machine. I wasn’t able to run “netperf”, as it didn’t see the cards as something it could put traffic through. So I upgraded the firmware on the cards, which several forums said would improve throughput and compatibility. Iwas then able to run netperf, with the following results:
root@raid:~# netperf -H 10.4.12.1 TCP STREAM TEST from 0.0.0.0 (0.0.0.0) port 0 AF_INET to 10.4.12.1 (10.4.12.1) port 0 AF_INET : demo
Recv Send Send Socket Socket Message Elapsed Size Size Size Time Throughput bytes bytes bytes secs. 10^6bits/sec 87380 16384 16384 10.00 7239.95
That’s over 7 gigabits/sec, or over 700MB/sec throughput between the two machines!
So, I now have an Infiniband Fabric working at home, with over 7 gigabit throughput between PCs. The stuff of high-end datacentres in my back room. The main thing is that you don’t need a switch, so a PC to PC 10-gigabit link CAN be achieved for under €110! Here’s the breakdown:
2 x Mellanox MHEA28-XTC infiniband HCA’s @ $34.99 + shippping = $113 (€85)
1 x 3m Molex SFF-8470 infiniband cable incl shipping = $29 (€22)
Total: $142 (€107)
The next step is to set up a raid array with several drives and stripe them so they all work in parallel, and maybe build it in such a way if one or two drives fail, it will still be recoverable (raid 5/6). More to come on that soon.
I’d been hearing good things about various types of battery packs recently, quantum, etc. Being the diy enthuasist (cheapskate) that I am, I thought that I’d research building an external battery pack for my strobes. These usually take the form of a 12v battery pack which connects into the external port of the flash (Canon’s, in my case), reducing recycle time, and giving longer battery life, depending on the type of battery used. I looked initially at the Canon pack, which takes 6 AA batteries. These come in at well over €100. I had a look around on Flickr and other sources, but nowhere could I find a circuit that would feed this external port on my Canon flashes.
Anyway, I had a look on eBay for some canon knock-offs, and I spotted an “iShoot” model for €23 (incl shipping from Hong Kong), which takes 8 AA batteries, so I ordered one. It duly arrived, but unfortunately it was DOA. The LED on the unit would not illuminate, and I saw no change in the recharge time of the flash. I contacted the seller, and he was really great about it. He said that the shipping back to Hong Kong was quite expensive, so I could keep the faulty unit, and he would ship me out a replacement. That arrived about a week later.
Better this time. Upon plugging it in to the flash, and powering the flash on, the LED lit up on the battery pack. With fully charged NiMH AA’s, I was now getting recycle times of just under 3 seconds with the flash set to full power.
I had a look at the internals of the faulty unit, and with the complexity of the circuit board I was looking at, there was no way I would be able to replicate that in a DIY fashion. So the easiest thing for me to do was to just simply order one pack for each of my flashes. At €23, that’s not too expensive, and certainly a lot cheaper than the Canon units. I used this unit over the weekend, and popped off several hundred shots at 1/4 power, and was still getting very fast recycle times. I’m definitely going to order more of these units for my next strobist shoot.