Technology Tales

Adventures in consumer and enterprise technology

A case of the reverse Midas touch?

18th May 2009

Last week, a power outage put my main home PC out of action. While it may have been recoverable if that silly accident of a few weeks back hadn't happened, a troubled rebuild is progressing. Despite the challenges, I somehow manage to remain hopeful that an avenue of exploration will yield some fruit. Even so, thoughts of throwing in the towel and calling in professionals rather than throwing good money after bad are gathering. The saga is causing me to question the sense of self building in place of buying something ready built. Saying that, they can have their off days too.

Meanwhile, I have been displaced onto the spare desktop PC and the laptop. In other words, my home computing needs are being fulfilled to a point, though the feeling of frustrated displacement and partial disconnection from my data remains; because I have been able to extricate most of my digital photos and my web building, things are far from being hopeless. With every disappointment, there remains an opportunity or two. Since the spare desktop runs Debian, I have been spending some time seeing if I can bend that to my will, which can be done, sometimes after a fashion.

A few posts should result from this period, not least regarding working with Debian. On the subject of hardware, I will not elaborate until the matter comes to a more permanent resolution. From past attempts (all were successful in the end), I know that the business of PC building can feel like a dark art: you are left there wondering why none of your efforts summon a working system to life work until it all comes together in the blink of an eye leaving you to wonder why all the effort was expended. The best analogy that I can offer is awaiting a bus or train; it often seems that the waiting takes longer than the journey. Restoring my home computing to what it was before is a mere triviality compared to what some people have to suffer, but resolution of a problem always puts a spring in my step.

Old tricks can still save money

14th May 2009

Recently, there was no sound coming out from my main PC and I began to suspect the sound card, an ageing Guillemot Maxi Sound Muse. Naturally, my thoughts turned to making an acquisition, only for those in stock at PC World seemed to have unknown or dubious Linux compatibility. However, I powered the machine down for the night and re-seated the card in its slot the following morning. That was enough to solve the problem and I retained my cash, never a bad thing and especially so in these times. It's always nice to know that and old trick still has its place.

Getting Fedora working in VirtualBox

12th May 2009

After a hiatus induced by disk errors seen on start up, I have gone having a go with Fedora again. In the world of real PC's, its place has been taken by Debian, so virtualisation was brought into play for my most recent explorations. I could have gone with 10, the current stable version, but curiosity got the better of me and I downloaded a pre-release version of 11 instead.

On my way to getting that instated, I encountered two issues. The first of these was boot failure with the message like this:

FATAL: INT18: BOOT FAILURE

As it turned out, that was easily sorted. I was performing the installation from a DVD image mounted as if it were a real DVD, and laziness or some other similar reason had me rebooting with it still mounted. Though there is an option to load the hard disk variant, it wasn't happening, resulting in the message that's above. A complete shutdown and replacement of the virtual DVD with a real one set matters to rights.

The next trick was to get Guest Additions added, but Fedora's 2.6.29 was not what VirtualBox was expecting, and it demanded the same ransom as Debian: gcc, make and kernel header files. Unfamiliarity had me firing up Fedora's software installation software, only to find that Synaptic seems to beat it hands down in the search department. Turning to Google dredged up the following command to be executed, which got me further:

yum install binutils gcc make patch libgomp glibc-headers glibc-devel kernel-headers kernel-devel

However, the installed kernel headers didn't match the kernel, but a reboot fixed that once the kernel was updated. Then, the Guest Additions installed themselves as intended, with necessary compilations to match the installed kernel.

The procedures that I have described here would, it seems, work for Fedora 10; well, they certainly have bequeathed me a working system. I have had a little poke and a beta of Firefox 3.5 is included, and I saw sign of OpenOffice 3.1 too. So, it looks very cutting edge, easily so in comparison with Ubuntu and Debian. Apart from one or niggles, it seems to run smoothly too. Firstly, don't use the command shutdown -h now to close the thing down, or you'll cause VirtualBox to choke. Using the usual means ensures that all goes well, though. The other irritation is that it doesn't connect to the network without a poke from me. Whether SELinux is to blame for this or not, I cannot tell, but it might be something for consideration by the powers than be. That these are the sorts of things that I have noticed should be telling you that I have no major cause for complaint. While I have mulled over a move to Fedora in the past and that option remains as strong as ever, Ubuntu is not forcing me to look at an alternative and the fact that I know how to achieve what I need is resulting in inertia anyway.

Some things don't mix...

10th May 2009

Now that the Release Candidate for Windows 7 is out, I have been giving it a whirl in a VirtualBox session and it, like the Beta that I had been trying too, feels a sold enough affair. I went for the complete installation route in place of the upgrade path. I was surprised to find that it bundled my old files into a single folder called Windows_old and that my old user bits and bobs were folded in with this too. There was nothing there that I wouldn't have missed, but this is a nice touch.

However, I have a spot of fixing to do after adding Kaspersky Internet Security 2009. Like the beta, mixing Windows 7 and Kaspersky seems not to be the way to a stable system. Whether this is down to the virtualisation aspect of the business is something that I don't know, but I have found that removing Kaspersky and replacing it made everything sing along together. Booting into Safe Mode and using msconfig to remove any incidences of Kaspersky being called at start up provides a partial restoration of service. Because the msiexec service isn't running, you need full mode before any software but pulling out any cause for execution of Kaspersky gets that back. I suppose that I could go and put Windows 7 on a real machine to see if Kaspersky causes problems there, but that's not a road that I really want to travel.

Where's WordPress 2.8?

7th May 2009

It now seems that WordPress 2.7 has been an unqualified success. The major changes that were made to the administration screens have been well received, and the grumblings that were extant about 2.5 and 2.6 seem to have dissipated too. Another observation is that security bugs have not been making their presence felt. All in all, it feels very much like assured progress and may explain why 2.8 has been taking its time in coming.

It's now pencilled in for the end of this month and looks as if it will be a polishing of what already works well. It seems to me that most of the changes are behind the scenes, but there is a new widgets interface that should be ever more user-friendly together with an automated theme installation and upgrade facility that is based on what is already in place for plugins (speaking of which, that interface has been tidied too). Another rough edge that has been removed is the whole business of time zones and daylight saving time. In summary, it seems to be a sharpening of a package that already works well anyway. I have been running it on another site without a whimper of drama, so that's probably saying something. Saying that, quite how they are going to get anyone to upgrade is another matter. For one thing, Lorelle VanFossen's overuse of the word "mandatory" cannot be likely to do it...

In a way, the subject of upgrade fatigue brings me to a recent poll run by Automattic's Jane Wells. Quite a number want to stick with what works, while others fancy a change. This split could be tricky to manage and might even encourage some not to upgrade at all and stick with what works for them. After all, there were two episodes of major upheaval last year and I cannot see everyone wishing to see that happen again. Continual evolutionary freshening would suit me better. Thankfully, any talk of changing the administration screens has been left for 2.9 now, and there's always the option of sticking with 2.8 if what is produced becomes a sufficient irritation. Well, it saves a leap to Habari or another alternative anyway...

Taming raw images with ImageMagick: A virtual workaround for Ubuntu 9.04

5th May 2009

While using a command line tool like ImageMagick for image processing may sound a really counter-intuitive thing to do, there's no need to do everything on a case by case interactive basis. Image resizing and format conversion come to mind here. Helper programs are used behind the scenes too, with Ghostscript being used to create Postscript files, for example.

The subject of helper programs brings me to an issue that has hampered me recently. While I am aware that there are tools like F-Spot available, I am also wont to use a combination of shell scripting (BASH & KSH), Perl and ImageMagick for organising my digital photos. My preference for using Raw camera files (DNG & CRW) means that ImageMagick cannot access these without a little helper. In the case of Ubuntu, it's UFRaw. However, Jaunty Jackalope appears to have seen UFRaw updated to a version that is incompatible with the included version of ImageMagick (6.4.5 as opposed to 3.5.2 at the time of writing). The result is that the command issued by ImageMagick to UFRaw - issue the command man ufraw-batch to see the details - is not accepted by the included version of the latter, 0.15 if you're interested. It appears that an older release of UFRaw accepted the output device ppm16 (16-bit PPM files) but this should now be specified as ppm for the output device and 16 for the output depth. In a nutshell, where the parameter output-type did the lot, you now need both output-type and output-depth.

While I thought of decoupling things by using UFRaw to create 16-bit PPM files for processing by ImageMagick, it was to no avail. The identify command wouldn't return the date on which the image was taken. Though I even changed the type to 8-bit JPEG's with added EXIF information, no progress was made. In the end, a mad plan came to mind: creating a VirtualBox VM running Debian. The logic was that if Debian deserves its reputation for solidity, dependencies like ImageMagick and UFRaw shouldn't be broken, and I wasn't wrong. To make it work well, I needed to see if I could get Guest Additions installed on Debian. Out of the box, the supported kernel version must be at least 2.6.27 and Debian's is 2.6.26, so additional work was on the cards. First, GCC, Make and the correct kernel header files need to be installed. Once those are in place, the installation works smoothly and a restart sets the goodies in motion. To make the necessary shared folder to be available, a command like the following was executed:

mount -t vboxfs [Shared Folder name] [mount point]

Once that deed was done and ImageMagick instated, the processing that I have been doing for new DSLR images was reinstated. Ironically, Debian's version of ImageMagick, 6.3.7, is even older than Ubuntu's, but it works and that's the main thing. Since there is an Ubuntu bug report for this on Launchpad, I hope that it gets fixed at some point in the near future. However, that may mean awaiting 9.10 or Karmic Koala, so I'm glad to have this workaround for now.

Ubuntu 9.04 and Tracker

30th April 2009

Shortly after it was released, I did the upgrade shuffle very painlessly, and it didn't take up so much time either. There was only one issue: Tracker falling over, complaining about corrupted indices. That got it removed from my system using apt-get remove (apt-get purge is another option, especially if you need to rid yourself of nefarious configuration files). After having a bit of a dig around the web, I found that I wasn't the only one seeing the problem. To me, it looks as if the upgrade to 9.04 doesn't work so well when it comes to Tracker, and it needs to be removed and reinstalled to ensure that all required dependencies are correctly set in place. Since I restored it on my system, all is working without complaint. Other than the Tracker issue, it has been a case of another uneventful upgrade. Though the evolutionary path that Ubuntu is following may disappoint anyone looking for excitement, no one would upgrade every six months if they knew that disruptive damage or upheaval might be caused. While I may do a clean installation at some point, that is well down the priority list right now.

Are ten seconds enough?

27th April 2009

Fasthosts, the hosting provider for what you find here, has, in their wisdom, decided to limit the execution time for ASP scripts to 15 seconds and 10 seconds for any others. I haven't used Perl sufficiently in this shared hosting set up to determine how that is affected. In contrast, I can share my experiences on the PHP side and you may have noticed occasional glitches. They have also disabled the set_time_limit PHP function, so you cannot easily address the matter yourself where you need to do it. You almost get the feeling that they don't trust the abilities, actions and oversight of their users. Personally, I reckon that the ten-second limit is too short and that something of the order of 20 or 30 seconds would be better. If it all gets too restrictive, I suppose that there are other providers, though I think that I would avoid resellers after a previous less than glorious experience. There's the dedicated server option too if I was feeling flush, not so likely given the economic times in which we live.

Self-hosted web analytics tracking

24th April 2009

It amazes me now to think how little tracking I used to do on my various web "experiments" only a few short years ago. However, there was a time when a mere web counter, perhaps displayed on web pages themselves, was enough to yield some level of satisfaction, or dissatisfaction in many a case. Things have come a long way since then, and we now seem to have analytics packages all around us. In fact, we don't even have to dig into our pockets to get our hands on the means to peruse this sort of information, either.

At this point, I need to admit that I am known to make use of a few simultaneously but thoughts about reducing their number are coming to mind, but there'll be more on that later. Given that this site is hosted using WordPress software, it should come as no surprise that Automattic's own plugin has been set into action to see how things are going. The main focus is on the total number of visits by day, week and month, with a breakdown showing what pages are doing well, together with an indication of how people came to the site and what links they followed while there. Don't go expecting details of your visitors like the software that they are using and the country where they are accessing the site with this minimalist option and satisfaction should head your way.

There is next to no way of discussing the subject of website analytics without mentioning Google's comprehensive offering in the area. You have to admit that it's comprehensive, with perhaps the only bugbear being the lack of live tracking. That need has been addressed very effectively by Woopra, even if its WordPress plugin will not work with IE6. Otherwise, you need the desktop application (being written in Java, it's a cross-platform affair and I have had it going in both Windows and Linux) but that works well too. Apart maybe from the lack of campaigns, Woopra supplies as good as all the information that its main competitor provides. It certainly does what I would need from it.

However, while they can be free as in beer, there are some costs associated with using external services like Google Analytics and Woopra. Their means of tracking your web pages for you is by executing a piece of JavaScript that needs to be added to every page. If you have everything set to use a common header or footer page, that shouldn't be too laborious, especially when there are plugins for publishing platforms like WordPress too. This way of working means that if anyone has JavaScript disabled or decides not to enable JavaScript for the requisite hosts while using the NoScript extension with Firefox, then your numbers are scuppered. Saying that, the same concerns probably any JavaScript code that you may want to execute, but there's another cost again: the calls to external websites can, even with the best attention in the world, slow down the loading of your own pages. When you add in latency caused by servers having to communicate across the web, it is not all about executing JavaScript code.

A self-hosted analytics package would avoid the latter, and I found one recently through Lifehacker: Piwik, formerly known as PHPMyVisites. Usefully, it turns out that it does next to everything that Google Analytics does. While I'd prefer that it used PHP for this, JavaScript is its means of tracking web pages too. Nevertheless, page loading is still faster than with Google Analytics and/or Woopra and Firefox/NoScript users would only have to allow JavaScript for one site too. If you have had experience with installing PHP/MySQL-powered publishing platforms like WordPress, Textpattern and such like, then putting Piwik in place is no ordeal. Though, you may find yourself changing folder access, uploading of the required files, the specification of database credentials and adding an administration user is all fairly standard stuff. After all that, I have the thing tracking this edifice as well as my outdoor activities (hillwalking/cycling/photography) web presence and I cannot say that I have any complaints, so we'll see how it goes from here.

A new feature request for Textpattern?

22nd April 2009

Having been doing some updates to articles in A Wanderer's Miscellany, an idea that makes life easier when working in Textpattern with old articles has come to mind. Currently, there is no way to navigate through pages in the administration area other than using the search or previous/next functionality. Thus, I have started thinking that being able to subset articles by section or category using dropdown menus would be a good way forward.

A search for a suitable plugin yielded nothing of immediate use (amazingly, no one has given it a go thus far); hence the thinking regarding a new feature request. Because there is a place on the Txp support forum for exactly this kind of thing, I am in the throes of plucking up the courage to go for it. Apparently, some code cutting of my own would grease the wheels for the progression of any such thing; a lack of confidence in that area could reduce my input to a bare feature request. While moving to an alternative platform might be an alternative proposition, I see little reason in doing so when what I have otherwise works well for what I want it to do. That Textpattern feature request might just come into being...

  • The content, images, and materials on this website are protected by copyright law and may not be reproduced, distributed, transmitted, displayed, or published in any form without the prior written permission of the copyright holder. All trademarks, logos, and brand names mentioned on this website are the property of their respective owners. Unauthorised use or duplication of these materials may violate copyright, trademark and other applicable laws, and could result in criminal or civil penalties.

  • All comments on this website are moderated and should contribute meaningfully to the discussion. We welcome diverse viewpoints expressed respectfully, but reserve the right to remove any comments containing hate speech, profanity, personal attacks, spam, promotional content or other inappropriate material without notice. Please note that comment moderation may take up to 24 hours, and that repeatedly violating these guidelines may result in being banned from future participation.

  • By submitting a comment, you grant us the right to publish and edit it as needed, whilst retaining your ownership of the content. Your email address will never be published or shared, though it is required for moderation purposes.