At the time of writing, if you do any of these searches on google:
charles dunston carphone warehouse
...you'll notice two patterns:
- My site is no. 1 on all three
- They're all badly spelled
At the time of writing, if you do any of these searches on google:
charles dunston carphone warehouse
...you'll notice two patterns:
Truncated! Read the rest by clicking the link below.
I've now set up alias to www.peterbe.com such that ww.peterbe.com and wwww.peterbe.com both redirect to www.peterbe.com.
We'll see what effect this might have and if it's worth. I guess 99% of all visitors on this site get it right but this tightens the "fool-proofness" even more. Google have one such alias set up on ww.google.com but not wwww.google.com
At the moment I'm not running Squid for this site but if experimentation time permits I'll have it running again soon. One thing I feel uneasy about is how to "manually" purge cached pages that needs to be updated. For example, if you read this page (and it's cached for one hour) and post a comment, then I'd like to re-cache this page with a purge. Setting a HTTP header could be something but that I would only be able to do on the page where you have this in the URL:
?msg=Comment+added
which, because of the presence of a querystring, is not necessarily cached anyway. The effect is that as soon as the "?msg=Comment+added" is removed from the URL, the viewer will see the page as it was before she posted her comment. squidclient
might be the solution. ...sort of.
Truncated! Read the rest by clicking the link below.
Finally I got my domain name back. What happened was that it expired without me being notified. The reason I wasn't notified was that the email address that Network Solutions use is ancient and I don't check it anymore. What I had to do was to send a signed fax with a photocopy of me driving license to Network Solutions in the states to tell them to change my email address. Once I've changed my email address I was able to log in and renew the service for three years.
What confused the whole thing was that apparently I thought I could transfer the domain name over to mydomain.com who I use to administer the domain name. The reason it didn't work was that the domain could not be transfered when it was pending deletion.
Long story short: I'm back. To all those of you who have emailed me on mail @peterbe.com you and have got a delivery-error-message, do you want to resend that important piece of email now?
I'm back! My dear little website is back up and running. This time on a different computer on a different network.
What happened was that the poor little old laptop that my computer was running on completely screwed itself up after a hard restart. Everything on its memory became totally random. When it managed to boot up I had several gigantic folders, some with equal name that couldn't be opened. My friend Jan Kokoska helped me eventually run a few disk-checking programs and eventually we could see my non-backedup files again. With a Linux LiveCD we managed to copy the data across to another computer and eventually it got up here on this server.
The culprit was faulty RAM. Jan did lots of tests on it with software and eventually we managed to isolate that the extra 128Mb memory I had in the computer was broken. We took it out and threw it in the bin.
Now this new server is one of Fry-IT's. It's a dual Xeon 2.4Ghz box and not on some silly 256Kbits/s connection that I had before. Let's see how it goes.
Now I've done something relatively useful with my PageRank algorithm written in Python that I'm actually quite proud of. It's not rocket science but at least I've managed to understand the Google PageRank algorithm and applied it to my own setup. This application is very simple and not so useful as one could hope but at least I prove to myself that it can be done.
I call it PlogRank. As you might have noticed, most blog items here on this site have on the left hand side, beneath the menu, a list of "Related blogs". These are from now on sorted by PlogRank! Cool, ha?
The "Related blogs" work by specific word matching. Every blog item has a list keywords that I define manually through the management interface. The selection of keywords is helped by another little database that filters out all typical words. E.g. "PageRank" is a particular word and "page" is not; so selecting these keywords is very easy for me.
Anyway. What I do now, once every week, is that I load a huge matrix of all connections between pages. If this blog item has a link to PageRank in Python then that page increases in PlogRank. It does not effect this page. I then feed this into the PageRanker program I've written which calculates the corresponding PageRank for each blog item. Easy! The whole calculation takes only a couple of seconds with 30 iterations. The calculation is actually only a small part of that time because reading from and writing to the database is the real bottleneck.
So, the end result is that every blog item that has related links will show these links in PlogRank-sorted order. Isn't that neat?
Did a quick analysis on all the referers to my web site. Referers being when web users click a link to my site from another site instead of manually typing in the URL. The result is not surprising but quite sad. About 5% of all referer visits to my web site is from other normal web pages. All the remaining is from search engine results such as Yahoo, Google etc or other obscure web services.
The sad truth is that very few people make a link to my site :(
The good thing is that my site must be very findable.
The most important conclusion is probably that people don't surf the web anymore. Instead they search it. I for one trust Google so much that I sometimes search instead of digging up the URL written down somewhere. This proves the importance of being findable on the web. You have to make your pages findable otherwise you don't get any hits. So, redesign your sites so that Google can index them accurately and avoid silly things like frames or images with text in them.
I have been experimenting recently with HTML optimization but haven't applied it yet. But I have applied this now to my stylsheets. The size gain is 33%! (1577 bytes to 1027 bytes) However, the speed gain involves also the time to perform the optimization so the speed gain will obviously be less than 33%. But the optimization takes, on this slow computer, 0.004 seconds so in approximate terms the speed gain is also 33%. This is on a stylesheet file with some but short and few comments.
The optimization script removes almost all unnecessary whitespace (newline characters included) and all comments. The code for python friends looks like this:
import re
css_comments = re.compile(r'/\*.*?\*/', re.MULTILINE|re.DOTALL)
def _css_slimmer(css):
css = css_comments.sub('', css)
css = re.sub(r'\s\s+', '', css)
css = re.sub(r'\s+{','{', css)
css = re.sub(r'\s}','}', css)
css = re.sub(r'}','}\n', css)
return css
The Wayback Machine has archived my websites since April 2001.
The last archive entry is of how this website looked like with the new design first. I.e. 7 months ago. I didn't know about this project but it looks promising yet limited. It's like a permanent cache.
Another fun one was my profile page from our intranet when I was working for Net4Any. (have patience and wait for the animated mug shot)
I've changed my RSS feed to be RSS 1.0 compliant. My feed has been invalid for couple of days so I thought I needed to upgrade.
RSS 1.0 also allows for a Subject meta data so my items will look better on Webforce.at