Older blog entries for jmason (starting at number 60)

Doc quotes:

The great teacher John Taylor Gatto said this about how he learned to truly teach:

I dropped the idea that I was an expert, whose job it was to fill the little heads with my expertise, and began to explore how I could remove those obstacles that prevented the inherent genius of children from gathering itself.

s/children/users/, for all those people who take the BOFH stories a bit literally.

10 Dec 2001 (updated 10 Dec 2001 at 02:40 UTC) »

thom: travelling? I was in Sydney for ~6 months, went to Melbourne for two weeks, stayed in Byron for 3 days, then stopped in Brisbane for 7 weeks... Now I'm off back to Sydney... Some traveller, me.

I can top that. Thailand, 3 weeks, Melbourne, er, 4 months ;). I am planning to get moving again soon though, otherwise my Official Traveller Certification will be rendered null and void. Also I'll be thoroughly pissed off if I don't get to try out my new diving sk1llz on the Great Barrier Reef while I'm over here.

gary: Saw Chris Raettig's journal the other day and was impressed; his rationale for an email based journal is appealing [..].

I use a mail-based system to update taint.org, my (other) blog. It works nicely -- the main point was to see how useful WebMake would be for blogging -- but I'm now reverting to using Advogato for diary stuff, keeping taint.org for interesting newsy snippets. Go figure. I reckon it's because Advogato's more of a diary-based community, whereas taint.org is kinda out on its own, and just doesn't seem like a good place to keep a proper journal.

ask: Jabber [..] Maybe it's just crappy clients, but it doesn't seem stable enough to run anything too critical on.

That's the same feeling I got about 6 months ago. :( Let's hope it polishes up some time soon. Open source != crappy packaging!


Oh -- another new hack I forgot to mention, which someone might find useful. MailMan-to-RSS, a little script (and a vaguely nice-looking demo site) which scrapes MailMan list archives and creates an RSS feed of the last 10 posted messages. Handy for following lists using any of the various portal systems (and Evolution!) that support RSS.

BTW as the page sez, if you would like to see a list RSS-ized, and don't have a server to host the RSS on, mail me and I'd be happy to add it to the scraped list on taint.org.

Oh look, thom is travelling 'round Oz too! cool. Must get back to the "travelling" part myself someday soon though.

BTW forgot to plug the Sub-Pixel Font Positioning on UNIX mini-HOWTO I wrote a few weeks ago. It really just ties up the loose ends in XFree86 4 doco, covers the implementation details, and gives UNIX-users something to point at when Windoze lusers rattle on about ClearType. ;) Sub-pixel positioning is a total beaut tweak for a laptop screen.

Kevin is talking about kittens. Why is everyone going on about cats this week? I miss my cat :(

I'm travelling 'round the world at the mo', and apparently the one thing people travelling always miss, is their pets. Statistically, I've backed this up with several drunken conversations with other pet-craving travellers.

Patches and Contributed Code

Here's an interesting one. I've written a few free-software apps in the past, and recently SpamAssassin has taken off. It's very much sysadmin-oriented, being a mail filter for spam which works well as a system-wide filter.

It's illustrated that there's a big difference in audiences, between app users and sysadmins; sysadmins will regularly hack the code to ''scratch their itch'' and send back a patch; whereas patches don't often come from users.


My ghod, it's been a while since I updated the diary. Things I've done since then:

  • wrote SpamAssassin, a mail filter to identify spam using text analysis. Using its rule base, it runs a wide range of heuristic tests on mail headers and body text to identify spam.

    This is pretty neat. It does a good job of differentiating spam from not-spam without too many false positives or negatives; and it's a proper Perl module, so it can be plugged into other mail delivery or filtering systems quite easily (at some stage ;).

    I've been using something similar for a long time, but I eventually decided to reinvent the wheel. The end result is pretty good so IMHO it was worth it.

  • Helped start up Ireland Offline, a new organisation campaigning to sort out Ireland's internet backwater status and bring fat pipes to the people. This is going well... lots of interest, press and support, and some great people involved.

  • Decided to move to Australia ;) Yep, despite getting involved in Ireland Offline, I'm heading off to Melbourne in a month's time. Haven't really figured out the job situation there, but hopefully it shouldn't be too tricky getting hold of one. If anyone reading is in a position to hire a UNIX guru (hey, I'm allowed to plug myself for this), give us a mail.

  • Sitescooper: not an awful lot of news here; Plucker support is pretty good now, and I've put its caching subsystem on a diet in preparation for a move to a new server for the Nightly Scoops site.

    The scoops page is an interesting situation. Every night, a cron job runs off and downloads pages from 136 sites (typically the ones that have clear-ish terms allowing redistribution of their content). The sitescooper script is run 5 times, for the 5 output formats that site provides. Since sitescooper caches these pages in a per-format cache (which allows it to run diffs on pages to see what's changed) as well as a shared cache (which ensures the network is only accessed once for each page), that was 6 copies of each page.

    The cache is expired every few days, removing pages older than a month or so. Still, it was running pretty big, all the same. I've now implemented a Singleton pattern for the cache usage, which brings it down to 1 ref-counted copy of each page, and 6 pointers. After a few weeks of this, the cache disk usage is running at about 120 megs, down from about 800.

    This unfortunately may still be too much for the poor overburdened colocated server I use, especially since I'll be on the other side of the world. :( As a result the list of sites on the page may need another diet. We'll see...

  • WebMake: lots of new stuff in the pipeline. It now supports plugins, which are library files that can define library functions for the inline perl code, and -- since I've added tag-definition support -- a plugin can also add new tags, for use either in the HTML input documents, or in the WebMake .wmk XML file itself. Who needs taglibs? ;)

    This has allowed lots of new features, without messing up the core. It's been in the released version for a while.

    However, a new new feature, not released yet, is IMHO neater. It's "edit-in-browser" support, which is long overdue.

    This is really just a CGI script and a set of modules, allowing a WebMake site to be managed in a web browser; the user logs in using traditional htpasswd authentication, picks a WebMake site (ie. a .wmk file), and can then pick bits of content from the file and edit them in a textbox. It also has a directory browser/file manager for the tags that load content from a directory tree, like contents and media.

    Once they're done editing, they can build the site (using WebMake, obviously), and -- the really neat bit -- check their changes into CVS.

    Since CVS support is built-in, this means that I can update my sites from anywhere in the world, with a web browser, or do it quickly at the command-line from anywhere I have the sites checked out -- at home, in work, etc. It also gives a bonus in that it makes site replication super-easy -- just cvs checkout and it's done. And it's free. CVS is cool.

    So I'm just documenting this up, grabbing screenshots etc., and then I'll release it.

Just certified Dave Brownell as a Master, seeing as he's one of those guys who just keeps cropping up in the most interesting projects.

Still need to do a proper diary update at some stage...

Sitescooper 3.0.2 released -- and about time too ;)

Been a long time since I updated the diary. There's a few reasons:

  • been busy :( -- trying to get up a head of steam to fight software patents in Europe -- Ireland is backing the move, so I'm trying to get some ILUG members (myself included) to fight it. Problem is, I don't know where to start, myself -- letterwriting and political campaigning are not my strong points :(

  • Also, I don't think recentlog.html is scaling, it's too difficult to follow the diaries. Generally if I check my diary the morning after posting, it's already scrolled off. This makes it very tricky to be bothered posting, if there's a 90pc chance no-one's going to read it... after all, who actually goes to a /person page to read their diaries? 's the tragedy of the commons, innit. ;)

But notwithstanding the latter point, I'll throw a few opinions into the ether on what I've read in other diaries. And might as well do an update on WebMake and sitescooper...

---- WebMake

Released 0.7. It works quite well, generates sitemaps, breadcrumb trails, back/forward navigation links, and other nifty metadata things. Not sure what needs to be done next... I have a few non-urgent plans:

generate RDF sitemaps

as suggested in Dan Bricklin's paper, URL on the WebMake todo list. This could be cool, esp. if it can be reused to generate RSS "what's new" lists for My Netscape, Scripting News, oreilly.net, etc.

access to stat() data on links

Allow automatic generation of file size info, by making file size a metadatum on a content item -- this'd be handy for download pages.

come up with an intermediate XML format for EtText

caolan suggested this one, and it's a goodie. If EtText generates an XML format instead of plain XHTML, it may be a neat way of (a) allowing more flexible styling of the HTML, (b) allowing other output formats (WML, DocBook, etc.), (c) some neat XSL tricks.

"edit-in-browser" functionality

Throw in a CGI which can parse and edit WebMake files and EtText, and you've got good ol' "edit-in-browser" as seen on Advogato, editthispage.com, blogger, etc.

Mebbe I'll just let it get stable first though.

---- Sitescooper

Not much here -- need to fix the NYT login problem (again). Lots of hassle with sites blocking us out of their "AvantGo versions"; AG are taking a strong line with the sites to block us out, it looks like. Nasty.

Mandrake caused a bit of a stink recently, with their announcement that Mandrake News and the Mandrake Forum would be made palm-readable with AvantGo, and not a mention of sitescooper or Plucker. So I've made a site file for MF, which AG still can't handle ;).

Michael Nordström from Plucker asked for the URL of their PDA-friendly version, but no response. hmm.

Maybe we should look into making a sitescooper-on-Mandrake RPM for their Cooker distro, and subvert from the inside ;)

---- Comments

lkcl --

i was going to have to send < and friends because of the break-ups in the data flow: jabber has a wrapper around data called a <stream>. this is where things start to get scary.

It's a nasty problem -- you could try using CDATA sections, which act as unreadable blocks of data, XML tags in there won't get parsed. Not sure how well libxml supports 'em though.

mrorganic mentioned:

Personal: got the QNX/RTP stuff loaded and working last night. I haven't done much with it yet, but I already know I like it better than anything I've gotten running on Linux. Photon makes X look like the buggy, bloated hack job that it is. I haven't made much use of PhAB yet (the GUI-builder for Photon), and reports indicate it is still unstable, but I'll probably play around with it a bit tonight and see what it's capable of.

I've always been a fan of OSes like VxWorks and QNX because they seem so much *cleaner* than other architectures.

I've been using QNX4 (the previous version before RTP) for the last year + 1/2. It's not much cleaner than Linux, it just has less functionality. And oh, the bugs, don't get me started ;)

BTW someone mentioned shouldexist.org. There's also halfbakery.com with a similar anti-patents concept.

51 older entries...

New Advogato Features

New HTML Parser: The long-awaited libxml2 based HTML parser code is live. It needs further work but already handles most markup better than the original parser.

Keep up with the latest Advogato features by reading the Advogato status blog.

If you're a C programmer with some spare time, take a look at the mod_virgule project page and help us with one of the tasks on the ToDo list!