taking the fun out of blogging

As a lawyer, I understand the need for policies, procedures, practices, etc. when running a business, managing vendors, employees, etc. Of course. Sure. That’s part of work – both my work and the work of my clients. But when I see an article entitled “Blogging Policies and Best Practices for Lawyers and Law Firms” well, gotta say, my eyes start glazing over.

Not that there’s anything particularly bad or wrong about the article. In fact, it offers some good advice on avoiding “ethical minefields”, creating “powerful marketing tools” and ensuring you realize a good return on your “investment”.

Ugh. To be perfectly honest one of the primary reasons I blog is not to realize a return on investment, or to create a powerful marketing tool, but rather just to offer casual observations (or ruminations) on my work or things related to my work. In other words, its a bit of fun, as compared, for example, to writing a formal research paper, journal article, or a 100 page outsourcing contract. For those types of writing, there are many, many rules, requirements and policies to remember and adhere to, amongst other considerations. And relatively speaking, its not quite as much writing that stuff as it is posting what are ostensibly meandering ramblings about the next new thing. Don’t get me wrong, its certainly interesting and challenging work, but its not the type of thing one typically does to relax.

I guess what I’m getting at is along the same lines as the previous post about making blogging part of someone’s job. Its kind of like saying that its part of your job to chat up your friends at work on a regular basis. Its kind of like saying that there should be internal policies governing who you go to lunch with, and what you talk about over lunch. In other words, to me, it seems to take all the fun out of it. It makes it seem like work. It puts you in the mindset that it is work. And, to be perfectly honest, I think it makes it less interesting, because you’re too worried about the time being put into it. Too worried about whether you’re writing for your “target market”. Too worried about “visualizing and addressing your market”. Too worried on making your blog sound “informal and conversational”. Too worried about this, that and the other thing, none of which have much to do with the subject matter of what you’re writing about.

Of course, this is just my take on blogging and what I hope to achieve (or perhaps rather not to achieve) by doing it.

silly lawsuit of the week

OK. Short version of the story in InformationWeek: Woman puts up a website. She puts a “webwrap” agreement at the bottom – i.e. basically a contract that says if you use the site then you agree to the contract. Still some question as to whether such a mechanism is binding, but anyway…

So the Internet Archive of course comes along and indexes her site. Which apparently is a violation of the webwrap. So she sues, representing herself, I believe. The court throws out everything on a preliminary motion by IA except for the breach of contract.

InformationWork observes that “Her suit asserts that the Internet Archive’s programmatic visitation of her site constitutes acceptance of her terms, despite the obvious inability of a Web crawler to understand those terms and the absence of a robots.txt file to warn crawlers away.” (my emphasis). They then conclude with this statement:

If a notice such as Shell’s is ultimately construed to represent just such a “meaningful opportunity” to an illiterate computer, the opt-out era on the Net may have to change. Sites that rely on automated content gathering like the Internet Archive, not to mention Google, will have to convince publishers to opt in before indexing or otherwise capturing their content. Either that or they’ll have to teach their Web spiders how to read contracts.

(my emphasis).

They already have – sort of. It’s called robots.txt – the thing referred to above. For those of you who haven’t heard of this, its a little file that you put on the top level of your site and which is the equivalent of a “no soliciation” sign on your door. Its been around for at least a decade (probably longer) and most (if not all) search engines

From the Internet Archive’s FAQ:

How can I remove my site’s pages from the Wayback Machine?

The Internet Archive is not interested in preserving or offering access to Web sites or other Internet documents of persons who do not want their materials in the collection. By placing a simple robots.txt file on your Web server, you can exclude your site from being crawled as well as exclude any historical pages from the Wayback Machine.

Internet Archive uses the exclusion policy intended for use by both academic and non-academic digital repositories and archivists. See our exclusion policy.

You can find exclusion directions at exclude.php. If you cannot place the robots.txt file, opt not to, or have further questions, email us at info at archive dot org.

standardized methods of communications – privacy policies, etc. – more. Question is, will people be required to use it, or simply disregard and act dumb?

Fair Use and the DMCA

An article in Wired News with the dramatic title of “Lawmakers Tout DMCA Killer” describes the most recent attempt to: (a) water down the protections afforded to content owners by the DMCA; (b) ensure the preservation of fair use rights on the part of users. As is usual, each side has its own rhetoric to describe what is happening, so in fairness I took the liberty of offering to readers of this blog the two alternative descriptions above. The nub:

The Boucher and Doolittle bill (.pdf), called the Fair Use Act of 2007, would free consumers to circumvent digital locks on media under six special circumstances.

Librarians would be allowed to bypass DRM technology to update or preserve their collections. Journalists, researchers and educators could do the same in pursuit of their work. Everyday consumers would get to “transmit work over a home or personal network” so long as movies, music and other personal media didn’t find their way on to the internet for distribution.

And then of course on the other side:

“The suggestion that fair use and technological innovation is endangered is ignoring reality,” said MPAA spokeswoman Gayle Osterberg. “This is addressing a problem that doesn’t exist.”

Osterberg pointed to a study the U.S. Copyright Office conducts every three years to determine whether fair use is being adversely affected. “The balance that Congress built into the DMCA is working.” The danger, Osterberg said, is in attempting to “enshrine exemptions” to copyright law.

To suggest that content owners have the right to be paid for their work is, for me, a  no-brainer. That being said, I wonder whether the DMCA and increasingly more complex and invasive DRM schemes will ultimately backfire – sure they protect the content, but they sure as heck are a pain in the ass – just my personal take on it. For example, I’d love to buy digital music, but having experienced the controls that iTunes imposes and suddenly having all my tracks disappear, I just don’t bother with it now. Not to mention the incredible hoops one needs to go through to display, say, Blu-ray on a computer – at least in its original, non-downgraded resolution – why bother with all of that at all?

I wonder whether this is, in a way, history repeating itself in a way. I am old enough to remember the early days of software protection – virtually every high-end game or application used fairly sophisticated techniques (like writing non-standard tracks on floppies in between standard tracks) in attempting to prevent piracy. Granted, these have never gone away altogether, particularly for super high end software that needs dongles and and the like, and of course recently there has been a resurgence in the levels of protection that have been layered on in Windows, but after the initial, almost universal lockdown of software long ago, there came a period where it seemed many (if not most) software developers just stopped using such measures.  At least that’s what seemed to happen. I’m not quite sure why, but I wonder if this same pattern will repeat with content rather than software. I suspect not. But hey, you never know.

In the meantime, off I go, reluctantly, in the cold, cold winter, to the nearest record shop to buy music the old fashioned way…


Vista Capable – Capable of Booting – And Not Much Else

Just a small quick one. Story in Computerworld about how 4GB is the optimal amount of RAM to run Windows Vista. Sure. Fine. Fair enough. Goes on to critique vendors like Dell who have “Vista Capable” machines. And what as does that mean, you ask?

For instance, Dell offers a Windows Vista Capable configuration that isn’t capable of much, according to what Dell says about it on its Web site: “Great for … Booting the Operating System, without running applications or games.”

I thought surely they must be leaving a bit out right? Nope. Not the case. That’s it. That’s all. Scroll down and see for yourself on the Dell site.

So, if you plan on buying a “Vista Capable” machine, enjoy, um, booting up your machine and, well, I guess, admiring the boot process as it boots.

ITAC – First Canadian Municipal Wireless Conference and Exhibition

Wow – lots happening the last week of May. Also forgot to mention previously the First Canadian Municipal Wireless Conference and Exhibition being organized by ITAC at the Direct Energy Conference Centre at the Canadian National Exhibition in Toronto, May 28-30, 2007:

Whether you live or work in a large urban municipality, a small rural town or village, the impact of wireless applications has already or will soon impact the quality of your life and the services you offer your community. If your organization engages in digital electronic services to customers, e.g., taxpayers, suppliers, emergency service providers, other levels of government, non-profit organizations and associations, you need to learn about the latest proven strategies to ensure the success of your wireless programs.

ITAC’s 1st Canadian Municipal Wireless Applications Conference and Exhibition will not only update you on the latest initiatives of Canadian Municipalities, but will provide you with real case study insights, proven strategies, commentary from leading wireless experts and techniques for deploying wireless applications in your communities. If you are currently engaged, or plan to be engaged, in a municipal wireless project, your attendance at this event is essential.

The Virtues and Evils of Open Source

Yes, I know, I’ve been behind lately. A ton of very interesting things to catch up on. But I’d like to put in one quick note about open source code. I recently came across an article, written last year by a lawyer, generally advising development companies not to use open source. I don’t quite recall where it was (if I did I’d link to it) but I do remember it being quite clear in stating that using open source is A Bad Thing and to avoid it altogether – not just to be careful, but rather to treat it as one would radioactive waste.

With respect, I don’t quite agree. I certainly advise my clients to take a great deal of caution in using open source code, particularly the GPL variety, and very particularly if they have a desire to keep some or all of their own secret, proprietary code secret and proprietary. That being said, I do have many, many clients who have used open source code to great advantage in various ways. Some have simply used existing open source code to avoid reinventing the wheel (and saving on costs), while taking care to keep viral elements out of their proprietary code. Others have been more aggressive with the open source model and have intentionally decided to use open source as the basis for their business model and making their very own code, or parts of it, either open source or subject to a dual-licensing model. As the Red Hats, JBosses, Sleepycats, MySQLs etc. etc. of the world have demonstrated, you can go open source and still have a pretty viable business. And, of course, there are the “old world” companies like IBM who have decided to go open source (in some limited ways – e.g. IBM’s DB2 Express-C thing).

Of course, this is not to suggest that anyone through caution to the wind and just start pulling down stuff from Sourceforge and whacking it into your product. Use of open source definitely requires some planning ahead and consideration of what the business model and value proposition of your business will be. Optimally, enlist the help of a lawyer who’s familiar with open source licenses to discuss what you plan to do and the packages you plan to use. Or, if that’s not feasible, try at least to read the applicable licenses yourself and ensure you comply with them, because if you don’t think that anyone will notice, or that no one will actually sue you, you may want to pay a visit to the GPL Violations Site and reconsider, in addition to the questions that will be asked of you when the due diligence starts on your next round of financing or, even worse, your (aborted) exit event. Can badly managed open source usage (and I emphasize badly managed, not simply open source usage) kill a deal? Definitely.

In short – I don’t think open source is necessarily a bad thing. In fact, it can be a pretty good thing, not just in the social good sense and all that, but also as a business. But it need to be used taking into account its terms of use and ensuring that its consistent with the strategy you plan to take.

If perhaps there’s one thing I’d recommend it would be for shops to make absolutely sure they have a disciplined approach in tracking where code comes from and the terms under which its being used and why its being used. That applies not only to open source stuff, but also, for example, your programmers taking neat snippets of code from Dr. Dobbs or something else, or coming across a nice little script somewhere on the Web and saying “Gee, that’s neat, let’s use it in our product”.

Anyway, if I remember where the article was I’ll update this to include a link.