Coding

The Problem With Implicit Variable Declarations

Originally posted: September 16th, 2008

Implicit variable declarations (ex: Visual Basic without "Option Explicit") can seem like a good idea if you look at them a certain way. I can understand what language designers are probably thinking when they decide to implement implicit declarations.

"Let the computer do the work for you."

That's a favorite mantra of mine, one of my cardinal rules. To a programmer, a computer with a compiler is like having an army of robots at your disposal. Why do things manually when you send send off a drone or a function or a script to do it for you while you get on with other things? So whenever possible, automate whatever you can. Efficiency: it's nice.

So I think I can understand what language designers are thinking when they decide to include the feature of implicit declarations. They're thinking "Hey, why should I have to explicitly point out that I'm going to use a variable? If I'm using the variable somewhere in the code, the computer should be able to figure out on it's own that the label is supposed to be a variable. Saves me the bother." Sounds good. Same result, less effort. Efficiency. Nice.

But there's an ugly assumption hidden in that reasoning. What that developer is really saying is "Hey computer, anytime you come across an undeclared label being used as a variable, just go ahead and assume it's a new variable". We all know what happens when we make assumptions, right? "ASSumptions will make an ass outta ya."

What do you really want to happen when you mistype a variable name? (Oh, sure, you'd never do anything like that? Riiight? It happens to the best. Deal with it.) There's two choices: A. The compiler grabs you, points at your error, and yells "Hey! You screwed this up! Fix it!". or B. The compiler pretends everything is ok, processes the bad code, and leaves you with a bug which, only if you're very luckly, will manifest itself immediately and in a way that makes the exact nature and location of the problem obvious. Hmm, which is a better way to write code...?

"But I might have really wanted it to be a new variable!" Ok fine. How about you go rewrite your "rm" command to never ask for confirmation just because, well hey, we can't let the potential downfall of accidentally loosing files force us to endure a much, much lesser inconvenience when we really do want to delete. Doesn't make much sense does it? Point being: Assumptions are bad. Bad enough even to outweigh convenience.

Language design, and heck, API design in general, are more than just programming. There is code involved, yes, but there's a large amount of psychology that needs to go into it as well. When you're writing an ordinary function, you're basically thinking "automation" (ie, that "army of robots"). But when you're designing an interface for programmers, even yourself, the real important thing suddenly becomes "How can I prevent the programmer from messing up?". It's like designing any interface, the weakest link is always the human factor. Even if it's the best programmer in the world, a highly unreliable computer will still make that human look like a giant pile of shoddy engineering. So it's your job, as the language/interface/API designer, to do whatever you can to minimize that risk of programmer error.

Another way to think of the issue is in terms of "good redundancy versus bad redundancy". There tends to be a lot of value placed on eliminating redundancy. Often this is good. But sometimes redundancy can improve reliability, which is a very important concern. Mandatory explicit variable declarations are one form of "good redundancy" that improves reliability. Walter Bright explains it best:

"Variable declarations are one [example of good redundancy in language design]. But since the compiler can figure the need for declarations from the context, declarations seem like prime redundancies that can be jettisoned. This is called implicit variable declaration. It sounds like a great idea, and it gets regularly enshrined into new languages. The problem is, the compiler cannot tell the difference between an intended new declaration and a typo - and the poor maintenance programmer can't tell, either. After a while, though, the lesson about redundancy is learned anew, and a special switch will be added to require explicit declaration."

- Walter Bright: Redundancy in Programming Languages

Visual Basic developers learned this lesson a long time ago. Even though VB supports implicit variable declarations, it's extremely rare to come across a professional VB developer that doesn't strongly recommend turning it off (with "Option Explicit") and religiously does so in their own code. It's a shame there are so many newer languages that haven't learned from this.

Read more


Type Systems and Indentations Are Weak

Originally posted: March 10th, 2008

I've complained about weak and dynamic typed languages before. But damnnit, I'm just so crotchety I'm gonna do it again:

Weak and dynamic typing are bad because type safety is good.

Weak and dynamic typing are bad because compile time errors are better than run time errors.

Obviously, weak typing and dynamic typing can be very useful in moderation. But like they say in booze commercials: only in moderation. Moderation, of course, means "Don't base the whole damn language on it".

I've also whined about semantically-meaningful indentation before. Guess what? I still feel crotchety. *Ahem*:

Semantically-Meaningful Indentation: It's bad. Don't do it. That means you, Python and SPIN.

Read more


I Want My IArithmetic!

Originally posted: March 10th, 2008

Here's a new language rant for me: If the .NET Framework has had an IComparable interface since version 1.1, why in the world does it still not have an IArithmetic? Or even operator constraints. We're now on version 3.5, and there's still no sign of this stuff. What in the world are they waiting for? Implementing IArithmetic would take what, about one person a couple of hours?

If you don't know what I'm complaining about, try this: Add two objects of type 'T' in a C# generic function...without using some esoteric work-around. Go ahead, I dare you. One hell of an oversight, innit?

Read more


Language Rants Part 2: Language-Enforced Indentation

Originally posted: August 5th, 2007

Using indentation to define scope is a bad, bad idea. I'm looking at you, Python and SPIN. Like weak/dynamic-typing, it's fine for trivial programs, but as soon as you use it for anything more substantial it becomes a maintenance nightmare.

Let me clarify one thing: Of course you should always properly indent your code. However, the problem is when a language is designed to interpret that indentation as having semantic meaning rather than just ignoring it.

"But, if code should always be indented, shouldn't the compiler enforce that?" Not exactly. We should (and do) have tools that optionally enforce proper indentation. But such tools belong in places like the editor or the compiler's warnings, not in the language definition itself.

Read more


Language Rants: Weak Typing and Misc.

Originally posted: July 4th, 2007

(Note from July 2010: Yes, I know I refer to dynamic typing as "weak typing" here. "Weak typing" includes more than just C's implicit reinterpret casts. Dynamic typing IS a form of weak typing: deal with it.)

Years of condemnation to the seething underbelly of the programming world known as "web application development" has driven me to this plea:

Please, for the love of god, no more weak-typed languages!!! I can't take it anymore! That includes you too, duck-typing. (Yes, I realize many of those languages optionally allow strong types, but that's just a bandage, and often an afterthought.)

Here's the generalized rule of thumb these weak-typed languages are breaking:

Anything that CAN be handled at compile-time, SHOULD be handled at compile-time.

(Obviously, I'm counting "parsing interpreted-language code" as compile-time here.)

Why is this a good rule of thumb? For one thing, optimization. If you offload certain calculations (such as arithmetic on numerical literals, ie '2 + 3′) to the compilation process, then your program doesn't have to, thus speeding it up. Fortunately, this sort of thing is already handled by most compilers, so you don't have to worry about it. (If you're wondering: yes, I have come across one notable exception: SPIN)

But there's another more important reason for this rule of thumb: Bugs. Bugs are bad. There are two types of bugs: compile-time bugs and run-time bugs. Compile-time bugs are always brought to your attention automatically, whenever your code is compiled. That makes them easy to fix. Good bugs. Or at least as good as bugs get. Run-time bugs only show themselves if you're lucky. These bugs hide and fester. That makes them hard to fix. Bad bugs.

I don't like weak typing because, while it might seem to eliminate type concerns, all it really does is change the type errors from compile-time (good) errors into run-time (bad) errors. This may not be a problem when you're whipping up "Lil' Coder's First PHP Page", but when you have a full-blown web application to maintain, you'll just waste your time hunting down bugs that could have been pointed out to you the moment they were introduced.

Languages that religiously force everything into the latest silver-bullet paradigm have also got to go. I'm looking at you, Smalltalk and Java (Clarification: I like OOP. I don't like religious OOP.) Here's an excellent explanation of one of the reasons I can't stand Java.

As long as I'm ranting on languages, VB6 is obnoxiously verbose. (Yes, I know VB6 has been replaced by VB.NET. But I'm still condemned to it anyway.) Behold:

In most C-based languages:

int var = 5;

In VB6:

Dim var As Integer
var = 5

How clunky. I have other examples of VB insanity (yea, I know, who doesn't?), but they'll have to wait for now.

Read more