As you may know, in JavaScript '' + null = "null"
and '' + undefined = "undefined"
(in most browsers I can test: Firefox, Chrome and IE). I would like to know the origin of this oddity (what the heck was in the head on Brendan Eich?!) and if there is any aim for changing it in a future version of ECMA. It's indeed pretty frustrating having to do 'sthg' + (var || '')
for concatenating Strings with variables and using a third party framework like Underscore or other for that is using a hammer for jelly nail pounding.
Edit:
To meet the criteria required by StackOverflow and clarify my question, it is a threefold one:
- What is the history behind the oddity that makes JS converting
null
orundefined
to their string value inString
concatenation? - Is there any chance for a change in this behavior in future ECMAScript versions?
- What is the prettiest way to concatenate
String
with potentialnull
orundefined
object without falling into this problem (getting some"undefined"
of"null"
in the middle of the String)? By the subjective criteria prettiest, I mean: short, clean and effective. No need to say that'' + (obj ? obj : '')
is not really pretty…
I would say the chances are very slim. And there are several reasons:
We already know what ES5 and ES6 look like
The future ES versions are already done or in draft. Neither one, afaik, changes this behavior. And the thing to keep in mind here is that it will take years for these standards to be established in browsers in the sense that you can write applications with these standards without relying on proxy tools that compile it to actual Javascript.
Just try to estimate the duration. Not even ES5 is fully supported by the majority of browsers out there and it will probably take another few years. ES6 is not even fully specified yet. Out of the blue, we are looking at at least another five years.
Browsers do their own things
Browsers are known to make their own decisions on certain topics. You don't know whether all browsers will fully support this feature in exactly the same way. Of course you would know once it is part of the standard, but as of now, even if it was announced to become part of ES7, it would only be speculation at best.
And browsers may make their own decision here especially because:
This change is breaking
One of the biggest things about standards is that they usually try to be backwards compatible. This is especially true for the web where the same code has to run on all kinds of envrionments.
If the standard introduces a new feature and it's not supported in old browsers, that's one thing. Tell your client to update their browser to use the site. But if you update your browser and suddenly half the internet breaks for you, that's a bug uhm-no.
Sure, this particular change is unlikely to break a lot of scripts. But that's usually a poor arguments because a standard is universal and has to take every chance into account. Just consider
as the instruction to switch to strict mode. It goes to show huw much effort a standard puts into trying to make everything compatible, because they could've made strict mode the default (and even only mode). But with this clever instruction, you allow old code to run without a change and still can take advantage of the new, stricter mode.
Another example for backwards compatibility: the
===
operator.==
is fundamentally flawed (though some people disagree) and it could've just changed its meaning. Instead,===
was introduced, allowing old code to still run without breaking; at the same time allowing new programs to be written using a more strict check.And for a standard to break compatibility, there has to be a very good reason. Which brings us to
There is just no good reason
Yes, it bugs you. That's understandable. But ultimately, it is nothing that can't be solved very easily. Use
||
, write a function – whatever. You can make it work at almost no cost. So what is really the benefit for investing all the time and effort into analyzing this change which we know is breaking anyway? I just don't see the point.Javascript has several weak points in its design. And it has increasingly become a bigger issue as the language became more and more important and powerful. But while there are very good reasons to change a lot of its design, other things just arent't meant to be changed.
Disclaimer: This answer is partly opinion-based.
To add null and '' they need to meet a minimum common type criterium which in this case is a string type.
null is converted to "null" for this reason and as they are string the two are concatenated.
The same happens with numbers:
4 + '' = '4'
as there is a string in there which can't be converted to any number, so the 4 will be converted to string instead.
There are several ways, and you partly mentioned them yourself. To make it short, the only clean way I can think of is a function:
Of course, you can (and should) change the actual implementation to suit your needs. And this is already why I think this method is superior: it introduced encapsulation.
Really, you only have to ask what the "prettiest" way is, if you don't have encapsulation. You ask yourself this question because you already know that you are going to get yourself into a place where you cannot change the implementation anymore, so you want it to be perfect right away. But that's the thing: requirements, views and even envrionments change. They evolve. So why not allow yourself to change the implementation with as little as adapting one line and perhaps one or two tests?
You could call this cheating, because it doesn't really answer how to implement the actual logic. But that's my point: it doesn't matter. Well, maybe a little. But really, there is no need to worry because of how simple it would be to change. And since it's not inlined, it also looks a lot prettier – whether or not you implement it this way or in a more sophisticated way.
If, throughout your code, you keep repeating the
||
inline, you run into two problems:And these are two points commonly known to be anti-patterns when it comes to high-quality software development.
Some people will say that this is too much overhead; they will talk about performance. It's non-sense. For one, this barely adds overhead. If this is what you are worried about, you chose the wrong language. Even jQuery uses functions. People need to get over micro-optimization.
The other thing is: you can use a code "compiler" = minifier. Good tools in this area will try to detect which statements to inline during the compilation step. This way, you keep your code clean and maintainable and can still get that last drop of performance if you still believe in it or really do have an environment where this matters.
Lastly, have some faith in browsers. They will optimize code and they do a pretty darn good job at it these days.
The ECMA Specification
Just to flesh out the reason it behaves this way in terms of the spec, this behavior has been present since version one. The definition there and in 5.1 are semantically equivalent, I'll show the 5.1 definitions.
Section 11.6.1: The Addition operator ( + )
So, if either value ends up being a
String
, thenToString
is used on both arguments (line 7) and those are concatenated (line 7a).ToPrimitive
returns all non-object values unchanged, sonull
andundefined
are untouched:Section 9.1 ToPrimitive
For all non-
Object
types, including bothNull
andUndefined
,[t]he result equals the input argument (no conversion).
SoToPrimitive
does nothing here.Finally, Section 9.8 ToString
Table 13 gives
"undefined"
for theUndefined
type and"null"
for theNull
type.Will it change? Is it even an "oddity"?
As others have pointed out, this is very unlikely to change as it would break backward compatibility (and bring no real benefit), even more so given that this behavior is the same since the 1997 version of the spec. I would also not really consider it an oddity.
If you were to change this behavior, would you change the definition of
ToString
fornull
andundefined
or would you special-case the addition operator for these values?ToString
is used many, many places throughout the spec and"null"
seems like an uncontroversial choice for representingnull
. Just to give a couple of examples, in Java"" + null
is the string"null"
and in Pythonstr(None)
is the string"None"
.Workaround
Others have given good workarounds, but I would add that I doubt you want to use
entity || ""
as your strategy since it resolvestrue
to"true"
butfalse
to""
. The array join in this answer has the more expected behavior, or you could change the implementation of this answer to checkentity == null
(bothnull == null
andundefined == null
are true).You can use
Array.prototype.join
to ignoreundefined
andnull
:According to the spec:
Given that,
In fact, in some cases, the current behavior makes sense.
For example, if the function above is called without arguments,
undefined + undefined = NaN
is probably better than+ = NaN
.In general, I think that if you want to insert some variables in a string, displaying
undefined
ornull
makes sense. Probably, Eich thought that too.Of course, there are cases in which ignoring those would be better, such as when joining strings together. But for those cases you can use
Array.prototype.join
.Most likely not.
Since there already is
Array.prototype.join
, modifying the behavior of string concatenation would only cause disadvantages, but no advantages. Moreover, it would break old codes, so it wouldn't be backwards compatible.Array.prototype.join
seems to be the simplest one. Whether it's the prettiest or not may be opinion-based.