Cool stuff! A suggestion: Avoid the term "dead code elimination". Yes, it is an established term, but it is established as a term for two distinct optimizations.
One form of "dead code" is code that can never be executed:
if (false) {
someDeadCode; // "dead"
}
The other meaning is code that can be executed but computes a value that can never be used:
x = someValue;
y = someOtherValue; // "dead"
return x;
I advise my students to use the terms "unreachable code" and "unused code", respectively.
On a Lisp compiler optimization tangent: It’s still relevant to SBCL and also generally interesting so the CMUCL advanced compiler manual section[1] is good reading.
Not a serious project, made purely because I had a class that mandated writing scheme for the homeworks.
I think the coolest thing to come out of that project was that I learned that it is possible to convert branching if statements to lisp constructs. That was a fun project :)
I admire your commitment towards not writing Scheme but I recommend giving it a go. Maybe use it as an opportunity to learn Vim or Emacs and have a go at structural editing. It'll change how you think about your code...
Guile scheme has a way to easily see the result of many of these optimisations since they are done on the source level. This means you see the result of inlining, DCE, constant propagation and partial evaluation. Extremely handy and helps even mediocre programmers like myself develop a good understanding of when optimisations are triggered.
The Lisp variant that the compiler supports at the moment only handles f64 numbers so I don't think this kind of issue is possible.
However, this is a very relevant point. If the goal is just shorter code (as opposed to a mix of shorter code and less run-time operations), then you need to check that folding strings (and similar types) actually makes the expression shorter to represent.
If these operations produce mutable strings, the conditions under which that would be allowed are fairly stringent. It's not worth doing; it's better for the Lisp to have constructs that allow the programmer easily stage the evaluation in the desired way.
Common Lisp has load-time-value. It's also easy to write a macro called macro-eval which evaluates its argument at macro time, and substitutes the result (as a quoted object).
Any situation where the program depends on the expression producing a new string each time it is evaluated, rather than returning the same string. The program may be modifying the string, on the assumption that nothing else has access to it, since it is brand new. The program could also be relying on the string having a unique identity, not comparing equal to any previously seen object. (E.g. assuming that each time the expression is evaluated, it produces an object that can serve as a unique key in an EQ hash table).
Any situation in which these behaviors cannot be ruled out (because the object escapes beyond the scope of analysis), the optimization cannot be applied.
Ah, well all JS strings are always immutable and only value-referable (you have no access to the underlying memory location), so that’s not a concern here.
What about the identity side of it? Does the JS specification say that an operation like "a" + "b" is not required to create a new object? Regardless of whether there is such a spec, you can write code that is sensitive to the difference.
It looks like JS doesn't expose equality operator which can distinguish different strings. Thus "abc" and "abc" are the same object, no matter how they are produced, even if under the hood they are separate instances.
Pretty much, though some would contest calling strings “objects”.
The spec indeed goes through some trouble to ensure they are pure value-types and do not exhibit any reference-like semantics, for instance by prohibiting their use as keys of WeakMaps and WeakSets - along with numbers, booleans, nullish values, and bignums.
In the context of a discussion on JS, the spec assigns a very specific meaning to the term “object”, one which value-types do not satisfy. That is, indeed, the entire point of this thread, and to suddenly switch terminology then ramble about how you’re better than “script kiddies” screams rhetorical incompetence.
It's clear that you don’t know JS. I was trying to help you understand one interesting design decision they make, but you’ve taken to personal attacks against the entire community in a way that I don’t find worth humoring. Good day.
I already understand the design space and all the implications of strings being immutable and indistinguishable objects if they are the same character sequences; where JS sits in that space is just trivia (which I easily looked up myself).
Indeed, I don't know JS well enough to have a rote memory of where it sits in every design space, or which words it specification does not use for what things. There are plenty of languages in whose specifications object has both the broader meaning, and the more specific OOP meaning. People don't get confused somehow.
Conversely, not everything we can say about a Javascript program necessarily has vocabulary in the Javascript spec; we're going to end up using outside words one way or another.
(I'm not interested in Javascript, actually; this is just a "drive by" for me, because I am interested in the submission topic. All that I will ever know about Javascript will always come from doing the minimal amount of research to answer some question out of some kind of necessity.)
"Javascript is a Lisp" would definitely be found in the Big Bumper Book of Divisive Things To Say To Programmers alongside its more famous entries, "Tabs or Spaces?", and "Vim or Emacs?".
One form of "dead code" is code that can never be executed:
The other meaning is code that can be executed but computes a value that can never be used: I advise my students to use the terms "unreachable code" and "unused code", respectively.