diff --git a/README.md b/README.md
index f9d2183..cc1d729 100644
--- a/README.md
+++ b/README.md
@@ -8,7 +8,7 @@
High-performance TypeScript compiler.
-## Goal
+## Goals
- Parser
- Type checking (as CLI and as library)
@@ -19,33 +19,30 @@ High-performance TypeScript compiler.
- (optional) RTTI in JavaScript
- (optional) type profiler
-The goal is to make TypeScript type checking as fast as possible and provide alongside with it a native library for other languages, so they can use TypeScript type information
-without the need for a JavaScript engine for all sorts of use cases like JSON-Schema replacement, ORM DSL, encoding information (like Protocol Buffers schema) and more.
+Additionally, making TypeScript type checking as fast as possible and providing a native library for other languages so they can use TypeScript type information without the need for a JavaScript engine enables all sorts of use cases like JSON-Schema replacement, ORM DSL, encoding information (like Protocol Buffers schema) and more.
-The goal is not to be a drop-in replacement for the entire official TypeScript compiler (tsc). TypeScript supports so much that is not always necessary.
-We focus on the more strict TS part which means TypeRunner won't support JSDoc and a lot of compiler options.
+### Non-goals
+
+- Drop-in replacement for the official TypeScript compiler
+- Runtime
+
+Being a drop-in replacement for the entire official TypeScript compiler (tsc) would require copying design limitations, bugs, and legacy decisions of tsc. Because of TypeScript's 10-year history there are many features that are unnecessary today but kept for compatibility. We focus on a stricter subset of TypeScript, which means TypeRunner won't support some features such as JSDoc and several compiler options.
## Status
-The source code in the initial version is really only a proof of concept. It consists of roughly 30k LoC and shows very promising results.
-The approach is a TS-to-bytecode compiler and then run the bytecode in a custom virtual machine.
-The data show that this approach can lead to a hundred- to several-thousand-fold improvement in speed.
+The source code in the initial version is really only a proof of concept. It consists of roughly 30k LoC and shows very promising results. The approach is to use a TypeScript-to-bytecode compiler and then to run the bytecode in a custom virtual machine. The data show that this approach can lead to a hundred- to several-thousand-fold improvement in speed.
![TypeRunner Debugger](./docs/typerunner-debugger.png)
-Once the project gets funding through the community, the development will continue.
+Once the project gets funding through the community, development will continue.
## Performance
-`TypeRunner cold` means the file was seen for the first time and has to be compiled to bytecode first.
-`TypeRunner warm` means the bytecode could be directly executed because it was cached. Generally it is assumed
-that only a few files change, for example if you have a project with 100 files and edit one and then rerun
-type checking, then only the changed one has the slower `cold` timing. Note that compilation has not yet been optimised
-(it still uses a slow memory allocator which can be improvement roughly tenfold).
+`TypeRunner cold` means the file was seen for the first time and has to be compiled to bytecode first. `TypeRunner warm` means the bytecode could be directly executed because it was cached. Usually only a few files change between compilations. For example, if you have a project with 100 files and edit one and rerun type checking, only the changed file has the slower `cold` timing.
+
+The `tsc` numbers below are after 10 iterations (the JavaScript engine V8 JIT optimises it early already), which leads to the most conservative comparisons. Running a single iteration is 10x slower, and a cold `tsc` start is even slower because of the initial bootstrap delay of several hundred milliseconds. Therefore you can safely assume that `tsc` is slower than what the numbers shown below.
-Note that `tsc` numbers are after 10 iterations (the JavaScript engine V8 JIT optimises it early already), which somewhat
-leads to a wrong conclusion. Only 1 iteration is 10x slower and a cold `tsc` start even slower because of the initial bootstrap
-delay of several hundred milliseconds. So you can generally assume that `tsc` is slower than the numbers shown below.
+Note that compilation has not yet been optimised (for example, it still uses a slow memory allocator, the optimisation of which can result in another tenfold improvement).
### Basic variables
@@ -107,15 +104,11 @@ TypeRunner warm: 0.839308334ms (417x faster)
### Why?
-TypeScript is cool and all, but also very slow and not usable in other languages. We want to make it fast and type computation accessible to all languages.
+TypeScript has one of the most versatile type systems, but the official tooling is very slow and the types are not usable in other languages. We want to make it fast and bring type computation accessible to all languages.
-We believe that TypeScript can be much faster. There are a lot of workarounds used these days to work around the slow type checking of TypeScript.
-Some "TypeScript runtimes" even entirely disable type checking (and thus the key selling point of TypeScript) per default and rely on suboptimal editor-only type checking
-(which is still slow). We think it is possible to build a TypeScript compiler that is so fast, that you can run TypeScript files directly with all its type checking features
-enabled and still having almost instant compile time.
+We believe that TypeScript can be much faster. Many workarounds exist to mitigate slow type checking. Some "TypeScript runtimes" even entirely disable type checking (and thus the key selling point of TypeScript) by default and rely on suboptimal editor-only type checking (which is still slow). We think it is possible to build a TypeScript compiler that is so fast that you can run TypeScript files directly with all its type checking features enabled and still have almost instant compile time.
-We also believe that TypeScript is more than just a linter. It is a language that is so powerful and can be used for so much more.
-Imagine for example to write a schema like the following and be able to validate data against it in any language, no matter if JavaScript, C++, C, Rust, or Go.
+We also believe that TypeScript is more than just a linter. It is a language that is so powerful and can be used for much more. Imagine writing a schema such as the following and be able to validate data against it in any language: JavaScript, C++, C, Rust, or Go.
```typescript
export interface User {
@@ -127,56 +120,38 @@ export type ReadUser = Omit
export type CreateUser = Omit;
```
-With TypeRunner this will be possible and thus has the potential to replace JSON schema with a much more powerful way of defining schemas.
-It doesn't stop with JSON schema however: ORM, GraphQL, gRPC, Protocol buffer, and many more technologies that rely on schema information can rely
-on the powerful TypeScript type system to increase usability and efficiency dramatically. This brings isomorphic TypeScript to a whole new level.
+With TypeRunner this will be possible and thus has the potential to replace JSON schema with a much more powerful way of defining schemas. It doesn't stop with JSON schema however: ORM, GraphQL, gRPC, Protocol Buffers, and many more technologies that rely on schema information could rely on the powerful TypeScript type system to increase usability and efficiency dramatically. This brings isomorphic TypeScript to a whole new level.
### Why doesn't Microsoft implement it like this?
-Microsoft had (or still has) as goal to get a good and wide adoption of TypeScript. That means to find product-market fit and improve UX, iteration speed is key to success.
-Writing high-performance code usually impacts these goals negatively.
+Microsoft has had the goal to create wide adoption of TypeScript. That means to find product-market fit and improve UX, iteration speed is key to success. Writing high-performance code usually impacts these goals negatively.
-Also, implementing features to make adoption easier and [dogfooding](https://en.wikipedia.org/wiki/Eating_your_own_dog_food) are important strategies to make sure
-the product you're building actually will be adopted and solves real problems. Since the TypeScript team doesn't have unlimited resources this means however that
-they have to focus on what brings growth. Performance is not something that brings much growth, especially if you take into consideration that the market
-expands every 4 years or so by 100%, which means that most developers are beginners and have other priorities than performance.
+Also, implementing features to make adoption easier and [dogfooding](https://en.wikipedia.org/wiki/Eating_your_own_dog_food) are important strategies to make sure the product you're building will actually be adopted and solves real problems. However, since the TypeScript team doesn't have unlimited resources, this means that they have to focus on what brings growth. Performance is not something that brings much growth, especially if you take into consideration that the market expands every four years or so by 100%, which means that most developers are beginners and have priorities other than performance.
-Strategically speaking it makes thus sense for Microsoft to develop TypeScript the way they do it, at least for the moment.
+Strategically speaking it makes sense for Microsoft to develop TypeScript the way they do it, at least for the moment.
### Why not a drop-in replacement for `tsc`?
-TypeScript offers a lot of functionality. Many things are targeted at a smooth transition from JavaScript to TypeScript
-by allowing not-so-strict syntax and alternative type declaration syntax like JSDoc. Although these features
-have a big impact on new TypeScript users they are not so much used in projects where TypeScript is used more strictly.
-TypeScript comes also with a fairly big variety of transpilers/transformers to generate JavaScript for all kind of ECMAScript standards.
-We think that although these features have a big value for new users and adoption, we want to focus more on the enterprise side of things, where
-performance often translates directly to money.
+TypeScript offers a lot of functionality. Many things are targeted at a smooth transition from JavaScript to TypeScript by allowing not-so-strict syntax and alternative type declaration syntax like JSDoc. Although these features have a big impact on new TypeScript users they are not so much used in projects where TypeScript is used more strictly. TypeScript comes also with a fairly big variety of transpilers/transformers to generate JavaScript for all the different versions of of ECMAScript standards. While we think that although these features have a big value for new users and adoption, we want to focus more on the enterprise side of things where performance often translates directly to money.
### What is currently supported?
-Only very basic type expressions are currently supported: primitives, variable declaration, (generic) function declaration,
-some type functions like type aliases, (distributive) conditional types, template literals, array/tuples, index access, union, rest, and a few other things.
-They are implemented in a very rough way just enough to proof that the language in its core can be implemented in a fast way.
+Only very basic type expressions are currently supported: primitives, variable declaration, (generic) function declaration, some type functions like type aliases, (distributive) conditional types, template literals, array/tuples, index access, union, rest, and a few other things. They are implemented in a very rough way just enough to proof that the language in its core can be implemented in a fast way.
### When will it be released?
-Currently, the development of TypeRunner is almost stalled and was more an experiment/proof of concept. To actually make this a reality, funding is necessary.
-In the next coming weeks, we will set up a kickstarter/funding project, which makes it possible for the community that is interested in high-speed TypeScript to
-make this a reality.
+Currently, the development of TypeRunner is almost stalled and was more an experiment/proof of concept. To actually make this a reality, funding is necessary. In the coming weeks, we will set up a kickstarter/funding project, which makes it possible for the community that is interested in high-speed TypeScript to make this a reality.
### Why C++ and not Rust?
-Because I know C++ much better than Rust. The market of good C++ developers is much bigger. TypeScript code also maps surprisingly good to C++, so porting
-the scanner, parser, and AST structure is actually rather easy, which allows back-porting features from TypeScript tsc to TypeRunner much better. I also find Rust ugly.
+Because I know C++ much better than Rust. The market of good C++ developers is much bigger. TypeScript code also maps surprisingly well to C++, so porting
+the scanner, parser, and AST structure is actually rather easy, which allows back-porting features from TypeScript tsc to TypeRunner much easier. I also find Rust ugly.
### Why not port TypeScript?
-Porting TypeScript alone to a faster language won't yield much performance difference since the V8 JIT optimises TypeScript compiler already good enough and the
-current slowdown comes mainly from architectural decisions made to work well in a JavaScript runtime (where optimisation strategies are very limited).
-To get a much faster compiler you have to do both, use a fast compiling language and use an architecture that utilizes all performance advantages of that language.
+Porting TypeScript alone to a faster language won't yield much performance difference since the V8 JIT optimises TypeScript compiler already, and the current slowdown comes mainly from architectural decisions made to work well in a JavaScript runtime (where optimisation strategies are very limited). To get a much faster compiler you have to do both: use a fast compiling language and use an architecture that utilizes all performance advantages of that language.
-That being said, in the current stage the actual TypeScript scanner/parser code is ported from the TypeScript compiler to make back-porting easier.
-It's also at the moment the most complicated and slowest code (with over 20k LoC), which will be improved in the future.
+That being said, in the current stage the actual TypeScript scanner/parser code is ported from the TypeScript compiler to make back-porting easier. It's also at the moment the most complicated and slowest code (with over 20k LoC), which will be improved in the future.
### What can I do to support this project?
@@ -184,22 +159,15 @@ Follow me on Twitter and fund it once the funding campaign is published: [twitte
### How does it work/Why is it so fast?
-It is so fast because it does the heavy-lifting only once per file and optimised the type computation/comparison in a custom virtual machine.
+It is fast because it does the heavy-lifting only once per file and the type computation/comparison is optimised in a custom virtual machine.
-The compiler consists of 3 stages: Parsing to AST, compiling to bytecode, and executing the bytecode in a virtual machine.
-The first two stages are the most expensive ones and have to be done for each file once. The resulting bytecode can then be cached on disk.
+The compiler consists of 3 stages: parsing to AST, compiling to bytecode, and executing the bytecode in a virtual machine. The first two stages are the most expensive and have to be done for each file once. The resulting bytecode can then be cached on disk.
-The compiler has a simple data-flow analysis implemented to generate more efficient bytecode
-(like tail call/tail rest detection, and detecting usage of potential usage of super instructions).
-It also tries to precompute as much as possible: for example all literal values will have in the bytecode header a precomputed 64bit hash,
-so that the virtual machine can compare literal types with a simple integer comparison and use it in hash-tables.
+The compiler has a simple data-flow analysis implemented to generate more efficient bytecode (like tail call/tail rest detection, and detecting usage of potential usage of super instructions). It also tries to precompute as much as possible. For example, all literal values will have in the bytecode header a precomputed 64-bit hash so that the virtual machine can compare literal types with a simple integer comparison and use it in hash-tables.
-The stack based virtual machine tries to be lightweight and high-performant as the most work happens here.
-It uses multiple memory pools with garbage collection, fix-sized type structure, tail call optimisation, tuple re-use (tail rest optimisation).
-For list structured linked lists and hash-tables (based on precomputed literal hash) are used to make lookups very fast.
+The stack based virtual machine tries to be lightweight and highly-performant as the most work happens here. It uses multiple memory pools with garbage collection, fix-sized type structure, tail call optimisation, and tuple re-use (tail rest optimisation). For list-structures, linked lists and hash tables (based on precomputed literal hash) are used to make lookups very fast.
-The instruction set consists of currently 83 instructions: https://github.com/marcj/TypeRunner/blob/master/src/checker/instructions.h
-and will likely grow in the future.
+The instruction set consists of currently 83 instructions: https://github.com/marcj/TypeRunner/blob/master/src/checker/instructions.h and will likely grow in the future.
## Development
@@ -228,4 +196,4 @@ $ make bench -j 8
$ ./bench ../tests/objectLiterals1.ts
```
-Now you find in the build folder some binaries you can execute.
\ No newline at end of file
+Binaries are created in the build folder, which can now be executed.