Skip to content

Commit

Permalink
Add bit packing to NodeImpl (#4651)
Browse files Browse the repository at this point in the history
Just a small packing optimization. We currently have 222 `NodeKinds`, so
this reduces us to just 30ish more we can add without needing to pack
more. However, if we did, there would be a couple options for bringing
the count down by reusing `NodeKinds` and disambiguating based on the
token kind (the 29 infix operators as an example). Or we could just undo
this.

I'm expecting this to yield a small improvement. I'll see if I can get
better numbers since my machine's not really reliable, but here are some
basic values.

Also suggesting to draw the use of `::RawEnumType` for `TokenKind`,
since bit packing appears to work without it. Hoping the `static_assert`
is easier for people to understand the size of the field.

With the change:

```
----------------------------------------------------------------------------------------------------------------------------
Benchmark                                                 Time             CPU   Iterations      Bytes      Lines     Tokens
----------------------------------------------------------------------------------------------------------------------------
BM_CompileAPIFileDenseDecls<Phase::Parse>/256         50399 ns        50359 ns        14336 104.588M/s 3.87217M/s 21.8629M/s
BM_CompileAPIFileDenseDecls<Phase::Parse>/1024       237823 ns       237629 ns         3072 136.721M/s 4.11986M/s 24.2058M/s
BM_CompileAPIFileDenseDecls<Phase::Parse>/4096       997645 ns       996771 ns          768 142.343M/s 4.04105M/s 23.9363M/s
BM_CompileAPIFileDenseDecls<Phase::Parse>/16384     4020308 ns      4018319 ns          192 152.041M/s 4.05966M/s 24.0874M/s
BM_CompileAPIFileDenseDecls<Phase::Parse>/65536    16691390 ns     16683058 ns           48 151.317M/s 3.92374M/s 23.2936M/s
BM_CompileAPIFileDenseDecls<Phase::Parse>/262144   75265735 ns     75233476 ns            8 135.842M/s 3.48421M/s 20.6862M/s
```

Without the change:
```
----------------------------------------------------------------------------------------------------------------------------
Benchmark                                                 Time             CPU   Iterations      Bytes      Lines     Tokens
----------------------------------------------------------------------------------------------------------------------------
BM_CompileAPIFileDenseDecls<Phase::Parse>/256         51515 ns        51480 ns        13312 102.312M/s 3.78789M/s  21.387M/s
BM_CompileAPIFileDenseDecls<Phase::Parse>/1024       241040 ns       240900 ns         3072 134.865M/s 4.06392M/s 23.8771M/s
BM_CompileAPIFileDenseDecls<Phase::Parse>/4096       985593 ns       984657 ns          768 144.094M/s 4.09077M/s 24.2308M/s
BM_CompileAPIFileDenseDecls<Phase::Parse>/16384     4109327 ns      4105496 ns          192 148.813M/s 3.97345M/s  23.576M/s
BM_CompileAPIFileDenseDecls<Phase::Parse>/65536    17459655 ns     17446006 ns           48   144.7M/s 3.75215M/s  22.275M/s
BM_CompileAPIFileDenseDecls<Phase::Parse>/262144   80802815 ns     80737489 ns            8 126.581M/s 3.24668M/s  19.276M/s
```

---------

Co-authored-by: Chandler Carruth <[email protected]>
  • Loading branch information
jonmeow and chandlerc authored Dec 17, 2024
1 parent c04d62a commit 08f2455
Show file tree
Hide file tree
Showing 8 changed files with 44 additions and 32 deletions.
5 changes: 2 additions & 3 deletions toolchain/lex/lex.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -1396,13 +1396,12 @@ auto Lexer::Finalize() -> void {
// many identifiers to fit an `IdentifierId` into a `token_payload_`, and
// likewise for `IntId` and so on. If we start adding any of those IDs prior
// to lexing, we may need to also limit the number of those IDs here.
if (buffer_.token_infos_.size() > TokenizedBuffer::MaxTokens) {
if (buffer_.token_infos_.size() > TokenIndex::Max) {
CARBON_DIAGNOSTIC(TooManyTokens, Error,
"too many tokens in source file; try splitting into "
"multiple source files");
// Subtract one to leave room for the `FileEnd` token.
token_emitter_.Emit(TokenIndex(TokenizedBuffer::MaxTokens - 1),
TooManyTokens);
token_emitter_.Emit(TokenIndex(TokenIndex::Max - 1), TooManyTokens);
// TODO: Convert tokens after the token limit to error tokens to avoid
// misinterpretation by consumers of the tokenized buffer.
}
Expand Down
6 changes: 6 additions & 0 deletions toolchain/lex/token_index.h
Original file line number Diff line number Diff line change
Expand Up @@ -24,6 +24,12 @@ namespace Carbon::Lex {
//
// All other APIs to query a `TokenIndex` are on the `TokenizedBuffer`.
struct TokenIndex : public IndexBase<TokenIndex> {
// The number of bits which must be allotted for `TokenIndex`.
static constexpr int Bits = 23;
// The maximum number of tokens that can be stored, including the FileStart
// and FileEnd tokens.
static constexpr int Max = 1 << Bits;

static constexpr llvm::StringLiteral Label = "token";
static const TokenIndex Invalid;
// Comments aren't tokenized, so this is the first token after FileStart.
Expand Down
16 changes: 6 additions & 10 deletions toolchain/lex/tokenized_buffer.h
Original file line number Diff line number Diff line change
Expand Up @@ -85,10 +85,6 @@ class TokenDiagnosticConverter : public DiagnosticConverter<TokenIndex> {
// `HasError` returning true.
class TokenizedBuffer : public Printable<TokenizedBuffer> {
public:
// The maximum number of tokens that can be stored in the buffer, including
// the FileStart and FileEnd tokens.
static constexpr int MaxTokens = 1 << 23;

// A comment, which can be a block of lines.
//
// This is the API version of `CommentData`.
Expand Down Expand Up @@ -281,7 +277,7 @@ class TokenizedBuffer : public Printable<TokenizedBuffer> {
class TokenInfo {
public:
// The kind for this token.
auto kind() const -> TokenKind { return TokenKind::Make(kind_); }
auto kind() const -> TokenKind { return kind_; }

// Whether this token is preceded by whitespace. We only store the preceding
// state, and look at the next token to check for trailing whitespace.
Expand Down Expand Up @@ -364,6 +360,7 @@ class TokenizedBuffer : public Printable<TokenizedBuffer> {

// Make sure we have enough payload bits to represent token-associated IDs.
static_assert(PayloadBits >= IntId::TokenIdBits);
static_assert(PayloadBits >= TokenIndex::Bits);

// Constructor for a TokenKind that carries no payload, or where the payload
// will be set later.
Expand Down Expand Up @@ -397,13 +394,12 @@ class TokenizedBuffer : public Printable<TokenizedBuffer> {
// Payload values are typically ID types for which we create at most one per
// token, so we ensure that `token_payload_` is large enough to fit any
// token index. Stores to this field may overflow, but we produce an error
// in `Lexer::Finalize` if the file has more than `MaxTokens` tokens, so
// this value never overflows if lexing succeeds.
TokenKind::RawEnumType kind_ : sizeof(TokenKind) * 8;
// in `Lexer::Finalize` if the file has more than `TokenIndex::Max` tokens,
// so this value never overflows if lexing succeeds.
TokenKind kind_;
static_assert(sizeof(kind_) == 1, "TokenKind must pack to 8 bits");
bool has_leading_space_ : 1;
unsigned token_payload_ : PayloadBits;
static_assert(MaxTokens <= 1 << PayloadBits,
"Not enough payload bits to store a token index");

// Separate storage for the byte offset, this is hot while lexing but then
// generally cold.
Expand Down
6 changes: 3 additions & 3 deletions toolchain/lex/tokenized_buffer_test.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -1160,9 +1160,9 @@ TEST_F(LexerTest, DiagnosticFileTooLarge) {
input += "{}\n";
}
EXPECT_CALL(consumer,
HandleDiagnostic(IsSingleDiagnostic(
DiagnosticKind::TooManyTokens, DiagnosticLevel::Error,
TokenizedBuffer::MaxTokens / 2, 1, _)));
HandleDiagnostic(IsSingleDiagnostic(DiagnosticKind::TooManyTokens,
DiagnosticLevel::Error,
TokenIndex::Max / 2, 1, _)));
compile_helper_.GetTokenizedBuffer(input, &consumer);
}

Expand Down
4 changes: 2 additions & 2 deletions toolchain/parse/context.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -45,8 +45,8 @@ auto Context::ReplacePlaceholderNode(int32_t position, NodeKind kind,
CARBON_CHECK(position >= 0 && position < tree_->size(),
"position: {0} size: {1}", position, tree_->size());
auto* node_impl = &tree_->node_impls_[position];
CARBON_CHECK(node_impl->kind == NodeKind::Placeholder);
*node_impl = {.kind = kind, .has_error = has_error, .token = token};
CARBON_CHECK(node_impl->kind() == NodeKind::Placeholder);
*node_impl = Tree::NodeImpl(kind, has_error, token);
}

auto Context::ConsumeAndAddOpenParen(Lex::TokenIndex default_token,
Expand Down
3 changes: 1 addition & 2 deletions toolchain/parse/context.h
Original file line number Diff line number Diff line change
Expand Up @@ -107,8 +107,7 @@ class Context {
kind != NodeKind::InvalidParseStart &&
kind != NodeKind::InvalidParseSubtree),
"{0} nodes must always have an error", kind);
tree_->node_impls_.push_back(
{.kind = kind, .has_error = has_error, .token = token});
tree_->node_impls_.push_back(Tree::NodeImpl(kind, has_error, token));
}

// Adds an invalid parse node.
Expand Down
2 changes: 1 addition & 1 deletion toolchain/parse/tree.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -23,7 +23,7 @@ auto Tree::postorder() const -> llvm::iterator_range<PostorderIterator> {

auto Tree::node_token(NodeId n) const -> Lex::TokenIndex {
CARBON_CHECK(n.is_valid());
return node_impls_[n.index].token;
return node_impls_[n.index].token();
}

auto Tree::Print(llvm::raw_ostream& output) const -> void {
Expand Down
34 changes: 23 additions & 11 deletions toolchain/parse/tree.h
Original file line number Diff line number Diff line change
Expand Up @@ -120,13 +120,13 @@ class Tree : public Printable<Tree> {
// full expected structure of the grammar.
auto node_has_error(NodeId n) const -> bool {
CARBON_DCHECK(n.is_valid());
return node_impls_[n.index].has_error;
return node_impls_[n.index].has_error();
}

// Returns the kind of the given parse tree node.
auto node_kind(NodeId n) const -> NodeKind {
CARBON_DCHECK(n.is_valid());
return node_impls_[n.index].kind;
return node_impls_[n.index].kind();
}

// Returns the token the given parse tree node models.
Expand Down Expand Up @@ -201,12 +201,24 @@ class Tree : public Printable<Tree> {

// The in-memory representation of data used for a particular node in the
// tree.
struct NodeImpl {
// The kind of this node. Note that this is only a single byte.
NodeKind kind;
class NodeImpl {
public:
explicit NodeImpl(NodeKind kind, bool has_error, Lex::TokenIndex token)
: kind_(kind), has_error_(has_error), token_index_(token.index) {
CARBON_DCHECK(token.index >= 0, "Unexpected token for node: {0}", token);
}

// We have 3 bytes of padding here that we can pack flags or other compact
// data into.
auto kind() const -> NodeKind { return kind_; }
auto set_kind(NodeKind kind) -> void { kind_ = kind; }
auto has_error() const -> bool { return has_error_; }
auto token() const -> Lex::TokenIndex {
return Lex::TokenIndex(token_index_);
}

private:
// The kind of this node. Note that this is only a single byte.
NodeKind kind_;
static_assert(sizeof(kind_) == 1, "TokenKind must pack to 8 bits");

// Whether this node is or contains a parse error.
//
Expand All @@ -221,20 +233,20 @@ class Tree : public Printable<Tree> {
// optional (and will depend on the particular parse implementation
// strategy). The goal is that you can rely on grammar-based structural
// invariants *until* you encounter a node with this set.
bool has_error;
bool has_error_ : 1;

// The token root of this node.
Lex::TokenIndex token;
unsigned token_index_ : Lex::TokenIndex::Bits;
};

static_assert(sizeof(NodeImpl) == 8,
static_assert(sizeof(NodeImpl) == 4,
"Unexpected size of node implementation!");

// Sets the kind of a node. This is intended to allow putting the tree into a
// state where verification can fail, in order to make the failure path of
// `Verify` testable.
auto SetNodeKindForTesting(NodeId node_id, NodeKind kind) -> void {
node_impls_[node_id.index].kind = kind;
node_impls_[node_id.index].set_kind(kind);
}

// Depth-first postorder sequence of node implementation data.
Expand Down

0 comments on commit 08f2455

Please sign in to comment.