Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

proc_macro/bridge: Eagerly pass TokenStream contents into the client #101419

Closed
wants to merge 1 commit into from

Conversation

mystor
Copy link
Contributor

@mystor mystor commented Sep 4, 2022

This change could improve performance, especially in cases like with the
cross-thread backend which have expensive RPC. It handles this by moving
more of the logic for TokenStream into the client, storing a
Rc<Vec<TokenTree>> directly, rather than converting to/from the
compiler representation over RPC for every change.

As this eagerly decomposes interpolated AST fragments, this ended up
changing the output of TokenStream::to_string(). To keep up
compatibility for now, formatting is still performed using RPC and
rustc_ast_pretty, however in the future we may look into decoupling
proc-macro formatting from rustc_ast_pretty, and doing it in the
client.

One side-effect of this was that the workaround for the
procedural-masquerade crate had to be updated to work with the new
approach, as we couldn't depend on the AST fragment printer to format
the input. To do this, a real , token is now inserted when passing the
enum to a proc-macro rather than depending on the pretty-printing
behaviour. From local tests it appears that the new output should be
compatible enough to work with older versions of procedural-masquerade.

The new conversion between the proc-macro and compiler representations
of a TokenStream is now handled by the RpcContext type provided by
the proc-macro server. This type is a distinct value which will be
fetched and stored by the server before commands are run, such that the
decoders can have access to relevant state. In the future, we may also
change the symbol handling code to work off of RpcContext rather than
being static methods on the Server trait.

@rustbot rustbot added the T-compiler Relevant to the compiler team, which will review and decide on the PR/issue. label Sep 4, 2022
@rust-highfive
Copy link
Contributor

r? @petrochenkov

(rust-highfive has picked a reviewer for you, use r? to override)

@rust-highfive rust-highfive added the S-waiting-on-review Status: Awaiting review from the assignee but also interested parties. label Sep 4, 2022
@mystor
Copy link
Contributor Author

mystor commented Sep 4, 2022

Posting this change as a work-in-progress change, which is not ready to merge. I think we'll want to run this through crater before we move too much further with it (assuming it passes CI), to make sure it doesn't cause breakage in the ecosystem.

r? @eddyb

@rust-highfive rust-highfive assigned eddyb and unassigned petrochenkov Sep 4, 2022
@mystor
Copy link
Contributor Author

mystor commented Sep 4, 2022

Another potential concern here is that the code for serializing and deserializing TokenStream with this patch is written in a recursive manner, as is the stringification/formatting code. I think the formatting code previously (from rustc_ast_pretty) was also recursive, but this could theoretically exhaust the stack for sufficiently nested TokenTrees.

That being said, it should be possible to rework things to avoid using the stack in the future, but keeping it this way for now makes everything much simpler.

@eddyb
Copy link
Member

eddyb commented Sep 4, 2022

Might be too early for this, but curious anyway:
@bors try @rust-timer queue

@rust-timer
Copy link
Collaborator

Awaiting bors try build completion.

@rustbot label: +S-waiting-on-perf

@rustbot rustbot added the S-waiting-on-perf Status: Waiting on a perf run to be completed. label Sep 4, 2022
@bors
Copy link
Collaborator

bors commented Sep 4, 2022

⌛ Trying commit 7a64bb64f92fa2c6741c3b4938804514e46b0097 with merge 0c7f4114a5582dfa18524643326a0e87bc6d62f3...

@rust-log-analyzer

This comment has been minimized.

@bors
Copy link
Collaborator

bors commented Sep 4, 2022

☀️ Try build successful - checks-actions
Build commit: 0c7f4114a5582dfa18524643326a0e87bc6d62f3 (0c7f4114a5582dfa18524643326a0e87bc6d62f3)

@rust-timer
Copy link
Collaborator

Queued 0c7f4114a5582dfa18524643326a0e87bc6d62f3 with parent a2cdcb3, future comparison URL.

@mystor
Copy link
Contributor Author

mystor commented Sep 4, 2022

Looks like I missed 2 proc macros in tests with changes to their output formats. Fortunately shouldn't be hard to update them.

@mystor
Copy link
Contributor Author

mystor commented Sep 5, 2022

It occurred to me that pretty-printing a large TokenStream might be rare enough nowadays that we can afford sending the entire TokenStream over IPC just to pretty-print it, at least as a stopgap, to avoid some of the pretty-printing changes from also happening in this patch.

I still have the other changes to do it more efficiently, but we can potentially leave that for a follow-up if the perf on this approach is OK.

This still requires the changes to the pretty-printing hack for procedural-masquerade because the interpolated node can't be used anymore.

@mystor
Copy link
Contributor Author

mystor commented Sep 5, 2022

cc #73933 (@Aaron1011 @petrochenkov), as that issue tracks removing the hack completely.

This change could improve performance, especially in cases like with the
cross-thread backend which have expensive RPC. It handles this by moving
more of the logic for TokenStream into the client, storing a
`Rc<Vec<TokenTree>>` directly, rather than converting to/from the
compiler representation over RPC for every change.

As this eagerly decomposes interpolated AST fragments, this ended up
changing the output of `TokenStream::to_string()`. To keep up
compatibility for now, formatting is still performed using RPC and
`rustc_ast_pretty`, however in the future we may look into decoupling
proc-macro formatting from `rustc_ast_pretty`, and doing it in the
client.

One side-effect of this was that the workaround for the
procedural-masquerade crate had to be updated to work with the new
approach, as we couldn't depend on the AST fragment printer to format
the input. To do this, a real `,` token is now inserted when passing the
enum to a proc-macro rather than depending on the pretty-printing
behaviour. From local tests it appears that the new output should be
compatible enough to work with older versions of procedural-masquerade.

The new conversion between the proc-macro and compiler representations
of a `TokenStream` is now handled by the `RpcContext` type provided by
the proc-macro server. This type is a distinct value which will be
fetched and stored by the server before commands are run, such that the
decoders can have access to relevant state. In the future, we may also
change the symbol handling code to work off of `RpcContext` rather than
being static methods on the `Server` trait.
@rust-timer
Copy link
Collaborator

Finished benchmarking commit (0c7f4114a5582dfa18524643326a0e87bc6d62f3): comparison URL.

Overall result: ❌✅ regressions and improvements - ACTION NEEDED

Benchmarking this pull request likely means that it is perf-sensitive, so we're automatically marking it as not fit for rolling up. While you can manually mark this PR as fit for rollup, we strongly recommend not doing so since this PR may lead to changes in compiler perf.

Next Steps: If you can justify the regressions found in this try perf run, please indicate this with @rustbot label: +perf-regression-triaged along with sufficient written justification. If you cannot justify the regressions please fix the regressions and do another perf run. If the next run shows neutral or positive results, the label will be automatically removed.

@bors rollup=never
@rustbot label: +S-waiting-on-review -S-waiting-on-perf +perf-regression

Instruction count

This is a highly reliable metric that was used to determine the overall result at the top of this comment.

mean1 range count2
Regressions ❌
(primary)
1.2% [0.3%, 4.3%] 35
Regressions ❌
(secondary)
0.9% [0.7%, 1.0%] 4
Improvements ✅
(primary)
- - 0
Improvements ✅
(secondary)
-15.9% [-43.7%, -1.0%] 11
All ❌✅ (primary) 1.2% [0.3%, 4.3%] 35

Max RSS (memory usage)

Results

This is a less reliable metric that may be of interest but was not used to determine the overall result at the top of this comment.

mean1 range count2
Regressions ❌
(primary)
4.3% [4.3%, 4.3%] 1
Regressions ❌
(secondary)
3.0% [2.6%, 3.5%] 2
Improvements ✅
(primary)
- - 0
Improvements ✅
(secondary)
-3.0% [-3.9%, -2.1%] 2
All ❌✅ (primary) 4.3% [4.3%, 4.3%] 1

Cycles

Results

This is a less reliable metric that may be of interest but was not used to determine the overall result at the top of this comment.

mean1 range count2
Regressions ❌
(primary)
3.0% [1.4%, 6.7%] 8
Regressions ❌
(secondary)
- - 0
Improvements ✅
(primary)
- - 0
Improvements ✅
(secondary)
-12.5% [-31.7%, -2.2%] 10
All ❌✅ (primary) 3.0% [1.4%, 6.7%] 8

Footnotes

  1. the arithmetic mean of the percent change 2 3

  2. number of relevant changes 2 3

@rustbot rustbot added perf-regression Performance regression. and removed S-waiting-on-perf Status: Waiting on a perf run to be completed. labels Sep 5, 2022
@mystor
Copy link
Contributor Author

mystor commented Sep 8, 2022

These results so far make sense. The runtime performance is better (as noticed by e.g. token-stream-stress which improved check by 43%), but I haven't manually optimized away inlining opportunities in the Extend and FromIterator implementations, so the optimizer spends substantially more time on calls to those functions in crates like syn and serde-derive, slowing down those builds. Once I've minimized the generic code in those functions those regressions should hopefully go away a bit.

@bors
Copy link
Collaborator

bors commented Oct 1, 2022

☔ The latest upstream changes (presumably #101986) made this pull request unmergeable. Please resolve the merge conflicts.

@jackh726 jackh726 added S-waiting-on-author Status: This is awaiting some action (such as code changes or more information) from the author. and removed S-waiting-on-review Status: Awaiting review from the assignee but also interested parties. labels Feb 10, 2023
@Dylan-DPC
Copy link
Member

@mystor any updates on this?

@mystor
Copy link
Contributor Author

mystor commented May 16, 2023

No, I unfortunately haven't had the time to dig into fixing this patch up & getting it ready to be looked into more.

bors added a commit to rust-lang-ci/rust that referenced this pull request Oct 6, 2023
[perf] Do not reuse internal `TokenStream` in `proc_macro::TokenStream`

Use a separate type instead, so that we need to convert from one `TokenStream` to another and back on proc macro boundaries.
I want to check how much it affects performance.

This PR also implements the suggestion to censor token jointness at proc macro boundary from rust-lang#114571 (since we now have such a boundary).
This PR is also semi-related to rust-lang#101419.

I don't like the result and will likely close this after the perf test.
@Dylan-DPC
Copy link
Member

Closing this as inactive. Feel free to reöpen this pr or create a new pr if you get the time to work on this. Thanks

@Dylan-DPC Dylan-DPC closed this Oct 7, 2023
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
perf-regression Performance regression. S-waiting-on-author Status: This is awaiting some action (such as code changes or more information) from the author. T-compiler Relevant to the compiler team, which will review and decide on the PR/issue.
Projects
None yet
Development

Successfully merging this pull request may close these issues.

10 participants