-
Notifications
You must be signed in to change notification settings - Fork 799
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Support semantic highlighting #18
Comments
Fully agree and we already discussed it since we would like to have it in VS Code as well. |
I'd love to have this for the PowerShell extension also. We provide the ability to create "dynamic keywords" for the purpose of writing domain-specific languages. Semantic highlighting would allow us to colorize those keywords in VS Code even though they aren't part of the PowerShell language spec. /cc @BrucePay |
I was looking for that as well. What's exactly the meaning of the current 'highlighting' for read/write/text (see document highlights)? I implemented that part but didn't see any for of visual feedback in the editor. |
DocumentHighlight is for "mark occurrences" |
@cdietrich sorry for hijacking the topic, but this is very unclear to me. How is document highlights used to realize a 'mark occurrences" feature? How does the read/write/text distinction fit in, and why does it only expect a single result in return? For mark occurrences, I'd expect to be able to return a collection, no? Would be great if that could be clarified also in |
@smarr yes you are right. this makes no sense. i asume the return type should be an array.
thus looks like a bug in the protocol |
@cdietrich thanks, I'll open a separate issue. |
Please find a proposal in PR #124 |
I’m guessing this is dead... anyone got any updates? |
I am working on a new proposal for semantic coloring. |
Is there a link where the updated protocol can be reviewed? Is it still primarily in the typescript server? |
@axelson the proposed version is here in terms of implementation. I am in the process of writing the markdown. https://github.com/microsoft/vscode-languageserver-node/blob/master/protocol/src/common/protocol.semanticTokens.proposed.ts#L1 |
And here is a first version of the spec. No word polish and no spell check :-) |
@dbaeumer When will we need to move our LSP server implementation over to this one from the current semantic tokens implementation supported in vscode? |
@kjeremy Are you relying on the implementation in the next version of the LSP libs ? |
@dbaeumer Our server-side implementation is in rust and based on https://github.com/gluon-lang/lsp-types which will need to be updated. The client-side opts in via https://github.com/rust-analyzer/rust-analyzer/blob/master/editors/code/src/client.ts#L151 |
@dbaeumer I really think a remark that clients are expected to cache locally with ranges would be quite important unless you want to make deltas de-facto required for servers. Right now, I can't see such a remark, it should probably go into "Implementation considerations." (Unless you don't think it should be optional @ deltas, but you did sound like you wanted to keep it in but as optional.) If you don't write that in, some clients will just not do such caching and then servers will be required to implement deltas to guarantee good performance or users will blame it on them, making deltas basically a must. Edit: also, right now the general concepts part is mixed with the pretty specific integer encoding in one block, I find that this is suboptimal for readability. Suggestion in detail to split up text:I think the General Concepts section would benefit from the integer encoding part split into a separate section placed separately. In detail, I am suggesting to split at the following start/end points: Split start at:
I would start to cut out parts for a new section after this, named Split end at:
So the new
... with the
This new section could be additionally prefixed with an introducing sentence like: |
I'm trying to figure out the delta behavior and have some questions:
I guess ultimately I'm trying to figure out what I need to hold onto to compute the deltas. I am assuming that each deltas really asks for the delta between "now" and the previous delta or full request. It does not appear to be spec'd that way however and it could be that a client asks for a delta between "now" and many revisions ago. |
@dbaeumer i'm reading through the initial draft spec (thanks for writing it up!). I have a handful of comments/questions/clarifications. What's the best way to provide feedback? would you like comments on a PR or something like that? |
@puremourning I think the comments are best provided as a PR. |
@kjeremy you are correct a delta for range makes no sense. I am pretty sure VS Code will basically drop a previous result ID. For the delta is should be always reported against the last result independent whether this was a full or a delta response. So in an easy implementation the id is simply incremented. I have clarified this in the spec. |
@etc0de thanks for the suggestions. I made them in the 3.16 version of the spec. |
If I understand it correctly, the basic idea is to let the client pull semantic tokens for all open files on a regular basis, the one having the focus with high frequency, the other files with low frequency (but automatically). That would allow for updating the semantic highlighting in all open files, even if file A has the focus and a change here does induce a change in highlighting in another file B - am I correct? If so, the current reference implementation in VSCode Insiders (1,49.0-insiders) does not implement automated pulls of semantic tokens of open files, right? Wouldn't it be good to describe the expected pull behavior in the 3.16 specs of LSP? |
Now that my LSP server is more stable, I decided to update the extension for VSCode and give the semantic highlighting feature a spin. On the server, I decided to first try out SemanticTokensOptions { range = false, full = true}. However, I see the following:
Well, yeah, that's right. There is no "textDocument/semanticTokens" message because it only mentions in the spec "textDocument/semanticTokens/range", "textDocument/semanticTokens/full", and "textDocument/semanticTokens/full/delta", but no "textDocument/semanticTokens". Is this spec up to date? I can debug VSCode, figure out what is it expecting, and program to the implementation, but it would be nice to know what's going on. |
@kaby76 I think the request is constructed by the vscode-languageclient package so the issue might be that the extension is not using the latest version of that package (7.0.0-next.9 is the one I've been using to test in the Dart extension and it seems to work well - though I've only implemented |
@MarFren adding this as a recommendation to the spec makes sense. A PR is welcome. However the LSP spec never enforces this. If a client decides to not do this it should still be fine. |
@dbaeumer in the spec for
It seems like the line is incomplete (or the last part shouldn't be there). If the return value is |
@DanTup That worked. I had to also set up a few other dependencies
use the "Insiders" VSCode, and change an import because LanguageClient was moved around:
But, "textDocument/semanticTokens/full" starting to work. Computing the start line/col deltas a little challenging. And VSCode does not work the same as the LSP client in VS2019 with edits. |
I had to make the
I've been refactoring my server work to collect the tokens using absolute data initially (line/cols, enum types) and then at the end do the conversion to the LSP format. This made it much simpler than when I was also doing things like splitting up multiline/nested tokens at the same time. The final conversion is now relatively simple: var lastLine = 0;
var lastColumn = 0;
_tokens.sort(
(t1, t2) => t1.line == t2.line
? t1.column.compareTo(t2.column)
: t1.line.compareTo(t2.line),
);
for (final token in _tokens) {
var relativeLine = token.line - lastLine;
var relativeColumn = relativeLine == 0
? token.column - lastColumn
: token.column;
encodedTokens.addAll([
relativeLine,
relativeColumn,
token.length,
semanticTokenLegend.indexForType(token.type),
semanticTokenLegend.bitmaskForModifiers(token.modifiers) ?? 0
]);
lastLine = token.line;
lastColumn = token.column;
} |
@DanTup Yes, my code looks more or less just like your code after I noticed initially that only the first symbol was being colored. I tried out some hardwired values, then understood what had to be done (i.e., sort + compute diffs). |
I will close the issue since SC is now part of the upcoming 3.16 spec. |
Like WebStorm and VS does: eg. Symbol is a type or parameter or namespace or unresolved ...
Textmate based grammars are hard to do this. Since we did support Diagnostics, why not support semantic highlighting?
The text was updated successfully, but these errors were encountered: