sammccall added a comment.

In D93531#2463126 <https://reviews.llvm.org/D93531#2463126>, @njames93 wrote:

> In D93531#2463052 <https://reviews.llvm.org/D93531#2463052>, @sammccall wrote:
>
>> This adds a bit of complexity, making the code here a fair amount harder to 
>> follow and verify the correctness of.
>>
>> - Do we have evidence that these allocations are causing a problem? (e.g. do 
>> we observe a significant decrease in RSS after the patch)? Naively I would 
>> expect these allocations to be basically unimportant compared to those of 
>> the JSON objects themselves.(And I don't particularly expect either of them 
>> to be significant - the comment on the other review was really just "0 
>> probably isn't the right arg to  malloc_trim if there's any allocation going 
>> on").
>
> It's not causing problems per say. but given the incoming json messages can 
> contain a whole file plus things like escape chars. its wise to allocate a 
> buffer that will grow to the largest json it receives but never shrink.

Well, it's also wise to measure before optimizing. The large requests (ones 
containing a whole file) typically cause us to parse a C++ TU. After we read 
this JSON message, we will immediately:

- allocate a std::string to hold the decoded file content payload as part of 
the json::Value
- copy that string into the DraftManager
- allocate a MemoryBuffer and copy the file into it, for parsing
- read all of the thousands of transitively included headers into 
newly-allocated MemoryBuffers
- run the clang parser and then the clangd indexer

I think the most likely case is that this change has no observable impact on 
performance or resource usage. If that's true, I don't want to land it - it 
makes the code harder to reason about for no detectable benefit.

It's definitely *plausible* there's something surprising going on with 
allocation patterns such that this makes a difference. The impact of 
`malloc_trim` shows that clangd's overall allocation patterns have surprising 
effects. But right now I don't see any reason to think that these allocations 
matter.

>> - there seem to be simpler ways to structure this avoiding allocations. 
>> JSONLineBuffer is effectively statically bounded, and can be 
>> `SmallString<32>` or so. The content buffer could simply be passed in if I'm 
>> reading right: `bool readRawMessage(std::string&)`? OutputBuffer probably 
>> does need to be a member variable though.
>
> Its not statically bounded unfortunately, the length is how ever long a line 
> in the json is, which can be infinite. the read line function uses a buffer 
> size of 1024 as an upperbound.

JSONLineBuffer is not used for reading JSON, but rather HTTP-style headers. The 
only headers supported are `Content-Length` and `Content-Type`, and I don't 
believe clients send Content-Type in practice (there's only one supported). In 
any case, you'll never see a line beyond 256 chars from a well-behaved client. 
("statically bounded" is too strong, but from a performance-tuning perspective 
it's enough).


Repository:
  rG LLVM Github Monorepo

CHANGES SINCE LAST ACTION
  https://reviews.llvm.org/D93531/new/

https://reviews.llvm.org/D93531

_______________________________________________
cfe-commits mailing list
cfe-commits@lists.llvm.org
https://lists.llvm.org/cgi-bin/mailman/listinfo/cfe-commits

Reply via email to