Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

JsonSerializer.Deserialize is intolerably slow in Blazor WebAssembly, but very fast in .NET Core integration test #40386

Closed
szalapski opened this issue Aug 5, 2020 · 51 comments
Assignees
Labels
arch-wasm WebAssembly architecture area-VM-meta-mono
Milestone

Comments

@szalapski
Copy link

In my Blazor app, I have a component that has a method like this. (I've replaced a call to GetFromJsonAsync with code from inside it, to narrow down the slow part.)

  private async Task GetData()
  {
      IsLoading = true;
      string url = $".../api/v1/Foo";  // will return a 1.5 MB JSON array
      var client = clientFactory.CreateClient("MyNamedClient");

      Console.WriteLine($"starting");

      List<Foo> results;

      Task<HttpResponseMessage> taskResponse = client.GetAsync(url, HttpCompletionOption.ResponseContentRead, default);

      var sw = Stopwatch.StartNew();
      using (HttpResponseMessage response = await taskResponse)
      {
        
        response.EnsureSuccessStatusCode();
        var content = response.Content!;

        if (content == null)
        {
          throw new ArgumentNullException(nameof(content));
        }
        
        string contentString = await content.ReadAsStringAsync();

        sw.Stop();
        Console.WriteLine($"Read string: {sw.Elapsed}");
        sw.Restart();

        results = System.Text.Json.JsonSerializer.Deserialize<List<Foo>>(contentString)!;
        //results = Newtonsoft.Json.JsonConvert.DeserializeObject<List<Foo>>(contentString); // comparable

      }

      sw.Stop();
      Console.WriteLine($"Deserialize: {sw.Elapsed}");
      
      StateHasChanged();
      IsLoading = false;

My download of 2-6 MB takes 1-6 seconds, but the rest of the operation (during which the UI is blocked) takes 10-30 seconds. Is this just slow deserialization in ReadFromJsonAsync (which calls System.Text.Json.JsonSerializer.Deserialize internally), or is there something else going on here? How can I improve the efficiency of getting this large set of data (though it isn't all that big, I think!)

I have commented out anything bound to Results to simplify, and instead I just have an indicator bound to IsLoading. This tells me there's no slowness in updating the DOM or rendering.

When I attempt the same set of code in an automated integration test, it only takes 3 seconds or so (the download time). Is WebAssembly really that slow at deserializing? If so, is the only solution to retrieve very small data sets everywhere on my site? This doesn't seem right to me. Can this slowness be fixed?

Here's the resulting browser console log from running the above code:

VM1131:1 Fetch finished loading: GET "https://localhost:5001/api/v1/Foo".
Read string: 00:00:05.5464300
Deserialize: 00:00:15.4109950
L: GC_MAJOR_SWEEP: major size: 3232K in use: 28547K
L: GC_MAJOR: (LOS overflow) time 18.49ms, stw 18.50ms los size: 2048K in use: 187K
L: GC_MINOR: (LOS overflow) time 0.33ms, stw 0.37ms promoted 0K major size: 3232K in use: 2014K los size: 2048K in use: 187K

Using Newtonsoft.Json (as in the commented-out line) instead of System.Text.Json gives very similar results.

For what it's worth, here's the Chrome performance graph. The green is the download and the orange is "perform microtasks", which I assume means WebAssembly work.

enter image description here

@szalapski
Copy link
Author

See also comments at https://stackoverflow.com/questions/63254162

@mkArtakMSFT mkArtakMSFT transferred this issue from dotnet/aspnetcore Aug 5, 2020
@Dotnet-GitSync-Bot Dotnet-GitSync-Bot added the untriaged New issue has not been triaged by the area owner label Aug 5, 2020
@Dotnet-GitSync-Bot
Copy link
Collaborator

I couldn't figure out the best area label to add to this issue. If you have write-permissions please help me learn by adding exactly one area label.

@mkArtakMSFT
Copy link
Member

Thanks for contacting us.
There has been a big push to optimize this further. You can learn more about the improvements #40318

@mkArtakMSFT
Copy link
Member

@steveharter FYI

@szalapski
Copy link
Author

Thanks. The performance is so poor that I am still skeptical that this is just a slow area--I still suspect that something is wrong with the way I am doing it. Would a deserialization of a few megabytes take 10-30 s?

@szalapski
Copy link
Author

szalapski commented Aug 5, 2020

needs tag area-System.Text.Json

@lewing
Copy link
Member

lewing commented Aug 5, 2020

Which version of Blazor are you using? You'll want to avoid creating a string from the content and use a Stream instead. If you are using the net5.0 you should look at the System.Net.Http.Json extensions.

@lewing lewing added the arch-wasm WebAssembly architecture label Aug 5, 2020
@szalapski
Copy link
Author

szalapski commented Aug 5, 2020

I'm using Blazor 3.2.0 with System.Text.Json 5.0.0-preview.7.

Yes, I used the extensions, but when I saw they were slow, I refactored to the code above so I could narrow the issue down to serialization. Here's the code before my performance refactoring.

private async Task GetData()
{
      IsLoading = true;
      string url = $".../api/v1/Foo";
      Results = await clientFactory.CreateClient("MyNamedClient").GetFromJsonAsync<List<Foo>>(url);
      IsLoading = false;
}

@steveharter
Copy link
Member

steveharter commented Aug 5, 2020

You'll want to avoid creating a string from the content and use a Stream instead.

Yes this will allow the deserializer to start before all of the data is read from the Stream and prevent the string alloc.

System.Text.Json should be ~2x faster for deserialization than Newtonsoft so it would be good to see your object model to see if you hit an area that is slow on STJ.

In either case, since both Newtonsoft and STJ are slow there is likely something else going on.

The Large object graph benchmark section in #40318 has deserialization perf of 372ms for a string of length 322K. This also includes a "polymorphic" mode due to using System.Object that causes deserialization to be much slower (almost 2x) than without it. Anyway, extrapolating 332K to your 1MB is a 3x factor, so I assume it would take about 372ms * 3 = ~1.1 seconds to deserialize (on my fast desktop in isolation).

Some thoughts:

  • Can you share your hardware specs? Memory\CPU
  • Is the test running in isolation on dedicated hardware or is it hosted?
  • Can you share your object model (your Foo type from the link)?
  • Is there also rendering going on (or other CPU tasks) that would affect perf significantly?
  • Change to Async\Stream mode as mentioned earlier.

@HenkHolterman
Copy link

HenkHolterman commented Aug 6, 2020

I posted an MCVE as answer on StackOverflow, based on the WeatherForecast page.
My results are much better: < 100ms download and < 4 seconds for deserialize.

See https://stackoverflow.com/q/63254162/

@szalapski
Copy link
Author

szalapski commented Aug 6, 2020

@steveharter ,

Yes, Intel Core i5 8350-U with 16 GB RAM. The test is running on my laptop. Foo is actually as follows, with minimal name changes only to protect the proprietary. Nothing significant on the CPU, this is my only focus when I am doing this.

I did start with a stream per the code just above --that was how I found this issue. I refactored to the code in the issue post just to narrow it down to slowness in deserialization.

    public class Foo
    {
        public int? FooId { get; set; }
        public int? FiscalYear { get; set; }
        public string? FundTypeCode { get; set; }
        public string? CategoryDescription { get; set; }
        public string? CustomerLevelCode { get; set; }
        public string? CustomerCode { get; set; }
        public string? CustomerDescription { get; set; }
        public string? ChangedFieldName { get; set; }
        public decimal OriginalAmount { get; set; }
        public decimal NewAmount { get; set; }
        public string? Comment { get; set; }
        public string? CreateUser { get; set; }
        public DateTime CreateDate { get; set; }
        public bool AdjustedBySystem { get; set; }
        public Guid? ChangeBatchNumber { get; set; }
        public string? FundDescription { get; set; }
        public string? ParentCustomerCode { get; set; }
        public string? ParentCustomerLevel { get; set; }
        public string? ParentCustomerDescription { get; set; }
        public decimal ChangedAmount { get { return NewAmount - OriginalAmount; } }
        public bool? CreatedFromNFile { get; set; }
        public string? Reason => (FundTypeCode == "N" ? (CreatedFromNfdaFile == true ? "From N File" : "N Manual") : "")}

    }

@steveharter
Copy link
Member

That model is simple and should be fast (no System.Object, non-generic collections or custom converters that could slow it down).

I suggest running the perf test that @HenkHolterman suggested in stackoverflow to compare against the baseline.

@szalapski
Copy link
Author

szalapski commented Aug 6, 2020

@steveharter , I tried it just as suggested. It indeed takes 7-12 seconds to return 17000 items (about 1.6 MB) of WeatherForecast. (Download time on localhost is about 20 ms.) using the default code, await Http.GetFromJsonAsync<WeatherForecast[]>("WeatherForecast"); So this seems consistent with the timings on my slightly more complex case in the original question. (FYI, this is on Blazor 3.2.0; I also updated System.Text.Json via NuGet to v 5.0.0-preview.7, but it didn't help much.)

(also, I tried to increase the payload to 5 MB and that took 23-27 seconds.)

@lewing
Copy link
Member

lewing commented Aug 6, 2020

Blazor in net5 should be considerably faster. Are you running this test from inside VS or from a published build?

@ghost
Copy link

ghost commented Aug 6, 2020

Tagging subscribers to this area: @CoffeeFlux
See info in area-owners.md if you want to be subscribed.

@szalapski
Copy link
Author

szalapski commented Aug 6, 2020

Running it from Visual Studio, "run without debugging" in Release configuration.

@lewing Do you mean just System.Text.Json should be faster? If so, I already have the latest preview.

Or are you suggesting I move the app to Blazor 5.0.0 latest preview?

@lewing
Copy link
Member

lewing commented Aug 6, 2020

@szalapski could you please try your timings with the a published app outside of VS? It looks like there is an issue where the runtime is always initialized in debug mode when run from inside VS.

Additionally if you update the app from 3.2 to 5.0 there are several interpreter optimizations and library improvements. Some are in preview7 some are in later builds.

@szalapski
Copy link
Author

szalapski commented Aug 6, 2020

Just tried outside of VS -- using dotnet run at the command line in Windows. No problems, very similar timings.

Also tried running the .exe after running dotnet publish --configuration Release, just to be sure. (I think this should be virtually the same as running dotnet run, right?) The timings were again similar.

@marek-safar marek-safar removed the untriaged New issue has not been triaged by the area owner label Aug 24, 2020
@marek-safar marek-safar added this to the 5.0.0 milestone Aug 24, 2020
@marek-safar marek-safar modified the milestones: 5.0.0, 6.0.0 Sep 3, 2020
@marek-safar
Copy link
Contributor

@lewing what are the next steps here?

@steveharter
Copy link
Member

what are the next steps here?

I assume no attempt to run on Blazor 5.0 yet? If so I think that should be next.

Both Newtonsoft and STJ are slow. This indicates a likely environmental or systemic issue, and not likely a (de)serialization issue.

The StackOverflow test runs <4 seconds for @HenkHolterman and 7-12 seconds for @szalapski. Different hardware and\or different Blazor versions could account for that 2x-3x slowness; would need a standard CPU benchmark and same Blazor version to actually compare apples-to-apples.

Also @szalapski on download perf you originally said:

My download of 2-6 MB takes 1-6 seconds

but with your latest test from StackFlow you said:

It indeed takes 7-12 seconds to return 17000 items (about 1.6 MB) of WeatherForecast. (Download time on localhost is about 20 ms.)

So download time went from 1-6 seconds for 2-6MB to 20ms for 1.6MB -- any thought on why that's the case?

@szalapski
Copy link
Author

szalapski commented Sep 4, 2020

The 1-6 seconds was over the internet, whereas the 20ms was running against a local web service. I just did that comparison to ensure that the download speed is not relevant--regardless of whether the download is 20 ms or 20,000 ms, the deserialization is quite slow.

I will try it on Blazor 5 preview 8 soon.

"The StackOverflow test runs <4 seconds for @HenkHolterman and 7-12 seconds for @szalapski. Different hardware and\or different Blazor versions could account for that 2x-3x slowness; would need a standard CPU benchmark and same Blazor version to actually compare apples-to-apples."

Why shouldn't it be on the order of tens of milliseconds? Are the optimizations we see in .NET Core just not possible in WebAssembly?

"Both Newtonsoft and STJ are slow. This indicates a likely environmental or systemic issue, and not likely a (de)serialization issue."

Wait, I thought all agreed that the slowness is in the deserialization code, not in a problem with my system or environment. You are saying that I have a problem that is not inherent to deserializing in WebAssembly? How can I diagnose that?

@steveharter
Copy link
Member

@rajeshaz09 I assume you've measured against 5.0 .NET since there have been gains.

I see that MessagePack claims to be ~13x faster deserializing than Json.NET (no benchmark for STJ) for the case of "large array of simple objects". So if STJ is 2x as fast as Json.NET here, the 7 seconds for STJ vs. 2 seconds for MessagePack seems consistent, although note that the benchmark is for standard .NET Core not under Blazor.

@rajeshaz09
Copy link

@steveharter
Thanks for the reply. Yes I am using .NET 5.
Difference between STJ and MessagePack is more visible on low config machine.

I have powerful dev machine. I didn't see much difference (Hardly 1 second) between STJ and MessagePack with HighPerformance power setting. But I can see significant gap if I use Balanced/PowerSaver setting.

MessagePack is temporary solution, once we satisfy with .NET 6, we will move back to JSON.

@lambrech
Copy link

lambrech commented May 4, 2021

I am not 100% sure but it seems very likely that this is related: I just tried to deserialize a 2.6MB json containing 10.000 simple POCOs.
In chrome the deserialization took took ~4 seconds (that's actually "good enough" for me, at least right now)
But in Firefox the same deserialization took ~35 seconds! That is a serious problem for me ...

FYI: I am using .NET 6 Preview 3 and System.Text.Json

@lewing lewing modified the milestones: 6.0.0, 7.0.0 Aug 8, 2021
@sxotney
Copy link

sxotney commented Aug 21, 2021

I have had a similar journey recently moving through different serialisers and finally arriving at Messagepack which has been good enough in interpreted WASM for current users. Performance Vs System.Text.Json is impressive

However, scope of our WASM app is definitely expanding and we have users looking to handle 100s of thousands of of objects to perform data manipulation/analysis in browser like Excel would chomp through on a normal desktop. Wait times for data loads of this size (they really aren't massive payloads delivered from the API) are at the point where it is difficult to satisfy users and Server Side Blazor is becoming the only option.

The Blazor/WASM community has generally always expressed that code runs at native speeds (until you learn that everything outside of the .net libraries is interpreted) and I had hoped AOT would make an enormous difference here, allowing Messagepack serialiser to run at native speed. Our initial benchmarks of rc1 are showing it to be slower in this area than interpreted mode.

Maybe it's my misunderstanding of how serialisation works - is it object construction in .Net itself being slow here and I shouldn't see any difference between AOT and interpreted builds? Either way, serialisation is painfully slow for what is really not that much data.

@ikeough
Copy link

ikeough commented Aug 30, 2021

First, and most importantly, thanks to the team working on Blazor and web assembly. We think this technology has a really bright future!

I'll add my support for @szalapski here. We have an .NET open source library that is used heavily in back end services run on AWS Lambda. We were excited with the possibility of running some of our code in our web application. Our initial attempts to compile and run web assembly from our library in .NET 6 preview 7 have been met with massive performance degradation.

I established a small benchmark that creates 1000 cubes using the library (the library is for creating 3d stuff with lots of Vector3 structs and Polygon), serializes them to JSON, then writes the resulting 3D model to glTF. I duplicated that code in a small Blazor app.

Running the Blazor code compiled using dotnet run -c release (non AOT) and viewing the console in Chrome shows:

00:00:07.2027000 for writing to gltf.

We found that AOT compilation (which takes nearly 15 minutes), increases the performance by 2x.

The benchmark containing the same code run on the desktop, shows the following for writing to gltf:

Method Mean Error StdDev Gen 0 Gen 1 Gen 2 Allocated
'Write all cubes to glb.' 105.8 ms 4.42 ms 2.31 ms 21000.0000 3000.0000 1000.0000 85.05 MB

It takes nearly 67x as long to run in web assembly. We have a similar performance degradation for serializing and deserializing JSON.

Some considerations as to what might be slow:

  • glTF creation involves the manipulation of List<byte>. We've seen guidance that suggests you shouldn't use IList<T> and we're not doing much of that. But perhaps reading and writing bytes is inherently slow?
  • JSON serialization uses Newtonsoft.Json.Net and a custom converter for deserializing to child classes. We've seen the recommendation to move to System.Text.Json and it's a hard pill to swallow because our code requires converters and makes liberal use of json.net attributes. We'd love to try and get this to work as is. The fact that writing to glTF, and potentially many other operations, is so slow suggests that optimizing for JSON may fix a small part of the problem but will not leave us with confidence that adoption of web assembly is a possibility.
  • We use several third party dlls that we had to compile with .NET 6 as well to even get the publishing of the Blazor project to work.

You can find our example Blazor project that has no UI but runs the wasm and reports to the console here: https://github.com/hypar-io/Elements/tree/wasm-perf/Elements.Wasm.
You can find the corresponding benchmark WasmComparison here:
https://github.com/hypar-io/Elements/tree/wasm-perf/Elements.Benchmarks

We're really excited for the effort to bring C# to web assembly and are happy to provide any further information necessary. It would be fantastic for these development efforts if there was a way to run a dotnet benchmark across the core CLR and web assembly to make an apples->apples comparison. For now we've had to build our own.

One more thing... This performance degradation is not everywhere. We can call methods in our library that do some pretty complicated geometry stuff and they run at near native speed. We have a couple of demos of interactive 3d geometry editing and display using Blazor wasm. It's just serialization and reading/writing bytes that seem to be a big issue. Also looping in @Gytaco who is doing some amazing work using c#->web assembly for geometry stuff.

@marek-safar
Copy link
Contributor

You can find the corresponding benchmark WasmComparison here:
https://github.com/hypar-io/Elements/tree/wasm-perf/Elements.Benchmarks

@SamMonoRT could we add it to interpreter benchmarks?

@90Kaboom
Copy link

90Kaboom commented Oct 30, 2021

If you face issue with JSON serialization performence , before trying to solve by refatoring your code, please check performeance in another browser, Blazor work realy fast on Edge, Opera, Chrome, but performance in Firefox is realy wick - slowdown serialization more than 10 times.

@sxotney
Copy link

sxotney commented Nov 2, 2021

@90Kaboom

Serialisation is slow across all browsers for Mono .Net. If the performance of Blazor is slow in a particular browser, that's more likely a wasm implementation issue for the team that maintain that browser as opposed to a Blazor/Mono .Net issue.

@Kevin-Lewis
Copy link

I see this is being targeted for .NET7. Blazor WASM has been great for the most part but this performance issue is making it really difficult to view Blazor as a viable option for some of the more data intensive projects I have coming up. I'll give MessagePack a try since it seems people have had some success with that,

@radical radical modified the milestones: 7.0.0, 8.0.0 Aug 12, 2022
@Chief4Master
Copy link

Any new news or suggestions (@szalapski )? We have the exact same problem. So we can not realize our application with Blazor.

@Cpt-Falcon
Copy link

Cpt-Falcon commented Feb 16, 2023

I just tried it with a 10MB json file and its unusably slow. 10MB isn't that much. Its tiny. Its taking over 2 minutes to load the initial page which doesn't make sense IMO. I'm using the best performance tricks too:

    Assembly powerAssembly = typeof(PowerService).Assembly;
    await using Stream? stream = powerAssembly.GetManifestResourceStream("PowerShared.PowerDocuments.json");

    ValueTask<XMLJsonWrapper?> powerXmlDocumentsTask = JsonSerializer.DeserializeAsync<XMLJsonWrapper>(stream, new JsonSerializerOptions()
    {
        DefaultIgnoreCondition = JsonIgnoreCondition.WhenWritingNull
    });

Its so slow and takes so long I can't even run a performance profiler either. THe performance profiler just bombs out and gets stuck.

I'm having other problems too where external .net 7 DLLs take forever to load.

There needs to be a way to quickly and efficiently load datasets into blazor WASM.

This is on .NET 7 by the way

@Cpt-Falcon
Copy link

Meanwhile react with an embedded file of the same size takes like 10 ms.

@marek-safar
Copy link
Contributor

/cc @lewing

@jirisykora83
Copy link

jirisykora83 commented Feb 17, 2023

Meanwhile react with an embedded file of the same size takes like 10 ms.

I have the exact same problem with even smaller (about 1.5MB uncompress) json it was absolutely unusable kind of slow. I even tried to write custom binary serialization & deserialization and it was better but still too slow to justify maintaining that code. I ended up using dotnet grpc. It was still slower than pure javascript (json parse) but it is the best current option in term of performance / maitainence cost.

I also try: https://github.com/salarcode/Bois
Custom solution was based on: MemoryStream& BinaryWriter

Edit:
Some number from my usecase (i7 12700K):

System.Text.Json: 237ms (size: 1278kb)
Bois: 42ms (size: 501kb)
Custom: 41ms (size: 441kb)

For context if I recreated tested model from in-memory original object (custom clone) it is about 5ms.

These are numbers on published version of app and it measure just Deserialization. I do not have the number for GRPC from deployed app) But in debug mode on localhost it was about the same with Bois/custom (70ms) so i expect in deployed app it will be about 40ms on my pc.

Don't have exact number from "normal" dotnet but it will be like few ms

@pragmaeuge
Copy link

Same issue with a 10Mb GeoJson file in .NET 7. This issue makes difficult to develop effective, WFS oriented, gis solutions with .NET Wasm.

@geometrikal
Copy link

Any progress on this? Running into same issue.

@Webreaper
Copy link

Apparently lots of performance improvements (as an indirect result of the new WASM JIT) in .Net 8 preview 2, so might be worth a look. I'd be trying it, but still no VS for Mac support....

@geometrikal
Copy link

@Webreaper Thanks, looking forward to .NET 8. Also I just compiled with AOT turned on in .NET 7 and the slowness disappeared so suggest trying that @pragmaeuge

@Webreaper
Copy link

The only thing you need to be careful of with AOT is this issue: #62149

@pragmaeuge
Copy link

@Webreaper Thanks, looking forward to .NET 8. Also I just compiled with AOT turned on in .NET 7 and the slowness disappeared so suggest trying that @pragmaeuge

Hi! I will take a look at it soon, thanks a lot.

@sofiageo
Copy link
Contributor

I deployed with .NET 8 preview 2 today and deserialization is much faster. Jiterpreter helps a lot, from about 4-5 seconds with .NET 7 to 1-2 seconds. The rest of the application works too which is nice :)

@lewing
Copy link
Member

lewing commented Aug 13, 2023

This issue is over three years old and while we will keep working on improving the json performance given the increases in execution speed we have measured from when this was opened to now I consider this issue closed. If you would like to report json performance issues with .NET 8 preview 7 or later please open a new issue.

@lewing lewing closed this as completed Aug 13, 2023
@ghost ghost locked as resolved and limited conversation to collaborators Sep 12, 2023
Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
arch-wasm WebAssembly architecture area-VM-meta-mono
Projects
None yet
Development

No branches or pull requests