@@ -724,7 +724,6 @@ void ov::npuw::CompiledModel::CompiledModelDesc::serialize(std::ostream& stream,
724724 write (stream, host_flash_attention);
725725 if (host_flash_attention.has_value ()) {
726726 write (stream, host_flash_attention.value ()._tile_size );
727- write (stream, host_flash_attention.value ()._kv_cache_size );
728727
729728 // Serialize compiled tile model
730729 if (host_flash_attention.value ()._compiled_tile_model ) {
@@ -843,7 +842,6 @@ void ov::npuw::CompiledModel::CompiledModelDesc::deserialize(std::istream& strea
843842 read (stream, host_flash_attention);
844843 if (host_flash_attention.has_value ()) {
845844 read (stream, host_flash_attention.value ()._tile_size );
846- read (stream, host_flash_attention.value ()._kv_cache_size );
847845
848846 bool has_compiled_model = false ;
849847 read (stream, has_compiled_model);
@@ -1794,8 +1792,8 @@ void ov::npuw::CompiledModel::compile_host_flash_attention_model(std::size_t id,
17941792 hfa.set_compiled_tile_model (std::move (compiled_tile_model));
17951793
17961794 LOG_INFO (" Successfully compiled host flash attention regular tile model" );
1797- std::cout << " HostFlashAttention tile model compiled on " << device << " (tile_size=" << hfa._tile_size
1798- << " , kv_cache_size= " << hfa. _kv_cache_size << " ) " << std::endl;
1795+ std::cout << " HostFlashAttention tile model compiled on " << device << " (tile_size=" << hfa._tile_size << " ) "
1796+ << std::endl;
17991797 } catch (const std::exception& ex) {
18001798 LOG_ERROR (" Failed to compile host flash attention tile model: " << ex.what ());
18011799 OPENVINO_THROW (" Host flash attention tile model compilation failed: " , ex.what ());
0 commit comments