You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
hi~感谢贵组开源了自己的工作,我在 infinite-bench 尝试用 A100 80 G 复现 meta-llama/Meta-Llama-3-8B-Instruct,发现速度很慢,请问是我硬件(比如 CPU / 内存)限制,还是算法本身就不是很快呢?我在论文中并未找到关于「时间 / 空间的具体分析」,只有 "In terms of efficiency, InfLLM achieves a 34% decrease in time consumption while using only 34% of the GPU memory compared to the full-attention models".
read kv_retrieval.jsonl
Pred kv_retrieval
2%|██▍ | 8/500 [07:18<7:22:35, 53.97s/it]
hi~感谢贵组开源了自己的工作,我在 infinite-bench 尝试用 A100 80 G 复现 meta-llama/Meta-Llama-3-8B-Instruct,发现速度很慢,请问是我硬件(比如 CPU / 内存)限制,还是算法本身就不是很快呢?我在论文中并未找到关于「时间 / 空间的具体分析」,只有 "In terms of efficiency, InfLLM achieves a 34% decrease in time consumption while using only 34% of the GPU memory compared to the full-attention models".
以下是相关配置(取自于仓库,并未做修改)
The text was updated successfully, but these errors were encountered: