-
Notifications
You must be signed in to change notification settings - Fork 285
Issues: InternLM/lmdeploy
[Benchmark] benchmarks on different cuda architecture with mo...
#815
opened Dec 11, 2023 by
lvhan028
Open
6
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
[Bug] same code A800 good but A10 stuck MiniCPM-Llama3-V-2_5
#1938
opened Jul 6, 2024 by
llmrainer
2 tasks
[Bug] unified_attention split kv for prefill with more workspace coredump
#1935
opened Jul 6, 2024 by
snippetzero
2 tasks done
[Bug] 想用vscode debug 代码的运行,发现debug到模型运行的时候直接返回结果,无法得知把处理好的输入送入模型得到输出的中间过程
#1933
opened Jul 5, 2024 by
AIFFFENG
2 tasks done
[Bug] Encount TCP error (Port Aready used) when deploy with PytorchEngine
awaiting response
#1925
opened Jul 5, 2024 by
Desein-Yang
2 tasks done
[Feature] Is there any plan to support for InternLM-XComposer2.5 inference?
#1920
opened Jul 4, 2024 by
Charles-Xie
[Bug] Failed to load InternVL-Chat-V1-5-Int8 quantized model. RuntimeError: Only Tensors of floating point and complex dtype can require gradients
#1907
opened Jul 3, 2024 by
jiajie-yang
2 tasks done
[Bug] Segmentation fault occurs and the machine with openEuler os was automatically reboots
#1905
opened Jul 3, 2024 by
jiajie-yang
2 tasks done
[Bug] Using the turbomind engine, prompting more than 10k tokens will result in garbage output.
#1896
opened Jul 2, 2024 by
dafu-wu
2 tasks done
[Bug] CUDA runtime error: an illegal memory access was encountered when 8bit kv quant was enabled
#1895
opened Jul 1, 2024 by
aabbccddwasd
2 tasks done
Previous Next
ProTip!
Type g i on any issue or pull request to go back to the issue listing page.