-
Notifications
You must be signed in to change notification settings - Fork 2
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Feature] update index automatically #2
Comments
fix the latest main build |
Hi @zhyncs I'm getting "from flashinfer.decode import _grouped_size_compiled_for_decode_kernels Even though I'm using this nightly version, is there a fix? I'm trying to run Lama3.2 with SGLang. |
@Iven2132 Please use SGLang latest release |
@zhyncs I treid using 0.4.0.post2 now getting 'ModuleNotFoundError: No module named 'zmq'' any if I install then get another one |
|
Getting 'ModuleNotFoundError: No module named 'torch'' But I've torch installed. What torch, CUDA would you recommend to use SGLANG to run Llama3.2 vision |
torch 2.5.1 cuda 12.4 works good for me |
how does your "https://flashinfer.ai/whl/cu121/torch2.5/flashinfer/" look like? for that |
You don’t need to specify torch 2.5 for flashinfer. torch 2.4 and torch 2.5 are ABI compatible |
I'm confused is this image config right??
Also, I can use Llama.3.2 Vision directly like what I'm doing in my code?
|
Now getting 'ModuleNotFoundError: No module named 'flashinfer'' |
You shouldn’t use 2.5 in the url |
I removed and used "pip install "sglang[all]" --find-links https://flashinfer.ai/whl/cu124/torch2.4/flashinfer/ --force-reinstall'" and got ModuleNotFoundError: No module named 'torch' |
You can install torch separately |
How would that look like I tried |
Why do you install with code, can you use command line |
I get
|
@zhyncs Getting AttributeError: module 'sglang' has no attribute 'Runtime'
|
here is the error. "ValueError: bad value(s) in fds_to_keep" |
@zhyncs here are the full log:
|
ref flashinfer-ai/whl#1 (review)
The text was updated successfully, but these errors were encountered: