-
Notifications
You must be signed in to change notification settings - Fork 45
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Phi-4 Conversion Failure #148
Comments
I won't post the code here, but I also tested the tokenizer via the RKLLM API, and it was producing correct output for encoding/decoding tokens. Also, is it possible to enable flash attention for optimizing? I know it is possible with rknn, but don't see an option in the LLM converter API. Thank you! |
Official PHI-4 on HF has been released.... any updates? |
Hi, there will be updates in the near future. |
Thanks @waydong ! Will the updates be in the same 1.1.x version of the library, i.e. 1.1.5, so it is backwards compatible? If not, will there be a similar |
Yes, we will maintain an interface(update_rkllm) to support easy model upgrades. |
@waydong Thank you! Will the updated library also require a new kernel module? If so, could you make dynamic instead of built-in? |
@waydong
Configuration: https://huggingface.co/c01zaut/phi-4/
Model weights: https://huggingface.co/NyxKrage/Microsoft_Phi-4/
Code:
Could you take a look and let me know what setting/parameter needs to be set/how to get proper debug output, so I can adjust the setting in the Phi3 config?
The text was updated successfully, but these errors were encountered: