-
Notifications
You must be signed in to change notification settings - Fork 113
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Android devices Support #133
Comments
the command i used: |
Wooow! This is so cool! Could you share specifications of phones? What a performance you get on a single phone vs 2 phones? Have you tried 3B model? BTW: you could add
You can ignore this warning. It's related to the memory assigned to the application. Distributed Llama tries to lock memory. If it cannot lock it, the application will still work, but probably a bit slower. However, I'm not entirely sure about this. |
Thank you for your reply! I will try to limit the generated length, I ran into a problem with insufficient memory when testing an Android phone, whenever I run it, the phone automatically shut down, I suspect because there is no limit max-seq-len, the two phones I use are oppo find x7 ultra, and redmi Note12 pro. I'll test the acceleration using both devices after I've eliminated the above issues! |
It would be great! If you want you can share your results also here. |
Hello! Thank you very much for your excellent work, which enables the distributed running of large models on heterogeneous devices! I was wondering if this project supports Android devices. I am currently using two Redmi phones and plan to cross-compile this project using Android NDK and then transfer the executable file and related weights to the device to run it. I have already tested communication between the two phones (TCP/Socket). I would like to know what I need to pay attention to in this process. I look forward to your answer! Thank you
The text was updated successfully, but these errors were encountered: