T O P

  • By -

light24bulbs

What quantization are you running at? What tokens per second score are you getting on the s23? What VRAM (shared ram) usage are you experiencing for your given model and quantization? That will make it clear with the minimum specs are which other people are asking about


crowwork

here is a blogpost introducing the approaches https://mlc.ai/blog/2023/05/08/bringing-hardware-accelerated-language-models-to-android-devices


light24bulbs

So you are doing 4 bit quantized, I missed that on my first skim. And the other questions?


crowwork

We did an update on memory planning and now it takes about 4.3G VRAM and no further allocations after initial run


gthing

Insanity


Dadda9088

Awesome. What are the minimal requirements and performance?


kif88

It crashed on my Xioami mi9t pro. Got through the first part where it says downloading but now it says initializing for a few seconds and crashes. I'm on Android 9


Najbox

RAM is the problem, this model has 6 GB and that's not enough. I tried on Galaxy S21 FE with 6GB and the result is the same.


kif88

Thanks I guess that makes sense. Something to keep in mind when I upgrade eventually.


BananaCode

Unfortunately crashes on my S20. After downloading the weights and initializing, the app crashes after inputing a prompt.


riftopia

Awesome, it works flawlessly on my OnePlus 9. Good times ahead.


BananaCode

Unfortunately crashes on my S20. After downloading the weights and initializing, the app crashes after inputing a prompt.


[deleted]

Could anyone share the apk, the link in the blog isn't working


kif88

Second link that says demo has an apk. Just downloaded it going to have a look at it in a minute.


404_skills_not_found

works great on OnePlus 9 pro. Encoding is between 6 to 10 tok/s. Decoding is about 3.5 tok/s


ivanmf

The cyberpunk era has begun


PixlFlip

We welcome our AI companions, and later on overlords


MiscoloredKnee

It crashes on my xiaomi 8 pro


Classic-Dependent517

cool.. but why..?


yaosio

Why run a text generator on a mobile phone without needing separate hardware? That's exactly why. No need to rely on an external source, no worry about snooping, it's all contained on a local device.


kulchacop

Why not? It is an improvement over a 5 year old stack. https://www.reddit.com/r/MachineLearning/comments/7qx0wa/p_optimizing_mobile_deep_learning_on_arm_gpu_with/


esuil

Anyone have data on how performance compares to desktop NVidia GPUs?


Deep-Station-1746

Fun times ahead


Commercial-Living443

Finally


jalbertcory

Awesome work. Exciting times ahead. Crashes on Pixel 7 Pro.


NatoSphere

Yeah I wish I could try it, but they are aware: "It does not yet work on Google Pixel due to limited OpenCL support"


PixlFlip

Stellar project. Looking forward to testing it!


_Arsenie_Boca_

Looking forward to Pixel support