Comment by aazo11

Comment by aazo11 21 hours ago

0 replies

This is a huge unlock for on-device inference. The download time of larger models makes local inference unusable for non-technical users.