You’re now watching this thread. If you’ve opted in to email or web notifications, you’ll be notified when there’s activity. Click again to stop watching or visit your profile to manage watched threads and notifications.
You’ve stopped watching this thread and will no longer receive emails or web notifications when there’s activity. Click again to start watching.
If Apple Neural Engine can support the 8bits/integer inference? Quantized weights to 8bits could reduce storage to quarter, but the inference speed did not change?