Streamlining LLM Inference at the Edge with TFLite

XNNPack, the default TensorFlow Lite CPU inference engine, has been updated to improve performance and memory management, allow cross-process collaboration, and simplify the user-facing API.


This content originally appeared on Google Developers Blog and was authored by Google Developers Blog

XNNPack, the default TensorFlow Lite CPU inference engine, has been updated to improve performance and memory management, allow cross-process collaboration, and simplify the user-facing API.


This content originally appeared on Google Developers Blog and was authored by Google Developers Blog


Print Share Comment Cite Upload Translate Updates
APA

Google Developers Blog | Sciencx (2024-08-13T17:26:50+00:00) Streamlining LLM Inference at the Edge with TFLite. Retrieved from https://www.scien.cx/2024/08/13/streamlining-llm-inference-at-the-edge-with-tflite/

MLA
" » Streamlining LLM Inference at the Edge with TFLite." Google Developers Blog | Sciencx - Tuesday August 13, 2024, https://www.scien.cx/2024/08/13/streamlining-llm-inference-at-the-edge-with-tflite/
HARVARD
Google Developers Blog | Sciencx Tuesday August 13, 2024 » Streamlining LLM Inference at the Edge with TFLite., viewed ,<https://www.scien.cx/2024/08/13/streamlining-llm-inference-at-the-edge-with-tflite/>
VANCOUVER
Google Developers Blog | Sciencx - » Streamlining LLM Inference at the Edge with TFLite. [Internet]. [Accessed ]. Available from: https://www.scien.cx/2024/08/13/streamlining-llm-inference-at-the-edge-with-tflite/
CHICAGO
" » Streamlining LLM Inference at the Edge with TFLite." Google Developers Blog | Sciencx - Accessed . https://www.scien.cx/2024/08/13/streamlining-llm-inference-at-the-edge-with-tflite/
IEEE
" » Streamlining LLM Inference at the Edge with TFLite." Google Developers Blog | Sciencx [Online]. Available: https://www.scien.cx/2024/08/13/streamlining-llm-inference-at-the-edge-with-tflite/. [Accessed: ]
rf:citation
» Streamlining LLM Inference at the Edge with TFLite | Google Developers Blog | Sciencx | https://www.scien.cx/2024/08/13/streamlining-llm-inference-at-the-edge-with-tflite/ |

Please log in to upload a file.




There are no updates yet.
Click the Upload button above to add an update.

You must be logged in to translate posts. Please log in or register.