Smart future

A site by Francesco Mugnai

Insights, news and reflections on a rapidly evolving world.

Approfondimento

How to self-host LLaMA 3.1 70B without spending a fortune

Autonomously hosting an LLM such as the 70-billion-parameter LLaMA 3.1 may seem challenging, but with the right hardware optimizations such as appropriate GPUs, quantization techniques and sharding, it can be done without spending a fortune. Hybrid cloud solutions offer a good trade-off between cost and flexibility while maintaining control over the data

Approfondimento

How to try FLUX on Mac thanks to a simple library

MFLUX brings FLUX to the Apple ecosystem through a careful port to Apple MLX. Key features include clean code, a minimalist approach without unnecessary configurations, and reduced dependencies. It supports FLUX.1-Schnell and FLUX.1-Dev models and is a breeze to install using pip. Image generation can be customized via command-line options, and it supports quantization to boost performance on Mac devices.

Note: These articles are written with the help of an AI assistant, and it couldn't be otherwise! 😃