This is detailed guide to calculate GPU Memory 💾 requirement for serving LLMs. Ps - it doesn't take context length into consideration Website 📖- https://lnkd.in/gf7YzqJt
Hello, How are do? Is to inference model? And... serves as a reference for fine tuning?
Data Engineer | Informatica
7moThanks for sharing! We can determine chunk sizes as per requirement from this right? After putting the context length also in the numerator...