Releasing LLongMA-2 13b, a Llama-2 model, trained at 8k context length using linear positional interpolation scaling. The model was trained in collaboration with @theemozilla of @NousResearch and @kaiokendev1. https://t.co/KGkEkEcFTx
— Enrico Shippole (@EnricoShippole) Jul 24, 2023
from Twitter https://twitter.com/EnricoShippole
July 24, 2023 at 09:07AM
via IFTTT
No comments:
Post a Comment