This is a solid and fast long-context model for summarization, Q&A etc. It's better than GPT-3.5x 16K on benchmarks... surprising for a model likely 3-4x smaller! https://t.co/SKdu4MjdUQ
— Vipul Ved Prakash (@vipulved) Aug 18, 2023
from Twitter https://twitter.com/vipulved
August 18, 2023 at 11:48AM
via IFTTT
No comments:
Post a Comment