Mistral Small but Mighty - Apache 2.0, Multimodal & Fast

Mistral Small but Mighty - Apache 2.0, Multimodal & Fast

Prompt Engineering

1 день назад

7,522 Просмотров

Ссылки и html тэги не поддерживаются


Комментарии:

@IqweoR
@IqweoR - 20.03.2025 01:34

Looks great. Can you include a bit of multi-language tests? Maybe use google translate on outputs, that would be awesome

Ответить
@user-nbfkxngjmyb
@user-nbfkxngjmyb - 19.03.2025 03:43

thanks for the great content and please keep up the good work

Ответить
@themarksmith
@themarksmith - 18.03.2025 20:32

Would have liked to see more info about the hardware needed to run these models at home - and your thoughts on models/hardware which can be run at home - but still a great video - cheers!

Ответить
@Dr.UldenWascht
@Dr.UldenWascht - 18.03.2025 18:19

It's a promising model and kudos to Mistral for the Apache license. I just wish they'd also trained smaller edge versions like Google has done with Gemma 3. I mean, the minimum GPU you need for his is a 3090. And even then you won't be able to fully utilize the 128k context window. That makes me wonder if Mistral is intentionally pushing people toward using its API for synthetic data gathering.

Ответить
@OnigoroshiZero
@OnigoroshiZero - 18.03.2025 17:13

I would have loved a 12b version, my 3080 10gb can't run a 24b model...

Ответить
@anonymouscommentator
@anonymouscommentator - 18.03.2025 15:41

im really interested in llm which fit in 24GB of vram. i think a comparison between llama3, phi4, gemma3 and this would be great!

Ответить
@vempatisaivishal5426
@vempatisaivishal5426 - 18.03.2025 14:38

great video man
because of your videos I finished one production standard project in our company
the one you explained context caching helped me a lot

Ответить
@nielseriksen3009
@nielseriksen3009 - 18.03.2025 14:37

Looking forward to the comparison between Mistral Small v3.1 and Gemma 3 27B

Ответить