- cross-posted to:
- fosai@lemmy.world
- technology@lemmit.online
- cross-posted to:
- fosai@lemmy.world
- technology@lemmit.online
souce has this title now :
Mistral CEO confirms ‘leak’ of new open source AI model nearing GPT-4 performance
2 excerpts :
Mistral co-founder and CEO Arthur Mensch took to X to clarify: “An over-enthusiastic employee of one of our early access customers leaked a quantised (and watermarked) version of an old model we trained and distributed quite openly…
To quickly start working with a few selected customers, we retrained this model from Llama 2 the minute we got access to our entire cluster — the pretraining finished on the day of Mistral 7B release. We’ve made good progress since — stay tuned!“Quantization in ML ((machine learning)) refers to a technique used to make it possible to run certain AI models on less powerful computers and chips by replacing specific long numeric sequences in a model’s architecture with shorter ones.
How would someone go about running these things locally?
LM Studio seems like the easiest option at this point.
Llama.cpp based on hardware.
DL from huggingface and run a command