QWEN CHAT API DEMO DISCORD
It is widely recognized that continuously scaling both data size and model size can lead to significant improvements in model intelligence. However, the research and industry community has limited experience in effectively scaling extremely large models, whether they are dense or Mixture-of-Expert (MoE) models. Many critical details regarding this scaling process were only disclosed with the recent release of DeepSeek V3. Concurrently, we are developing Qwen2.
Yeah, that’s kind of AI companies’ definition of open source… Other companies just have “open” in their name for historical reasons. The FSF doesn’t really agree ( https://www.fsf.org/news/fsf-is-working-on-freedom-in-machine-learning-applications ) and neither do I. It’s “open weight”. Or I’d need to see the datasets and training scripts as well.
Yeah, “open weight” seems a more appropriate label. It still seems better than a fully proprietary system, but calling it open source without clarification is misleading.
Isn’t every software binary open source then? Since you can open it in a hex editor and modify it
But tou don’t have permission to do. And hacking a binary is much more difficult than specializing a model, for instance.
Yeah, that’s kind of AI companies’ definition of open source… Other companies just have “open” in their name for historical reasons. The FSF doesn’t really agree ( https://www.fsf.org/news/fsf-is-working-on-freedom-in-machine-learning-applications ) and neither do I. It’s “open weight”. Or I’d need to see the datasets and training scripts as well.
Yeah, “open weight” seems a more appropriate label. It still seems better than a fully proprietary system, but calling it open source without clarification is misleading.