QWEN CHAT API DEMO DISCORD
It is widely recognized that continuously scaling both data size and model size can lead to significant improvements in model intelligence. However, the research and industry community has limited experience in effectively scaling extremely large models, whether they are dense or Mixture-of-Expert (MoE) models. Many critical details regarding this scaling process were only disclosed with the recent release of DeepSeek V3. Concurrently, we are developing Qwen2.
Yeah, that’s kind of AI companies’ definition of open source… Other companies just have “open” in their name for historical reasons. The FSF doesn’t really agree ( https://www.fsf.org/news/fsf-is-working-on-freedom-in-machine-learning-applications ) and neither do I. It’s “open weight”. Or I’d need to see the datasets and training scripts as well.
Yeah, “open weight” seems a more appropriate label. It still seems better than a fully proprietary system, but calling it open source without clarification is misleading.
Yeah, that’s kind of AI companies’ definition of open source… Other companies just have “open” in their name for historical reasons. The FSF doesn’t really agree ( https://www.fsf.org/news/fsf-is-working-on-freedom-in-machine-learning-applications ) and neither do I. It’s “open weight”. Or I’d need to see the datasets and training scripts as well.
Yeah, “open weight” seems a more appropriate label. It still seems better than a fully proprietary system, but calling it open source without clarification is misleading.