Tried the "Dolphin 2.5 Mixtral 8X7B" LLM. It's an uncensored model, so it can tell you how to cook meth or make a nuclear bomb. But I don't have enough knowledge to let it tell me how to do it precisely.
Anyway, I noticed the LLM have different performance on different languages. Multi lang support is fairly normal for LLM nowadays, but English is significantly better than expensive languages like Chinese and Japanese. Not only on speed (one token can be one English word, or one Chinese character), but also the result. The model is fairly shy and doesn't spill out a lot in Chinese. If you force it to generate more details, it's just repeat the same thing. However in English, it gives a more detailed answer by default.
@AmpBenzScientist I have no idea of how to make nuclear-related stuff, so when it gives a general step for me, I don't know what to ask.
But generally speaking, the model is still a model. It doesn't know how to use CVE to sneak into a system. I copied the CVE page to it and asked how I can use it. The answer is a mixture, part from CVE, part from hallucination.
Also, when asking in Chinese, the result itself is kind of odd. Looks like the model didn't read too much Chinese corpus.
Context: I'm using llama.cpp with the Q5_K_M model from https://huggingface.co/TheBloke/dolphin-2.5-mixtral-8x7b-GGUF
@skyblond That's an interesting result.