Other lanuage ablity
Outside of training in the base model, was this trained with examples in other languages than just English? To me seems there might be a improvement in Japanese but not sure
Actually more testing it seems worse in some cases? I asked it to translate a song and it just outputted this when I corrected it and told it not to use romaji
[1]
[2]
[3]
And as sanity check I tested on Huggingchat
Translated Japanese Lyrics:
[Verse 1]
[Verse 2]
[Chorus]
[Bridge]
[Chorus]
This reminds me of what Llama 3 not to be confused with Llama 3.1 would do annoyingly but 3.1 fixed, I suspected that Japanese data was literally find and replaced in the whole training dataset for some reason. But when it doesn't do that, it's quite good maybe? I am not sure why the heck Meta thought that was a good idea, but besides the point 3.1 didn't do this anymore I thought