[Feature] Adding support for Mixtral and Gemma models#1247
[Feature] Adding support for Mixtral and Gemma models#1247martinakaduc wants to merge 6 commits intohaotian-liu:mainfrom
Conversation
|
@martinakaduc I see some of my comments in here--did you figure out the necessary changes to the |
@muhark Yes, I have referenced your code. I also figured out the problem with I have successfully trained my llava_gemma for Vietnamese! 😄 (https://huggingface.co/ura-hcmut/GemSUraV-7B) |
|
@martinakaduc Sorry to bother. I see some warnings when pretrain Mixtral 8*7B, just like #1417. Did you face the same problem? Thanks in advance! |
|
why did you use |
This pull request intends to add support for Mixtral and Gemma as LLM backbone.