It was trained on 1T tokens of text and code that was curated by MosaicML’s data team. MPT-7B Base is a decoder-style transformer with 6.7B parameters. Today, we are releasing the base MPT model and three other finetuned variants that demonstrate the many ways of building on this base model: MPT-7B Base: We rigorously evaluated MPT on a range of benchmarks, and MPT met the high quality bar set by LLaMA-7B.
0 Comments
Leave a Reply. |
AuthorWrite something about yourself. No need to be fancy, just an overview. ArchivesCategories |