Meta is launching Chameleon, a new advanced multi-modal large language model.

From Nasdaq: 2024-05-22 15:41:25

Meta Platforms is advancing in the artificial intelligence game with the development of Chameleon, a cutting-edge multi-modal large language model. Chameleon is designed to integrate information more effectively than previous models, using a unique architecture that combines images, code, text, and other inputs.

With an ‘early-fusion token-based mixed-modal’ architecture, Chameleon can seamlessly reason over and generate interleaved image and text sequences without modality-specific components. The model is trained in two stages using a massive dataset of 4.4 trillion tokens, image-text combinations, and interwoven text and images.

Researchers at Meta Platforms trained two versions of Chameleon – one with 7 billion parameters and another with 34 billion parameters – for over 5 million hours on Nvidia A100 80GB GPUs. This development comes as competitors like OpenAI and Microsoft have also recently introduced their own advanced AI models.

Please note that the views expressed here are those of the author and do not necessarily reflect those of Nasdaq, Inc.



Read more at Nasdaq: Meta To Launch Chameleon Multi-modal LLM