Apple has introduced MM1, a multimodal language learning mode available in 3 billion, 7 billion, and 30 billion parameter sizes. Based on the company's expertise in advanced technological solutions, MM1 can recognize images and natural language reasoning.
The model is developed based on expert architecture and the top-2 gating method. The models are claimed to answer questions based on an image and text, answer text-based questions about an image, and answer scientific questions.
Reportedly, Apple is working on an LLM framework called “Ajax” and a chatbot known as “Apple GPT."
By using this site, you agree to allow SPEEDA Edge and our partners to use cookies for analytics and personalization. Visit our privacy policy for more information about our data collection practices.