Gemma 3 - The NEW Gemma Family Members Have Arrived!!!
AI Summary
Overview of the new Gemma 3 model family released by Google, featuring four models: 1B, 4B, 12B, and 27B.
- Multimodal Functionality: The 4B, 12B, and 27B models can process both text and images; the 1B model is text-only.
- Context Length: The 1B model supports 32,000 tokens, while the others support 128,000 tokens, a significant increase over previous models.
- Training Details: Each model has been trained on a vast number of tokens, with the 27B model trained on 14 trillion tokens. Enhanced with knowledge distillation and better data filtering methods.
- Performance: The 4B model is competitive with the previous 27B model, enhancing capabilities in math and reasoning.
- Use Cases: Supports multilingual tasks and allows for custom fine-tuning. Models show proficiency in visual question answering and OCR tasks.
- Integration: Users can utilize these models via the Transformer’s library, highlighting new pipeline functionalities.
- Future Prospects: Expected availability on platforms like Google Cloud and AI studio. Research potentials with smaller models for experimentation in diverse languages and tasks.