Groq Unveils Exclusive Access to Meta's Llama 4 in Saudi Arabia

Groq Unveils Exclusive Access to Meta's Llama 4 in Saudi Arabia



In a significant step forward for artificial intelligence in the Middle East, Groq has announced the exclusive market launch of Meta's Llama 4 Scout and Maverick models. Available exclusively on GroqCloud™, these advanced AI models are set to revolutionize the accessibility of high-level machine learning technology for developers in the region.

This milestone is part of Saudi Arabia's ongoing efforts to establish itself as a key player in cutting-edge AI infrastructure. In February, the activation of the largest inference cluster in the region at a data center in Dammam highlighted the kingdom's commitment to advancing its technological capabilities. This switch has enabled Groq to provide Llama 4 globally, creating an opportunity for developers to tap into the potential of these state-of-the-art models right from the outset.

Tareq Amin, a vocal proponent of technological progression in the region, remarked, "Integrating Llama 4 with Groq's technology is a crucial step toward Saudi Arabia's technological leadership." Jonathan Ross, CEO and founder of Groq, echoed this sentiment by stating, "We built Groq to eliminate compute costs entirely. Together with our partners, we're bringing Llama 4 to the region with high-performance inference that operates faster, at a lower cost, and without compromise."

Pricing Overview


The availability pricing is structured as follows for Llama 4 on GroqCloud™:
  • - Llama 4 Scout: $0.11 per million input tokens and $0.34 per million output tokens, averaging out to about $0.13 per mixed rate.
  • - Llama 4 Maverick: $0.50 per million input tokens and $0.77 per million output tokens, resulting in a mixed rate of roughly $0.53.

Model Details


Llama 4 represents the latest family of accessible models from Meta, featuring a Mixture of Experts (MoE) architecture and native multimodality.
  • - Llama 4 Scout (17Bx16E): A versatile model ideal for tasks ranging from summarization to coding. It operates at an impressive rate of over 625 tokens per second on Groq's platform.
  • - Llama 4 Maverick (17Bx128E): A larger and more powerful model optimized for multilingual and multimodal tasks, making it particularly well-suited for applications in assistance, chats, and creative functions. Notably, it supports 12 languages, including Arabic, expanding its usability across diverse markets.

Getting Started with Llama 4


Accessing Llama 4 is straightforward through several channels:
  • - GroqChat
  • - GroqCloud Console
  • - Groq API (model IDs are available in the console)

You can start using Llama 4 today at no initial cost through the GroqCloud console, with options to upgrade for seamless rate limiting and increased throughput.

About Groq


Groq is redefining AI inference through unparalleled pricing and performance. Its custom-designed LPU and cloud infrastructure execute robust models promptly and reliably while maintaining the lowest cost per token–without compromise. With over a million developers utilizing Groq, the platform empowers quicker development cycles and smarter scaling for advanced AI initiatives.

For further information on Groq and its groundbreaking offerings, feel free to reach out or visit their official website for more details.

Topics Consumer Technology)

【About Using Articles】

You can freely use the title and article content by linking to the page where the article is posted.
※ Images cannot be used.

【About Links】

Links are free to use.