GPT and Gemini aren’t the only LLMs available, nor are they the largest. Many other LLMs come from different vendors, some with a much larger number of parameters, like Megatron-Turing from NVIDIA/Microsoft (530 billion parameters) or PaLM from Google (540 billion parameters). Others are tailored for specific tasks, like Codex from OpenAI and GitHub Copilot from GitHub/Microsoft, which specialize in writing code.
For the Chinese language, there are several models: PanGu-α from Huawei (200 billion parameters), Z-Code from Alibaba (70 billion parameters), and Wu Dao 2.0 from Beijing Academy (1,750 billion parameters). However, only PanGu-α is available through an API.
Returning to Gemini, it has three versions: Gemini Nano with 6 billion parameters, Gemini Pro with 60 billion, and Gemini Ultra with 540 billion, but the free version is only available for Gemini Nano.
GPT-5 is projected to be released in July 2024 and will have almost three times the number of parameters as GPT-3 (500 billion compared to 175 billion). Note that the number of parameters for GPT-4 hasn’t been publicly disclosed (it’s likely larger than 175 billion, but possibly the same). You can now speak with GPT-4o (released in May 2024) almost as fluently as with a human, request it to read you a poem, or ask it to create and sing a song. While it is not yet an artificial general intelligence (AGI), its current capabilities are impressive.
As you can see, the landscape of LLMs is quite complex and diverse, so choosing the right model can be a challenging task. At Leobit, we build R&D projects for each of the major vendors: OpenAI, Microsoft, NVIDIA, AWS, and Meta in order to better consult our customers regarding the preferred choices for their specific needs.
