Model Overview
Description
SeaLLM-7B-v2.5 is a state-of-the-art multilingual large language model for Southeast Asian (SEA) languages that can task solve, mathematically reason, write, answer questions, and respond to safety-related inquiries. It is the most significant upgrade since SeaLLM-13B, with half the size, outperforming performance across diverse multilingual tasks, from world knowledge, math reasoning, and instruction following. This model is for demonstration purposes and not-for-production usage.
Third-Party Community Consideration
This model is not owned or developed by NVIDIA. This model has been developed and built to a third-party’s requirements for this application and use case; see link to the SeaLLM Model Card.
License and Terms of use
GOVERNING TERMS: Your use of this API is governed by the NVIDIA API Trial Service Terms of Use; and the use of this model is governed by the NVIDIA AI Foundation Models Community License and SeaLLMs License.
Model Developer: Language Technology Lab at Alibaba DAMO Academy
Model Release Date: April 3, 2024.
Model Architecture
- Architecture Type: Transformer
- Network Architecture: Gemma
SeaLLM-7B-v2.5 was built on top of Gemma-7b, and underwent large scale supervised fine-tuning (SFT) and carefully designed alignment.
Input
- Input Type: Text
- Input Format: String
- Input Parameters: max_tokens, temperature, top_p, stop, frequency_penalty, presence_penalty, seed
Output
- Output Type: Text
- Output Format: String
Software Integration:
- Supported Hardware Platform(s): Lovelace
[Preferred/Supported] Operating System(s):
- Linux
Model Version
SeaLLM-7B-v2.5
Inference
Engine: Triton + TensorRT-LLM
Test Hardware: L40S