seallms / seallm-7b-v2.5

Model Overview

Description

SeaLLM-7B-v2.5 is a state-of-the-art multilingual large language model for Southeast Asian (SEA) languages that can task solve, mathematically reason, write, answer questions, and respond to safety-related inquiries. It is the most significant upgrade since SeaLLM-13B, with half the size, outperforming performance across diverse multilingual tasks, from world knowledge, math reasoning, and instruction following. This model is for demonstration purposes and not-for-production usage.

Third-Party Community Consideration

This model is not owned or developed by NVIDIA. This model has been developed and built to a third-party’s requirements for this application and use case; see link to the SeaLLM Model Card.

License and Terms of use

GOVERNING TERMS: Your use of this API is governed by the NVIDIA API Trial Service Terms of Use; and the use of this model is governed by the NVIDIA AI Foundation Models Community License and SeaLLMs License.

Model Developer: Language Technology Lab at Alibaba DAMO Academy

Model Release Date: April 3, 2024.

Model Architecture

  • Architecture Type: Transformer
  • Network Architecture: Gemma

SeaLLM-7B-v2.5 was built on top of Gemma-7b, and underwent large scale supervised fine-tuning (SFT) and carefully designed alignment.

Input

  • Input Type: Text
  • Input Format: String
  • Input Parameters: max_tokens, temperature, top_p, stop, frequency_penalty, presence_penalty, seed

Output

  • Output Type: Text
  • Output Format: String

Software Integration:

  • Supported Hardware Platform(s): Lovelace

[Preferred/Supported] Operating System(s):

  • Linux

Model Version

SeaLLM-7B-v2.5

Inference

Engine: Triton + TensorRT-LLM

Test Hardware: L40S