API Catalog QSG

Within this Quick Start Guide, you will interact with the live chat model API endpoint and get results from an NVIDIA DGX-hosted cloud compute environment.

πŸ“’ NOTE: For this guide, we will reference the llama-3-8B-instruct model. Update the model name to suit your requirements. For example, for a llama3-70b-instruct model.

  1. Access the NVIDIA API Catalog within your web browser.

  2. Within the search bar, type llama3 and select the llama3-8B-instruct model:

This will bring you to the model page. On the left pane is the preview tab, here you can ask NIM LLM a question using the API endpoint.

The sample question provided is β€œWrite a limerick about the wonders of GPU computing.” This technology is using a hosted API endpoint on NVIDIA DGX Cloud.

  1. Click on the sample question and review the response.

You can also test out the model programmatically with the hosted API endpoint. On the right side, the Python tab shows the sample code snippets.

  1. Select Get API Key.


  2. Enter your corporate email, this will sign you up for a trail license of AI Enterprise.

    If you do not have a corporate email, you can use a personal email which will sign you up for Developer Account.


An API key has been generated for you, this key is used for authenticating your API call.


  1. Copy Key.

Do not share it or store it in a place where others can see or copy it.

  1. Click Copy Code

  2. Paste code into your local terminal.

  3. Update your API key within code snippet.


πŸ“˜

Need help?

If you want to learn more about how you can self-host LLM model by deploying a localized NIM on your own infrastructure, click here.


What’s Next