Skip to content

Model Catalog

Our comprehensive model catalog provides a diverse array of models for your selection. To configure your agents to leverage any of these models, please refer to our project configuration guidelines. Below, you will find a list of the models currently supported. We are dedicated to the continuous enhancement and expansion of our model catalog, so please visit this page regularly for the latest updates.

Large Languge Models (LLMs)

The list of LLMs that we support are:

  • meta-llama/Llama-3.1-70B-Instruct
  • meta-llama/Llama-3.1-70B-Instruct-NIM (powered by NIM)
  • meta-llama/Llama-3.1-8B-Instruct

Configuring LLMs for Your Project

To utilize any of these LLMs in your project, simply update the llm_config within the base_config or within the config section of any utility agents in your project's YAML configuration file. Ensure that the model parameter of the llm_config is set to one of the names listed above.

Using LLMs through Our Inference API

You can also directly use any of the models listed above through our inference API. See an example below:

import os

from openai import OpenAI

from air import login

auth = login(
    account=str(os.getenv("ACCOUNT")), # your account 
    api_key=str(os.getenv("API_KEY")), # your API key
)
base_url = os.getenv("AIREFINERY_ADDRESS", "")

client = OpenAI(**auth.openai(base_url=base_url))

# Create a chat request  
response = client.chat.completions.create(
    messages=[{"role": "user", "content": "What is the capital of France?"}],
    model="meta-llama/Llama-3.1-70B-Instruct", # an LLM from the list  above
)

Vision Language Models (VLMs)

The list of VLMs that we support are:

  • meta-llama/Llama-3.2-90B-Vision-Instruct

To utilize any of these VLMs, simply update the vlm_config within the base_config or within the config section of any utility agents (which support images e.g.,ImageUnderstandingAgent) in your project's YAML configuration file. Ensure that the model parameter of the vlm_config is set to one of the names listed above.

Embedding Models

The list of models that we support for embedding your data are as follows:

  • intfloat/e5-mistral-7b-instruct
  • nvidia/nv-embedqa-mistral-7b-v2
  • intfloat/multilingual-e5-large

Using Embedding Models in Your Project

To utilize any of these embedding models in your project, simply update the embedding_config within the base_config or within the aisearch_config section of the ResearchAgent. Ensure that the model_name parameter of the embedding_config is set to one of the names listed above.

Embedding Your Data Using Our Inference API

You can also directly use any of the models listed above to embed your data using our inference API. See an example below:

import os

from openai import OpenAI

from air import login

auth = login(
    account=str(os.getenv("ACCOUNT")), # your account 
    api_key=str(os.getenv("API_KEY")), # your API key
)
base_url = os.getenv("AIREFINERY_ADDRESS", "")

client = OpenAI(**auth.openai(base_url=base_url))

# Create an embedding request  
response = client.embeddings.create(  
    input=["What is the capital of France?"],  
    model="nvidia/nv-embedqa-mistral-7b-v2",  # required
    encoding_format="float",  # required
    extra_body={"input_type": "query", "truncate": "NONE"}  # extra_body is required for "nvidia/nv-embedqa-mistral-7b-v2" model
    # where "input_type" can be either "query" or "passage"
)  

Compression Models

The list of prompt compression models that we support are:

  • llmlingua/bert

To utilize any of these prompt compression models in your project, simply update the compression_config within the base_config of your project. To learn more about prompt compression, see this tutorial. Ensure that the model parameter of the compression_config is set to one of the names listed above.

Reranker Models

The list of reranker models that we support are:

  • BAAI/bge-reranker-large
  • BAAI/bge-reranker-v2-m3

To utilize any of these reranker models in your project, simply update the reranker_config within the base_config of your project. To learn more about reranking, see this tutorial. Ensure that the model parameter of the reranker_config is set to one of the names listed above.