To configure the Display settings, you can follow the following steps. This information is also available by the company hosting the model.Provider: The organization that built and trained the model. This doesn’t necessarily align with the company you consume the model from. For example, you can use Microsoft Azure to use OpenAI models in the EU, but the provider will still be OpenAI.Model name: The name of the model.Hosting provider: Where you consume the model. For example, GPT-4o can be hosted by Microsoft Azure.Region: Shows the user where the model is hosted. This can be set to the US or the EU.Ranking: To give users an indication of how the model performs speed- and quality-wise, you can add a ranking from 1 to 5. Smaller models, like Claude 3 Haiku, GPT-4o mini or Llama 3.1 8B, are faster but don’t have the highest quality. The top models, GPT-4o or Claude 3.5, have high output quality.Knowledge cutoff: When the model training data ended. Most models have a knowledge cutoff at the end of 2023.Image analysis: Indicates whether the model can analyze images. This information is available from the model provider and the model hoster. Please only enable this setting if the model supports vision/image analysis. Models that allow image analysis are GPT-4o, GPT-4o mini, Claude, and Gemini models.
To set up the Model Configuration, select the SDK you are using. You will find information on the configuration of the model provider (e.g., Azure or AWS):SDK: The kit or library Langdock needs to use the model you added.Base URL: To send prompts to the corresponding endpoint of your model.Model ID: The name of the model in your configuration (this might not be the “official” model name, like GPT-4o).API key: Allows your users to authenticate using the model from within Langdock when they send prompts.Context Size: The number of tokens the model can process in its context window. Please use the exact value of the model to ensure the context management in Langdock works correctly.
Maximum messages in 3 hours: Allows you to influence usage/costs and limit messages per user. This setting is optional.Input and output token pricing: Allows you to set the token pricing of the individual model to monitor usage and costs.Visible to everyone: You can set the model to be visible to everyone in the workspace. If this option is disabled, the model is only visible to admins and cannot be used by other users. This allows you to test the model before launching it to the entire workspace.Maintenance mode: Can be activated to show users in the interface that the model might not work as expected. This is useful if you are changing some configuration or there is a temporary issue with the model from your model provider.
We recommend testing the model before making it visible to everyone. Send a message to the model and see if there is a response generated by the model. If you run into any issues, contact support@langdock.com
Mistral from Azure: Make sure to select “Mistral” as the SDK.Claude from AWS Bedrock: The Base URL needs to contain the “access key” / “Zugriffsschlüssel”.Flux from Replicate: The base URL field needs to have the full model path, not just the base URL. For Flux 1.1 Pro this is: https://api.replicate.com/v1/models/black-forest-labs/flux-1.1-pro/predictions