The more context and details you add, the better your response because the model understands precisely what you expect. Do not miss our Prompt Engineering Guide to learn how to write great prompts.

The models can call another model in the background to generate images. The current image model in Langdock is Dall-E3 by OpenAI. Image generation uses the following steps:

  1. The model you selected chooses the image generation tool and writes a prompt to the image model in the background.
  2. The image model generates the image based on the prompt and returns it to the main model and you as the user.

You can select any language model for image generation. Each model sends prompts to the underlying image generation model differently, so feel free to try different models and see how the generated images differ.

Here’s a known limitation we’re working on:

  • Text in images has mistakes / is written in non-existing letters:

    This happens because Dall-E3 was trained on real images that included text. The model generates objects that look similar to what it learned, but it can’t write full, correct sentences yet. Instead, it tries to mimic letters from the alphabet, leading to incorrect spelling or non-existing letters. This is a current limitation of image generation models that OpenAI is actively improving in upcoming versions.