Uploaded image for project: 'translate5'
  1. translate5
  2. TRANSLATE-4829

Max number of segments in request to LLM

XMLWordPrintable

    • Icon: New Feature New Feature
    • Resolution: Fixed
    • translate5 - 7.29.0
    • None
    • AI
    • High
    • Hide
      [🆕 Feature] AI-Plugin:
      * updated model-list & model capabilities to match OpenAI latest changes
      * improved token-limit calculations
      * added model configuration to customize batch-size for pretranslation and disable batching all together
      Show
      [🆕 Feature] AI-Plugin: * updated model-list & model capabilities to match OpenAI latest changes * improved token-limit calculations * added model configuration to customize batch-size for pretranslation and disable batching all together

      Problem

      The larger the text, that is send to an LLM, the more unspecific gets the info to the LLM, what it should do/what this is about. That means, the more segments you send, the worse is the quality.

      On the contrary, the less segments are send the higher are the costs because the overhead (system messages, examples) stay the same.

      Solution

      Users need to be able to define themselves, how many segments are send in one request (be it translation or quality estimation) to the LLM.

      The max should still be defined by the max number of tokens, that can be send in one request. So this should still be checked before the request is send. And if needed, the payload should be split across multiple requests (will only be valid for very high number of configured segments/request usually).

      The config should be an AI language resource property and its default is a system config. The default of the default should be 6.

            axelbecher Axel Becher
            marcmittag Marc Mittag [Administrator]
            Leon Kiz
            Votes:
            0 Vote for this issue
            Watchers:
            1 Start watching this issue

              Created:
              Updated:
              Resolved: