Skip to content

Lemonade Chat Model node#

Use the Lemonade Chat Model node to run chat-capable language models managed by a Lemonade server from within n8n. This node functions as a LangChain-compatible chat model root node and is suitable for chat-style workloads. It lets you select a model hosted on your Lemonade server, and control generation behavior using common sampling and decoding options.

On this page, you'll find a list of the node parameters, and available options to refine generation.

Credentials

You can find authentication information for this node here.

Parameter resolution in sub-nodes

Sub-nodes behave differently to other nodes when processing multiple items using an expression.

Most nodes, including root nodes, take any number of items as input, process these items, and output the results. You can use expressions to refer to input items, and the node resolves the expression for each item in turn. For example, given an input of five name values, the expression {{ $json.name }} resolves to each name in turn.

In sub-nodes, the expression always resolves to the first item. For example, given an input of five name values, the expression {{ $json.name }} always resolves to the first name.

Node parameters#

Model#

The model which will generate the completion. Models are loaded and managed through the Lemonade server. This parameter is required. Select the model name made available by your Lemonade server (for example, a model alias like "gpt-4", or any custom model name exposed by Lemonade).

Models are provided by the Lemonade server; if you don't see the model you expect, verify your Lemonade server configuration and credentials.

Node options#

Use these options to further refine the node's behavior.

Sampling Temperature#

Controls the randomness of the generated text. Lower values make the output more focused and deterministic, while higher values make it more diverse and random.

Property Value
Type number
Required no
Default 0.7

Top P#

Controls which words the model can choose from when generating text. Lower values progressively remove the least likely options, so the model can only pick from a smaller, higher-confidence pool.

Property Value
Type number
Required no
Default 1

Frequency Penalty#

Adjusts the penalty for tokens that have already appeared in the generated text. Positive values discourage repetition, negative values encourage it.

Property Value
Type number
Required no
Default 0

Presence Penalty#

Adjusts the penalty for tokens based on their presence in the generated text so far. Positive values penalize tokens that have already appeared, encouraging diversity.

Property Value
Type number
Required no
Default 0

Max Tokens to Generate#

The maximum number of tokens to generate. Set to -1 for no limit. Be cautious when setting this to a large value, as it can lead to long outputs.

Property Value
Type number
Required no
Default -1

Stop Sequences#

Comma-separated list of sequences where the model will stop generating text. Use this to define explicit termination strings for responses.

Property Value
Type string
Required no
Default ""

Templates and examples#

AI agent chat

by n8n Team

View template details
Building Your First WhatsApp Chatbot

by Jimleuk

View template details
Scrape and summarize webpages with AI

by n8n Team

View template details
Browse Lemonade Chat Model integration templates, or search all templates

Refer to Lemonade Server's documentation for more information about the service.

View n8n's Advanced AI documentation.

This page was