Skip to main content

Documentation Index

Fetch the complete documentation index at: https://docs.iblueprint.ai/llms.txt

Use this file to discover all available pages before exploring further.

iBlueprint Blueprints are built from nodes, and each node type is designed for a specific kind of work. Choosing the right node type keeps your Blueprint readable and ensures you get the right behavior from the execution engine. The sections below group node types by purpose so you can find the right one quickly.
These nodes call AI models to generate, transform, or analyse content. They are the core of most Blueprints.

prompt — Text generation

The prompt node sends a message to a language model and returns the generated text. Use it any time you need the model to write, summarise, classify, extract, translate, or reason over text.Example configuration
{
  "model": "gpt-4o",
  "systemPrompt": "You are a helpful assistant that writes concise summaries.",
  "userPrompt": "Summarise the following article in three bullet points:\n\n{{article_text}}"
}
FieldDescription
modelThe model to use, e.g. gpt-4o, claude-3-5-sonnet, gemini-1.5-pro
systemPromptSets the model’s persona and constraints
userPromptThe actual request sent to the model; supports {{variables}}
Use {{previousNode.output}} in userPrompt to feed the result of the previous node directly into your prompt.
The image_generation node sends a text prompt to an image model (such as DALL·E or Stable Diffusion) and returns a generated image. Use it to create illustrations, product mockups, or visual assets on the fly.Key config fields: prompt, model, size, quality, n (number of images).
The image_editing node takes an input image and applies model-guided edits based on a text instruction. Use it to retouch photos, apply styles, or modify regions of an image.Key config fields: image (URL or base64), instruction, model, mask.
The image_variations node produces alternative versions of a source image while preserving its overall composition. Use it to explore creative directions or generate multiple options.Key config fields: image, n, size, model.
The vision node passes one or more images to a multimodal model alongside a text prompt. Use it to describe images, extract text from screenshots, or answer questions about visual content.Key config fields: images (array of URLs or base64), prompt, model.
The ocr node runs optical character recognition on an image and returns the extracted text. Use it to digitise scanned documents, receipts, or screenshots before feeding the text into a prompt node.Key config fields: image, language.
The speech_to_text node transcribes an audio file into text. Use it to process voice recordings, meeting audio, or podcast episodes before running downstream analysis.Key config fields: audio (URL or base64), model, language.
The text_to_speech node synthesises spoken audio from a text string. Use it to build audio summaries, voice interfaces, or accessibility features.Key config fields: text, voice, model, speed.
The video_analysis node sends a video (or keyframes) to a multimodal model for analysis. Use it to summarise video content, detect events, or extract structured information from recordings.Key config fields: video (URL), prompt, model.
The data_extraction node instructs a model to extract structured fields from unstructured text and return them as JSON. Use it to parse emails, reports, or any free-form document into a schema you define.Key config fields: input, schema (JSON Schema), model.