Docs
Overview

Overview

Weavel automates prompt and LLM engineering. With Ape, LLM monitoring, batch testing and evaluations, we help you build better LLM applications with less effort.

Core features

Ape, the first AI Prompt Engineer

Ape is an AI Prompt Engineer that continuously improves your prompts by leveraging the latest research in prompt engineering.

  • Prompt Playground: An interactive playground to play with prompts and see how they perform.
  • Start Without a Dataset: Annotate data from the playground to build a dataset.
  • Automated Evaluation: Automatically generate evaluation code based on your dataset and task.
  • Use Production Data: Log production data and Ape will curate a dataset for you.
  • Human-in-the-loop: Interactive Human-in-the-loop UI that helps Ape to follow your guidance.

Monitoring

Monitor LLM calls and use real-world data to enhance your prompts, improving performance and efficiency.

  • Basic LLM Monitoring: 1-line logging integration with your LLM application.
  • Generation View: View all input and output of LLM calls. Use it to build datasets to improve your prompts.

Advanced Usage: Use Weavel to log client and server-side events to monitor any LLM application.

  • Async SDK for Python and TypeScript. REST API for custom integrations.
  • Session View: Recreate user sessions (from user interactions to server-logs) in conversational format.
  • User View: Calendar to view new and returning users on a daily basis.

Evaluation

  • Evaluation Metrics: Pre-defined evaluation metrics to run batch testing.
  • Batch Testing: SDK-based batch testing & evaluation for any function in your LLM application.

Analytics

This feature is currently in beta. Please contact us to enable it for your account.

  • Semantic Analysis: Extract topic, intent and sentiment from user messages
  • Reports: Find out which semantic events are correlated with your KPIs (e.g. retention, engagement)

Get started

Why Use Weavel?

When building applications powered by LLMs, fine-tuning prompts to get the desired output can feel like an endless loop of trial and error. This process often involves hours of tweaking and testing prompts to achieve the right results, which can be frustrating and time-consuming.

Many developers face similar challenges when working with LLMs:

  • Difficulty in Measuring Output Quality: Evaluating the quality of AI-generated content often requires manual assessments, which can be subjective and inefficient.
  • Time-Consuming Prompt Engineering: Finding the perfect prompt can take hours of experimentation, as even minor changes can lead to significantly different outputs.

How Weavel Helps

To address these challenges, we developed Weavel, a comprehensive tool designed to make prompt engineering and LLM interaction management easier and more efficient.

Key Features

Within Weavel, you have access to Ape, our AI-powered prompt engineering feature that simplifies the entire process. Here’s how Weavel, with Ape, can help you:

  1. Logging: Automatically log all your inputs and outputs with a single line of code. Weavel ensures that you capture every detail of your interactions.
  2. Dataset Curation: Easily select and manage the best input-output examples to refine your LLM’s performance.
  3. Evaluation Templates: Use ready-made templates within Weavel to assess and improve your prompts, saving you hours of manual work.
  4. AI Prompt Engineer (Ape): Ape is a cutting-edge AI-powered prompt engineer within Weavel. It continuously improves your prompts by referencing the latest research in prompt engineering, automatically iterating on them to enhance output quality.

With Weavel and its Ape feature, you can streamline the prompt engineering process, making it faster and more efficient to create effective prompts and manage LLM applications.

Follow us

Join our discord server to get updates and announcements. Join us (opens in a new tab)
To follow our journey, follow us on Linkedin. Follow us (opens in a new tab)