Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

docs(app): AGE-1228 changelog 0.27.0 #2220

Merged
merged 4 commits into from
Nov 12, 2024
Merged
Changes from 1 commit
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
34 changes: 34 additions & 0 deletions docs/blog/main.mdx
Original file line number Diff line number Diff line change
Expand Up @@ -9,6 +9,40 @@ import Image from "@theme/IdealImage";

<section class="changelog">

### Observability and Prompt Management

_6 November 2024_

**v0.27.0**

This release is one of our biggest yet—one changelog hardly does it justice.

**First up: Observability**

We’ve had observability in beta for a while, but now it’s been completely rewritten, with a brand-new UI and fully **open-source code**.

The new Observability SDK is compatible with OpenTelemetry (Otel) and semantic conventions (semconv). This means you get a lot of integrations right out of the box, like LangChain, OpenAI, and more.

We’ll publish a full blog post soon, but here’s a quick look at what the new observability offers:

A redesigned UI that lets you visualize nested traces, making it easier to understand what’s happening behind the scenes.

The web UI lets you filter traces by name, cost, and other attributes—you can even search through them easily.

The SDK is Otel-compatible, and we’ve already tested integrations for OpenAI, LangChain, LiteLLM, and Instructor, with guides available for each. In most cases, adding a few lines of code will have you seeing traces directly in Agenta.

**Next: Prompt Management**

We’ve completely rewritten the prompt management SDK, giving you full CRUD capabilities for prompts and configurations. This includes creating, updating, reading history, deploying new versions, and deleting old ones. You can find the documentation for this here: [link]

**And finally: LLM-as-a-Judge Overhaul**

We’ve made significant upgrades to the LLM-as-a-Judge evaluator. It now supports prompts with multiple messages and has access to all variables in a test case. You can also switch models (currently supporting OpenAI and Anthropic). These changes make the evaluator much more flexible, and we’re seeing faster evaluations as a result.

We’re excited about these improvements and believe they’ll make a real difference for your workflows. Let us know what you think, and keep an eye out for the upcoming blog post with more details!

---

### New Application Management View and Various Improvements

_22 October 2024_
Expand Down