Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[TLM] Add TLMCalibrated class #326

Merged
merged 7 commits into from
Sep 25, 2024
Merged
Show file tree
Hide file tree
Changes from 4 commits
Commits
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
4 changes: 4 additions & 0 deletions cleanlab_studio/errors.py
Original file line number Diff line number Diff line change
Expand Up @@ -96,6 +96,10 @@ class TlmPartialSuccess(APIError):
pass


class TlmNotCalibratedError(HandledError):
pass


class UnsupportedVersionError(HandledError):
def __init__(self) -> None:
super().__init__(
Expand Down
23 changes: 22 additions & 1 deletion cleanlab_studio/studio/studio.py
Original file line number Diff line number Diff line change
Expand Up @@ -29,7 +29,7 @@
init_dataset_source,
telemetry,
)
from cleanlab_studio.utils import tlm_lite
from cleanlab_studio.utils import tlm_lite, tlm_calibrated

from . import enrichment, inference, trustworthy_language_model

Expand Down Expand Up @@ -652,6 +652,27 @@ def TLMLite(
verbose=verbose,
)

def TLMCalibrated(
self,
quality_preset: TLMQualityPreset = "medium",
*,
options: Optional[trustworthy_language_model.TLMOptions] = None,
timeout: Optional[float] = None,
verbose: Optional[bool] = None,
) -> tlm_calibrated.TLMCalibrated:
"""
Instantiate a version of the Trustworthy Language Model that you can calibrate using existing ratings for example prompt-response pairs.
For more details, see the documentation of:
[cleanlab_studio.utils.tlm_calibrated.TLMCalibrated](../utils.tlm_calibrated/#class-tlmcalibrated)
"""
return tlm_calibrated.TLMCalibrated(
self._api_key,
quality_preset,
options=options,
timeout=timeout,
verbose=verbose,
)

def poll_cleanset_status(self, cleanset_id: str, timeout: Optional[int] = None) -> bool:
"""
This method has been deprecated, instead use: `wait_until_cleanset_ready()`
Expand Down
4 changes: 4 additions & 0 deletions cleanlab_studio/studio/trustworthy_language_model.py
Original file line number Diff line number Diff line change
Expand Up @@ -204,6 +204,9 @@ def __init__(
if "log" in options_dict.keys() and len(options_dict["log"]) > 0:
self._return_log = True

if "custom_eval_criteria" in options_dict.keys():
jas2600 marked this conversation as resolved.
Show resolved Hide resolved
self._return_log = True

# explicitly specify the default model
self._options = {**{"model": _TLM_DEFAULT_MODEL}, **options_dict}

Expand Down Expand Up @@ -787,6 +790,7 @@ class TLMOptions(TypedDict):
num_consistency_samples: NotRequired[int]
use_self_reflection: NotRequired[bool]
log: NotRequired[List[str]]
custom_eval_criteria: NotRequired[List[Dict[str, Any]]]


def is_notebook() -> bool:
Expand Down
176 changes: 176 additions & 0 deletions cleanlab_studio/utils/tlm_calibrated.py
Original file line number Diff line number Diff line change
@@ -0,0 +1,176 @@
"""
TLM Calibrated is a variant of the Trustworthy Language Model (TLM) that facilitates the calibration of trustworthiness scores
using existing ratings for prompt-response pairs, which allows for better alignment of the TLM scores in specialized-use cases.

**This module is not meant to be imported and used directly.**
Instead, use [`Studio.TLMCalibrated()`](/reference/python/studio/#method-tlmcalibrated) to instantiate a [TLMCalibrated](#class-tlmcalibrated) object,
and then you can use the methods like [`get_trustworthiness_score()`](#method-get_trustworthiness_score) documented on this page.
"""

from __future__ import annotations
from typing import Optional, Sequence, Union, List

import numpy as np
import numpy.typing as npt
import pandas as pd
from sklearn.ensemble import RandomForestRegressor
from sklearn.exceptions import NotFittedError
from sklearn.utils.validation import check_is_fitted

from cleanlab_studio.errors import ValidationError, TlmNotCalibratedError
from cleanlab_studio.internal.types import TLMQualityPreset
from cleanlab_studio.studio.trustworthy_language_model import TLM, TLMOptions, TLMScore


class TLMCalibrated:
def __init__(
self,
api_key: str,
quality_preset: TLMQualityPreset,
*,
options: Optional[TLMOptions] = None,
timeout: Optional[float] = None,
verbose: Optional[bool] = None,
) -> None:
"""
Use `Studio.TLMCalibrated()` instead of this method to initialize a TLMCalibrated object.
lazydocs: ignore
"""
self._api_key = api_key

if quality_preset not in {"base", "low", "medium"}:
raise ValidationError(
f"Invalid quality preset: {quality_preset}. TLMCalibrated only supports 'base', 'low' and 'medium' presets."
)
self._quality_preset = quality_preset

self._options = options
self._timeout = timeout if timeout is not None and timeout > 0 else None
self._verbose = verbose

custom_eval_criteria_list = (
self._options.get("custom_eval_criteria", []) if self._options else []
)

# number of custom eval critera + 1 to account for the default TLM trustworthiness score
self._num_features = len(custom_eval_criteria_list) + 1
self._rf_model = RandomForestRegressor(monotonic_cst=[1] * self._num_features)

self._tlm = TLM(
self._api_key,
quality_preset=self._quality_preset,
options=self._options,
timeout=self._timeout,
verbose=self._verbose,
)

def fit(self, tlm_scores: List[TLMScore], ratings: Sequence[float]) -> None:
"""
Callibrate the model using TLM scores obtained from a previous `TLM.get_trustworthiness_score()` call
using the provided numeric ratings.

Args:
tlm_scores (List[TLMScore]): list of [TLMScore](../trustworthy_language_model/#class-tlmscore) object obtained
from a previous `TLM.get_trustworthiness_score()` call
ratings (Sequence[float]): sequence of numeric ratings corresponding to each prompt-response pair,
the length of this sequence must match the length of the `tlm_scores`.
"""
if len(tlm_scores) != len(ratings):
raise ValidationError(
"The list of ratings must be of the same length as the list of TLM scores."
)

tlm_scores_df = pd.DataFrame(tlm_scores)
extracted_scores = self._extract_tlm_scores(tlm_scores_df)

if extracted_scores.shape[1] != self._num_features:
raise ValidationError(
f"TLMCalibrated has {self._num_features - 1} custom evaluation criteria defined, "
f"however the tlm_scores provided have {extracted_scores.shape[1] - 1} custom evaluation scores. "
"Please make sure the number of custom evaluation criterias match."
)

# using pandas so that NaN values are handled correctly
ratings_series = pd.Series(ratings)
ratings_normalized = (ratings_series - ratings_series.min()) / (
ratings_series.max() - ratings_series.min()
)

self._rf_model.fit(extracted_scores, ratings_normalized.values)

def get_trustworthiness_score(
jas2600 marked this conversation as resolved.
Show resolved Hide resolved
self, prompt: Union[str, Sequence[str]], response: Union[str, Sequence[str]]
) -> Union[TLMScoreWithCalibration, List[TLMScoreWithCalibration]]:
"""
Computes the calibrated trustworthiness score for arbitrary given prompt-response pairs,
make sure that the model has been calibrated by calling the `.fit()` method before using this method.

Similar to [`TLM.get_trustworthiness_score()`](../trustworthy_language_model/#method-get_trustworthiness_score),
view documentation there for expected input arguments and outputs.
"""
try:
check_is_fitted(self._rf_model)
except NotFittedError:
raise TlmNotCalibratedError(
"TLMCalibrated has to be calibrated before scoring new data, use the .fit() method to calibrate the model."
)

tlm_scores = self._tlm.get_trustworthiness_score(prompt, response)

is_single_query = isinstance(tlm_scores, dict)
if is_single_query:
assert not isinstance(tlm_scores, list)
tlm_scores = [tlm_scores]
tlm_scores_df = pd.DataFrame(tlm_scores)

extracted_scores = self._extract_tlm_scores(tlm_scores_df)

tlm_scores_df["calibrated_score"] = self._rf_model.predict(extracted_scores)

if is_single_query:
return tlm_scores_df.to_dict(orient="records")[0]

return tlm_scores_df.to_dict(orient="records")

def _extract_tlm_scores(self, tlm_scores_df: pd.DataFrame) -> npt.NDArray[np.float64]:
"""
Transform a DataFrame containing TLMScore objects into a 2D numpy array,
where each column represents different scores including trustworthiness score and any custom evaluation criteria.

Args:
tlm_scores_df: DataFrame constructed using a list of TLMScore objects.

Returns:
np.ndarray: 2D numpy array where each column corresponds to different scores.
The first column is the trustworthiness score, followed by any custom evaluation scores if present.
"""
tlm_log = tlm_scores_df.get("log", None)

# if custom_eval_criteria is present in the log, use it as features
if tlm_log is not None and "custom_eval_criteria" in tlm_log.iloc[0]:
custom_eval_scores = np.array(
tlm_scores_df["log"]
.apply(lambda x: [criteria["score"] for criteria in x["custom_eval_criteria"]])
.tolist()
)
all_scores = np.hstack(
[tlm_scores_df["trustworthiness_score"].values.reshape(-1, 1), custom_eval_scores]
)
# otherwise use the TLM trustworthiness score as the only feature
else:
all_scores = tlm_scores_df["trustworthiness_score"].values.reshape(-1, 1)

return all_scores


class TLMScoreWithCalibration(TLMScore):
"""
A typed dict similar to [TLMScore](../trustworthy_language_model/#class-tlmscore) but containing an extra key `calibrated_score`.
View [TLMScore](../trustworthy_language_model/#class-tlmscore) for the description of the other keys in this dict.

Attributes:
calibrated_score (float, optional): score between 0 and 1 that has been calibrated to the provided ratings.
A higher score indicates a higher confidence that the response is correct/trustworthy,
"""

calibrated_score: Optional[float]
1 change: 1 addition & 0 deletions setup.py
Original file line number Diff line number Diff line change
Expand Up @@ -63,6 +63,7 @@
"openpyxl>=3.0.0,!=3.1.0",
"validators>=0.20.0",
"matplotlib>=3.4.0",
"scikit-learn",
jas2600 marked this conversation as resolved.
Show resolved Hide resolved
],
entry_points="""
[console_scripts]
Expand Down
Loading