@@ -101,12 +101,11 @@ def prompt(
101101 "claude-3.5-sonnet", "claude-3.5-haiku", "claude-3-haiku", "nova-micro",
102102 "nova-lite", "nova-pro"}, default = "gpt-4.1-mini"): Underlying base LLM to use
103103 (better models yield better results, faster models yield faster results). -
104- Models still in beta: "gpt-5", "gpt-5-mini", "gpt-5-nano", "o3", "o1",
105- "o4-mini", "o3-mini", "o1-mini", "gpt-4.5-preview", "claude-opus-4-0",
106- "claude-sonnet-4-0", "claude-3.7-sonnet", "claude-3.5-haiku". - Recommended
107- models for accuracy: "gpt-5", "gpt-4.1", "o4-mini", "o3", "claude-opus-4-0",
108- "claude-sonnet-4-0". - Recommended models for low latency/costs: "gpt-4.1-nano",
109- "nova-micro".
104+ Models still in beta: "o3", "o1", "o4-mini", "o3-mini", "o1-mini",
105+ "gpt-4.5-preview", "claude-opus-4-0", "claude-sonnet-4-0", "claude-3.7-sonnet",
106+ "claude-3.5-haiku". - Recommended models for accuracy: "gpt-5", "gpt-4.1",
107+ "o4-mini", "o3", "claude-opus-4-0", "claude-sonnet-4-0". - Recommended models
108+ for low latency/costs: "gpt-4.1-nano", "nova-micro".
110109
111110 log (list[str], default = []): optionally specify additional logs or metadata that TLM should return.
112111 For instance, include "explanation" here to get explanations of why a response is scored with low trustworthiness.
@@ -245,12 +244,11 @@ def score(
245244 "claude-3.5-sonnet", "claude-3.5-haiku", "claude-3-haiku", "nova-micro",
246245 "nova-lite", "nova-pro"}, default = "gpt-4.1-mini"): Underlying base LLM to use
247246 (better models yield better results, faster models yield faster results). -
248- Models still in beta: "gpt-5", "gpt-5-mini", "gpt-5-nano", "o3", "o1",
249- "o4-mini", "o3-mini", "o1-mini", "gpt-4.5-preview", "claude-opus-4-0",
250- "claude-sonnet-4-0", "claude-3.7-sonnet", "claude-3.5-haiku". - Recommended
251- models for accuracy: "gpt-5", "gpt-4.1", "o4-mini", "o3", "claude-opus-4-0",
252- "claude-sonnet-4-0". - Recommended models for low latency/costs: "gpt-4.1-nano",
253- "nova-micro".
247+ Models still in beta: "o3", "o1", "o4-mini", "o3-mini", "o1-mini",
248+ "gpt-4.5-preview", "claude-opus-4-0", "claude-sonnet-4-0", "claude-3.7-sonnet",
249+ "claude-3.5-haiku". - Recommended models for accuracy: "gpt-5", "gpt-4.1",
250+ "o4-mini", "o3", "claude-opus-4-0", "claude-sonnet-4-0". - Recommended models
251+ for low latency/costs: "gpt-4.1-nano", "nova-micro".
254252
255253 log (list[str], default = []): optionally specify additional logs or metadata that TLM should return.
256254 For instance, include "explanation" here to get explanations of why a response is scored with low trustworthiness.
@@ -405,12 +403,11 @@ async def prompt(
405403 "claude-3.5-sonnet", "claude-3.5-haiku", "claude-3-haiku", "nova-micro",
406404 "nova-lite", "nova-pro"}, default = "gpt-4.1-mini"): Underlying base LLM to use
407405 (better models yield better results, faster models yield faster results). -
408- Models still in beta: "gpt-5", "gpt-5-mini", "gpt-5-nano", "o3", "o1",
409- "o4-mini", "o3-mini", "o1-mini", "gpt-4.5-preview", "claude-opus-4-0",
410- "claude-sonnet-4-0", "claude-3.7-sonnet", "claude-3.5-haiku". - Recommended
411- models for accuracy: "gpt-5", "gpt-4.1", "o4-mini", "o3", "claude-opus-4-0",
412- "claude-sonnet-4-0". - Recommended models for low latency/costs: "gpt-4.1-nano",
413- "nova-micro".
406+ Models still in beta: "o3", "o1", "o4-mini", "o3-mini", "o1-mini",
407+ "gpt-4.5-preview", "claude-opus-4-0", "claude-sonnet-4-0", "claude-3.7-sonnet",
408+ "claude-3.5-haiku". - Recommended models for accuracy: "gpt-5", "gpt-4.1",
409+ "o4-mini", "o3", "claude-opus-4-0", "claude-sonnet-4-0". - Recommended models
410+ for low latency/costs: "gpt-4.1-nano", "nova-micro".
414411
415412 log (list[str], default = []): optionally specify additional logs or metadata that TLM should return.
416413 For instance, include "explanation" here to get explanations of why a response is scored with low trustworthiness.
@@ -549,12 +546,11 @@ async def score(
549546 "claude-3.5-sonnet", "claude-3.5-haiku", "claude-3-haiku", "nova-micro",
550547 "nova-lite", "nova-pro"}, default = "gpt-4.1-mini"): Underlying base LLM to use
551548 (better models yield better results, faster models yield faster results). -
552- Models still in beta: "gpt-5", "gpt-5-mini", "gpt-5-nano", "o3", "o1",
553- "o4-mini", "o3-mini", "o1-mini", "gpt-4.5-preview", "claude-opus-4-0",
554- "claude-sonnet-4-0", "claude-3.7-sonnet", "claude-3.5-haiku". - Recommended
555- models for accuracy: "gpt-5", "gpt-4.1", "o4-mini", "o3", "claude-opus-4-0",
556- "claude-sonnet-4-0". - Recommended models for low latency/costs: "gpt-4.1-nano",
557- "nova-micro".
549+ Models still in beta: "o3", "o1", "o4-mini", "o3-mini", "o1-mini",
550+ "gpt-4.5-preview", "claude-opus-4-0", "claude-sonnet-4-0", "claude-3.7-sonnet",
551+ "claude-3.5-haiku". - Recommended models for accuracy: "gpt-5", "gpt-4.1",
552+ "o4-mini", "o3", "claude-opus-4-0", "claude-sonnet-4-0". - Recommended models
553+ for low latency/costs: "gpt-4.1-nano", "nova-micro".
558554
559555 log (list[str], default = []): optionally specify additional logs or metadata that TLM should return.
560556 For instance, include "explanation" here to get explanations of why a response is scored with low trustworthiness.
0 commit comments