Picture by Editor
# Introduction
Python decorators are tailored options which can be designed to assist simplify advanced software program logic in a wide range of functions, together with LLM-based ones. Coping with LLMs usually includes dealing with unpredictable, sluggish—and ceaselessly costly—third-party APIs, and interior decorators have rather a lot to supply for making this activity cleaner by wrapping, as an illustration, API calls with optimized logic.
Let’s check out 5 helpful Python decorators that can make it easier to optimize your LLM-based functions with out noticeable further burden.
The accompanying examples illustrate the syntax and strategy to utilizing every decorator. They’re generally proven with out precise LLM use, however they’re code excerpts in the end designed to be a part of bigger functions.
# 1. In-memory Caching
This resolution comes from Python’s functools commonplace library, and it’s helpful for costly features like these utilizing LLMs. If we had an LLM API name within the perform outlined under, wrapping it in an LRU (Least Lately Used) decorator provides a cache mechanism that stops redundant requests containing equivalent inputs (prompts) in the identical execution or session. That is a sublime option to optimize latency points.
This instance illustrates its use:
from functools import lru_cache
import time
@lru_cache(maxsize=100)
def summarize_text(textual content: str) -> str:
print("Sending text to LLM...")
time.sleep(1) # A simulation of community delay
return f"Summary of {len(text)} characters."
print(summarize_text("The quick brown fox.")) # Takes one second
print(summarize_text("The quick brown fox.")) # Prompt
# 2. Caching On Persistent Disk
Talking of caching, the exterior library diskcache takes it a step additional by implementing a persistent cache on disk, specifically through a SQLite database: very helpful for storing outcomes of time-consuming features similar to LLM API calls. This manner, outcomes could be shortly retrieved in later calls when wanted. Think about using this decorator sample when in-memory caching will not be enough as a result of the execution of a script or utility could cease.
import time
from diskcache import Cache
# Creating a light-weight native SQLite database listing
cache = Cache(".local_llm_cache")
@cache.memoize(expire=86400) # Cached for twenty-four hours
def fetch_llm_response(immediate: str) -> str:
print("Calling expensive LLM API...") # Exchange this by an precise LLM API name
time.sleep(2) # API latency simulation
return f"Response to: {prompt}"
print(fetch_llm_response("What is quantum computing?")) # 1st perform name
print(fetch_llm_response("What is quantum computing?")) # Prompt load from disk occurs right here!
# 3. Community-resilient Apps
Since LLMs could usually fail on account of transient errors in addition to timeouts and “502 Bad Gateway” responses on the Web, utilizing a community resilience library like tenacity together with the @retry decorator might help intercept these widespread community failures.
The instance under illustrates this implementation of resilient habits by randomly simulating a 70% likelihood of community error. Strive it a number of instances, and ultimately you will note this error arising: completely anticipated and meant!
import random
from tenacity import retry, wait_exponential, stop_after_attempt, retry_if_exception_type
class RateLimitError(Exception): cross
# Retrying as much as 4 instances, ready 2, 4, and eight seconds between every try
@retry(
wait=wait_exponential(multiplier=2, min=2, max=10),
cease=stop_after_attempt(4),
retry=retry_if_exception_type(RateLimitError)
)
def call_flaky_llm_api(immediate: str):
print("Attempting to call API...")
if random.random() < 0.7: # Simulating a 70% likelihood of API failure
increase RateLimitError("Rate limit exceeded! Backing off.")
return "Text has been successfully generated!"
print(call_flaky_llm_api("Write a haiku"))
# 4. Shopper-side Throttling
This mixed decorator makes use of the ratelimit library to manage the frequency of calls to a (often extremely demanded) perform: helpful to keep away from client-side limits when utilizing exterior APIs. The next instance does so by defining Requests Per Minute (RPM) limits. The supplier will reject prompts from a shopper utility when too many concurrent prompts are launched.
from ratelimit import limits, sleep_and_retry
import time
# Strictly implementing a 3-call restrict per 10-second window
@sleep_and_retry
@limits(calls=3, interval=10)
def generate_text(immediate: str) -> str:
print(f"[{time.strftime('%X')}] Processing: {prompt}")
return f"Processed: {prompt}"
# First 3 print instantly, the 4th pauses, thereby respecting the restrict
for i in vary(5):
generate_text(f"Prompt {i}")
# 5. Structured Output Binding
The fifth decorator on the checklist makes use of the magentic library along side Pydantic to offer an environment friendly interplay mechanism with LLMs through API, and acquire structured responses. It simplifies the method of calling LLM APIs. This course of is necessary for coaxing LLMs to return formatted knowledge like JSON objects in a dependable style. The decorator would deal with underlying system prompts and Pydantic-led parsing, optimizing the utilization of tokens consequently and serving to maintain a cleaner codebase.
To do that instance out, you will have an OpenAI API key.
# IMPORTANT: An OPENAI_API_KEY set is required to run this simulated instance
from magentic import immediate
from pydantic import BaseModel
class CapitalInfo(BaseModel):
capital: str
inhabitants: int
# A decorator that simply maps the immediate to the Pydantic return kind
@immediate("What is the capital and population of {country}?")
def get_capital_info(nation: str) -> CapitalInfo:
... # No perform physique wanted right here!
information = get_capital_info("France")
print(f"Capital: {info.capital}, Population: {info.population}")
# Wrapping Up
On this article, we listed and illustrated 5 Python decorators primarily based on numerous libraries that tackle explicit significance when used within the context of LLM-based functions to simplify logic, make processes extra environment friendly, or enhance community resilience, amongst different elements.
Iván Palomares Carrascosa is a frontrunner, author, speaker, and adviser in AI, machine studying, deep studying & LLMs. He trains and guides others in harnessing AI in the true world.



