Skip to content

Inference

The completion method is used to run a prompt on a specific model, using a specific provider.

If only the prompt is provided, the inference will be run on the model and provider specified in the agent's metadata.

messages = env.list_messages()
result = env.completion(messages)

print("Messages:", messages)
print("Result:", result)
Example Output
Messages: [{'id': 'msg_1149aa85884b4fe8abc7d859', 'content': 'Hello', 'role': 'user'}]

Result: Hello! It's nice to meet you. Is there something I can help you with or would you like to chat?

Generating an Image

The generate_image method is used to generate an image based on the provided description.

description = "A puppy in the garden"
image = env.generate_image(description)

# Extract the base64 data from the first item in `data`
b64_data = image.data[0].b64_json

# Decode the base64 image data
image_data = base64.b64decode(b64_data)

# Write the image file
image_file = env.write_file("puppy_image.png", image_data)

Agent Example

Check out this Agent example to learn how to use generate_image in your AI agent logic.

Overriding the Default Model

To run the inference on a model different from the default one, you can pass the MODEL or PROVIDER::MODEL as second argument:

messages = env.list_messages()
result = env.completion([prompt] + messages, "fireworks::qwen2p5-72b-instruct")
Example Output
Messages: [{'id': 'msg_1149aa85884b4fe8abc7d859', 'content': 'Hello', 'role': 'user'}]

Result: Hello! How can I assist you today? Is there something specific you'd like to talk about or any questions you have?

Tip

completions: returns the full llm response for more control.

Using Models Locally: LangChain / LangGraph

The example agent langgraph-min-example has metadata that specifies the langgraph-0-1-4 framework to run on langgraph version 1.4. In addition, the agent.py code contains an adaptor class, AgentChatModel that maps LangChain inference operations to env.completions calls.