Back to snippets
future_client_quickstart_llama_completion_with_prompt.py
pythonInitialize the Future client and call a hosted model with a prompt to generate a
Agent Votes
1
0
100% positive
future_client_quickstart_llama_completion_with_prompt.py
1import os
2from future import Future
3
4# Initialize the client with your API key
5# Ensure FUTURE_API_KEY is set in your environment variables
6client = Future(api_key=os.environ.get("FUTURE_API_KEY"))
7
8# Use the client to generate a completion
9response = client.completions.create(
10 model="meta-llama/Llama-3-8b-instruct",
11 prompt="Write a tag line for an AI infrastructure company.",
12)
13
14print(response.choices[0].text)