Tidy Integration of Large Language Models


[Up] [Top]

Documentation for package ‘tidyllm’ version 0.2.0

Help Pages

azure_openai Send LLM Messages to an OpenAI Chat Completions endpoint on Azure
chatgpt ChatGPT Wrapper (Deprecated)
check_claude_batch Check Batch Processing Status for Claude API
check_openai_batch Check Batch Processing Status for OpenAI Batch API
claude Interact with Claude AI models via the Anthropic API
df_llm_message Convert a Data Frame to an LLMMessage Object
fetch_claude_batch Fetch Results for a Claude Batch
fetch_openai_batch Fetch Results for an OpenAI Batch
generate_callback_function Generate API-Specific Callback Function for Streaming Responses
get_reply Get Assistant Reply as Text
get_reply_data Get Data from an Assistant Reply by parsing structured JSON responses
get_user_message Retrieve a User Message by Index
groq Send LLM Messages to the Groq Chat API
groq_transcribe Transcribe an Audio File Using Groq transcription API
initialize_api_env Initialize or Retrieve API-specific Environment
last_reply Get the Last Assistant Reply as Text
last_reply_data Get the Last Assistant Reply as Text
last_user_message Retrieve the Last User Message
list_claude_batches List Claude Batch Requests
list_openai_batches List OpenAI Batch Requests
LLMMessage Large Language Model Message Class
llm_message Create or Update Large Language Model Message Object
mistral Send LLMMessage to Mistral API
mistral_embedding Generate Embeddings Using Mistral API
ollama Interact with local AI models via the Ollama API
ollama_download_model Download a model from the Ollama API
ollama_embedding Generate Embeddings Using Ollama API
ollama_list_models Retrieve and return model information from the Ollama API
openai Send LLM Messages to the OpenAI Chat Completions API
openai_embedding Generate Embeddings Using OpenAI API
parse_duration_to_seconds This internal function parses duration strings as returned by the OpenAI API
pdf_page_batch Batch Process PDF into LLM Messages
perform_api_request Perform an API request to interact with language models
ratelimit_from_header Extract rate limit information from API response headers
rate_limit_info Get the current rate limit information for all or a specific API
send_claude_batch Send a Batch of Messages to Claude API
send_openai_batch Send a Batch of Messages to OpenAI Batch API
tidyllm_schema Create a JSON schema for structured outputs
update_rate_limit Update the standard API rate limit info in the hidden .tidyllm_rate_limit_env environment