#!/usr/bin/env bash set -e # @meta dotenv # @env DRY_RUN Dry run mode # @cmd Test first running # @env AICHAT_CONFIG_DIR=tmp/test-init-config # @arg args~ test-init-config() { unset OPENAI_API_KEY mkdir -p "$AICHAT_CONFIG_DIR" config_file="$AICHAT_CONFIG_DIR/config.yaml" if [[ -f "$config_file" ]]; then rm -f "$config_file" fi cargo run -- "$@" } # @cmd Test running with AICHAT_PLATFORM environment variable # @env AICHAT_PLATFORM! # @arg args~ test-platform-env() { cargo run -- "$@" } # @cmd Test function calling # @option --model[?`_choice_model`] # @option --preset[=default|weather|multi-weathers] # @flag -S --no-stream # @arg text~ test-function-calling() { args=(--role %functions%) if [[ -n "$argc_model" ]]; then args+=("--model" "$argc_model") fi if [[ -n "$argc_no_stream" ]]; then args+=("-S") fi if [[ -z "$argc_text" ]]; then case "$argc_preset" in multi-weathers) text="what is the weather in London and Pairs?" ;; weather|*) text="what is the weather in London?" ;; esac else text="${argc_text[*]}" fi cargo run -- "${args[@]}" "$text" } # @cmd Test clients # @arg clients+[`_choice_client`] test-clients() { for c in "${argc_clients[@]}"; do echo "### $c stream" aichat -m "$c" 1 + 2 = ? echo "### $c non-stream" aichat -m "$c" -S 1 + 2 = ? done } # @cmd Test proxy server # @option -m --model[?`_choice_model`] # @flag -S --no-stream # @arg text~ test-server() { args=() if [[ -n "$argc_no_stream" ]]; then args+=("-S") fi argc chat-llm "${args[@]}" \ --api-base http://localhost:8000/v1 \ --model "${argc_model:-default}" \ "$@" } OPENAI_COMPATIBLE_PLATFORMS=( \ openai,gpt-3.5-turbo,https://api.openai.com/v1 \ anyscale,meta-llama/Meta-Llama-3-8B-Instruct,https://api.endpoints.anyscale.com/v1 \ deepinfra,meta-llama/Meta-Llama-3-8B-Instruct,https://api.deepinfra.com/v1/openai \ deepseek,deepseek-chat,https://api.deepseek.com \ fireworks,accounts/fireworks/models/llama-v3-8b-instruct,https://api.fireworks.ai/inference/v1 \ groq,llama3-8b-8192,https://api.groq.com/openai/v1 \ mistral,mistral-small-latest,https://api.mistral.ai/v1 \ moonshot,moonshot-v1-8k,https://api.moonshot.cn/v1 \ openrouter,meta-llama/llama-3-8b-instruct,https://openrouter.ai/api/v1 \ octoai,meta-llama-3-8b-instruct,https://text.octoai.run/v1 \ perplexity,llama-3-8b-instruct,https://api.perplexity.ai \ together,meta-llama/Llama-3-8b-chat-hf,https://api.together.xyz/v1 \ zhipuai,glm-4-0520,https://open.bigmodel.cn/api/paas/v4 \ lingyiwanwu,yi-large,https://api.lingyiwanwu.com/v1 \ ) # @cmd Chat with any LLM api # @flag -S --no-stream # @arg platform_model![?`_choice_platform_model`] # @arg text~ chat() { if [[ "$argc_platform_model" == *':'* ]]; then model="${argc_platform_model##*:}" argc_platform="${argc_platform_model%:*}" else argc_platform="${argc_platform_model}" fi for platform_config in "${OPENAI_COMPATIBLE_PLATFORMS[@]}"; do if [[ "$argc_platform" == "${platform_config%%,*}" ]]; then api_base="${platform_config##*,}" break fi done if [[ -n "$api_base" ]]; then env_prefix="$(echo "$argc_platform" | tr '[:lower:]' '[:upper:]')" api_key_env="${env_prefix}_API_KEY" api_key="${!api_key_env}" if [[ -z "$model" ]]; then model="$(echo "$platform_config" | cut -d, -f2)" fi if [[ -z "$model" ]]; then model_env="${env_prefix}_MODEL" model="${!model_env}" fi argc chat-openai-compatible \ --api-base "$api_base" \ --api-key "$api_key" \ --model "$model" \ "${argc_text[@]}" else argc chat-$argc_platform "${argc_text[@]}" fi } # @cmd List models by openai-compatible api # @arg platform![`_choice_platform`] models() { for platform_config in "${OPENAI_COMPATIBLE_PLATFORMS[@]}"; do if [[ "$argc_platform" == "${platform_config%%,*}" ]]; then api_base="${platform_config##*,}" break fi done if [[ -n "$api_base" ]]; then env_prefix="$(echo "$argc_platform" | tr '[:lower:]' '[:upper:]')" api_key_env="${env_prefix}_API_KEY" api_key="${!api_key_env}" _openai_models else argc models-$argc_platform fi } # @cmd Chat with openai-compatible api # @option --api-base! $$ # @option --api-key! $$ # @option -m --model! $$ # @flag -S --no-stream # @arg text~ chat-openai-compatible() { _openai_chat "$@" } # @cmd List models by openai-compatible api # @option --api-base! $$ # @option --api-key! $$ models-openai-compatible() { _openai_models } # @cmd Chat with gemini api # @env GEMINI_API_KEY! # @option -m --model=gemini-1.0-pro-latest $GEMINI_MODEL # @flag -S --no-stream # @arg text~ chat-gemini() { method="streamGenerateContent" if [[ -n "$argc_no_stream" ]]; then method="generateContent" fi _wrapper curl -i "https://generativelanguage.googleapis.com/v1beta/models/${argc_model}:${method}?key=${GEMINI_API_KEY}" \ -i -X POST \ -H 'Content-Type: application/json' \ -d "$(_build_body gemini "$@")" } # @cmd List gemini models # @env GEMINI_API_KEY! models-gemini() { _wrapper curl "https://generativelanguage.googleapis.com/v1beta/models?key=${GEMINI_API_KEY}" \ -H 'Content-Type: application/json' \ } # @cmd Chat with claude api # @env CLAUDE_API_KEY! # @option -m --model=claude-3-haiku-20240307 $CLAUDE_MODEL # @flag -S --no-stream # @arg text~ chat-claude() { _wrapper curl -i https://api.anthropic.com/v1/messages \ -X POST \ -H 'content-type: application/json' \ -H 'anthropic-version: 2023-06-01' \ -H 'anthropic-beta: tools-2024-05-16' \ -H "x-api-key: $CLAUDE_API_KEY" \ -d "$(_build_body claude "$@")" } # @cmd Chat with cohere api # @env COHERE_API_KEY! # @option -m --model=command-r $COHERE_MODEL # @flag -S --no-stream # @arg text~ chat-cohere() { _wrapper curl -i https://api.cohere.ai/v1/chat \ -X POST \ -H 'Content-Type: application/json' \ -H "Authorization: Bearer $COHERE_API_KEY" \ -d "$(_build_body cohere "$@")" } # @cmd List cohere models # @env COHERE_API_KEY! models-cohere() { _wrapper curl https://api.cohere.ai/v1/models \ -H "Authorization: Bearer $COHERE_API_KEY" \ } # @cmd Chat with reka api # @env REKA_API_KEY! # @option -m --model=reka-flash $REKA_MODEL # @flag -S --no-stream # @arg text~ chat-reka() { _wrapper curl -i https://api.reka.ai/v1/chat \ -X POST \ -H 'Content-Type: application/json' \ -H "X-Api-Key: $REKA_API_KEY" \ -d "$(_build_body reka "$@")" } # @cmd List reka models # @env REKA_API_KEY! models-reka() { _wrapper curl https://api.reka.ai/v1/models \ -H "X-Api-Key: $REKA_API_KEY" \ } # @cmd Chat with ollama api # @option -m --model=codegemma $OLLAMA_MODEL # @flag -S --no-stream # @arg text~ chat-ollama() { _wrapper curl -i http://localhost:11434/api/chat \ -X POST \ -H 'Content-Type: application/json' \ -d "$(_build_body ollama "$@")" } # @cmd Chat with vertexai api # @env require-tools gcloud # @env VERTEXAI_PROJECT_ID! # @env VERTEXAI_LOCATION! # @option -m --model=gemini-1.0-pro $VERTEXAI_GEMINI_MODEL # @flag -S --no-stream # @arg text~ chat-vertexai() { api_key="$(gcloud auth print-access-token)" func="streamGenerateContent" if [[ -n "$argc_no_stream" ]]; then func="generateContent" fi url=https://$VERTEXAI_LOCATION-aiplatform.googleapis.com/v1/projects/$VERTEXAI_PROJECT_ID/locations/$VERTEXAI_LOCATION/publishers/google/models/$argc_model:$func _wrapper curl -i $url \ -X POST \ -H "Authorization: Bearer $api_key" \ -H 'Content-Type: application/json' \ -d "$(_build_body vertexai "$@")" } # @cmd Chat with vertexai-claude api # @env require-tools gcloud # @env VERTEXAI_PROJECT_ID! # @env VERTEXAI_LOCATION! # @option -m --model=claude-3-haiku@20240307 $VERTEXAI_CLAUDE_MODEL # @flag -S --no-stream # @arg text~ chat-vertexai-claude() { api_key="$(gcloud auth print-access-token)" url=https://$VERTEXAI_LOCATION-aiplatform.googleapis.com/v1/projects/$VERTEXAI_PROJECT_ID/locations/$VERTEXAI_LOCATION/publishers/anthropic/models/$argc_model:streamRawPredict _wrapper curl -i $url \ -X POST \ -H "Authorization: Bearer $api_key" \ -H 'Content-Type: application/json' \ -d "$(_build_body vertexai-claude "$@")" } # @cmd Chat with bedrock api # @meta require-tools aws # @option -m --model=mistral.mistral-7b-instruct-v0:2 $BEDROCK_MODEL # @env AWS_REGION=us-east-1 chat-bedrock() { file="$(mktemp)" case "$argc_model" in mistral.* | meta.*) body='{"prompt":"'"$*"'"}' ;; anthropic.*) body="$(_build_body bedrock-claude "$@")" ;; *) _die "Invalid model: $argc_model" ;; esac _wrapper aws bedrock-runtime invoke-model \ --model-id $argc_model \ --region $AWS_REGION \ --body "$(echo "$body" | base64)" \ "$file" cat "$file" } # @cmd Chat with cloudflare api # @env CLOUDFLARE_API_KEY! # @option -m --model=@cf/meta/llama-3-8b-instruct $CLOUDFLARE_MODEL # @flag -S --no-stream # @arg text~ chat-cloudflare() { url="https://api.cloudflare.com/client/v4/accounts/$CLOUDFLARE_ACCOUNT_ID/ai/run/$argc_model" _wrapper curl -i "$url" \ -X POST \ -H "Authorization: Bearer $CLOUDFLARE_API_KEY" \ -d "$(_build_body cloudflare "$@")" } # @cmd Chat with replicate api # @env REPLICATE_API_KEY! # @option -m --model=meta/meta-llama-3-8b-instruct $REPLICATE_MODEL # @flag -S --no-stream # @arg text~ chat-replicate() { url="https://api.replicate.com/v1/models/$argc_model/predictions" res="$(_wrapper curl -s "$url" \ -X POST \ -H "Authorization: Bearer $REPLICATE_API_KEY" \ -H "Content-Type: application/json" \ -d "$(_build_body replicate "$@")" \ )" echo "$res" if [[ -n "$argc_no_stream" ]]; then prediction_url="$(echo "$res" | jq -r '.urls.get')" while true; do output="$(_wrapper curl -s -H "Authorization: Bearer $REPLICATE_API_KEY" "$prediction_url")" prediction_status=$(printf "%s" "$output" | jq -r .status) if [ "$prediction_status"=="succeeded" ]; then echo "$output" break fi if [ "$prediction_status"=="failed" ]; then exit 1 fi sleep 2 done else stream_url="$(echo "$res" | jq -r '.urls.stream')" _wrapper curl -i --no-buffer "$stream_url" \ -H "Accept: text/event-stream" \ fi } # @cmd Chat with ernie api # @meta require-tools jq # @env ERNIE_API_KEY! # @option -m --model=ernie-tiny-8k $ERNIE_MODEL # @flag -S --no-stream # @arg text~ chat-ernie() { auth_url="https://aip.baidubce.com/oauth/2.0/token?grant_type=client_credentials&client_id=$ERNIE_API_KEY&client_secret=$ERNIE_SECRET_KEY" ACCESS_TOKEN="$(curl -fsSL "$auth_url" | jq -r '.access_token')" url="https://aip.baidubce.com/rpc/2.0/ai_custom/v1/wenxinworkshop/chat/$argc_model?access_token=$ACCESS_TOKEN" _wrapper curl -i "$url" \ -X POST \ -d "$(_build_body ernie "$@")" } # @cmd Chat with qianwen api # @env QIANWEN_API_KEY! # @option -m --model=qwen-turbo $QIANWEN_MODEL # @flag -S --no-stream # @arg text~ chat-qianwen() { stream_args="-H X-DashScope-SSE:enable" parameters_args='{"incremental_output": true}' if [[ -n "$argc_no_stream" ]]; then stream_args="" parameters_args='{}' fi url=https://dashscope.aliyuncs.com/api/v1/services/aigc/text-generation/generation _wrapper curl -i "$url" \ -X POST \ -H "Authorization: Bearer $QIANWEN_API_KEY" \ -H 'Content-Type: application/json' $stream_args \ -d "$(_build_body qianwen "$@")" } _argc_before() { stream="true" if [[ -n "$argc_no_stream" ]]; then stream="false" fi } _openai_chat() { api_base="${api_base:-"$argc_api_base"}" api_key="${api_key:-"$argc_api_key"}" _wrapper curl -i $curl_args "$api_base/chat/completions" \ -X POST \ -H "Content-Type: application/json" \ -H "Authorization: Bearer $api_key" \ -d "$(_build_body openai "$@")" } _openai_models() { api_base="${api_base:-"$argc_api_base"}" api_key="${api_key:-"$argc_api_key"}" _wrapper curl $curl_args "$api_base/models" \ -H "Authorization: Bearer $api_key" \ } _choice_model() { aichat --list-models } _choice_platform_model() { _choice_platform _choice_model } _choice_platform() { _choice_client _choice_openai_compatible_platform } _choice_client() { printf "%s\n" openai gemini claude cohere ollama azure-openai vertexai bedrock cloudflare replicate ernie qianwen moonshot } _choice_openai_compatible_platform() { for platform_config in "${OPENAI_COMPATIBLE_PLATFORMS[@]}"; do echo "${platform_config%%,*}" done } _build_body() { kind="$1" if [[ "$#" -eq 1 ]]; then file="${BODY_FILE:-"tmp/body/$1.json"}" if [[ -f "$file" ]]; then cat "$file" | \ sed \ -e 's/"model": ".*"/"model": "'"$argc_model"'"/' \ -e 's/"stream": \(true\|false\)/"stream": '$stream'/' \ fi else shift case "$kind" in openai|ollama) echo '{ "model": "'$argc_model'", "messages": [ { "role": "user", "content": "'"$*"'" } ], "stream": '$stream' }' ;; cohere) echo '{ "model": "'$argc_model'", "message": "'"$*"'", "stream": '$stream' }' ;; reka) echo '{ "model": "'$argc_model'", "messages": [ { "role": "user", "content": "'"$*"'" } ], "stream": '$stream' }' ;; claude) echo '{ "model": "'$argc_model'", "messages": [ { "role": "user", "content": "'"$*"'" } ], "max_tokens": 4096, "stream": '$stream' }' ;; vertexai-claude|bedrock-claude) echo '{ "anthropic_version": "vertex-2023-10-16", "messages": [ { "role": "user", "content": "'"$*"'" } ], "max_tokens": 4096, "stream": '$stream' }' ;; gemini|vertexai) echo '{ "contents": [{ "role": "user", "parts": [ { "text": "'"$*"'" } ] }], "safetySettings":[{"category":"HARM_CATEGORY_HARASSMENT","threshold":"BLOCK_ONLY_HIGH"},{"category":"HARM_CATEGORY_HATE_SPEECH","threshold":"BLOCK_ONLY_HIGH"},{"category":"HARM_CATEGORY_SEXUALLY_EXPLICIT","threshold":"BLOCK_ONLY_HIGH"},{"category":"HARM_CATEGORY_DANGEROUS_CONTENT","threshold":"BLOCK_ONLY_HIGH"}] }' ;; ernie|cloudflare) echo '{ "messages": [ { "role": "user", "content": "'"$*"'" } ], "stream": '$stream' }' ;; replicate) echo '{ "stream": '$stream', "input": { "prompt": "'"$*"'" } }' ;; qianwen) echo '{ "model": "'$argc_model'", "parameters": '"$parameters_args"', "input":{ "messages": [ { "role": "user", "content": "'"$*"'" } ] } }' ;; *) _die "Unsupported build body for $kind" ;; esac fi } _wrapper() { if [[ "$DRY_RUN" == "true" ]] || [[ "$DRY_RUN" == "1" ]]; then echo "$@" else "$@" fi } _die() { echo $* exit 1 } # See more details at https://github.com/sigoden/argc eval "$(argc --argc-eval "$0" "$@")"