#!/usr/bin/env bash set -e # @meta dotenv # @env DRY_RUN Dry run mode # @cmd Test first running # @env AICHAT_CONFIG_DIR=tmp/test-init-config # @arg args~ test-init-config() { unset OPENAI_API_KEY mkdir -p "$AICHAT_CONFIG_DIR" config_file="$AICHAT_CONFIG_DIR/config.yaml" if [[ -f "$config_file" ]]; then rm -f "$config_file" fi cargo run -- "$@" } # @cmd Test running with AICHAT_PLATFORM environment variable # @env AICHAT_PLATFORM! # @arg args~ test-platform-env() { cargo run -- "$@" } # @cmd Test clients # @arg clients+[`_choice_client`] test-clients() { for c in "${argc_clients[@]}"; do echo "### $c stream" aichat -m "$c" 1 + 2 = ? echo "### $c non-stream" aichat -m "$c" -S 1 + 2 = ? done } # @cmd Test proxy server # @option -m --model[`_choice_model`] # @flag -S --no-stream # @arg text~ test-server() { args=() if [[ -n "$argc_no_stream" ]]; then args+=("-S") fi argc chat-llm "${args[@]}" \ --api-base http://localhost:8000/v1 \ --model "${argc_model:-default}" \ "$@" } OPENAI_COMPATIBLE_PLATFORMS=( \ openai,gpt-3.5-turbo,https://api.openai.com/v1 \ anyscale,meta-llama/Meta-Llama-3-8B-Instruct,https://api.endpoints.anyscale.com/v1 \ deepinfra,meta-llama/Meta-Llama-3-8B-Instruct,https://api.deepinfra.com/v1/openai \ fireworks,accounts/fireworks/models/llama-v3-8b-instruct,https://api.fireworks.ai/inference/v1 \ groq,llama3-8b-8192,https://api.groq.com/openai/v1 \ mistral,mistral-small-latest,https://api.mistral.ai/v1 \ moonshot,moonshot-v1-8k,https://api.moonshot.cn/v1 \ openrouter,meta-llama/llama-3-8b-instruct,https://openrouter.ai/api/v1 \ octoai,meta-llama-3-8b-instruct,https://text.octoai.run/v1 \ perplexity,llama-3-8b-instruct,https://api.perplexity.ai \ together,meta-llama/Llama-3-8b-chat-hf,https://api.together.xyz/v1 \ ) # @cmd Chat with any LLM api # @flag -S --no-stream # @arg platform_model![?`_choice_platform_model`] # @arg text~ chat() { if [[ "$argc_platform_model" == *':'* ]]; then model="${argc_platform_model##*:}" argc_platform="${argc_platform_model%:*}" else argc_platform="${argc_platform_model}" fi for platform_config in "${OPENAI_COMPATIBLE_PLATFORMS[@]}"; do if [[ "$argc_platform" == "${platform_config%%,*}" ]]; then api_base="${platform_config##*,}" break fi done if [[ -n "$api_base" ]]; then env_prefix="$(echo "$argc_platform" | tr '[:lower:]' '[:upper:]')" api_key_env="${env_prefix}_API_KEY" api_key="${!api_key_env}" if [[ -z "$model" ]]; then model="$(echo "$platform_config" | cut -d, -f2)" fi if [[ -z "$model" ]]; then model_env="${env_prefix}_MODEL" model="${!model_env}" fi argc chat-openai-compatible \ --api-base "$api_base" \ --api-key "$api_key" \ --model "$model" \ "${argc_text[@]}" else argc chat-$argc_platform "${argc_text[@]}" fi } # @cmd List models by openai-compatible api # @arg platform![`_choice_platform`] models() { for platform_config in "${OPENAI_COMPATIBLE_PLATFORMS[@]}"; do if [[ "$argc_platform" == "${platform_config%%,*}" ]]; then api_base="${platform_config##*,}" break fi done if [[ -n "$api_base" ]]; then env_prefix="$(echo "$argc_platform" | tr '[:lower:]' '[:upper:]')" api_key_env="${env_prefix}_API_KEY" api_key="${!api_key_env}" _openai_models else argc models-$argc_platform fi } # @cmd Chat with openai-compatible api # @option --api-base! $$ # @option --api-key! $$ # @option -m --model! $$ # @flag -S --no-stream # @arg text~ chat-openai-compatible() { _openai_chat "$@" } # @cmd List models by openai-compatible api # @option --api-base! $$ # @option --api-key! $$ models-openai-compatible() { _openai_models } # @cmd Chat with gemini api # @env GEMINI_API_KEY! # @option -m --model=gemini-1.0-pro-latest $GEMINI_MODEL # @flag -S --no-stream # @arg text~ chat-gemini() { method="streamGenerateContent" if [[ -n "$argc_no_stream" ]]; then method="generateContent" fi _wrapper curl -i "https://generativelanguage.googleapis.com/v1beta/models/${argc_model}:${method}?key=${GEMINI_API_KEY}" \ -i -X POST \ -H 'Content-Type: application/json' \ -d '{ "safetySettings":[{"category":"HARM_CATEGORY_HARASSMENT","threshold":"BLOCK_ONLY_HIGH"},{"category":"HARM_CATEGORY_HATE_SPEECH","threshold":"BLOCK_ONLY_HIGH"},{"category":"HARM_CATEGORY_SEXUALLY_EXPLICIT","threshold":"BLOCK_ONLY_HIGH"},{"category":"HARM_CATEGORY_DANGEROUS_CONTENT","threshold":"BLOCK_ONLY_HIGH"}], "contents": '"$(_build_msg_gemini $*)"' }' } # @cmd List gemini models # @env GEMINI_API_KEY! models-gemini() { _wrapper curl "https://generativelanguage.googleapis.com/v1beta/models?key=${GEMINI_API_KEY}" \ -H 'Content-Type: application/json' \ } # @cmd Chat with claude api # @env CLAUDE_API_KEY! # @option -m --model=claude-3-haiku-20240307 $CLAUDE_MODEL # @flag -S --no-stream # @arg text~ chat-claude() { _wrapper curl -i https://api.anthropic.com/v1/messages \ -X POST \ -H 'content-type: application/json' \ -H 'anthropic-version: 2023-06-01' \ -H "x-api-key: $CLAUDE_API_KEY" \ -d '{ "model": "'$argc_model'", "messages": '"$(_build_msg $*)"', "max_tokens": 4096, "stream": '$stream' } ' } # @cmd Chat with cohere api # @env COHERE_API_KEY! # @option -m --model=command-r $COHERE_MODEL # @flag -S --no-stream # @arg text~ chat-cohere() { _wrapper curl -i https://api.cohere.ai/v1/chat \ -X POST \ -H 'Content-Type: application/json' \ -H "Authorization: Bearer $COHERE_API_KEY" \ --data '{ "model": "'$argc_model'", "message": "'"$*"'", "stream": '$stream' } ' } # @cmd List cohere models # @env COHERE_API_KEY! models-cohere() { _wrapper curl https://api.cohere.ai/v1/models \ -H "Authorization: Bearer $COHERE_API_KEY" \ } # @cmd Chat with ollama api # @option -m --model=codegemma $OLLAMA_MODEL # @flag -S --no-stream # @arg text~ chat-ollama() { _wrapper curl -i http://localhost:11434/api/chat \ -X POST \ -H 'Content-Type: application/json' \ -d '{ "model": "'$argc_model'", "stream": '$stream', "messages": '"$(_build_msg $*)"' }' } # @cmd Chat with vertexai-gemini api # @env require-tools gcloud # @env VERTEXAI_PROJECT_ID! # @env VERTEXAI_LOCATION! # @option -m --model=gemini-1.0-pro $VERTEXAI_GEMINI_MODEL # @flag -S --no-stream # @arg text~ chat-vertexai-gemini() { api_key="$(gcloud auth print-access-token)" func="streamGenerateContent" if [[ -n "$argc_no_stream" ]]; then func="generateContent" fi url=https://$VERTEXAI_LOCATION-aiplatform.googleapis.com/v1/projects/$VERTEXAI_PROJECT_ID/locations/$VERTEXAI_LOCATION/publishers/google/models/$argc_model:$func _wrapper curl -i $url \ -X POST \ -H "Authorization: Bearer $api_key" \ -H 'Content-Type: application/json' \ -d '{ "contents": '"$(_build_msg_gemini $*)"', "generationConfig": {} }' } # @cmd Chat with vertexai-claude api # @env require-tools gcloud # @env VERTEXAI_PROJECT_ID! # @env VERTEXAI_LOCATION! # @option -m --model=claude-3-haiku@20240307 $VERTEXAI_CLAUDE_MODEL # @flag -S --no-stream # @arg text~ chat-vertexai-claude() { api_key="$(gcloud auth print-access-token)" url=https://$VERTEXAI_LOCATION-aiplatform.googleapis.com/v1/projects/$VERTEXAI_PROJECT_ID/locations/$VERTEXAI_LOCATION/publishers/anthropic/models/$argc_model:streamRawPredict _wrapper curl -i $url \ -X POST \ -H "Authorization: Bearer $api_key" \ -H 'Content-Type: application/json' \ -d '{ "anthropic_version": "vertex-2023-10-16", "messages": '"$(_build_msg $*)"', "max_tokens": 4096, "stream": '$stream' }' } # @cmd Chat with bedrock api # @meta require-tools aws # @option -m --model=mistral.mistral-7b-instruct-v0:2 $BEDROCK_MODEL # @env AWS_REGION=us-east-1 chat-bedrock() { file="$(mktemp)" case "$argc_model" in mistral.* | meta.*) body='{"prompt":"'"$*"'"}' ;; anthropic.*) body='{ "anthropic_version": "vertex-2023-10-16", "messages": '"$(_build_msg $*)"', "max_tokens": 4096 }' ;; *) _die "Invalid model: $argc_model" ;; esac _wrapper aws bedrock-runtime invoke-model \ --model-id $argc_model \ --region $AWS_REGION \ --body "$(echo "$body" | base64)" \ "$file" cat "$file" } # @cmd Chat with cloudflare api # @env CLOUDFLARE_API_KEY! # @option -m --model=@cf/meta/llama-3-8b-instruct $CLOUDFLARE_MODEL # @flag -S --no-stream # @arg text~ chat-cloudflare() { url="https://api.cloudflare.com/client/v4/accounts/$CLOUDFLARE_ACCOUNT_ID/ai/run/$argc_model" _wrapper curl -i "$url" \ -X POST \ -H "Authorization: Bearer $CLOUDFLARE_API_KEY" \ -d '{ "messages": '"$(_build_msg $*)"', "stream": '$stream' }' } # @cmd Chat with replicate api # @env REPLICATE_API_KEY! # @option -m --model=meta/meta-llama-3-8b-instruct $REPLICATE_MODEL # @flag -S --no-stream # @arg text~ chat-replicate() { url="https://api.replicate.com/v1/models/$argc_model/predictions" res="$(_wrapper curl -s "$url" \ -X POST \ -H "Authorization: Bearer $REPLICATE_API_KEY" \ -H "Content-Type: application/json" \ -d '{ "stream": '$stream', "input": { "prompt": "'"$*"'" } }')" echo "$res" if [[ -n "$argc_no_stream" ]]; then prediction_url="$(echo "$res" | jq -r '.urls.get')" while true; do output="$(_wrapper curl -s -H "Authorization: Bearer $REPLICATE_API_KEY" "$prediction_url")" prediction_status=$(printf "%s" "$output" | jq -r .status) if [ "$prediction_status"=="succeeded" ]; then echo "$output" break fi if [ "$prediction_status"=="failed" ]; then exit 1 fi sleep 2 done else stream_url="$(echo "$res" | jq -r '.urls.stream')" _wrapper curl -i --no-buffer "$stream_url" \ -H "Accept: text/event-stream" \ fi } # @cmd Chat with ernie api # @meta require-tools jq # @env ERNIE_API_KEY! # @option -m --model=ernie-tiny-8k $ERNIE_MODEL # @flag -S --no-stream # @arg text~ chat-ernie() { auth_url="https://aip.baidubce.com/oauth/2.0/token?grant_type=client_credentials&client_id=$ERNIE_API_KEY&client_secret=$ERNIE_SECRET_KEY" ACCESS_TOKEN="$(curl -fsSL "$auth_url" | jq -r '.access_token')" url="https://aip.baidubce.com/rpc/2.0/ai_custom/v1/wenxinworkshop/chat/$argc_model?access_token=$ACCESS_TOKEN" _wrapper curl -i "$url" \ -X POST \ -d '{ "messages": '"$(_build_msg $*)"', "stream": '$stream' }' } # @cmd Chat with qianwen api # @env QIANWEN_API_KEY! # @option -m --model=qwen-turbo $QIANWEN_MODEL # @flag -S --no-stream # @arg text~ chat-qianwen() { stream_args="-H X-DashScope-SSE:enable" parameters_args='{"incremental_output": true}' if [[ -n "$argc_no_stream" ]]; then stream_args="" parameters_args='{}' fi url=https://dashscope.aliyuncs.com/api/v1/services/aigc/text-generation/generation _wrapper curl -i "$url" \ -X POST \ -H "Authorization: Bearer $QIANWEN_API_KEY" \ -H 'Content-Type: application/json' $stream_args \ -d '{ "model": "'$argc_model'", "parameters": '"$parameters_args"', "input":{ "messages": '"$(_build_msg $*)"' } }' } _argc_before() { stream="true" if [[ -n "$argc_no_stream" ]]; then stream="false" fi } _openai_chat() { api_base="${api_base:-"$argc_api_base"}" api_key="${api_key:-"$argc_api_key"}" _wrapper curl -i $curl_args "$api_base/chat/completions" \ -X POST \ -H "Content-Type: application/json" \ -H "Authorization: Bearer $api_key" \ --data '{ "model": "'$argc_model'", "messages": '"$(_build_msg $*)"', "stream": '$stream' } ' } _openai_models() { api_base="${api_base:-"$argc_api_base"}" api_key="${api_key:-"$argc_api_key"}" _wrapper curl $curl_args "$api_base/models" \ -H "Authorization: Bearer $api_key" \ } _choice_model() { aichat --list-models } _choice_platform_model() { _choice_platform _choice_model } _choice_platform() { _choice_client _choice_openai_compatible_platform } _choice_client() { printf "%s\n" openai gemini claude cohere ollama azure-openai vertexai bedrock cloudflare replicate ernie qianwen moonshot } _choice_openai_compatible_platform() { for platform_config in "${OPENAI_COMPATIBLE_PLATFORMS[@]}"; do echo "${platform_config%%,*}" done } _build_msg() { if [[ $# -eq 0 ]]; then cat tmp/messages.json else echo ' [ { "role": "user", "content": "'"$*"'" } ] ' fi } _build_msg_gemini() { if [[ $# -eq 0 ]]; then cat tmp/messages.gemini.json else echo ' [{ "role": "user", "parts": [ { "text": "'"$*"'" } ] }] ' fi } _wrapper() { if [[ "$DRY_RUN" == "true" ]] || [[ "$DRY_RUN" == "1" ]]; then echo "$@" else "$@" fi } _die() { echo $* exit 1 } # See more details at https://github.com/sigoden/argc eval "$(argc --argc-eval "$0" "$@")"