Authorization#
SkyServe provides robust authorization capabilities at the replica level, allowing you to control access to service endpoints with API keys.
Setup API Keys#
SkyServe relies on the authorization of the service running on underlying service replicas, e.g., the inference engine. We take the vLLM inference engine as an example, which supports static API key authorization with an argument --api-key
.
We define a SkyServe service spec for serving Llama-3 chatbot with vLLM and an API key. In the example YAML below, we define the authorization token as an environment variable, AUTH_TOKEN
, and pass it to both the service field to enable readiness_probe
to access the replicas and the vllm entrypoint to start services on replicas with the API key.
# auth.yaml
envs:
MODEL_NAME: meta-llama/Meta-Llama-3-8B-Instruct
HF_TOKEN: # TODO: Fill with your own huggingface token, or use --env to pass.
AUTH_TOKEN: # TODO: Fill with your own auth token (a random string), or use --env to pass.
service:
readiness_probe:
path: /v1/models
headers:
Authorization: Bearer $AUTH_TOKEN
replicas: 2
resources:
accelerators: {L4, A10g, A10, L40, A40, A100, A100-80GB}
ports: 8000
setup: |
pip install vllm==0.4.0.post1 flash-attn==2.5.7 gradio openai
# python -c "import huggingface_hub; huggingface_hub.login('${HF_TOKEN}')"
run: |
export PATH=$PATH:/sbin
python -m vllm.entrypoints.openai.api_server \
--model $MODEL_NAME --trust-remote-code \
--gpu-memory-utilization 0.95 \
--host 0.0.0.0 --port 8000 \
--api-key $AUTH_TOKEN
To deploy the service, run the following command:
HF_TOKEN=xxx AUTH_TOKEN=yyy sky serve up auth.yaml -n auth --env HF_TOKEN --env AUTH_TOKEN
To send a request to the service endpoint, a service client need to include the static API key in a request’s header:
$ ENDPOINT=$(sky serve status --endpoint auth)
$ AUTH_TOKEN=yyy
$ curl http://$ENDPOINT/v1/chat/completions \
-H "Content-Type: application/json" \
-H "Authorization: Bearer $AUTH_TOKEN" \
-d '{
"model": "meta-llama/Meta-Llama-3-8B-Instruct",
"messages": [
{
"role": "system",
"content": "You are a helpful assistant."
},
{
"role": "user",
"content": "Who are you?"
}
],
"stop_token_ids": [128009, 128001]
}' | jq
Example output
{
"id": "cmpl-cad2c1a2a6ee44feabed0b28be294d6f",
"object": "chat.completion",
"created": 1716819147,
"model": "meta-llama/Meta-Llama-3-8B-Instruct",
"choices": [
{
"index": 0,
"message": {
"role": "assistant",
"content": "I'm so glad you asked! I'm LLaMA, an AI assistant developed by Meta AI that can understand and respond to human input in a conversational manner. I'm here to help you with any questions, tasks, or topics you'd like to discuss. I can provide information on a wide range of subjects, from science and history to entertainment and culture. I can also assist with language-related tasks such as language translation, text summarization, and even writing and proofreading. My goal is to provide accurate and helpful responses to your inquiries, while also being friendly and engaging. So, what's on your mind? How can I assist you today?"
},
"logprobs": null,
"finish_reason": "stop",
"stop_reason": 128009
}
],
"usage": {
"prompt_tokens": 26,
"total_tokens": 160,
"completion_tokens": 134
}
}
A service client without an API key will not be able to access the service and get a 401 Unauthorized
error:
$ curl http://$ENDPOINT/v1/models
{"error": "Unauthorized"}
$ curl http://$ENDPOINT/v1/models -H "Authorization: Bearer random-string"
{"error": "Unauthorized"}