Hugging Face
Models
Datasets
Spaces
Posts
Docs
Enterprise
Pricing
Log In
Sign Up
2
randy lynn
9x25dillon
Follow
0 followers
ยท
2 following
AI & ML interests
SQL-BASED integration
Recent Activity
updated
a model
about 1 hour ago
9x25dillon/DS-R1-Distill-Qwen-32BSQL-INT
new
activity
about 2 hours ago
9x25dillon/DS-R1-Distill-Qwen-32BSQL-INT:
Rename README.md to { "script_id": 1, "parameter_id": 1 }# Install vLLM from pip: pip install vllm Copy # Load and run the model: vllm serve "deepseek-ai/DeepSeek-R1-Distill-Qwen-32B" Copy # Call the server using curl: curl -X POST "http://localhost:8000/v1/chat/completions" \ -H "Content-Type: application/json" \ --data '{ "model": "deepseek-ai/DeepSeek-R1-Distill-Qwen-32B", "messages": [ { "role": "user", "content": "What is the capital of France?" } ] }' Use Docker images Copy # Deploy with docker on Linux: docker run --runtime nvidia --gpus all \ --name my_vllm_container \ -v ~/.cache/huggingface:/root/.cache/huggingface \ --env "HUGGING_FACE_HUB_TOKEN=<secret>" \ -p 8000:8000 \ --ipc=host \ vllm/vllm-openai:latest \ --model deepseek-ai/DeepSeek-R1-Distill-Qwen-32B Copy # Load and run the model: docker exec -it my_vllm_container bash -c "vllm serve deepseek-ai/DeepSeek-R1-Distill-Qwen-32B" Copy # Call the server using curl: curl -X POST "http://localhost:8000/v1/chat/completions" \ -H "Content-Type: application/json" \ --data '{ "model": "deepseek-ai/DeepSeek-R1-Distill-Qwen-32B", "messages": [ { "role": "user", "content": "What is the capital of France?" } ] }' Quick Links Read the vLLM documentation
published
a model
about 3 hours ago
9x25dillon/DS-R1-Distill-Qwen-32BSQL-INT
View all activity
Organizations
None yet
9x25dillon
's activity
All
Models
Datasets
Spaces
Papers
Collections
Community
Posts
Upvotes
Likes
updated
a model
about 1 hour ago
9x25dillon/DS-R1-Distill-Qwen-32BSQL-INT
Updated
about 1 hour ago
New activity in
9x25dillon/DS-R1-Distill-Qwen-32BSQL-INT
about 2 hours ago
Rename README.md to { "script_id": 1, "parameter_id": 1 }# Install vLLM from pip: pip install vllm Copy # Load and run the model: vllm serve "deepseek-ai/DeepSeek-R1-Distill-Qwen-32B" Copy # Call the server using curl: curl -X POST "http://localhost:8000/v1/chat/completions" \ -H "Content-Type: application/json" \ --data '{ "model": "deepseek-ai/DeepSeek-R1-Distill-Qwen-32B", "messages": [ { "role": "user", "content": "What is the capital of France?" } ] }' Use Docker images Copy # Deploy with docker on Linux: docker run --runtime nvidia --gpus all \ --name my_vllm_container \ -v ~/.cache/huggingface:/root/.cache/huggingface \ --env "HUGGING_FACE_HUB_TOKEN=<secret>" \ -p 8000:8000 \ --ipc=host \ vllm/vllm-openai:latest \ --model deepseek-ai/DeepSeek-R1-Distill-Qwen-32B Copy # Load and run the model: docker exec -it my_vllm_container bash -c "vllm serve deepseek-ai/DeepSeek-R1-Distill-Qwen-32B" Copy # Call the server using curl: curl -X POST "http://localhost:8000/v1/chat/completions" \ -H "Content-Type: application/json" \ --data '{ "model": "deepseek-ai/DeepSeek-R1-Distill-Qwen-32B", "messages": [ { "role": "user", "content": "What is the capital of France?" } ] }' Quick Links Read the vLLM documentation
#1 opened about 2 hours ago by
9x25dillon
published
a model
about 3 hours ago
9x25dillon/DS-R1-Distill-Qwen-32BSQL-INT
Updated
about 1 hour ago
New activity in
deepseek-ai/DeepSeek-R1-Distill-Qwen-32B
about 9 hours ago
sharing something maybe beneficial ?
#13 opened about 12 hours ago by
9x25dillon
sharing something maybe beneficial ?
#13 opened about 12 hours ago by
9x25dillon
New activity in
deepseek-ai/DeepSeek-R1-Distill-Qwen-32B
about 12 hours ago
sharing something maybe beneficial ?
#13 opened about 12 hours ago by
9x25dillon