Spaces:
Sleeping
Sleeping
Commit History
Doc: Idea for speed improvements and intermediate results display, grouping future directions
968cab3
Documentation changes
21ce4d4
Add description to app
9f20b49
Luke Stanley
commited on
Add cached examples
acc8b42
Luke Stanley
commited on
Documentation: Add image
2dac454
Luke Stanley
commited on
Documentation: Update future directions
d72193c
Luke Stanley
commited on
Add HuggingFace Space demo link
03b6491
Luke Stanley
commited on
Reduce max_iterations value in chill.py
6bfaa63
Luke Stanley
commited on
Comment out llama-cpp-python installation command in Docker for HuggingFace Space
56e7667
Luke Stanley
commited on
Switch to serverless worker by default (PR #2 from lukestanley/serverless_json_llm)
a054519
unverified
Luke Stanley
commited on
Revert expected serverless output metadata stripper
c013599
Luke Stanley
commited on
Documents serverless motivation and testing instructions
5da2aef
Luke Stanley
commited on
Avoid unneeded imports, make serverless output more sensible, removing some debugging and comments
469f650
Luke Stanley
commited on
Fix RUNPOD_ENDPOINT_ID environment variable
ce5ad5f
Luke Stanley
commited on
Add more serverless GPU endpoint setup instruction detail
b51ce5c
Luke Stanley
commited on
Document serverless setup
f2e80c9
Luke Stanley
commited on
Rename serverless test file, set default model to Phi 2 for test, removed jq install, and env vars that are set the same in utils.py already, ignore .cache in git
83e4d57
Luke Stanley
commited on
Introduces worker mode env var
56e785c
Luke Stanley
commited on
Make GPU detection and llama-cpp-python re-installation conditional
434144a
Luke Stanley
commited on
Initialise global variables in improvement_loop function
e30b729
Luke Stanley
commited on
Ensure N_GPU_LAYERS is int
9475016
Luke Stanley
commited on
Expose json typed LLM interface for RunPod
976ea17
Luke Stanley
commited on
RunPod Mixtral JSON output test
233efeb
Luke Stanley
commited on
Add hello world RunPod setup
feeb679
Luke Stanley
commited on
Update default GPU layer, temperature values
e327a9e
Add env vars to set GPU layer count and context size, make verbose
e01e28e
Fix gif link since LFS related gif binary purge due to HF requirments
0945e5b
Add n_gpu_layers parameter to Llama initialization
88e6118
Fix: Move n_ctx parameter to model setup!
358cd20
Fix check for LLM_MODEL_PATH to avoid load error
ff938c3
Correct Space metadata
f5a3b9d
Add HuggingFace space metadata
994c606
Adds Gradio app wrapper and Dockerfile
c355718
Luke Stanley
commited on
Auto-downloads model if env var is not set
74d6e52
Luke Stanley
commited on
Make llm_stream_sans_network actually stream to stdout
a0f49a0
Luke Stanley
commited on
Default to in-memory LLM interface
ddb0d91
Luke Stanley
commited on