Skip to content

Commit

Permalink
Update repositories
Browse files Browse the repository at this point in the history
  • Loading branch information
sasha0552 authored Mar 25, 2024
1 parent c14b128 commit de3395b
Show file tree
Hide file tree
Showing 11 changed files with 11 additions and 53 deletions.
2 changes: 1 addition & 1 deletion airootfs/home/tori/automatic
2 changes: 1 addition & 1 deletion airootfs/home/tori/axolotl
2 changes: 1 addition & 1 deletion airootfs/home/tori/llama.cpp
2 changes: 1 addition & 1 deletion airootfs/home/tori/vllm
Submodule vllm updated 247 files
Original file line number Diff line number Diff line change
@@ -1,6 +1,6 @@
--- a/examples/server/server.cpp
+++ b/examples/server/server.cpp
@@ -94,7 +94,7 @@ struct server_task_multi {
@@ -95,7 +95,7 @@ struct server_task_multi {

struct slot_params {
bool stream = true;
Expand All @@ -9,7 +9,7 @@

uint32_t seed = -1; // RNG seed
int32_t n_keep = 0; // number of tokens to keep from initial prompt
@@ -825,7 +825,7 @@ struct server_context {
@@ -827,7 +827,7 @@ struct server_context {
}

slot.params.stream = json_value(data, "stream", false);
Expand All @@ -18,14 +18,3 @@
slot.params.n_predict = json_value(data, "n_predict", default_params.n_predict);
slot.sparams.top_k = json_value(data, "top_k", default_sparams.top_k);
slot.sparams.top_p = json_value(data, "top_p", default_sparams.top_p);
--- a/examples/server/utils.hpp
+++ b/examples/server/utils.hpp
@@ -353,7 +353,7 @@ static json oaicompat_completion_params_parse(
llama_sampling_params default_sparams;
llama_params["model"] = json_value(body, "model", std::string("unknown"));
llama_params["prompt"] = format_chat(model, chat_template, body["messages"]);
- llama_params["cache_prompt"] = json_value(body, "cache_prompt", false);
+ llama_params["cache_prompt"] = json_value(body, "cache_prompt", true);
llama_params["temperature"] = json_value(body, "temperature", 0.0);
llama_params["top_k"] = json_value(body, "top_k", default_sparams.top_k);
llama_params["top_p"] = json_value(body, "top_p", 1.0);

This file was deleted.

8 changes: 0 additions & 8 deletions airootfs/root/customize_airootfs/scripts/0100-vllm-patches.sh

This file was deleted.

Original file line number Diff line number Diff line change
Expand Up @@ -11,8 +11,8 @@ pushd "vllm"
export MAX_JOBS=1

# define supported architectures
export TORCH_CUDA_ARCH_LIST="6.0 6.1 7.0 7.5 8.0 8.6 8.9 9.0"

export TORCH_CUDA_ARCH_LIST="6.0 6.1 7.0 7.5 8.0 8.6 8.9 9.0+PTX"
""
# cuda home directory
export CUDA_HOME=/opt/cuda

Expand Down

0 comments on commit de3395b

Please sign in to comment.