|
@@ -23,11 +23,12 @@ def parse_global_args():
|
|
|
parser.add_argument('--max_gpu_memory', type=json.loads, help="Max gpu memory allocated for the LLM")
|
|
|
parser.add_argument('--eval_device', type=str, help="Evaluation device")
|
|
|
parser.add_argument('--max_new_tokens', type=int, default=256, help="The maximum number of new tokens for generation")
|
|
|
- parser.add_argument("--log_mode", type=str, default="console", choices=["console", "file"])
|
|
|
- parser.add_argument("--use_backend", type=str, choices=["ollama", "vllm"])
|
|
|
-
|
|
|
+ # parser.add_argument("--log_mode", type=str, default="console", choices=["console", "file"])
|
|
|
+ parser.add_argument("--scheduler_log_mode", type=str, default="console", choices=["console", "file"], help="Log mode for the scheduler")
|
|
|
+ parser.add_argument("--agent_log_mode", type=str, default="console", choices=["console", "file"], help="Log mode for the agents")
|
|
|
+ parser.add_argument("--llm_kernel_log_mode", type=str, default="console", choices=["console", "file"], help="Log mode for the LLM kernel")
|
|
|
+ parser.add_argument("--use_backend", type=str, choices=["ollama", "vllm"], help="Backend to use for running the LLM kernel")
|
|
|
return parser
|
|
|
-
|
|
|
def extract_before_parenthesis(s: str) -> str:
|
|
|
match = re.search(r'^(.*?)\([^)]*\)', s)
|
|
|
return match.group(1) if match else s
|