diff --git a/06_gpu_and_ml/dreambooth/dreambooth_app.py b/06_gpu_and_ml/dreambooth/dreambooth_app.py index 23878c899..f52e248b7 100644 --- a/06_gpu_and_ml/dreambooth/dreambooth_app.py +++ b/06_gpu_and_ml/dreambooth/dreambooth_app.py @@ -1,6 +1,5 @@ # --- # deploy: true -# runtimes: ["runc"] # --- # # # Pet Art Dreambooth with Hugging Face and Gradio diff --git a/06_gpu_and_ml/falcon_bitsandbytes.py b/06_gpu_and_ml/falcon_bitsandbytes.py index 80cdbc538..504cf256c 100644 --- a/06_gpu_and_ml/falcon_bitsandbytes.py +++ b/06_gpu_and_ml/falcon_bitsandbytes.py @@ -1,6 +1,5 @@ # --- # args: ["--prompt", "How do planes work?"] -# runtimes: ["runc"] # --- # # Run Falcon-40B with bitsandbytes # diff --git a/06_gpu_and_ml/falcon_gptq.py b/06_gpu_and_ml/falcon_gptq.py index de118fb4c..6a405c4d4 100644 --- a/06_gpu_and_ml/falcon_gptq.py +++ b/06_gpu_and_ml/falcon_gptq.py @@ -1,6 +1,3 @@ -# --- -# runtimes: ["runc"] -# --- # # Run Falcon-40B with AutoGPTQ # # In this example, we run a quantized 4-bit version of Falcon-40B, the first open-source large language diff --git a/06_gpu_and_ml/openllama.py b/06_gpu_and_ml/openllama.py index f2a2353fd..d3d307808 100644 --- a/06_gpu_and_ml/openllama.py +++ b/06_gpu_and_ml/openllama.py @@ -1,6 +1,3 @@ -# --- -# runtimes: ["runc"] -# --- # # Run OpenLLaMa on an A100 GPU # # In this example, we run [OpenLLaMa](https://github.com/openlm-research/open_llama),