diff --git a/05_scheduling/hackernews_alerts.py b/05_scheduling/hackernews_alerts.py index f6e7bdee6..70b7f7b58 100644 --- a/05_scheduling/hackernews_alerts.py +++ b/05_scheduling/hackernews_alerts.py @@ -1,6 +1,5 @@ # --- # lambda-test: false -# tags: ["featured"] # --- # # Run cron jobs in the cloud to search Hacker News diff --git a/06_gpu_and_ml/blender/blender_video.py b/06_gpu_and_ml/blender/blender_video.py index a7c41596e..3b8c353e0 100644 --- a/06_gpu_and_ml/blender/blender_video.py +++ b/06_gpu_and_ml/blender/blender_video.py @@ -1,6 +1,5 @@ # --- # output-directory: "/tmp/render" -# tags: ["use-case-image-video-3d"] # --- # # Render a video with Blender on many GPUs or CPUs in parallel # diff --git a/06_gpu_and_ml/controlnet/controlnet_gradio_demos.py b/06_gpu_and_ml/controlnet/controlnet_gradio_demos.py index cef2483c1..a627865d3 100644 --- a/06_gpu_and_ml/controlnet/controlnet_gradio_demos.py +++ b/06_gpu_and_ml/controlnet/controlnet_gradio_demos.py @@ -1,6 +1,5 @@ # --- # cmd: ["modal", "serve", "06_gpu_and_ml/controlnet/controlnet_gradio_demos.py"] -# tags: ["use-case-image-video-3d", "featured"] # --- # # # Play with the ControlNet demos diff --git a/06_gpu_and_ml/dreambooth/diffusers_lora_finetune.py b/06_gpu_and_ml/dreambooth/diffusers_lora_finetune.py index 6cf027775..471b7cb60 100644 --- a/06_gpu_and_ml/dreambooth/diffusers_lora_finetune.py +++ b/06_gpu_and_ml/dreambooth/diffusers_lora_finetune.py @@ -1,6 +1,5 @@ # --- # deploy: true -# tags: ["use-case-image-video-3d", "use-case-finetuning", "featured"] # --- # # Create a character LoRA for Flux with Hugging Face and Gradio diff --git a/06_gpu_and_ml/flan_t5/flan_t5_finetune.py b/06_gpu_and_ml/flan_t5/flan_t5_finetune.py index 264c08da0..805d8d5e1 100644 --- a/06_gpu_and_ml/flan_t5/flan_t5_finetune.py +++ b/06_gpu_and_ml/flan_t5/flan_t5_finetune.py @@ -1,8 +1,3 @@ -# --- -# runtimes: ["runc", "gvisor"] -# tags: ["use-case-finetuning", "use-case-lm-inference"] -# --- -# # # Finetuning Flan-T5 # # Example by [@anishpdalal](https://github.com/anishpdalal) diff --git a/06_gpu_and_ml/langchains/potus_speech_qanda.py b/06_gpu_and_ml/langchains/potus_speech_qanda.py index 469ecbb01..3e35e1e46 100644 --- a/06_gpu_and_ml/langchains/potus_speech_qanda.py +++ b/06_gpu_and_ml/langchains/potus_speech_qanda.py @@ -1,7 +1,6 @@ # --- # deploy: true # args: ["--query", "How many oil barrels were released from reserves?"] -# tags: ["featured"] # --- # # Retrieval-augmented generation (RAG) for question-answering with LangChain diff --git a/06_gpu_and_ml/llm-serving/llama_cpp.py b/06_gpu_and_ml/llm-serving/llama_cpp.py index d192021ee..37de36cf5 100644 --- a/06_gpu_and_ml/llm-serving/llama_cpp.py +++ b/06_gpu_and_ml/llm-serving/llama_cpp.py @@ -1,6 +1,3 @@ -# --- -# tags: ["use-case-lm-inference"] -# --- # # Run `llama.cpp` on Modal # [`llama.cpp`](https://github.com/ggerganov/llama.cpp) is a C++ library for running LLM inference. diff --git a/06_gpu_and_ml/llm-serving/sgl_vlm.py b/06_gpu_and_ml/llm-serving/sgl_vlm.py index cf686a73c..88fbe4b78 100644 --- a/06_gpu_and_ml/llm-serving/sgl_vlm.py +++ b/06_gpu_and_ml/llm-serving/sgl_vlm.py @@ -1,6 +1,3 @@ -# --- -# tags: ["use-case-lm-inference", "use-case-image-video-3d"] -# --- # # Run LLaVA-Next on SGLang for Visual QA # # Vision-Language Models (VLMs) are like LLMs with eyes: diff --git a/06_gpu_and_ml/llm-serving/trtllm_llama.py b/06_gpu_and_ml/llm-serving/trtllm_llama.py index 571902e0f..9bd622cc3 100644 --- a/06_gpu_and_ml/llm-serving/trtllm_llama.py +++ b/06_gpu_and_ml/llm-serving/trtllm_llama.py @@ -1,6 +1,5 @@ # --- # deploy: true -# tags: ["use-case-lm-inference"] # --- # # Serverless TensorRT-LLM (LLaMA 3 8B) diff --git a/06_gpu_and_ml/llm-serving/vllm_inference.py b/06_gpu_and_ml/llm-serving/vllm_inference.py index 884f7297f..b4672c269 100644 --- a/06_gpu_and_ml/llm-serving/vllm_inference.py +++ b/06_gpu_and_ml/llm-serving/vllm_inference.py @@ -1,7 +1,6 @@ # --- # cmd: ["modal", "serve", "06_gpu_and_ml/llm-serving/vllm_inference.py"] # pytest: false -# tags: ["use-case-lm-inference", "featured"] # --- # # Run OpenAI-compatible LLM inference with LLaMA 3.1-8B and vLLM diff --git a/06_gpu_and_ml/llm-structured/instructor_generate.py b/06_gpu_and_ml/llm-structured/instructor_generate.py index 19294385c..9dbb52678 100644 --- a/06_gpu_and_ml/llm-structured/instructor_generate.py +++ b/06_gpu_and_ml/llm-structured/instructor_generate.py @@ -1,6 +1,5 @@ # --- # output-directory: "/tmp/instructor_generate" -# tags: ["use-case-lm-inference"] # --- # # Structured Data Extraction using `instructor` diff --git a/06_gpu_and_ml/llm-structured/outlines_generate.py b/06_gpu_and_ml/llm-structured/outlines_generate.py index f7084d571..ff72ab7bd 100644 --- a/06_gpu_and_ml/llm-structured/outlines_generate.py +++ b/06_gpu_and_ml/llm-structured/outlines_generate.py @@ -1,6 +1,3 @@ -# --- -# tags: ["use-case-lm-inference"] -# --- # # Enforcing JSON outputs on LLMs # [Outlines](https://github.com/outlines-dev/outlines) is a tool that lets you control the generation of language models to make their output more predictable. diff --git a/06_gpu_and_ml/obj_detection_webcam/webcam.py b/06_gpu_and_ml/obj_detection_webcam/webcam.py index df7697d85..4e5a606bd 100644 --- a/06_gpu_and_ml/obj_detection_webcam/webcam.py +++ b/06_gpu_and_ml/obj_detection_webcam/webcam.py @@ -1,7 +1,6 @@ # --- # cmd: ["modal", "serve", "06_gpu_and_ml/obj_detection_webcam/webcam.py"] # deploy: true -# tags: ["use-case-image-video-3d", "featured"] # --- # # Real-time object detection via webcam # diff --git a/06_gpu_and_ml/stable_diffusion/flux.py b/06_gpu_and_ml/stable_diffusion/flux.py index 649584c0a..d065adc22 100644 --- a/06_gpu_and_ml/stable_diffusion/flux.py +++ b/06_gpu_and_ml/stable_diffusion/flux.py @@ -1,7 +1,6 @@ # --- # output-directory: "/tmp/flux" # args: ["--no-compile"] -# tags: ["use-case-image-video-3d", "featured"] # --- # # Run Flux fast on H100s with `torch.compile` diff --git a/06_gpu_and_ml/stable_diffusion/image_to_image.py b/06_gpu_and_ml/stable_diffusion/image_to_image.py index 399f63a3d..0c77af1d4 100644 --- a/06_gpu_and_ml/stable_diffusion/image_to_image.py +++ b/06_gpu_and_ml/stable_diffusion/image_to_image.py @@ -1,6 +1,5 @@ # --- # output-directory: "/tmp/stable-diffusion" -# tags: ["use-case-image-video-3d"] # --- # # Transform images with SDXL Turbo diff --git a/06_gpu_and_ml/stable_diffusion/text_to_image.py b/06_gpu_and_ml/stable_diffusion/text_to_image.py index ce384860f..ef0bb8be6 100644 --- a/06_gpu_and_ml/stable_diffusion/text_to_image.py +++ b/06_gpu_and_ml/stable_diffusion/text_to_image.py @@ -1,7 +1,6 @@ # --- # output-directory: "/tmp/stable-diffusion" # args: ["--prompt", "A 1600s oil painting of the New York City skyline"] -# tags: ["use-case-image-video-3d"] # --- # # Run Stable Diffusion 3.5 Large Turbo as a CLI, API, and web UI diff --git a/06_gpu_and_ml/tensorflow/tensorflow_tutorial.py b/06_gpu_and_ml/tensorflow/tensorflow_tutorial.py index dc909a5d4..99d7829d7 100644 --- a/06_gpu_and_ml/tensorflow/tensorflow_tutorial.py +++ b/06_gpu_and_ml/tensorflow/tensorflow_tutorial.py @@ -1,7 +1,5 @@ # --- # args: ["--just-run"] -# runtimes: ["runc", "gvisor"] -# tags: ["use-case-image-video-3d"] # --- # # TensorFlow tutorial # diff --git a/06_gpu_and_ml/yolo/finetune_yolo.py b/06_gpu_and_ml/yolo/finetune_yolo.py index 473d9d7bf..71915b431 100644 --- a/06_gpu_and_ml/yolo/finetune_yolo.py +++ b/06_gpu_and_ml/yolo/finetune_yolo.py @@ -1,6 +1,5 @@ # --- # args: ["--no-quick-check"] -# tags: ["use-case-image-video-3d", "use-case-finetuning"] # --- # # Fine-tune open source YOLO models for object detection # diff --git a/09_job_queues/doc_ocr_jobs.py b/09_job_queues/doc_ocr_jobs.py index bc1d18047..c6434e245 100644 --- a/09_job_queues/doc_ocr_jobs.py +++ b/09_job_queues/doc_ocr_jobs.py @@ -1,6 +1,5 @@ # --- # deploy: true -# tags: ["use-case-image-video-3d", "use-case-job-queues-batch", "featured"] # --- # # # Document OCR job queue diff --git a/10_integrations/cloud_bucket_mount_loras.py b/10_integrations/cloud_bucket_mount_loras.py index 36bf8af7a..64fc358ee 100644 --- a/10_integrations/cloud_bucket_mount_loras.py +++ b/10_integrations/cloud_bucket_mount_loras.py @@ -1,7 +1,6 @@ # --- # output-directory: "/tmp/stable-diffusion-xl" # deploy: true -# tags: ["use-case-image-video-3d"] # --- # # LoRAs Galore: Create a LoRA Playground with Modal, Gradio, and S3 # diff --git a/10_integrations/covid_datasette.py b/10_integrations/covid_datasette.py index cea70db08..ce3cc5fdf 100644 --- a/10_integrations/covid_datasette.py +++ b/10_integrations/covid_datasette.py @@ -1,7 +1,5 @@ # --- # deploy: true -# runtimes: ["runc", "gvisor"] -# tags: ["use-case-job-queues-batch"] # --- # # Publish interactive datasets with Datasette # diff --git a/10_integrations/dbt/dbt_duckdb.py b/10_integrations/dbt/dbt_duckdb.py index 2144b687b..486270e6b 100644 --- a/10_integrations/dbt/dbt_duckdb.py +++ b/10_integrations/dbt/dbt_duckdb.py @@ -1,6 +1,5 @@ # --- # deploy: true -# tags: ["use-case-job-queues-batch"] # --- # # Build your own data warehouse with DuckDB, DBT, and Modal # diff --git a/10_integrations/s3_bucket_mount.py b/10_integrations/s3_bucket_mount.py index e58d739db..bb1514209 100644 --- a/10_integrations/s3_bucket_mount.py +++ b/10_integrations/s3_bucket_mount.py @@ -1,6 +1,5 @@ # --- # output-directory: "/tmp/s3_bucket_mount" -# tags: ["use-case-job-queues-batch", "featured"] # --- # # Analyze NYC yellow taxi data with DuckDB on Parquet files from S3 # diff --git a/13_sandboxes/codelangchain/agent.py b/13_sandboxes/codelangchain/agent.py index ebdbc29da..b67c45203 100644 --- a/13_sandboxes/codelangchain/agent.py +++ b/13_sandboxes/codelangchain/agent.py @@ -1,6 +1,5 @@ # --- # cmd: ["modal", "run", "13_sandboxes.codelangchain.agent", "--question", "Use gpt2 and transformers to generate text"] -# tags: ["featured", "use-case-sandboxed-code-execution"] # pytest: false # env: {"MODAL_AUTOMOUNT": "True"} # --- diff --git a/13_sandboxes/jupyter_sandbox.py b/13_sandboxes/jupyter_sandbox.py index 1c95160c7..2170f7257 100644 --- a/13_sandboxes/jupyter_sandbox.py +++ b/13_sandboxes/jupyter_sandbox.py @@ -1,6 +1,5 @@ # --- # cmd: ["python", "13_sandboxes/jupyter_sandbox.py"] -# tags: ["use-case-sandboxed-code-execution"] # pytest: false # --- diff --git a/13_sandboxes/safe_code_execution.py b/13_sandboxes/safe_code_execution.py index 93ac727f7..ac09fb8fa 100644 --- a/13_sandboxes/safe_code_execution.py +++ b/13_sandboxes/safe_code_execution.py @@ -1,6 +1,5 @@ # --- # cmd: ["python", "13_sandboxes/safe_code_execution.py"] -# tags: ["use-case-sandboxed-code-execution"] # pytest: false # --- diff --git a/13_sandboxes/simple_code_interpreter.py b/13_sandboxes/simple_code_interpreter.py index 03ffe31d1..8f870012d 100644 --- a/13_sandboxes/simple_code_interpreter.py +++ b/13_sandboxes/simple_code_interpreter.py @@ -1,6 +1,5 @@ # --- # cmd: ["python", "13_sandboxes/simple_code_interpreter.py"] -# tags: ["use-case-sandboxed-code-execution"] # pytest: false # ---