Important
Collection
5 items • Updated
• 1
model_id stringlengths 16 38 | vram float64 0 4.97k | scripts listlengths 0 2 | code_urls listlengths 0 2 | execution_urls listlengths 0 2 |
|---|---|---|---|---|
PaddlePaddle/PaddleOCR-VL | 2.32 | [
"# /// script\n# requires-python = \">=3.12\"\n# dependencies = [\n# \"torch\",\n# \"torchvision\",\n# \"transformers\",\n# \"accelerate\",\n# \"peft\",\n# ]\n# ///\n\ntry:\n # See https://www.paddleocr.ai/latest/version3.x/pipeline_usage/PaddleOCR-VL.html to installation\n \n from padd... | [
"https://huggingface.co/datasets/model-metadata/custom_code_py_files/raw/main/PaddlePaddle_PaddleOCR-VL_0.py"
] | [
"https://huggingface.co/datasets/model-metadata/custom_code_execution_files/raw/main/PaddlePaddle_PaddleOCR-VL_0.txt"
] |
nvidia/omnivinci | 0 | [
"# /// script\n# requires-python = \">=3.12\"\n# dependencies = [\n# \"torch\",\n# \"torchvision\",\n# \"transformers\",\n# \"accelerate\",\n# \"peft\",\n# ]\n# ///\n\ntry:\n # Use a pipeline as a high-level helper\n from transformers import pipeline\n \n pipe = pipeline(\"feature-ex... | [
"https://huggingface.co/datasets/model-metadata/custom_code_py_files/raw/main/nvidia_omnivinci_0.py",
"https://huggingface.co/datasets/model-metadata/custom_code_py_files/raw/main/nvidia_omnivinci_1.py"
] | [
"https://huggingface.co/datasets/model-metadata/custom_code_execution_files/raw/main/nvidia_omnivinci_0.txt",
"https://huggingface.co/datasets/model-metadata/custom_code_execution_files/raw/main/nvidia_omnivinci_1.txt"
] |
deepseek-ai/DeepSeek-OCR | 8.08 | [
"# /// script\n# requires-python = \">=3.12\"\n# dependencies = [\n# \"torch\",\n# \"torchvision\",\n# \"transformers\",\n# \"accelerate\",\n# \"peft\",\n# ]\n# ///\n\ntry:\n # ⚠️ Type of model/library unknown.\n \n # Feel free to open a Pull request \n # for integration of the hug... | [
"DO NOT EXECUTE"
] | [
"WAS NOT EXECUTED"
] |
inclusionAI/LLaDA2.0-flash-preview | 249.14 | [
"# /// script\n# requires-python = \">=3.12\"\n# dependencies = [\n# \"torch\",\n# \"torchvision\",\n# \"transformers\",\n# \"accelerate\",\n# \"peft\",\n# ]\n# ///\n\ntry:\n # Use a pipeline as a high-level helper\n from transformers import pipeline\n \n pipe = pipeline(\"text-gener... | [
"https://huggingface.co/datasets/model-metadata/custom_code_py_files/raw/main/inclusionAI_LLaDA2.0-flash-preview_0.py",
"https://huggingface.co/datasets/model-metadata/custom_code_py_files/raw/main/inclusionAI_LLaDA2.0-flash-preview_1.py"
] | [
"https://huggingface.co/datasets/model-metadata/custom_code_execution_files/raw/main/inclusionAI_LLaDA2.0-flash-preview_0.txt",
"https://huggingface.co/datasets/model-metadata/custom_code_execution_files/raw/main/inclusionAI_LLaDA2.0-flash-preview_1.txt"
] |
facebook/MobileLLM-Pro | 2.63 | [
"# /// script\n# requires-python = \">=3.12\"\n# dependencies = [\n# \"torch\",\n# \"torchvision\",\n# \"transformers\",\n# \"accelerate\",\n# \"peft\",\n# ]\n# ///\n\ntry:\n # Use a pipeline as a high-level helper\n from transformers import pipeline\n \n pipe = pipeline(\"text-gener... | [
"https://huggingface.co/datasets/model-metadata/custom_code_py_files/raw/main/facebook_MobileLLM-Pro_0.py",
"https://huggingface.co/datasets/model-metadata/custom_code_py_files/raw/main/facebook_MobileLLM-Pro_1.py"
] | [
"https://huggingface.co/datasets/model-metadata/custom_code_execution_files/raw/main/facebook_MobileLLM-Pro_0.txt",
"https://huggingface.co/datasets/model-metadata/custom_code_execution_files/raw/main/facebook_MobileLLM-Pro_1.txt"
] |
nvidia/llama-embed-nemotron-8b | 18.17 | [] | [] | [] |
inclusionAI/LLaDA2.0-mini-preview | 39.36 | [
"# /// script\n# requires-python = \">=3.12\"\n# dependencies = [\n# \"torch\",\n# \"torchvision\",\n# \"transformers\",\n# \"accelerate\",\n# \"peft\",\n# ]\n# ///\n\ntry:\n # Use a pipeline as a high-level helper\n from transformers import pipeline\n \n pipe = pipeline(\"text-gener... | [
"https://huggingface.co/datasets/model-metadata/custom_code_py_files/raw/main/inclusionAI_LLaDA2.0-mini-preview_0.py",
"https://huggingface.co/datasets/model-metadata/custom_code_py_files/raw/main/inclusionAI_LLaDA2.0-mini-preview_1.py"
] | [
"https://huggingface.co/datasets/model-metadata/custom_code_execution_files/raw/main/inclusionAI_LLaDA2.0-mini-preview_0.txt",
"https://huggingface.co/datasets/model-metadata/custom_code_execution_files/raw/main/inclusionAI_LLaDA2.0-mini-preview_1.txt"
] |
inclusionAI/Ling-1T | 2,420.73 | [
"# /// script\n# requires-python = \">=3.12\"\n# dependencies = [\n# \"torch\",\n# \"torchvision\",\n# \"transformers\",\n# \"accelerate\",\n# \"peft\",\n# ]\n# ///\n\ntry:\n # Use a pipeline as a high-level helper\n from transformers import pipeline\n \n pipe = pipeline(\"text-gener... | [
"https://huggingface.co/datasets/model-metadata/custom_code_py_files/raw/main/inclusionAI_Ling-1T_0.py",
"https://huggingface.co/datasets/model-metadata/custom_code_py_files/raw/main/inclusionAI_Ling-1T_1.py"
] | [
"https://huggingface.co/datasets/model-metadata/custom_code_execution_files/raw/main/inclusionAI_Ling-1T_0.txt",
"https://huggingface.co/datasets/model-metadata/custom_code_execution_files/raw/main/inclusionAI_Ling-1T_1.txt"
] |
moonshotai/Kimi-K2-Instruct-0905 | 4,971.07 | [
"# /// script\n# requires-python = \">=3.12\"\n# dependencies = [\n# \"torch\",\n# \"torchvision\",\n# \"transformers\",\n# \"accelerate\",\n# \"peft\",\n# ]\n# ///\n\ntry:\n # Use a pipeline as a high-level helper\n from transformers import pipeline\n \n pipe = pipeline(\"text-gener... | [
"https://huggingface.co/datasets/model-metadata/custom_code_py_files/raw/main/moonshotai_Kimi-K2-Instruct-0905_0.py",
"https://huggingface.co/datasets/model-metadata/custom_code_py_files/raw/main/moonshotai_Kimi-K2-Instruct-0905_1.py"
] | [
"https://huggingface.co/datasets/model-metadata/custom_code_execution_files/raw/main/moonshotai_Kimi-K2-Instruct-0905_0.txt",
"https://huggingface.co/datasets/model-metadata/custom_code_execution_files/raw/main/moonshotai_Kimi-K2-Instruct-0905_1.txt"
] |
rednote-hilab/dots.ocr | 7.36 | [
"# /// script\n# requires-python = \">=3.12\"\n# dependencies = [\n# \"torch\",\n# \"torchvision\",\n# \"transformers\",\n# \"accelerate\",\n# \"peft\",\n# ]\n# ///\n\ntry:\n # integration status unknown.\n \n # Please clone model and use locally.\n \n # Also feel free to open a... | [
"https://huggingface.co/datasets/model-metadata/custom_code_py_files/raw/main/rednote-hilab_dots.ocr_0.py"
] | [
"https://huggingface.co/datasets/model-metadata/custom_code_execution_files/raw/main/rednote-hilab_dots.ocr_0.txt"
] |
inclusionAI/Ring-1T | 4,841.45 | [
"# /// script\n# requires-python = \">=3.12\"\n# dependencies = [\n# \"torch\",\n# \"torchvision\",\n# \"transformers\",\n# \"accelerate\",\n# \"peft\",\n# ]\n# ///\n\ntry:\n # Use a pipeline as a high-level helper\n from transformers import pipeline\n \n pipe = pipeline(\"text-gener... | [
"https://huggingface.co/datasets/model-metadata/custom_code_py_files/raw/main/inclusionAI_Ring-1T_0.py",
"https://huggingface.co/datasets/model-metadata/custom_code_py_files/raw/main/inclusionAI_Ring-1T_1.py"
] | [
"https://huggingface.co/datasets/model-metadata/custom_code_execution_files/raw/main/inclusionAI_Ring-1T_0.txt",
"https://huggingface.co/datasets/model-metadata/custom_code_execution_files/raw/main/inclusionAI_Ring-1T_1.txt"
] |
inclusionAI/Ring-flash-linear-2.0-128k | 504.54 | [
"# /// script\n# requires-python = \">=3.12\"\n# dependencies = [\n# \"torch\",\n# \"torchvision\",\n# \"transformers\",\n# \"accelerate\",\n# \"peft\",\n# ]\n# ///\n\ntry:\n # Use a pipeline as a high-level helper\n from transformers import pipeline\n \n pipe = pipeline(\"text-gener... | [
"https://huggingface.co/datasets/model-metadata/custom_code_py_files/raw/main/inclusionAI_Ring-flash-linear-2.0-128k_0.py",
"https://huggingface.co/datasets/model-metadata/custom_code_py_files/raw/main/inclusionAI_Ring-flash-linear-2.0-128k_1.py"
] | [
"https://huggingface.co/datasets/model-metadata/custom_code_execution_files/raw/main/inclusionAI_Ring-flash-linear-2.0-128k_0.txt",
"https://huggingface.co/datasets/model-metadata/custom_code_execution_files/raw/main/inclusionAI_Ring-flash-linear-2.0-128k_1.txt"
] |
deepseek-ai/DeepSeek-R1 | 1,657.55 | [
"# /// script\n# requires-python = \">=3.12\"\n# dependencies = [\n# \"torch\",\n# \"torchvision\",\n# \"transformers\",\n# \"accelerate\",\n# \"peft\",\n# ]\n# ///\n\ntry:\n # Use a pipeline as a high-level helper\n from transformers import pipeline\n \n pipe = pipeline(\"text-gener... | [
"https://huggingface.co/datasets/model-metadata/custom_code_py_files/raw/main/deepseek-ai_DeepSeek-R1_0.py",
"https://huggingface.co/datasets/model-metadata/custom_code_py_files/raw/main/deepseek-ai_DeepSeek-R1_1.py"
] | [
"https://huggingface.co/datasets/model-metadata/custom_code_execution_files/raw/main/deepseek-ai_DeepSeek-R1_0.txt",
"https://huggingface.co/datasets/model-metadata/custom_code_execution_files/raw/main/deepseek-ai_DeepSeek-R1_1.txt"
] |