<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>nodepedia</title><link>https://nodepedia.com/tags/meta/</link><description>Compare GPU cloud pricing across providers. Daily-updated spot and on-demand prices for H100, A100, RTX 4090, and more. Free tools and guides.</description><language>en-us</language><lastBuildDate>Sun, 19 Apr 2026 08:36:51 +0000</lastBuildDate><atom:link href="https://nodepedia.com/tags/meta/index.xml" rel="self" type="application/rss+xml"/><item><title>CodeLlama-13b-hf</title><link>https://nodepedia.com/models/codellama-13b-hf/</link><pubDate>Sun, 19 Apr 2026 05:26:21 +0000</pubDate><guid>https://nodepedia.com/models/codellama-13b-hf/</guid><description>CodeLlama-13b-hf specs, VRAM requirements, and which GPUs can run it.</description></item><item><title>CodeLlama-13b-Instruct-hf</title><link>https://nodepedia.com/models/codellama-13b-instruct-hf/</link><pubDate>Sun, 19 Apr 2026 05:26:21 +0000</pubDate><guid>https://nodepedia.com/models/codellama-13b-instruct-hf/</guid><description>CodeLlama-13b-Instruct-hf specs, VRAM requirements, and which GPUs can run it.</description></item><item><title>CodeLlama-34b-hf</title><link>https://nodepedia.com/models/codellama-34b-hf/</link><pubDate>Sun, 19 Apr 2026 05:26:21 +0000</pubDate><guid>https://nodepedia.com/models/codellama-34b-hf/</guid><description>CodeLlama-34b-hf specs, VRAM requirements, and which GPUs can run it.</description></item><item><title>CodeLlama-34b-Instruct-hf</title><link>https://nodepedia.com/models/codellama-34b-instruct-hf/</link><pubDate>Sun, 19 Apr 2026 05:26:21 +0000</pubDate><guid>https://nodepedia.com/models/codellama-34b-instruct-hf/</guid><description>CodeLlama-34b-Instruct-hf specs, VRAM requirements, and which GPUs can run it.</description></item><item><title>CodeLlama-7b-Instruct-hf</title><link>https://nodepedia.com/models/codellama-7b-instruct-hf/</link><pubDate>Sun, 19 Apr 2026 05:26:21 +0000</pubDate><guid>https://nodepedia.com/models/codellama-7b-instruct-hf/</guid><description>CodeLlama-7b-Instruct-hf specs, VRAM requirements, and which GPUs can run it.</description></item><item><title>CodeLlama-7b-Python-hf</title><link>https://nodepedia.com/models/codellama-7b-python-hf/</link><pubDate>Sun, 19 Apr 2026 05:26:21 +0000</pubDate><guid>https://nodepedia.com/models/codellama-7b-python-hf/</guid><description>CodeLlama-7b-Python-hf specs, VRAM requirements, and which GPUs can run it.</description></item><item><title>Llama 3.1 70B</title><link>https://nodepedia.com/models/llama-3-1-70b/</link><pubDate>Sun, 19 Apr 2026 05:26:21 +0000</pubDate><guid>https://nodepedia.com/models/llama-3-1-70b/</guid><description>Llama 3.1 70B specs, VRAM requirements, and which GPUs can run it. The sweet spot for local reasoning.</description></item><item><title>Llama 3.1 8B</title><link>https://nodepedia.com/models/llama-3-1-8b/</link><pubDate>Sun, 19 Apr 2026 05:26:21 +0000</pubDate><guid>https://nodepedia.com/models/llama-3-1-8b/</guid><description>Llama 3.1 8B specs, VRAM requirements, and which GPUs can run it. The go-to small model for local inference.</description></item><item><title>Llama-2-7b-hf</title><link>https://nodepedia.com/models/llama-2-7b-hf/</link><pubDate>Sun, 19 Apr 2026 05:26:21 +0000</pubDate><guid>https://nodepedia.com/models/llama-2-7b-hf/</guid><description>Llama-2-7b-hf specs, VRAM requirements, and which GPUs can run it.</description></item><item><title>Llama-3.1-405B-FP8</title><link>https://nodepedia.com/models/llama-3-1-405b-fp8/</link><pubDate>Sun, 19 Apr 2026 05:26:21 +0000</pubDate><guid>https://nodepedia.com/models/llama-3-1-405b-fp8/</guid><description>Llama-3.1-405B-FP8 specs, VRAM requirements, and which GPUs can run it.</description></item><item><title>Llama-3.1-405B-Instruct</title><link>https://nodepedia.com/models/llama-3-1-405b-instruct/</link><pubDate>Sun, 19 Apr 2026 05:26:21 +0000</pubDate><guid>https://nodepedia.com/models/llama-3-1-405b-instruct/</guid><description>Llama-3.1-405B-Instruct specs, VRAM requirements, and which GPUs can run it.</description></item><item><title>Llama-3.1-405B-Instruct-FP8</title><link>https://nodepedia.com/models/llama-3-1-405b-instruct-fp8/</link><pubDate>Sun, 19 Apr 2026 05:26:21 +0000</pubDate><guid>https://nodepedia.com/models/llama-3-1-405b-instruct-fp8/</guid><description>Llama-3.1-405B-Instruct-FP8 specs, VRAM requirements, and which GPUs can run it.</description></item><item><title>Llama-3.1-70B-Instruct</title><link>https://nodepedia.com/models/llama-3-1-70b-instruct/</link><pubDate>Sun, 19 Apr 2026 05:26:21 +0000</pubDate><guid>https://nodepedia.com/models/llama-3-1-70b-instruct/</guid><description>Llama-3.1-70B-Instruct specs, VRAM requirements, and which GPUs can run it.</description></item><item><title>Llama-3.1-8B-Instruct</title><link>https://nodepedia.com/models/llama-3-1-8b-instruct/</link><pubDate>Sun, 19 Apr 2026 05:26:21 +0000</pubDate><guid>https://nodepedia.com/models/llama-3-1-8b-instruct/</guid><description>Llama-3.1-8B-Instruct specs, VRAM requirements, and which GPUs can run it.</description></item><item><title>Llama-3.2-1B</title><link>https://nodepedia.com/models/llama-3-2-1b/</link><pubDate>Sun, 19 Apr 2026 05:26:21 +0000</pubDate><guid>https://nodepedia.com/models/llama-3-2-1b/</guid><description>Llama-3.2-1B specs, VRAM requirements, and which GPUs can run it.</description></item><item><title>Llama-3.2-1B-Instruct</title><link>https://nodepedia.com/models/llama-3-2-1b-instruct/</link><pubDate>Sun, 19 Apr 2026 05:26:21 +0000</pubDate><guid>https://nodepedia.com/models/llama-3-2-1b-instruct/</guid><description>Llama-3.2-1B-Instruct specs, VRAM requirements, and which GPUs can run it.</description></item><item><title>Llama-3.2-3B</title><link>https://nodepedia.com/models/llama-3-2-3b/</link><pubDate>Sun, 19 Apr 2026 05:26:21 +0000</pubDate><guid>https://nodepedia.com/models/llama-3-2-3b/</guid><description>Llama-3.2-3B specs, VRAM requirements, and which GPUs can run it.</description></item><item><title>Llama-3.2-3B-Instruct</title><link>https://nodepedia.com/models/llama-3-2-3b-instruct/</link><pubDate>Sun, 19 Apr 2026 05:26:21 +0000</pubDate><guid>https://nodepedia.com/models/llama-3-2-3b-instruct/</guid><description>Llama-3.2-3B-Instruct specs, VRAM requirements, and which GPUs can run it.</description></item><item><title>Llama-3.3-70B-Instruct</title><link>https://nodepedia.com/models/llama-3-3-70b-instruct/</link><pubDate>Sun, 19 Apr 2026 05:26:21 +0000</pubDate><guid>https://nodepedia.com/models/llama-3-3-70b-instruct/</guid><description>Llama-3.3-70B-Instruct specs, VRAM requirements, and which GPUs can run it.</description></item><item><title>Llama-Guard-3-1B</title><link>https://nodepedia.com/models/llama-guard-3-1b/</link><pubDate>Sun, 19 Apr 2026 05:26:21 +0000</pubDate><guid>https://nodepedia.com/models/llama-guard-3-1b/</guid><description>Llama-Guard-3-1B specs, VRAM requirements, and which GPUs can run it.</description></item><item><title>Llama-Guard-3-8B</title><link>https://nodepedia.com/models/llama-guard-3-8b/</link><pubDate>Sun, 19 Apr 2026 05:26:21 +0000</pubDate><guid>https://nodepedia.com/models/llama-guard-3-8b/</guid><description>Llama-Guard-3-8B specs, VRAM requirements, and which GPUs can run it.</description></item><item><title>Llama-Guard-3-8B-INT8</title><link>https://nodepedia.com/models/llama-guard-3-8b-int8/</link><pubDate>Sun, 19 Apr 2026 05:26:21 +0000</pubDate><guid>https://nodepedia.com/models/llama-guard-3-8b-int8/</guid><description>Llama-Guard-3-8B-INT8 specs, VRAM requirements, and which GPUs can run it.</description></item><item><title>LlamaGuard-7b</title><link>https://nodepedia.com/models/llamaguard-7b/</link><pubDate>Sun, 19 Apr 2026 05:26:21 +0000</pubDate><guid>https://nodepedia.com/models/llamaguard-7b/</guid><description>LlamaGuard-7b specs, VRAM requirements, and which GPUs can run it.</description></item><item><title>Meta-Llama-3-70B</title><link>https://nodepedia.com/models/meta-llama-3-70b/</link><pubDate>Sun, 19 Apr 2026 05:26:21 +0000</pubDate><guid>https://nodepedia.com/models/meta-llama-3-70b/</guid><description>Meta-Llama-3-70B specs, VRAM requirements, and which GPUs can run it.</description></item><item><title>Meta-Llama-3-70B-Instruct</title><link>https://nodepedia.com/models/meta-llama-3-70b-instruct/</link><pubDate>Sun, 19 Apr 2026 05:26:21 +0000</pubDate><guid>https://nodepedia.com/models/meta-llama-3-70b-instruct/</guid><description>Meta-Llama-3-70B-Instruct specs, VRAM requirements, and which GPUs can run it.</description></item><item><title>Meta-Llama-3-8B</title><link>https://nodepedia.com/models/meta-llama-3-8b/</link><pubDate>Sun, 19 Apr 2026 05:26:21 +0000</pubDate><guid>https://nodepedia.com/models/meta-llama-3-8b/</guid><description>Meta-Llama-3-8B specs, VRAM requirements, and which GPUs can run it.</description></item><item><title>Meta-Llama-3-8B-Instruct</title><link>https://nodepedia.com/models/meta-llama-3-8b-instruct/</link><pubDate>Sun, 19 Apr 2026 05:26:21 +0000</pubDate><guid>https://nodepedia.com/models/meta-llama-3-8b-instruct/</guid><description>Meta-Llama-3-8B-Instruct specs, VRAM requirements, and which GPUs can run it.</description></item><item><title>Meta-Llama-Guard-2-8B</title><link>https://nodepedia.com/models/meta-llama-guard-2-8b/</link><pubDate>Sun, 19 Apr 2026 05:26:21 +0000</pubDate><guid>https://nodepedia.com/models/meta-llama-guard-2-8b/</guid><description>Meta-Llama-Guard-2-8B specs, VRAM requirements, and which GPUs can run it.</description></item></channel></rss>