{"id":4059,"date":"2025-07-10T15:40:23","date_gmt":"2025-07-10T15:40:23","guid":{"rendered":"https:\/\/salarydistribution.com\/machine-learning\/2025\/07\/10\/how-to-run-coding-assistants-for-free-on-rtx-ai-pcs-and-workstations\/"},"modified":"2025-07-10T15:40:23","modified_gmt":"2025-07-10T15:40:23","slug":"how-to-run-coding-assistants-for-free-on-rtx-ai-pcs-and-workstations","status":"publish","type":"post","link":"https:\/\/salarydistribution.com\/machine-learning\/2025\/07\/10\/how-to-run-coding-assistants-for-free-on-rtx-ai-pcs-and-workstations\/","title":{"rendered":"How to Run Coding Assistants for Free on RTX AI PCs and Workstations"},"content":{"rendered":"<div>\n\t\t<span class=\"bsf-rt-reading-time\"><span class=\"bsf-rt-display-label\"><\/span> <span class=\"bsf-rt-display-time\"><\/span> <span class=\"bsf-rt-display-postfix\"><\/span><\/span><\/p>\n<p>Coding assistants or copilots \u2014 AI-powered assistants that can suggest, explain and debug code \u2014 are fundamentally changing how software is developed for both experienced and novice developers.<\/p>\n<p>Experienced developers use these assistants to stay focused on complex coding tasks, reduce repetitive work and explore new ideas more quickly. Newer coders \u2014 like students and AI hobbyists \u2014 benefit from coding assistants that accelerate learning by describing different implementation approaches or explaining what a piece of code is doing and why.<\/p>\n<p>Coding assistants can run in cloud environments or locally. Cloud-based coding assistants can be run anywhere but offer some limitations and require a subscription. Local coding assistants remove these issues but require performant hardware to operate well.<\/p>\n<p><a target=\"_blank\" href=\"https:\/\/www.nvidia.com\/en-us\/geforce\/laptops\/50-series\/\" rel=\"noopener\">NVIDIA GeForce RTX GPUs<\/a> provide the necessary hardware acceleration to run local assistants effectively.<\/p>\n<h2><b>Code, Meet Generative AI<\/b><\/h2>\n<p>Traditional software development includes many mundane tasks such as reviewing documentation, researching examples, setting up boilerplate code, authoring code with appropriate syntax, tracing down bugs and documenting functions. These are essential tasks that can take time away from problem solving and software design. Coding assistants help streamline such steps.<\/p>\n<p>Many AI assistants are linked with popular integrated development environments (IDEs) like Microsoft Visual Studio Code or JetBrains\u2019 Pycharm, which embed AI support directly into existing workflows.<\/p>\n<p>There are two ways to run coding assistants: in the cloud or locally.<\/p>\n<p>Cloud-based coding assistants require source code to be sent to external servers before responses are returned. This approach can be laggy and impose usage limits. Some developers prefer to keep their code local, especially when working with sensitive or proprietary projects. Plus, many cloud-based assistants require a paid subscription to unlock full functionality, which can be a barrier for students, hobbyists and teams that need to manage costs.<\/p>\n<p>Coding assistants run in a local environment, enabling cost-free access with:<\/p>\n<figure id=\"attachment_83025\" aria-describedby=\"caption-attachment-83025\" class=\"wp-caption alignnone\"><img decoding=\"async\" loading=\"lazy\" class=\"wp-image-83025 size-medium\" src=\"https:\/\/blogs.nvidia.com\/wp-content\/uploads\/2025\/07\/raig-07-09-infographic-960x384.jpg\" alt=\"\" width=\"960\" height=\"384\"><figcaption id=\"caption-attachment-83025\" class=\"wp-caption-text\">Coding assistants running locally on RTX offer numerous advantages.<\/figcaption><\/figure>\n<h2><b>Get Started With Local Coding Assistants<\/b><\/h2>\n<p>Tools that make it easy to run coding assistants locally include:<\/p>\n<ul>\n<li><a target=\"_blank\" href=\"https:\/\/docs.continue.dev\/getting-started\/install\" rel=\"noopener\"><b>Continue.dev<\/b><\/a> \u2014 An open-source extension for the VS Code IDE that connects to local large language models (LLMs) via Ollama, LM Studio or custom endpoints. This tool offers in-editor chat, autocomplete and debugging assistance with minimal setup. <a target=\"_blank\" href=\"https:\/\/docs.continue.dev\/customize\/model-providers\/ollama\" rel=\"noopener\">Get started with Continue.dev<\/a> using the Ollama backend for local RTX acceleration.<\/li>\n<li><a target=\"_blank\" href=\"https:\/\/www.tabbyml.com\/\" rel=\"noopener\"><b>Tabby<\/b><\/a> \u2014 A secure and transparent coding assistant that\u2019s compatible across many IDEs with the ability to run AI on NVIDIA RTX GPUs. This tool offers code completion, answering queries, inline chat and more. <a target=\"_blank\" href=\"https:\/\/tabby.tabbyml.com\/docs\/quick-start\/installation\/windows\/\" rel=\"noopener\">Get started with Tabby<\/a> on NVIDIA RTX AI PCs.<\/li>\n<li><a target=\"_blank\" href=\"https:\/\/github.com\/OpenInterpreter\/open-interpreter?tab=readme-ov-file\" rel=\"noopener\"><b>OpenInterpreter<\/b><\/a> \u2014 Experimental but rapidly evolving interface that combines LLMs with command-line access, file editing and agentic task execution. Ideal for automation and devops-style tasks for developers. <a target=\"_blank\" href=\"https:\/\/docs.openinterpreter.com\/guides\/running-locally\" rel=\"noopener\">Get started with OpenInterpreter<\/a> on NVIDIA RTX AI PCs.<\/li>\n<li><a target=\"_blank\" href=\"https:\/\/lmstudio.ai\/\" rel=\"noopener\"><b>LM Studio<\/b><\/a> \u2014 A graphical user interface-based runner for local LLMs that offers chat, context window management and system prompts. Optimal for testing coding models interactively before IDE deployment. <a href=\"https:\/\/blogs.nvidia.com\/blog\/rtx-ai-garage-lmstudio-llamacpp-blackwell\/\">Get started with LM Studio<\/a> on NVIDIA RTX AI PCs.<\/li>\n<li><a target=\"_blank\" href=\"https:\/\/github.com\/ollama\/ollama?tab=readme-ov-file\" rel=\"noopener\"><b>Ollama<\/b><\/a> \u2014 A local AI model inferencing engine that enables fast, private inference of models like Code Llama, StarCoder2 and DeepSeek. It integrates seamlessly with tools like Continue.dev.<\/li>\n<\/ul>\n<p>These tools support models served through frameworks like Ollama or llama.cpp, and many are now optimized for GeForce RTX and <a target=\"_blank\" href=\"http:\/\/www.nvidia.com\/rtx-pro\" rel=\"noopener\">NVIDIA RTX PRO GPUs<\/a>.<\/p>\n<h2><b>See AI-Assisted Learning on RTX in Action<\/b><\/h2>\n<p>Running on a GeForce RTX-powered PC, <a target=\"_blank\" href=\"https:\/\/docs.continue.dev\/getting-started\/install\" rel=\"noopener\">Continue.dev<\/a> paired with the Gemma 12B Code LLM helps explain existing code, explore search algorithms and debug issues \u2014 all entirely on device. Acting like a virtual teaching assistant, the assistant provides plain-language guidance, context-aware explanations, inline comments and suggested code improvements tailored to the user\u2019s project.<\/p>\n<p><a href=\"https:\/\/blogs.nvidia.com\/wp-content\/uploads\/2025\/07\/Demo-of-Continue.dev-with-Gemma-12B-Code-LLM.gif\"><img decoding=\"async\" loading=\"lazy\" class=\"aligncenter size-full wp-image-83002\" src=\"https:\/\/blogs.nvidia.com\/wp-content\/uploads\/2025\/07\/Demo-of-Continue.dev-with-Gemma-12B-Code-LLM.gif\" alt=\"\" width=\"800\" height=\"544\"><\/a><\/p>\n<p>This workflow highlights the advantage of local acceleration: the assistant is always available, responds instantly and provides personalized support, all while keeping the code private on device and making the learning experience immersive.<\/p>\n<p><a href=\"https:\/\/blogs.nvidia.com\/wp-content\/uploads\/2025\/07\/Continue.dev-running-Gemma-12B-Code-LLM.gif\"><img decoding=\"async\" loading=\"lazy\" class=\"aligncenter size-full wp-image-83005\" src=\"https:\/\/blogs.nvidia.com\/wp-content\/uploads\/2025\/07\/Continue.dev-running-Gemma-12B-Code-LLM.gif\" alt=\"\" width=\"800\" height=\"544\"><\/a><\/p>\n<p>That level of responsiveness comes down to GPU acceleration. Models like Gemma 12B are compute-heavy, especially when they\u2019re processing long prompts or working across multiple files. Running them locally without a GPU can feel sluggish \u2014 even for simple tasks. With RTX GPUs, Tensor Cores accelerate inference directly on the device, so the assistant is fast, responsive and able to keep up with an active development workflow.<\/p>\n<figure id=\"attachment_83008\" aria-describedby=\"caption-attachment-83008\" class=\"wp-caption aligncenter\"><a href=\"https:\/\/blogs.nvidia.com\/wp-content\/uploads\/2025\/07\/Coding-assistant-performance.png\"><img decoding=\"async\" loading=\"lazy\" class=\"size-full wp-image-83008\" src=\"https:\/\/blogs.nvidia.com\/wp-content\/uploads\/2025\/07\/Coding-assistant-performance.png\" alt=\"\" width=\"1070\" height=\"654\"><\/a><figcaption id=\"caption-attachment-83008\" class=\"wp-caption-text\">Coding assistants running on the Meta Llama 3.1-8B model experience 5-6x faster throughput on RTX-powered laptops versus on CPU. Data measured uses the average tokens per second at BS = 1, ISL\/OSL = 2000\/100, with the Llama-3.1-8B model quantized to int4.<\/figcaption><\/figure>\n<p>Whether used for academic work, coding bootcamps or personal projects, RTX AI PCs are enabling developers to build, learn and iterate faster with AI-powered tools.<\/p>\n<p>For those just getting started \u2014 especially students building their skills or experimenting with <a target=\"_blank\" href=\"https:\/\/www.nvidia.com\/en-us\/glossary\/generative-ai\/\" rel=\"noopener\">generative AI<\/a> \u2014 NVIDIA GeForce RTX 50 Series laptops feature specialized AI technologies that accelerate top applications for learning, creating and gaming, all on a single system. Explore RTX laptops <a target=\"_blank\" href=\"https:\/\/www.nvidia.com\/en-us\/geforce\/campaigns\/back-to-school\/\" rel=\"noopener\">ideal for back-to-school season<\/a>.<\/p>\n<p>And to encourage AI enthusiasts and developers to experiment with local AI and extend the capabilities of their RTX PCs, NVIDIA is hosting a <a target=\"_blank\" href=\"https:\/\/developer.nvidia.com\/g-assist-plug-in-hackathon\" rel=\"noopener\">Plug and Play: Project G-Assist Plug-In Hackathon<\/a> \u2014 running virtually through Wednesday, July 16. Participants can create custom plug-ins for <a target=\"_blank\" href=\"https:\/\/www.nvidia.com\/en-us\/software\/nvidia-app\/g-assist\/\" rel=\"noopener\">Project G-Assist<\/a>, an experimental AI assistant designed to respond to natural language and extend across creative and development tools. It\u2019s a chance to win prizes and showcase what\u2019s possible with RTX AI PCs.<\/p>\n<p>Join NVIDIA\u2019s <a target=\"_blank\" href=\"https:\/\/discord.gg\/taH4gkMt\" rel=\"noopener\">Discord server<\/a> to connect with community developers and AI enthusiasts for discussions on what\u2019s possible with RTX AI.<\/p>\n<p><i>Each week, the <\/i><a href=\"https:\/\/blogs.nvidia.com\/blog\/tag\/rtx-ai-garage\/\"><i>RTX AI Garage<\/i><\/a> <i>blog series features community-driven AI innovations and content for those looking to learn more about NVIDIA NIM microservices and AI Blueprints, as well as building <\/i><a target=\"_blank\" href=\"https:\/\/www.nvidia.com\/en-us\/glossary\/ai-agents\/\" rel=\"noopener\"><i>AI agents<\/i><\/a><i>, creative workflows, digital humans, productivity apps and more on AI PCs and workstations.\u00a0<\/i><\/p>\n<p><i>Plug in to NVIDIA AI PC on <\/i><a target=\"_blank\" href=\"https:\/\/www.facebook.com\/NVIDIA.AI.PC\/\" rel=\"noopener\"><i>Facebook<\/i><\/a><i>, <\/i><a target=\"_blank\" href=\"https:\/\/www.instagram.com\/nvidia.ai.pc\/\" rel=\"noopener\"><i>Instagram<\/i><\/a><i>, <\/i><a target=\"_blank\" href=\"https:\/\/www.tiktok.com\/@nvidia_ai_pc\" rel=\"noopener\"><i>TikTok<\/i><\/a><i> and <\/i><a target=\"_blank\" href=\"https:\/\/x.com\/NVIDIA_AI_PC\" rel=\"noopener\"><i>X<\/i><\/a><i> \u2014 and stay informed by subscribing to the <\/i><a target=\"_blank\" href=\"https:\/\/www.nvidia.com\/en-us\/ai-on-rtx\/?modal=subscribe-ai\" rel=\"noopener\"><i>RTX AI PC newsletter<\/i><\/a><i>.<\/i><\/p>\n<p><i>Follow NVIDIA Workstation on <\/i><a target=\"_blank\" href=\"https:\/\/www.linkedin.com\/showcase\/3761136\/\" rel=\"noopener\"><i>LinkedIn<\/i><\/a><i> and <\/i><a target=\"_blank\" href=\"https:\/\/x.com\/NVIDIAworkstatn\" rel=\"noopener\"><i>X<\/i><\/a><i>.\u00a0<\/i><\/p>\n<p><i>See <\/i><a target=\"_blank\" href=\"https:\/\/www.nvidia.com\/en-eu\/about-nvidia\/terms-of-service\/\" rel=\"noopener\"><i>notice<\/i><\/a><i> regarding software product information.<\/i><\/p>\n<\/p><\/div>\n","protected":false},"excerpt":{"rendered":"<p>https:\/\/blogs.nvidia.com\/blog\/rtx-ai-garage-coding-assistants\/<\/p>\n","protected":false},"author":0,"featured_media":4060,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":[],"categories":[3],"tags":[],"_links":{"self":[{"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/posts\/4059"}],"collection":[{"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/types\/post"}],"replies":[{"embeddable":true,"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/comments?post=4059"}],"version-history":[{"count":0,"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/posts\/4059\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/media\/4060"}],"wp:attachment":[{"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/media?parent=4059"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/categories?post=4059"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/tags?post=4059"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}