[{"data":1,"prerenderedAt":746},["ShallowReactive",2],{"navigation":3,"\u002Fblog\u002Fusing-ollama-and-continue-as-github-copilot-alternative":204,"\u002Fblog\u002Fusing-ollama-and-continue-as-github-copilot-alternative-surround":741},[4,8,12,16,20,24,28,32,36,40,44,48,52,56,60,64,68,72,76,80,84,88,92,96,100,104,108,112,116,120,124,128,132,136,140,144,148,152,156,160,164,168,172,176,180,184,188,192,196,200],{"title":5,"path":6,"stem":7},"You do not have time to not have tests","\u002Fblog\u002Fyou-do-not-have-time-to-not-have-tests","2.blog\u002F20211217.you-do-not-have-time-to-not-have-tests",{"title":9,"path":10,"stem":11},"Migrate Vue 2 with Vuetify and Jest to Vite and Vitest","\u002Fblog\u002Fmigrate-vue-2-with-vuetify-and-jest-to-vite-and-vitest","2.blog\u002F20220109.migrate-vue-2-with-vuetify-and-jest-to-vite-and-vitest",{"title":13,"path":14,"stem":15},"I am a Dark Matter Developer","\u002Fblog\u002Fi-am-a-dark-matter-developer","2.blog\u002F20220626.i-am-a-dark-matter-developer",{"title":17,"path":18,"stem":19},"Why using Conventional commits is useful","\u002Fblog\u002Fusing-conventional-commits","2.blog\u002F20240623.using-conventional-commits",{"title":21,"path":22,"stem":23},"Why you should make a toolbox repository","\u002Fblog\u002Fwhy-you-should-make-a-toolbox-repository","2.blog\u002F20240630.Why-you-should-make-a-toolbox-repository",{"title":25,"path":26,"stem":27},"Apache Airflow Part 1 - Why and Goals for a near Serverless ELT","\u002Fblog\u002Fapache-airflow-part-1-why-and-goals","2.blog\u002F20240710.apache-airflow-part-1-why-and-goals",{"title":29,"path":30,"stem":31},"Oh My Zsh on your server","\u002Fblog\u002Foh-my-zsh-on-your-server","2.blog\u002F20240711.oh-my-zsh-on-your-server",{"title":33,"path":34,"stem":35},"Fire tablet and YouTube Kids","\u002Fblog\u002Ffire-tablet-and-youtube-kids","2.blog\u002F20240714.fire-tablet-and-youtube-kids",{"title":37,"path":38,"stem":39},"Using Ollama and Continue as a GitHub Copilot Alternative","\u002Fblog\u002Fusing-ollama-and-continue-as-github-copilot-alternative","2.blog\u002F20240723.using-ollama-and-continue-as-github-copilot-alternative",{"title":41,"path":42,"stem":43},"Debugging Local Packages Made Easy with pnpm","\u002Fblog\u002Fdebugging-local-packages-with-pnpm-link","2.blog\u002F20250422.debugging local-packages-with-pnpm-link",{"title":45,"path":46,"stem":47},"Two Weeks with Cloudflare AI and Tools","\u002Fblog\u002Ftwo-weeks-with-cloudflare-ai-and-tools","2.blog\u002F20250509.two-weeks-with-cloudflare-aI-and-tools",{"title":49,"path":50,"stem":51},"Adding Prompts to VS Code - How I Learned to Stop Worrying and Love AI Context","\u002Fblog\u002Fadding-prompts-to-vscode","2.blog\u002F20250528.adding-prompts-to-vscode",{"title":53,"path":54,"stem":55},"My Best Practices","\u002Fblog\u002Fmy-best-practicies","2.blog\u002F20250607.my-best-practicies",{"title":57,"path":58,"stem":59},"Creating my own CLI Tool - Towles Tool","\u002Fblog\u002Ftowles-tool","2.blog\u002F20250607.towles-tool",{"title":61,"path":62,"stem":63},"Software Development Best Practices & ITIL","\u002Fblog\u002Fsoftware-engineering-and-itil-best-practices","2.blog\u002F20250612.software-engineering-and-itil-best-practices",{"title":65,"path":66,"stem":67},"Voice to Text","\u002Fblog\u002Fvoice-to-text","2.blog\u002F20250622.voice-to-text",{"title":69,"path":70,"stem":71},"Setting Up ComfyUI - A Better Alternative to Fooocus","\u002Fblog\u002Fcomfy-ui-setup","2.blog\u002F20250628.comfy-ui-setup",{"title":73,"path":74,"stem":75},"Voice to System","\u002Fblog\u002Fvoice-to-system","2.blog\u002F20250705.voice-to-system",{"title":77,"path":78,"stem":79},"Tips for Claude Code","\u002Fblog\u002Ftips-for-claude-code","2.blog\u002F20250713.tips-for-claude-code",{"title":81,"path":82,"stem":83},"Review That AI Code: Why I Read Every Line Generated Code","\u002Fblog\u002Freview-that-ai-code","2.blog\u002F20250720.review-that-ai-code",{"title":85,"path":86,"stem":87},"My Context Engineering Journey: From Dev Scripts to AI Collaboration","\u002Fblog\u002F20250803-1.my-context-engineering-journey","2.blog\u002F20250803-1.my-context-engineering-journey",{"title":89,"path":90,"stem":91},"Context Engineering at Scale: Enterprise Lessons and the Future of Development","\u002Fblog\u002F20250803-2.context-engineering-at-scale","2.blog\u002F20250803-2.context-engineering-at-scale",{"title":93,"path":94,"stem":95},"Check That Your Tools and Linters Do Not Burn Tokens","\u002Fblog\u002Fcheck-that-your-tools-and-linters-do-not-burn-tokens","2.blog\u002F20250806.check-that-your-tools-and-linters-do-not-burn-tokens",{"title":97,"path":98,"stem":99},"Markdown + AI: The Communication Protocol That Changes Everything","\u002Fblog\u002Fmarkdown-plus-ai-the-communication-protocol-that-changes-everything","2.blog\u002F20250814.markdown-plus-ai-the-communication-protocol-that-changes-everything",{"title":101,"path":102,"stem":103},"Finally: Type-Safe AI in Production (And Why I'm Here For It)","\u002Fblog\u002Ffinally-type-safe-ai-in-production-and-why-im-here-for-it","2.blog\u002F20250819.finally-type-safe-ai-in-production-and-why-im-here-for-it",{"title":105,"path":106,"stem":107},"Dotfiles: Masterpiece or Late Stage Picasso?","\u002Fblog\u002Fdotfiles-masterpiece-or-late-stage-picasso","2.blog\u002F20250822.dotfiles-masterpiece-or-late-stage-picasso",{"title":109,"path":110,"stem":111},"Beyond API Wrappers: Building State-Driven MCP Servers for Long-Horizon Agent Orchestration","\u002Fblog\u002Fbeyond-api-wrappers-mcp-servers","2.blog\u002F20250907.beyond-api-wrappers-mcp-servers",{"title":113,"path":114,"stem":115},"Why Vertical Integration Wins: A Software Engineer's Case for Owning Your Stack","\u002Fblog\u002Fwhy-i-bought-tesla-model-3-vertical-integration","2.blog\u002F20250928.why-i-bought-tesla-model-3-vertical-integration",{"title":117,"path":118,"stem":119},"The Min-Maxer's Trifecta: Building Tools for the Game You Actually Play","\u002Fblog\u002Fmin-maxer-trifecta","2.blog\u002F20251004.min-maxer-trifecta",{"title":121,"path":122,"stem":123},"Read The Source: Learning by Cutting Out The Middleman and RTFM","\u002Fblog\u002Fread-the-source","2.blog\u002F20251010.read-the-source",{"title":125,"path":126,"stem":127},"The Exponential Shift: Why AI Progress Feels Different Now","\u002Fblog\u002Fthe-exponential-shift","2.blog\u002F20251015.the-exponential-shift",{"title":129,"path":130,"stem":131},"Plan Mode for Your Problems, Edit Mode for Claude's","\u002Fblog\u002Fplan-mode-problems-edit-mode-solutions","2.blog\u002F20251019.plan-mode-problems-edit-mode-solutions",{"title":133,"path":134,"stem":135},"AWS Aurora DSQL Looked Perfect Until I Needed the Connection String","\u002Fblog\u002Faws-aurora-dsql-postgres-serverless-authentication","2.blog\u002F20251028.aws-aurora-dsql-postgres-serverless-authentication",{"title":137,"path":138,"stem":139},"Switchback: Browser History for Your Thoughts","\u002Fblog\u002Fswitchback-second-order-reasoning","2.blog\u002F20251205.switchback-second-order-reasoning",{"title":141,"path":142,"stem":143},"AI Pairing: Notes to Self","\u002Fblog\u002Fai-pairing-notes-to-self","2.blog\u002F20251216.ai-pairing-notes-to-self",{"title":145,"path":146,"stem":147},"I've Been Sleeping on Zellij","\u002Fblog\u002Fsleeping-on-zellij","2.blog\u002F20251229.sleeping-on-zellij",{"title":149,"path":150,"stem":151},"Implementing a Ralph Wiggum Loop: The Secret is Session Markers","\u002Fblog\u002Fimplementing-ralph-wiggum-loop-for-autonomous-ai-coding","2.blog\u002F20260114.implementing-ralph-wiggum-loop-for-autonomous-ai-coding",{"title":153,"path":154,"stem":155},"Goodhart's Law Ate My Context Window","\u002Fblog\u002Fgoodharts-law-ate-my-context-window","2.blog\u002F20260119.goodharts-law-ate-my-context-window",{"title":157,"path":158,"stem":159},"Claude Code's Hidden Multi-Agent System Is Real","\u002Fblog\u002Fclaude-code-hidden-multi-agent-system","2.blog\u002F20260124.claude-code-hidden-multi-agent-system",{"title":161,"path":162,"stem":163},"Free Printable Math Sheets for Kids — Number Chart, Skip Counting, Multiplication, and More","\u002Fblog\u002Ffree-printable-number-chart-and-coin-sheets","2.blog\u002F20260214.free-printable-number-chart-and-coin-sheets",{"title":165,"path":166,"stem":167},"We Are Near the End of the Exponential","\u002Fblog\u002Fnear-the-end-of-the-exponential","2.blog\u002F20260214.near-the-end-of-the-exponential",{"title":169,"path":170,"stem":171},"Free Printable Language Arts Sheets for Kids — Sight Words, Parts of Speech, Homophones, and More","\u002Fblog\u002Ffree-printable-sight-words-and-grammar-sheets","2.blog\u002F20260215.free-printable-sight-words-and-grammar-sheets",{"title":173,"path":174,"stem":175},"Interactive Code Execution with Artifacts","\u002Fblog\u002Finteractive-code-execution-with-artifacts","2.blog\u002F20260215.interactive-code-execution-with-artifacts",{"title":177,"path":178,"stem":179},"Free Printable Telling Time Worksheet for Kids — Clock Reference & Practice Sheet","\u002Fblog\u002Ffree-printable-telling-time-worksheet","2.blog\u002F20260216.free-printable-telling-time-worksheet",{"title":181,"path":182,"stem":183},"Claude Code Skills: Teaching AI Your Playbook","\u002Fblog\u002Fclaude-code-skills-guide","2.blog\u002F20260221.claude-code-skills-guide",{"title":185,"path":186,"stem":187},"Building a Multi-Agent Loan Approval System with Human-in-the-Loop","\u002Fblog\u002Fmulti-agent-loan-approval-human-in-the-loop","2.blog\u002F20260225.multi-agent-loan-approval-human-in-the-loop",{"title":189,"path":190,"stem":191},"The Inception of AI Infrastructure: Bottlenecks All the Way Down","\u002Fblog\u002Fbiggest-bottleneck-scaling-ai-compute","2.blog\u002F20260313.biggest-bottleneck-scaling-ai-compute",{"title":193,"path":194,"stem":195},"What I Tell Teams About Claude Code","\u002Fblog\u002Fwhat-i-tell-teams-about-claude-code","2.blog\u002F20260314.what-i-tell-teams-about-claude-code",{"title":197,"path":198,"stem":199},"The Hardest Part of AI Isn't the AI","\u002Fblog\u002Fthe-hardest-part-of-ai-isnt-the-ai","2.blog\u002F20260327.the-hardest-part-of-ai-isnt-the-ai",{"title":201,"path":202,"stem":203},"Claude Code Hooks: The Capability I Left on the Table","\u002Fblog\u002Fclaude-code-hooks-capability-left-on-the-table","2.blog\u002F20260401.claude-code-hooks-capability-left-on-the-table",{"id":205,"title":37,"authors":206,"badge":212,"body":214,"date":731,"description":732,"extension":733,"image":734,"meta":737,"navigation":427,"path":38,"seo":738,"status":739,"stem":39,"__hash__":740},"posts\u002F2.blog\u002F20240723.using-ollama-and-continue-as-github-copilot-alternative.md",[207],{"name":208,"to":209,"avatar":210},"Chris Towles","https:\u002F\u002Ftwitter.com\u002FChris_Towles",{"src":211},"\u002Fimages\u002Fctowles-profile-512x512.png",{"label":213},"AI",{"type":215,"value":216,"toc":723},"minimark",[217,221,224,241,244,259,264,271,274,277,282,285,293,309,334,346,360,367,370,384,387,391,394,400,403,453,460,677,687,690,694,700,706,712,716,719],[218,219,220],"p",{},"Unless you've just awoken from a coma, you've likely tried using a Large Language Model (LLM) in your IDE.",[218,222,223],{},"Some hate it, while others love it.",[218,225,226,227,234,235,240],{},"You probably use ",[228,229,233],"a",{"href":230,"rel":231},"https:\u002F\u002Fgithub.com\u002Ffeatures\u002Fcopilot",[232],"nofollow","GitHub Copilot"," for your coding needs, and perhaps even ",[228,236,239],{"href":237,"rel":238},"https:\u002F\u002Fdocs.aws.amazon.com\u002Fcodewhisperer\u002Flatest\u002Fuserguide\u002Fwhisper-legacy.html",[232],"Code Whisperer"," if you're an AWS user. But what if you can't afford it, or your company doesn't permit the use of these services? To be clear, using these services involves sending a copy of your text and code to them for processing, and they use your data to train their models.",[218,242,243],{},"Anyway, back on topic...",[218,245,246,247,252,253,258],{},"Due to such limitations at work, I've been using ",[228,248,251],{"href":249,"rel":250},"https:\u002F\u002Follama.com\u002F",[232],"Ollama"," and ",[228,254,257],{"href":255,"rel":256},"https:\u002F\u002Fwww.continue.dev\u002F",[232],"Continue"," as alternatives to Copilot, and it has been working quite well so far!",[260,261,263],"h2",{"id":262},"ollama-run-a-local-llm","Ollama - Run a Local LLM",[218,265,266,267,270],{},"If you haven't heard of it, ",[228,268,251],{"href":249,"rel":269},[232]," is an open-source alternative to GPT models like ChatGPT, focusing on privacy and control. It's designed to run locally (or privately) without sending your data to third parties. Woot!",[218,272,273],{},"After installing Ollama, you can run a local instance of an LLM and interact with it through your terminal or IDE plugins.",[218,275,276],{},"For this guide, let's install one model for prompting and another for autocompletion.",[278,279,281],"h3",{"id":280},"pull-the-models","Pull the Models",[218,283,284],{},"First, keep in mind that models are constantly changing, so you need to ensure your models are up-to-date. Also, consider the amount of GPU and memory available on your machine.",[218,286,287,288,292],{},"As of writing this guide, ",[289,290,291],"code",{},"llama3.1"," has just been released and is great. However, check the following resources for the latest recommended models:",[294,295,296,303],"ul",{},[297,298,299],"li",{},[228,300,301],{"href":301,"rel":302},"https:\u002F\u002Follama.com\u002Flibrary",[232],[297,304,305],{},[228,306,307],{"href":307,"rel":308},"https:\u002F\u002Fdocs.continue.dev\u002Fsetup\u002Fselect-model#chat",[232],[310,311,316],"pre",{"className":312,"code":313,"language":314,"meta":315,"style":315},"language-bash shiki shiki-themes material-theme-lighter material-theme material-theme-palenight","ollama pull llama3.1:8b\n","bash","",[289,317,318],{"__ignoreMap":315},[319,320,323,327,331],"span",{"class":321,"line":322},"line",1,[319,324,326],{"class":325},"sBMFI","ollama",[319,328,330],{"class":329},"sfazB"," pull",[319,332,333],{"class":329}," llama3.1:8b\n",[218,335,336,337,340,341,345],{},"For autocompletion, ",[289,338,339],{},"deepseek-coder-v2"," is an excellent model, but check ",[228,342,343],{"href":343,"rel":344},"https:\u002F\u002Fdocs.continue.dev\u002Fsetup\u002Fselect-model#autocomplete",[232]," to see if there are any updates or newer models available.",[310,347,349],{"className":312,"code":348,"language":314,"meta":315,"style":315},"ollama pull deepseek-coder-v2:16b\n",[289,350,351],{"__ignoreMap":315},[319,352,353,355,357],{"class":321,"line":322},[319,354,326],{"class":325},[319,356,330],{"class":329},[319,358,359],{"class":329}," deepseek-coder-v2:16b\n",[218,361,362],{},[363,364],"img",{"alt":365,"src":366},"ollama-pull-llama-3-1","\u002Fimages\u002Fblog\u002Follama-pull-llama-3-1.png",[218,368,369],{},"Once the models are downloaded, you can run them locally:",[310,371,373],{"className":312,"code":372,"language":314,"meta":315,"style":315},"ollama run llama3.1:8b\n",[289,374,375],{"__ignoreMap":315},[319,376,377,379,382],{"class":321,"line":322},[319,378,326],{"class":325},[319,380,381],{"class":329}," run",[319,383,333],{"class":329},[218,385,386],{},"Next, we'll configure VS Code with the Continue extension to use these models for autocompletion and prompting.",[260,388,390],{"id":389},"continue-llm-extension-for-vs-code","Continue - LLM Extension for VS Code",[218,392,393],{},"Install the Continue Extension from the VS Code Marketplace.",[218,395,396],{},[228,397,398],{"href":398,"rel":399},"https:\u002F\u002Fmarketplace.visualstudio.com\u002Fitems?itemName=Continue.continue",[232],[218,401,402],{},"Once installed, we'll modify its configuration to use the models we just downloaded. Open the configuration file:",[310,404,406],{"className":312,"code":405,"language":314,"meta":315,"style":315},"# On Linux\u002FmacOS\ncode ~\u002F.continue\u002Fconfig.json\n\n# On Windows (using Command Prompt)\n# code %USERPROFILE%\\.continue\\config.json\n# or (using PowerShell)\n# code $env:USERPROFILE\\.continue\\config.json\n",[289,407,408,414,422,429,435,441,447],{"__ignoreMap":315},[319,409,410],{"class":321,"line":322},[319,411,413],{"class":412},"sHwdD","# On Linux\u002FmacOS\n",[319,415,417,419],{"class":321,"line":416},2,[319,418,289],{"class":325},[319,420,421],{"class":329}," ~\u002F.continue\u002Fconfig.json\n",[319,423,425],{"class":321,"line":424},3,[319,426,428],{"emptyLinePlaceholder":427},true,"\n",[319,430,432],{"class":321,"line":431},4,[319,433,434],{"class":412},"# On Windows (using Command Prompt)\n",[319,436,438],{"class":321,"line":437},5,[319,439,440],{"class":412},"# code %USERPROFILE%\\.continue\\config.json\n",[319,442,444],{"class":321,"line":443},6,[319,445,446],{"class":412},"# or (using PowerShell)\n",[319,448,450],{"class":321,"line":449},7,[319,451,452],{"class":412},"# code $env:USERPROFILE\\.continue\\config.json\n",[218,454,455,456,459],{},"Add the following two entries into your ",[289,457,458],{},"~\u002F.continue\u002Fconfig.json"," file:",[310,461,465],{"className":462,"code":463,"language":464,"meta":315,"style":315},"language-json shiki shiki-themes material-theme-lighter material-theme material-theme-palenight","{\n  \"models\": [\n    \u002F\u002F ... other models if any ...\n    {\n      \"title\": \"Llama 3.1 8B (Ollama)\", \u002F\u002F Descriptive title\n      \"provider\": \"ollama\",\n      \"model\": \"llama3.1:8b\"\n    }\n    \u002F\u002F ...\n  ],\n  \"tabAutocompleteModel\": {\n    \"title\": \"Deepseek Coder V2 16B (Ollama)\", \u002F\u002F Descriptive title\n    \"provider\": \"ollama\",\n    \"model\": \"deepseek-coder-v2:16b\"\n  }\n  \u002F\u002F ... rest of config ...\n}\n","json",[289,466,467,473,491,496,501,527,547,566,572,578,584,599,622,641,659,665,671],{"__ignoreMap":315},[319,468,469],{"class":321,"line":322},[319,470,472],{"class":471},"sMK4o","{\n",[319,474,475,478,482,485,488],{"class":321,"line":416},[319,476,477],{"class":471},"  \"",[319,479,481],{"class":480},"spNyl","models",[319,483,484],{"class":471},"\"",[319,486,487],{"class":471},":",[319,489,490],{"class":471}," [\n",[319,492,493],{"class":321,"line":424},[319,494,495],{"class":412},"    \u002F\u002F ... other models if any ...\n",[319,497,498],{"class":321,"line":431},[319,499,500],{"class":471},"    {\n",[319,502,503,506,509,511,513,516,519,521,524],{"class":321,"line":437},[319,504,505],{"class":471},"      \"",[319,507,508],{"class":325},"title",[319,510,484],{"class":471},[319,512,487],{"class":471},[319,514,515],{"class":471}," \"",[319,517,518],{"class":329},"Llama 3.1 8B (Ollama)",[319,520,484],{"class":471},[319,522,523],{"class":471},",",[319,525,526],{"class":412}," \u002F\u002F Descriptive title\n",[319,528,529,531,534,536,538,540,542,544],{"class":321,"line":443},[319,530,505],{"class":471},[319,532,533],{"class":325},"provider",[319,535,484],{"class":471},[319,537,487],{"class":471},[319,539,515],{"class":471},[319,541,326],{"class":329},[319,543,484],{"class":471},[319,545,546],{"class":471},",\n",[319,548,549,551,554,556,558,560,563],{"class":321,"line":449},[319,550,505],{"class":471},[319,552,553],{"class":325},"model",[319,555,484],{"class":471},[319,557,487],{"class":471},[319,559,515],{"class":471},[319,561,562],{"class":329},"llama3.1:8b",[319,564,565],{"class":471},"\"\n",[319,567,569],{"class":321,"line":568},8,[319,570,571],{"class":471},"    }\n",[319,573,575],{"class":321,"line":574},9,[319,576,577],{"class":412},"    \u002F\u002F ...\n",[319,579,581],{"class":321,"line":580},10,[319,582,583],{"class":471},"  ],\n",[319,585,587,589,592,594,596],{"class":321,"line":586},11,[319,588,477],{"class":471},[319,590,591],{"class":480},"tabAutocompleteModel",[319,593,484],{"class":471},[319,595,487],{"class":471},[319,597,598],{"class":471}," {\n",[319,600,602,605,607,609,611,613,616,618,620],{"class":321,"line":601},12,[319,603,604],{"class":471},"    \"",[319,606,508],{"class":325},[319,608,484],{"class":471},[319,610,487],{"class":471},[319,612,515],{"class":471},[319,614,615],{"class":329},"Deepseek Coder V2 16B (Ollama)",[319,617,484],{"class":471},[319,619,523],{"class":471},[319,621,526],{"class":412},[319,623,625,627,629,631,633,635,637,639],{"class":321,"line":624},13,[319,626,604],{"class":471},[319,628,533],{"class":325},[319,630,484],{"class":471},[319,632,487],{"class":471},[319,634,515],{"class":471},[319,636,326],{"class":329},[319,638,484],{"class":471},[319,640,546],{"class":471},[319,642,644,646,648,650,652,654,657],{"class":321,"line":643},14,[319,645,604],{"class":471},[319,647,553],{"class":325},[319,649,484],{"class":471},[319,651,487],{"class":471},[319,653,515],{"class":471},[319,655,656],{"class":329},"deepseek-coder-v2:16b",[319,658,565],{"class":471},[319,660,662],{"class":321,"line":661},15,[319,663,664],{"class":471},"  }\n",[319,666,668],{"class":321,"line":667},16,[319,669,670],{"class":412},"  \u002F\u002F ... rest of config ...\n",[319,672,674],{"class":321,"line":673},17,[319,675,676],{"class":471},"}\n",[678,679,680],"blockquote",{},[218,681,682,683,686],{},"Note: The config file does not support comments. Ensure you maintain the correct JSON format. Remove the example comments (",[289,684,685],{},"\u002F\u002F ...",") before saving.",[218,688,689],{},"Now, restart VS Code and try out the new configuration!",[260,691,693],{"id":692},"try-it-out","Try It Out",[218,695,696,697,699],{},"On the activity bar, click the Continue icon. At the bottom of the prompt input area, select the ",[289,698,518],{}," model (or the title you chose) from the dropdown menu. Then, try typing a prompt and see if you get helpful suggestions!",[218,701,702],{},[363,703],{"alt":704,"src":705},"Continue-extension-prompt-example","\u002Fimages\u002Fblog\u002FContinue-extension-prompt-example-vscode.png",[218,707,708,709,711],{},"Now, edit any code file and start typing to see if the autocompletion works with your ",[289,710,615],{}," model!",[260,713,715],{"id":714},"conclusion","Conclusion",[218,717,718],{},"Congratulations! You've successfully integrated custom LLMs into VS Code using the Continue Extension. This setup runs the LLMs locally on your machine, offering a free and private alternative to cloud-based AI assistants.",[720,721,722],"style",{},"html pre.shiki code .sBMFI, html code.shiki .sBMFI{--shiki-light:#E2931D;--shiki-default:#FFCB6B;--shiki-dark:#FFCB6B}html pre.shiki code .sfazB, html code.shiki .sfazB{--shiki-light:#91B859;--shiki-default:#C3E88D;--shiki-dark:#C3E88D}html .light .shiki span {color: var(--shiki-light);background: var(--shiki-light-bg);font-style: var(--shiki-light-font-style);font-weight: var(--shiki-light-font-weight);text-decoration: var(--shiki-light-text-decoration);}html.light .shiki span {color: var(--shiki-light);background: var(--shiki-light-bg);font-style: var(--shiki-light-font-style);font-weight: var(--shiki-light-font-weight);text-decoration: var(--shiki-light-text-decoration);}html .default .shiki span {color: var(--shiki-default);background: var(--shiki-default-bg);font-style: var(--shiki-default-font-style);font-weight: var(--shiki-default-font-weight);text-decoration: var(--shiki-default-text-decoration);}html .shiki span {color: var(--shiki-default);background: var(--shiki-default-bg);font-style: var(--shiki-default-font-style);font-weight: var(--shiki-default-font-weight);text-decoration: var(--shiki-default-text-decoration);}html .dark .shiki span {color: var(--shiki-dark);background: var(--shiki-dark-bg);font-style: var(--shiki-dark-font-style);font-weight: var(--shiki-dark-font-weight);text-decoration: var(--shiki-dark-text-decoration);}html.dark .shiki span {color: var(--shiki-dark);background: var(--shiki-dark-bg);font-style: var(--shiki-dark-font-style);font-weight: var(--shiki-dark-font-weight);text-decoration: var(--shiki-dark-text-decoration);}html pre.shiki code .sHwdD, html code.shiki .sHwdD{--shiki-light:#90A4AE;--shiki-light-font-style:italic;--shiki-default:#546E7A;--shiki-default-font-style:italic;--shiki-dark:#676E95;--shiki-dark-font-style:italic}html pre.shiki code .sMK4o, html code.shiki .sMK4o{--shiki-light:#39ADB5;--shiki-default:#89DDFF;--shiki-dark:#89DDFF}html pre.shiki code .spNyl, html code.shiki .spNyl{--shiki-light:#9C3EDA;--shiki-default:#C792EA;--shiki-dark:#C792EA}",{"title":315,"searchDepth":416,"depth":416,"links":724},[725,728,729,730],{"id":262,"depth":416,"text":263,"children":726},[727],{"id":280,"depth":424,"text":281},{"id":389,"depth":416,"text":390},{"id":692,"depth":416,"text":693},{"id":714,"depth":416,"text":715},"2024-07-23","A guide on how Ollama and Continue can serve as an alternative to GitHub Copilot.","md",{"src":735,"alt":736},"\u002Fimages\u002Fblog\u002Follama-and-continue.png","A llama looking at a screen with code on it",{},{"title":37,"description":732},"published","Qp4uEhMY8tm63cPE3tUGTKBZKXSPtMOh1FuUut-FbEs",[742,744],{"title":33,"path":34,"stem":35,"description":743,"status":739,"children":-1},"Getting YouTube and YouTube Kids to work on Amazon Fire for kids",{"title":41,"path":42,"stem":43,"description":745,"status":739,"children":-1},"Debug your local Node.js packages like a pro (and with less pain) using pnpm link.",1776221196452]