 {"id":519831,"date":"2026-02-05T09:34:08","date_gmt":"2026-02-05T16:34:08","guid":{"rendered":"https:\/\/jorgep.com\/blog\/?p=519831"},"modified":"2026-02-08T11:54:38","modified_gmt":"2026-02-08T18:54:38","slug":"llm-usage-stats-on-my-development-spree","status":"publish","type":"post","link":"https:\/\/jorgep.com\/blog\/llm-usage-stats-on-my-development-spree\/","title":{"rendered":"LLM Usage Stats on My Development Spree"},"content":{"rendered":"\n<div class=\"wp-block-columns has-theme-palette-7-background-color has-background is-layout-flex wp-container-core-columns-is-layout-9d6595d7 wp-block-columns-is-layout-flex\">\n<div class=\"wp-block-column is-layout-flow wp-block-column-is-layout-flow\">\n<p>Part of: <strong> <a href=\"https:\/\/jorgep.com\/blog\/series-ai-learnings\/\">AI Learning Series Here<\/a><\/strong><\/p>\n\n\n<style>.kadence-column395113_43ef2d-d5 > .kt-inside-inner-col,.kadence-column395113_43ef2d-d5 > .kt-inside-inner-col:before{border-top-left-radius:0px;border-top-right-radius:0px;border-bottom-right-radius:0px;border-bottom-left-radius:0px;}.kadence-column395113_43ef2d-d5 > .kt-inside-inner-col{column-gap:var(--global-kb-gap-sm, 1rem);}.kadence-column395113_43ef2d-d5 > .kt-inside-inner-col{flex-direction:column;}.kadence-column395113_43ef2d-d5 > .kt-inside-inner-col > .aligncenter{width:100%;}.kadence-column395113_43ef2d-d5 > .kt-inside-inner-col:before{opacity:0.3;}.kadence-column395113_43ef2d-d5{position:relative;}@media all and (max-width: 1024px){.kadence-column395113_43ef2d-d5 > .kt-inside-inner-col{flex-direction:column;justify-content:center;}}@media all and (max-width: 767px){.kadence-column395113_43ef2d-d5 > .kt-inside-inner-col{flex-direction:column;justify-content:center;}}<\/style>\n<div class=\"wp-block-kadence-column kadence-column395113_43ef2d-d5\"><div class=\"kt-inside-inner-col\"><style>.wp-block-kadence-advancedheading.kt-adv-heading510545_6813a5-28, .wp-block-kadence-advancedheading.kt-adv-heading510545_6813a5-28[data-kb-block=\"kb-adv-heading510545_6813a5-28\"]{font-size:var(--global-kb-font-size-sm, 0.9rem);font-style:normal;}.wp-block-kadence-advancedheading.kt-adv-heading510545_6813a5-28 mark.kt-highlight, .wp-block-kadence-advancedheading.kt-adv-heading510545_6813a5-28[data-kb-block=\"kb-adv-heading510545_6813a5-28\"] mark.kt-highlight{font-style:normal;color:#f76a0c;-webkit-box-decoration-break:clone;box-decoration-break:clone;padding-top:0px;padding-right:0px;padding-bottom:0px;padding-left:0px;}.wp-block-kadence-advancedheading.kt-adv-heading510545_6813a5-28 img.kb-inline-image, .wp-block-kadence-advancedheading.kt-adv-heading510545_6813a5-28[data-kb-block=\"kb-adv-heading510545_6813a5-28\"] img.kb-inline-image{width:150px;vertical-align:baseline;}<\/style>\n<p class=\"kt-adv-heading510545_6813a5-28 wp-block-kadence-advancedheading\" data-kb-block=\"kb-adv-heading510545_6813a5-28\">Quick Links:&nbsp;<a href=\"https:\/\/jorgep.com\/blog\/resources-for-learning-ai\/\">Resources for Learning AI<\/a> | <a href=\"https:\/\/jorgep.com\/blog\/keeping-up-with-ai\/\">Keep up with AI<\/a> | <a href=\"https:\/\/jorgep.com\/blog\/list-of-ai-tools\/\" data-type=\"post\" data-id=\"402818\">List of AI Tools<\/a><\/p>\n<\/div><\/div>\n<\/div>\n\n\n\n<div class=\"wp-block-column is-layout-flow wp-block-column-is-layout-flow\"><div class=\"wp-block-template-part\"><style>.wp-block-kadence-advancedheading.kt-adv-heading395113_c650df-47, .wp-block-kadence-advancedheading.kt-adv-heading395113_c650df-47[data-kb-block=\"kb-adv-heading395113_c650df-47\"]{text-align:center;font-size:var(--global-kb-font-size-md, 1.25rem);line-height:60px;font-style:normal;background-color:#f5a511;}.wp-block-kadence-advancedheading.kt-adv-heading395113_c650df-47 mark.kt-highlight, .wp-block-kadence-advancedheading.kt-adv-heading395113_c650df-47[data-kb-block=\"kb-adv-heading395113_c650df-47\"] mark.kt-highlight{font-style:normal;color:#f76a0c;-webkit-box-decoration-break:clone;box-decoration-break:clone;padding-top:0px;padding-right:0px;padding-bottom:0px;padding-left:0px;}.wp-block-kadence-advancedheading.kt-adv-heading395113_c650df-47 img.kb-inline-image, .wp-block-kadence-advancedheading.kt-adv-heading395113_c650df-47[data-kb-block=\"kb-adv-heading395113_c650df-47\"] img.kb-inline-image{width:150px;vertical-align:baseline;}<\/style>\n<p class=\"kt-adv-heading395113_c650df-47 wp-block-kadence-advancedheading\" data-kb-block=\"kb-adv-heading395113_c650df-47\">Subscribe to <a href=\"https:\/\/go.35s.be\/jtb\" target=\"_blank\" rel=\"noreferrer noopener\"><strong>JorgeTechBits  newsletter<\/strong><\/a><\/p>\n<\/div><\/div>\n<\/div>\n\n\n<style>.wp-block-kadence-advancedheading.kt-adv-heading519190_4a1b6f-84, .wp-block-kadence-advancedheading.kt-adv-heading519190_4a1b6f-84[data-kb-block=\"kb-adv-heading519190_4a1b6f-84\"]{font-size:var(--global-kb-font-size-sm, 0.9rem);font-style:normal;}.wp-block-kadence-advancedheading.kt-adv-heading519190_4a1b6f-84 mark.kt-highlight, .wp-block-kadence-advancedheading.kt-adv-heading519190_4a1b6f-84[data-kb-block=\"kb-adv-heading519190_4a1b6f-84\"] mark.kt-highlight{font-style:normal;color:#f76a0c;-webkit-box-decoration-break:clone;box-decoration-break:clone;padding-top:0px;padding-right:0px;padding-bottom:0px;padding-left:0px;}.wp-block-kadence-advancedheading.kt-adv-heading519190_4a1b6f-84 img.kb-inline-image, .wp-block-kadence-advancedheading.kt-adv-heading519190_4a1b6f-84[data-kb-block=\"kb-adv-heading519190_4a1b6f-84\"] img.kb-inline-image{width:150px;vertical-align:baseline;}<\/style>\n<p class=\"kt-adv-heading519190_4a1b6f-84 wp-block-kadence-advancedheading\" data-kb-block=\"kb-adv-heading519190_4a1b6f-84\">AI Disclaimer I love exploring new technology, and that includes using AI to help with research and editing! My digital &#8220;team&#8221; includes tools like Google Gemini, Notebook LM, Microsoft Copilot, Perplexity.ai, Claude.ai, and others as needed. They help me gather insights and polish content\u2014so you get the best, most up-to-date information possible.<\/p>\n\n\n\n<p>As technologies, we&#8217;re constantly seeking tools that enhance our productivity, accelerate our workflows, and ultimately, help us bring ideas to life faster. Over a period of about 45 days, I embarked on an intense development sprint, leveraging AI models and a specialized coding assistant, <a href=\"https:\/\/go.35s.be\/KiloCode\">Kilo Code<\/a>, to build <strong>four distinct applications now successfully deployed in production<\/strong>. This journey wasn&#8217;t just about speed; it was a profound exploration into how modern AI tools fundamentally change the development paradigm.<\/p>\n\n\n\n<p><\/p>\n\n\n\n<p>Please let me say that although I started my carrier as a developer, I am not a professional developer, and have not been for a long long time.    I consider myself an amateur one, a regular user that has many ideas and needs.     Having a Coding partner has been a fantastic thing over the past year: See my blogs:&nbsp;<a href=\"https:\/\/jorgep.com\/blog\/vibe-coding-beyond-ai-code-completion\/\">Vibe Coding: Beyond AI Code Completion<\/a>&nbsp;and&nbsp;<a href=\"https:\/\/jorgep.com\/blog\/from-vibe-coding-to-coding-partner\/\">From Vibe Coding to Coding Partner<\/a>.<\/p>\n\n\n\n<p><\/p>\n\n\n\n<p>I wrote about coding assistants in blog post:   <a href=\"https:\/\/jorgep.com\/blog\/the-evolution-of-ai-coding-assistants-as-of-december-2025\/\" data-type=\"post\" data-id=\"519688\">The Evolution of AI Coding Assistants (as of December 2025)<\/a>.   Since then, as I get more and more familiar with <a href=\"https:\/\/go.35s.be\/KiloCode\" data-type=\"link\" data-id=\"https:\/\/kilo.ai\/\" target=\"_blank\" rel=\"noreferrer noopener\">Kilo Code<\/a>, I am using Cursor less and less.    Do not get me wrong, in my opinion, it is the best platform out there, but three things are happening to it: 1) the quotas are getting much more restricted and 2) I wanted to understand more how it worked behind the scenes (Context, tokens, limitations)  and 3) the capability of much cheaper, and sometimes free, open source models.<\/p>\n\n\n\n<p><\/p>\n\n\n\n<p>This post will peel back the layers of my experience, analyze the raw usage data, and extract key insights into model efficiency, cost, and the invaluable role AI played in this rapid development cycle. <\/p>\n\n\n\n<p><\/p>\n\n\n\n<p><\/p>\n\n\n\n<h3 class=\"wp-block-heading\">The Challenge: Four Apps, Two Months, AI as My Co-Pilot<\/h3>\n\n\n\n<p>I did not have a specific goal in mind, but I wanted to see the capabilities and learn how will I interact and what I can creates.   The result: four separate applications from concept and ideation to production within a strict 60-day timeframe. Each application served a unique business need, ranging from data processing utilities to user-facing interfaces. To achieve this, I turned to a blend of powerful AI models and an integral coding assistant.  The plan is to write about them future article but a you can see one in action on my artist site:  <a href=\"https:\/\/doodlingjorge.com\">https:\/\/doodlingjorge.com<\/a>     &#8211; Everything on it was written interacting with the agent and prompts.   <\/p>\n\n\n\n<p>My primary toolkit included:<\/p>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>Kilo Code (coding assistant):<\/strong> This was my constant companion, providing real-time code suggestions, error detection, and boilerplate generation.<\/li>\n\n\n\n<li><strong>Diverse AI Models:<\/strong> I utilized a variety of large language models (LLMs) for tasks like high-level architectural guidance, complex algorithm design, data analysis, and even content generation for application UIs.<\/li>\n<\/ul>\n\n\n\n<p>Let&#8217;s dive into the data to see which models were my unsung heroes and where the efficiencies truly lay.<\/p>\n\n\n\n<h3 class=\"wp-block-heading\">The Raw Data: A Glimpse into My AI Usage<\/h3>\n\n\n\n<p>Over the 45-day period (late December 2025 to early February 2026), my AI usage generated a comprehensive log. This log tracks every interaction, its cost, and the volume of information processed (tokens).<\/p>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>Total Input Tokens:<\/strong> 461,133,739<\/li>\n\n\n\n<li><strong>Total Output Tokens:<\/strong> 5,069,969<\/li>\n\n\n\n<li><strong>Grand Total Tokens:<\/strong> <strong>466,203,708<\/strong><\/li>\n<\/ul>\n\n\n\n<p>Update:   Please see companion post:   <a href=\"https:\/\/jorgep.com\/blog\/the-cost-of-460-million-tokens-understanding-tokens-token-types\/\" data-type=\"post\" data-id=\"519854\">The Cost of 460 Million Tokens \u2013 Understanding Tokens, Token Types<\/a><\/p>\n\n\n\n<p><\/p>\n\n\n\n<p>Here&#8217;s the summary of my total usage per model:<\/p>\n\n\n\n<figure class=\"wp-block-table\"><table><thead><tr><td><strong>Model<\/strong><\/td><td><strong>Total Cost<\/strong><\/td><td><strong>Total Requests<\/strong><\/td><td><strong>Input Tokens<\/strong><\/td><td><strong>Output Tokens<\/strong><\/td><td><strong>Cache Hits<\/strong><\/td><\/tr><\/thead><tbody><tr><td><strong>anthropic\/claude-3-opus-4.5<\/strong><\/td><td>$25.23<\/td><td>105<\/td><td>8.9 M<\/td><td>83,540<\/td><td>3.5 M<\/td><\/tr><tr><td><strong>anthropic\/claude-haiku-4.5<\/strong><\/td><td>$0.00<\/td><td>131<\/td><td>6.9 M<\/td><td>49,617<\/td><td>6.1 M<\/td><\/tr><tr><td><strong>anthropic\/claude-sonnet-4.5<\/strong><\/td><td>$34.02<\/td><td>163<\/td><td>13.5 M<\/td><td>110,131<\/td><td>917,949<\/td><\/tr><tr><td><strong>arcee-ai\/finity-large-preview<\/strong><\/td><td>$0.00<\/td><td>28<\/td><td>1.1 M<\/td><td>12,648<\/td><td>0<\/td><\/tr><tr><td><strong>codestral-2508<\/strong><\/td><td>$5.30<\/td><td>20,443<\/td><td>12.0 M<\/td><td>1.7 M<\/td><td>0<\/td><\/tr><tr><td><strong>giga-potato<\/strong><\/td><td>$0.00<\/td><td>1,444<\/td><td>64.1 M<\/td><td>448,849<\/td><td>36.6 M<\/td><\/tr><tr><td><strong>google\/gemini-2.5-flash<\/strong><\/td><td>$0.01<\/td><td>6<\/td><td>29,663<\/td><td>146<\/td><td>13,545<\/td><\/tr><tr><td><strong>google\/gemini-3-flash-preview<\/strong><\/td><td>$4.80<\/td><td>385<\/td><td>32.5 M<\/td><td>165,304<\/td><td>14.8 M<\/td><\/tr><tr><td><strong>google\/gemini-3-pro-preview<\/strong><\/td><td>$0.96<\/td><td>24<\/td><td>1.7 M<\/td><td>10,391<\/td><td>1.4 M<\/td><\/tr><tr><td><strong>minimax\/minimax-m2.1<\/strong><\/td><td>$9.38<\/td><td>2,752<\/td><td>115.8 M<\/td><td>1.2 M<\/td><td>71.0 M<\/td><\/tr><tr><td><strong>mistral\/devstral-2<\/strong><\/td><td>$0.00<\/td><td>398<\/td><td>25.5 M<\/td><td>128,477<\/td><td>0<\/td><\/tr><tr><td><strong>mistral\/devstral-2512-free<\/strong><\/td><td>$0.00<\/td><td>46<\/td><td>2.1 M<\/td><td>15,625<\/td><td>0<\/td><\/tr><tr><td><strong>moonshotai\/kimi-k2.5<\/strong><\/td><td>$0.90<\/td><td>42<\/td><td>2.7 M<\/td><td>10,315<\/td><td>92,917<\/td><\/tr><tr><td><strong>openai\/gpt-5.1-codex-mini<\/strong><\/td><td>$0.21<\/td><td>86<\/td><td>2.8 M<\/td><td>51,978<\/td><td>2.6 M<\/td><\/tr><tr><td><strong>openai\/gpt-5.2<\/strong><\/td><td>$8.07<\/td><td>154<\/td><td>12.5 M<\/td><td>83,034<\/td><td>11.5 M<\/td><\/tr><tr><td><strong>openai\/gpt-5.2-codex<\/strong><\/td><td>$22.75<\/td><td>499<\/td><td>49.3 M<\/td><td>293,736<\/td><td>24.1 M<\/td><\/tr><tr><td><strong>openai\/gpt-5.1-codex<\/strong><\/td><td>$3.41<\/td><td>65<\/td><td>9.2 M<\/td><td>21,758<\/td><td>7.4 M<\/td><\/tr><tr><td><strong>xai\/grok-code-fast-1<\/strong><\/td><td>$0.00<\/td><td>105<\/td><td>1.6 M<\/td><td>61,081<\/td><td>1.0 M<\/td><\/tr><tr><td><strong>z-ai\/glm-4.7<\/strong><\/td><td>$1.30<\/td><td>10<\/td><td>424,145<\/td><td>124,724<\/td><td>248,240<\/td><\/tr><tr><td><strong>z-ai\/glm-4-7<\/strong><\/td><td>$0.00<\/td><td>10<\/td><td>287,482<\/td><td>10,920<\/td><td>75,671<\/td><\/tr><\/tbody><\/table><\/figure>\n\n\n\n<hr class=\"wp-block-separator has-alpha-channel-opacity\"\/>\n\n\n\n<h3 class=\"wp-block-heading\">Unpacking &#8220;Efficiency&#8221;: What Does the Data Really Say?<\/h3>\n\n\n\n<p>To truly understand my AI usage, we need to go beyond raw costs and requests. &#8220;Efficiency&#8221; in the context of LLMs involves several factors:<\/p>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>Cache Hits:<\/strong> These are tokens (parts of your prompt or context) that the model provider already has stored, leading to significantly reduced costs and faster response times. High cache hits mean you&#8217;re effectively reusing information.<\/li>\n\n\n\n<li><strong>Tokens per Dollar:<\/strong> This metric directly translates to how much &#8220;work&#8221; (in terms of processed text) you&#8217;re getting for your money. A higher number indicates better economic efficiency.<\/li>\n\n\n\n<li><strong>Total Tokens Processed:<\/strong> The sheer volume of data (input, output, and cached context) handled by a model.<\/li>\n<\/ul>\n\n\n\n<p>Here\u2019s an efficiency breakdown, sorted by the bang-for-buck:  <\/p>\n\n\n\n<figure class=\"wp-block-table\"><table><thead><tr><td><strong>Model<\/strong><\/td><td><strong>Total Tokens<\/strong><\/td><td><strong>Total Cost<\/strong><\/td><td><strong>Efficiency (Tokens per $1)<\/strong><\/td><\/tr><\/thead><tbody><tr><td><strong>Giga-potato<\/strong><\/td><td>101.1 M<\/td><td>$0.00<\/td><td><strong>Free \/ Infinite<\/strong><\/td><\/tr><tr><td><strong>Devstral-2<\/strong><\/td><td>25.6 M<\/td><td>$0.00<\/td><td><strong>Free \/ Infinite<\/strong><\/td><\/tr><tr><td><strong>Claude Haiku 4.5<\/strong><\/td><td>13.0 M<\/td><td>$0.00<\/td><td><strong>Free \/ Infinite<\/strong> (promo?)<\/td><\/tr><tr><td><strong>GPT-5.1 Codex Mini<\/strong><\/td><td>5.4 M<\/td><td>$0.21<\/td><td><strong>25.9 Million<\/strong><\/td><\/tr><tr><td><strong>Minimax-M2.1<\/strong><\/td><td>188.0 M<\/td><td>$9.38<\/td><td><strong>20.0 Million<\/strong><\/td><\/tr><tr><td><strong>Gemini 3 Flash<\/strong><\/td><td>47.4 M<\/td><td>$4.80<\/td><td><strong>9.8 Million<\/strong><\/td><\/tr><tr><td><strong>GPT-5.1 Codex<\/strong><\/td><td>16.6 M<\/td><td>$3.41<\/td><td><strong>4.8 Million<\/strong><\/td><\/tr><tr><td><strong>Gemini 3 Pro<\/strong><\/td><td>3.1 M<\/td><td>$0.96<\/td><td><strong>3.2 Million<\/strong><\/td><\/tr><tr><td><strong>GPT-5.2 Codex<\/strong><\/td><td>73.6 M<\/td><td>$22.75<\/td><td><strong>3.2 Million<\/strong><\/td><\/tr><tr><td><strong>Codestral-2508<\/strong><\/td><td>13.7 M<\/td><td>$5.30<\/td><td><strong>2.5 Million<\/strong><\/td><\/tr><tr><td><strong>Claude 3 Opus<\/strong><\/td><td>12.4 M<\/td><td>$25.23<\/td><td><strong>0.49 Million<\/strong><\/td><\/tr><tr><td><strong>Claude Sonnet 4.5<\/strong><\/td><td>14.5 M<\/td><td>$34.02<\/td><td><strong>0.42 Million<\/strong><\/td><\/tr><\/tbody><\/table><\/figure>\n\n\n\n<p>A note about <strong>Claude Haiku 4.5<\/strong>:    It is not free, but looks like I got a free ride for a little while.     <\/p>\n\n\n\n<p><\/p>\n\n\n\n<figure class=\"wp-block-image size-large\"><img loading=\"lazy\" decoding=\"async\" width=\"1024\" height=\"665\" src=\"https:\/\/jorgep.com\/blog\/wp-content\/uploads\/image-154-1024x665.png\" alt=\"\" class=\"wp-image-519843\" srcset=\"https:\/\/jorgep.com\/blog\/wp-content\/uploads\/image-154-1024x665.png 1024w, https:\/\/jorgep.com\/blog\/wp-content\/uploads\/image-154-300x195.png 300w, https:\/\/jorgep.com\/blog\/wp-content\/uploads\/image-154-768x498.png 768w, https:\/\/jorgep.com\/blog\/wp-content\/uploads\/image-154.png 1131w\" sizes=\"auto, (max-width: 1024px) 100vw, 1024px\" \/><\/figure>\n\n\n\n<h3 class=\"wp-block-heading\">Key Takeaways from My 45-Day Sprint<\/h3>\n\n\n\n<ol start=\"1\" class=\"wp-block-list\">\n<li><strong>The Rise of the Coding Assistant (Kilo Code &amp; Codestral):<\/strong> My most frequent interactions were with <code>codestral-2508<\/code>, clocking over 20,000 requests. This highlights the indispensable role of Kilo Code as my coding assistant. For rapid application development, nothing beats immediate, context-aware code suggestions and generation. While its &#8220;Tokens per $1&#8221; wasn&#8217;t the highest, its sheer volume of assistance was critical. It acted like a hyper-efficient junior developer, constantly churning out snippets, fixing typos, and completing boilerplate code, freeing me to focus on architectural decisions and complex logic.<\/li>\n\n\n\n<li><strong>Strategic Use of &#8220;Free&#8221; &amp; High-Efficiency Models:<\/strong> Models with $0.00 cost (like <code>Claude Haiku 4.5<\/code>, <code>Giga-potato<\/code>, and <code>Mistral\/Devstral-2<\/code>) were invaluable. These were likely used for less critical tasks, internal experiments, or benefit from promotional\/free tiers. <code>GPT-5.1 Codex Mini<\/code> and <code>Minimax-M2.1<\/code> stood out as incredibly efficient paid options, providing millions of tokens per dollar. These became my workhorses for drafting documentation, generating larger code blocks, and processing extensive text data where cost-efficiency was paramount.<\/li>\n\n\n\n<li><strong>The &#8220;Premium&#8221; Powerhouses (and their Price Tag):<\/strong> The Anthropic Claude models (<code>Opus<\/code> and <code>Sonnet<\/code>) were my most expensive per token. These models excel in complex reasoning, nuanced understanding, and handling extensive context windows. I reserved them for critical tasks:\n<ul class=\"wp-block-list\">\n<li><strong>Architectural Brainstorming:<\/strong> Discussing high-level design patterns and trade-offs.<\/li>\n\n\n\n<li><strong>Complex Algorithm Design:<\/strong> Getting detailed breakdowns and alternative approaches for intricate logic.<\/li>\n\n\n\n<li><strong>Refactoring Guidance:<\/strong> Seeking advanced suggestions for improving code quality and performance.<\/li>\n\n\n\n<li><strong>Deep Error Analysis:<\/strong> When standard debugging failed, these models often provided breakthrough insights.The higher cost was justified by the quality and depth of their responses, which saved me significant time on intellectually demanding problems.<\/li>\n<\/ul>\n<\/li>\n\n\n\n<li><strong>The Power of Caching:<\/strong> Models like <code>GPT-5.2<\/code> and <code>Minimax-M2.1<\/code> showed strong &#8220;Cache Hit %.&#8221; This indicates effective reuse of repetitive prompts, system instructions, or large document contexts. Optimizing for caching is a subtle but powerful way to reduce costs, especially when iterating on similar problems or using consistent background context.<\/li>\n\n\n\n<li><strong>Diverse Models for Diverse Tasks:<\/strong> My data clearly shows that no single model did it all.\n<ul class=\"wp-block-list\">\n<li><strong>Codestral (via Kilo Code):<\/strong> The rapid-fire coding assistant.<\/li>\n\n\n\n<li><strong>Gemini Flash\/Pro:<\/strong> Good for balanced performance and cost, likely used for general task assistance.<\/li>\n\n\n\n<li><strong>GPT Codex variants:<\/strong> Strong for code-related queries, debugging, and understanding existing codebases.<\/li>\n\n\n\n<li><strong>Claude Opus\/Sonnet:<\/strong> The &#8220;thought partners&#8221; for complex problem-solving.<\/li>\n<\/ul>\n<\/li>\n<\/ol>\n\n\n\n<h3 class=\"wp-block-heading\">The Impact: Speed, Quality, and a New Workflow<\/h3>\n\n\n\n<p>Leveraging AI in this structured manner allowed me to achieve unprecedented development velocity. I estimate that Kilo Code and the complementary LLMs collectively boosted my productivity by <strong>at least 3x<\/strong> compared to traditional methods.<\/p>\n\n\n\n<ul class=\"wp-block-list\">\n<li><strong>Faster Prototyping:<\/strong> AI quickly generated initial structures and components.<\/li>\n\n\n\n<li><strong>Reduced Debugging Time:<\/strong> AI-powered analysis helped pinpoint issues rapidly.<\/li>\n\n\n\n<li><strong>Access to Expertise:<\/strong> I could tap into a vast knowledge base instantly, akin to having a team of experts at my fingertips.<\/li>\n\n\n\n<li><strong>Consistent Quality:<\/strong> AI helped maintain coding standards and consistency across applications.<\/li>\n<\/ul>\n\n\n\n<h3 class=\"wp-block-heading\">Looking Ahead<\/h3>\n\n\n\n<p>My 60-day sprint highlights a paradigm shift in software development. AI isn&#8217;t just a tool; it&#8217;s becoming an integral part of the development team. For those looking to accelerate their projects and build robust applications faster, I highly recommend:<\/p>\n\n\n\n<ol start=\"1\" class=\"wp-block-list\">\n<li><strong>Embracing a dedicated coding assistant<\/strong> like Kilo Code for daily coding tasks.<\/li>\n\n\n\n<li><strong>Strategically choosing LLMs<\/strong> based on the complexity and cost-sensitivity of the task.<\/li>\n\n\n\n<li><strong>Optimizing for cache hits<\/strong> where possible to maximize efficiency.<\/li>\n\n\n\n<li><strong>Context limits and usage MATTERS:<\/strong> Watch Context Size!<\/li>\n\n\n\n<li><strong>Different LLMs have different capabilities<\/strong>:  Some are good at complex tasks others are not.   Use the right LLM for the task at hand!<\/li>\n<\/ol>\n\n\n\n<h2 class=\"wp-block-heading\">Was it Worth It?<\/h2>\n\n\n\n<p>SHORT ANSWER:  <strong>Absolutely.<\/strong> For an investment of just $120, I gained a deep understanding of both how to interact with the technology and the mechanics behind it. To date, I\u2019ve built four production apps that are already integrated into my own workflows and some already being used by one of my client. One of these applications alone slashed my analysis time from two to three hours per week to just 10 minutes. That kind of efficiency is where the real value lies\u2014the tools will easily pay for themselves.<\/p>\n\n\n\n<p><\/p>\n\n\n\n<p><\/p>\n\n\n\n<p><\/p>\n\n\n\n<p>We are witnessing a fundamental shift in technology: the democratization of software. We\u2019ve reached a point where <strong>everyday people can build sophisticated, functional apps without writing a single line of code.<\/strong> The future of development is no longer gated by syntax; it is intelligent, collaborative, and incredibly fast. I\u2019m excited to continue exploring these frontiers and pushing the boundaries of what\u2019s possible with AI.<\/p>\n\n\n\n<div style=\"font-family: Verdana, Geneva, sans-serif; font-size: 11px; line-height: 1.6; color: #333;\">\n    <p>\n        <strong>Disclaimer:<\/strong> \n        <em>I personally love to share my learnings, thoughts, and ideas; I get great satisfaction knowing someone has read and benefited from an article. This content is created entirely on my own time and in a personal capacity. The views expressed here are mine alone and do not represent the positions or opinions of my employer.<\/em>\n    <\/p>\n    <p>\n        In my professional role, I serve as a Workforce Transformation Solutions Principal for \n        <a href=\"https:\/\/www.dell.com\/en-us\/work\/learn\/by-service-type-deployment\" style=\"color: #007db8; font-weight: bold; text-decoration: none;\">Dell Technology Services<\/a>. \n        I am passionate about guiding organizations through complex technology transitions and \n        <a href=\"https:\/\/www.delltechnologies.com\/en-us\/what-we-do\/workforce-transformation.htm\" style=\"color: #007db8; font-weight: bold; text-decoration: none;\">Workforce Transformation<\/a>. \n        <a href=\"https:\/\/www.delltechnologies.com\/en-us\/index.htm\" style=\"color: #007db8; font-weight: bold; text-decoration: none;\">Learn more at Dell Technologies<\/a>.\n    <\/p>\n    <hr style=\"border: 0; border-top: 1px solid #ddd; margin: 12px 0;\">\n<\/div>\n","protected":false},"excerpt":{"rendered":"<p>As technologies, we&#8217;re constantly seeking tools that enhance our productivity, accelerate our workflows, and ultimately, help us bring ideas to life faster. Over a period of about 45 days, I embarked on an intense development sprint, leveraging AI models and a specialized coding assistant, Kilo Code, to build four distinct applications now successfully deployed in&#8230;<\/p>\n","protected":false},"author":2,"featured_media":519838,"comment_status":"closed","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"_kad_blocks_custom_css":"","_kad_blocks_head_custom_js":"","_kad_blocks_body_custom_js":"","_kad_blocks_footer_custom_js":"","ngg_post_thumbnail":0,"episode_type":"","audio_file":"","podmotor_file_id":"","podmotor_episode_id":"","cover_image":"","cover_image_id":"","duration":"","filesize":"","filesize_raw":"","date_recorded":"","explicit":"","block":"","itunes_episode_number":"","itunes_title":"","itunes_season_number":"","itunes_episode_type":"","_kad_post_transparent":"","_kad_post_title":"","_kad_post_layout":"","_kad_post_sidebar_id":"","_kad_post_content_style":"","_kad_post_vertical_padding":"","_kad_post_feature":"","_kad_post_feature_position":"","_kad_post_header":false,"_kad_post_footer":false,"_kad_post_classname":"","footnotes":""},"categories":[441],"tags":[471,930,871,1017,989],"class_list":["post-519831","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-tech-talk","tag-ai","tag-ai-series","tag-genai","tag-tokens","tag-vibe-coding"],"taxonomy_info":{"category":[{"value":441,"label":"Tech Talk"}],"post_tag":[{"value":471,"label":"AI"},{"value":930,"label":"AI Series"},{"value":871,"label":"GenAi"},{"value":1017,"label":"Tokens"},{"value":989,"label":"Vibe Coding"}]},"featured_image_src_large":["https:\/\/jorgep.com\/blog\/wp-content\/uploads\/FeaturedImage-Jorgep-UsageStatsLLM.png",1024,512,false],"author_info":{"display_name":"Jorge Pereira","author_link":"https:\/\/jorgep.com\/blog\/author\/jorge\/"},"comment_info":0,"category_info":[{"term_id":441,"name":"Tech Talk","slug":"tech-talk","term_group":0,"term_taxonomy_id":451,"taxonomy":"category","description":"","parent":0,"count":678,"filter":"raw","cat_ID":441,"category_count":678,"category_description":"","cat_name":"Tech Talk","category_nicename":"tech-talk","category_parent":0}],"tag_info":[{"term_id":471,"name":"AI","slug":"ai","term_group":0,"term_taxonomy_id":481,"taxonomy":"post_tag","description":"","parent":0,"count":147,"filter":"raw"},{"term_id":930,"name":"AI Series","slug":"ai-series","term_group":0,"term_taxonomy_id":940,"taxonomy":"post_tag","description":"","parent":0,"count":152,"filter":"raw"},{"term_id":871,"name":"GenAi","slug":"genai","term_group":0,"term_taxonomy_id":881,"taxonomy":"post_tag","description":"","parent":0,"count":83,"filter":"raw"},{"term_id":1017,"name":"Tokens","slug":"tokens","term_group":0,"term_taxonomy_id":1027,"taxonomy":"post_tag","description":"","parent":0,"count":3,"filter":"raw"},{"term_id":989,"name":"Vibe Coding","slug":"vibe-coding","term_group":0,"term_taxonomy_id":999,"taxonomy":"post_tag","description":"","parent":0,"count":13,"filter":"raw"}],"_links":{"self":[{"href":"https:\/\/jorgep.com\/blog\/wp-json\/wp\/v2\/posts\/519831","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/jorgep.com\/blog\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/jorgep.com\/blog\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/jorgep.com\/blog\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/jorgep.com\/blog\/wp-json\/wp\/v2\/comments?post=519831"}],"version-history":[{"count":12,"href":"https:\/\/jorgep.com\/blog\/wp-json\/wp\/v2\/posts\/519831\/revisions"}],"predecessor-version":[{"id":519860,"href":"https:\/\/jorgep.com\/blog\/wp-json\/wp\/v2\/posts\/519831\/revisions\/519860"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/jorgep.com\/blog\/wp-json\/wp\/v2\/media\/519838"}],"wp:attachment":[{"href":"https:\/\/jorgep.com\/blog\/wp-json\/wp\/v2\/media?parent=519831"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/jorgep.com\/blog\/wp-json\/wp\/v2\/categories?post=519831"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/jorgep.com\/blog\/wp-json\/wp\/v2\/tags?post=519831"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}