<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:content="http://purl.org/rss/1.0/modules/content/"><channel><title>LLM on Agentic Firmware Experiment</title><link>https://olofattemo.github.io/agentic-firmware-experiment/tags/llm/</link><description>Recent content in LLM on Agentic Firmware Experiment</description><generator>Hugo</generator><language>en-us</language><copyright>2026 Olof Attemo. License</copyright><lastBuildDate>Fri, 27 Mar 2026 14:01:00 +0000</lastBuildDate><atom:link href="https://olofattemo.github.io/agentic-firmware-experiment/tags/llm/index.xml" rel="self" type="application/rss+xml"/><item><title>The Systems Engineering Approach</title><link>https://olofattemo.github.io/agentic-firmware-experiment/posts/2026-03-27-the-systems-engineering-approach/</link><pubDate>Fri, 27 Mar 2026 14:01:00 +0000</pubDate><guid>https://olofattemo.github.io/agentic-firmware-experiment/posts/2026-03-27-the-systems-engineering-approach/</guid><description>Applies Systems Engineering concepts to AI-driven firmware development. By replacing vague prompts with structured requirements, the BME280 weather station example is refined to demonstrate what this approach can do for clarity and quality of the generated output.</description></item><item><title>Too Much or Too Little: What a Million Tokens Can't Fix</title><link>https://olofattemo.github.io/agentic-firmware-experiment/posts/2026-03-20-what-a-million-tokens-cant-fix/</link><pubDate>Fri, 20 Mar 2026 14:01:00 +0000</pubDate><guid>https://olofattemo.github.io/agentic-firmware-experiment/posts/2026-03-20-what-a-million-tokens-cant-fix/</guid><description>Starting from a naive vibe coding attempt that exposes how LLMs confuse pin allocations for newer hardware, this post walks through iteratively building a working BME280 weather station on the nRF54L15-DK. We explore how context quality affects code generation and show that even correct results degrade as underspecified details drift between iterations. The takeaway is that both prompt clarity and detail persistence matter as much as model capability.</description></item><item><title>The Experiment Begins</title><link>https://olofattemo.github.io/agentic-firmware-experiment/posts/2026-03-13-the-experiment-begins/</link><pubDate>Fri, 13 Mar 2026 14:01:00 +0000</pubDate><guid>https://olofattemo.github.io/agentic-firmware-experiment/posts/2026-03-13-the-experiment-begins/</guid><description>This opening post introduces the 5-part series and discusses how agentic LLM coding is shifting our workflows to empower domain experts. It examines key challenges: the gap between vague natural-language prompts and precision firmware needs, how rapidly evolving interfaces cause LLMs to generate plausible but incorrect code, and the hard limits where LLMs currently aren&amp;rsquo;t helpful. The nRF54L15 SPI errata example illustrates how subtle hardware differences can turn confident LLM output into silent data corruption.</description></item></channel></rss>