Trying out a new format this week. Instead of links, here’s some shorts. Too little for a post, too much for a link. This week we have
- Power variation in AI data centres
- Using Claude Code for Terraform
- Adopting new CLI tools
- Historical maps
Power variation in AI data centres
(Link requires registration and full article isn’t free, but there’s a massive preview which is all I have access to)
A new to me issue with AI workloads - training runs are very, very spiky in their power use.
Google have released this graph
and Meta have this in the Llama 3 paper
During training, tens of thousands of GPUs may increase or decrease power consumption at the same time, for example, due to all GPUs waiting for checkpointing or collective communications to finish, or the startup or shutdown of the entire training job. When this happens, it can result in instant fluctuations of power consumption across the datacenter on the order of tens of megawatts, stretching the limits of the power grid. This is an ongoing challenge for us as we scale training for future, even larger Llama models.
So companies are looking at massive battery banks to smooth the grid load
Using Claude Code for Terraform
Because I’m a rock and roll sort of person, I spent last night trying a coding interview question on Terraform and AWS using Claude Code in mostly vibe mode and I’ve [published the results](https://github.com/ArthurClune/claude-terraform-experiment, including prompts.
Various plan files used are commited in the repo when used and then deleted afterwards, so it should be straightfoward to follow the evolution. I’ve left this as Claude generated except for a few tidy ups to docs, so you’ll see where it’s opened up to 0.0.0.0/0 etc (see the prompts file). This is a test for how Claude does, not an attempt to produce production code!
This was a relatively expensive use of Claude, presumably because of all the tool use and large inputs/outputs from checkov. About $18.
The bar for coding for an quick interview test (this exercise is expected to take at most 1-2 hours, and that’s what I spend on it) is getting very high indeed.
I’m still wary of using LLM generated terraform - the potential blast radius is massive! Even with good PRs and code review, the sheer volume of code that can be generated can easily lead to ’lgtm’ type ‘reviews’
Adopting new CLI tools
I’ve been using a linux/posix cli for a very long time. But every so often it’s good to change it up. In the last six months I’ve
-
Changed from iTerm2 to Ghostty. I was skeptical that I’d notice any difference, but it does actually feel noticably faster
-
Changed from powerline prompts to Starship. No new functionality, but it’s quicker
-
Adopted fzf and integrated it with zsh so now file completion and history have fuzzy matching
The combined effect is to make things feel fresh, new and faster
Historical maps
Open Street Map based, but with borders, railways etc over time