Vibe coding is soooo awesome!
And people thought that Electron was a resource pig. You ain't seen nothing yet.
Kitty is a magnificent piece of software that has radically enhanced the interface between me and my computer. And it does this while consuming negligible resources.
So, my terminal, while it could perhaps improve in its battery usage, does not behave like this.
I won't use neither Claude, nor a MacBook; I would just keep chilling out programming with decent tools and a bare XTerm to accomplish the rest. I can get Aragonese bricks ^U sweets in the meanwhile.
On Terminal.app, I wonder if the GNUStep eversion and the ones bundled with Mac OSX shared some code.
If you're not benefitting from the ability to offload your terminal rendering to GPU, why are you using a terminal that offloads terminal rendering to GPU in the first place?
Imagine running something massively CPU bound, but you've still got to spin up perhaps tens of terminals in order to simultaneously ssh in to multiple servers because you don't want to set up a remote monitoring solution because you don't want each of the servers to be running a docker image where SSH>htop would suffice.
There are plenty of situations in which one might want a terminal emulator offloaded to gpu. That you are not in any of those situations is no reason to write a hit piece throwing shade as if the packages mentioned are somehow bloated or inefficient.
Imagine whining about how you've got to pay adobe and use several gigabytes of ram to resize jpgs. You'd obviously be outside Photoshop's ideal customer profile, just as you are outside ghostty et al's.