Hacker Newsnew | past | comments | ask | show | jobs | submit | TimTheTinker's commentslogin

These "AI rewrite" projects are beginning to grate on me.

Sure, if you have a complete test suite for a library or CLI tool, it is possible to prompt Claude Opus 4.6 such that it creates a 100% passing, "more performant", drop-in replacement. However, if the original package is in its training data, it's very likely to plagiarize the original source.

Also, who actually wants to use or maintain a large project that no one understands and that doesn't have a documented history of thoughtful architectural decisions and the context behind them? No matter how tightly you structure AI work, probabilistic LLM logorrhea cannot reliably adopt or make high-level decisions/principles, apply them, or update them as new data arrives. If you think otherwise, you're believing an illusion - truly.

A large software project's source code and documentation are the empirical ground-truth encoding of a ton of decisions made by many individuals and teams -- decisions that need to be remembered, understood, and reconsidered in light of new information. AI has no ability to consider these types of decisions and their accompanying context, whether they are past, present, or future -- and is not really able to coherently communicate them in a way that can be trusted to be accurate.

That's why I can't and won't trust fully AI-written software beyond small one-off-type tools until AI gains two fundamentally new capabilities:

(1) logical reasoning that can weigh tradeoffs and make accountable decisions in terms of ground-truth principles accurately applied to present circumstances, and

(2) ability to update those ground-truth principles coherently and accurately based on new, experiential information -- this is real "learning"


> Sure, if you have a complete test suite for a library or CLI tool

And this is a huge "if". Having 100% test coverage does not mean you've accounted for every possible edge or corner case. Additionally, there's no guarantee that every bugfix implemented adequate test coverage to ensure the bug doesn't get reintroduced. Finally, there are plenty of poorly written tests out there that increase the test coverage without actually testing anything.

This is why any sort of big rewrite carries some level of risk. Tests certainly help mitigate this risk, but you can never be 100% sure that your big rewrite didn't introduce new problems. This is why code reviews are important, especially if the code was AI generated.


You raised very good points, however, what you typed negatively affects the shell game (as to what "AI" companies are often really doing) and partial pyramid scheme.

People seem not to realize that AI companies can not only plagiarize someone's original source code, but any source code that people connected to it are feeding and uploading to it. The shell game is taking Tom's code (with a few changes) and feeding it to Bill (based on prompts given). Both Tom and Bill are paying fees to the AI company, yet don't realize their code (along with many others) can be spit back at them.

You, the humans, are doing a lot of the work, and many don't realize it. Because Tom is not realizing someone has or is working on something similar. The AI company is connecting Tom and Bill together, without either of them realizing it. If they type in the right prompt, the search then feeds back that info. It's not the only thing going on or only way things work, but it is part of it, that is often not publicly acknowledged.


> Sure, if you have a complete test suite for a library or CLI tool, it is possible to prompt Claude Opus 4.6 such that it creates a 100% passing, "more performant", drop-in replacement.

This was the "validation" used for determining how much progress was made at a given point in time. Re training data concerns, this was done and shipped to be open source (under GPLv2) so there's no abuse of open source work here imo

Re the tradeoffs you highlight - these are absolutely true and fair. I don't expect or want anyone to just use ziggit because it's new. The places where there performance gains (ie internally with `bun install` or as a better WASM binary alternative) are places that I do have interest or use in myself

_However_, if I could interest you in one thing. ziggit when compiled into a release build on my arm-based Mac, showed 4-10x faster performance than git's CLI for the core workflows I use in my git development


I suppose "Project X has been used productively by Y developers for Z amount of time" is a decent-enough endorsement (in this case, ziggit used by you).

But after the massive one-off rewrite, what are the chances that (a) humans will want to do any personal effort on reading it, documenting it, understanding it, etc., or that (b) future work by either agents or humans is going to be consistently high-quality?

Beyond a certain level of complexity, "high-quality work" is not just about where a codebase is right now, it's where it's going and how much its maintainers can be trusted to keep it moving in the right direction - a trust that only people with a name, reputation, observable values/commitments, and track record can earn.


Perhaps there's a future where "add a new feature" means "add tests for that feature and re-implement the whole project from scratch in AI".

But that approach would create significant instability. You can't write tests that will cover every possible edge case. That's why good thinking & coding, not good tests, is the foundation of good software.


> the latest wifi drivers for my brand new wifi 7 motherboard were too flaky

A GL.iNet travel router in WiFi to ethernet bridge mode is an excellent stopgap until Linux support arrives. It also has the benefits of (a) taking with you on trips for safer/easier internet use (use your home SSID, even auto-VPN traffic if you want) and (b) letting you plug in other wired-only devices adjacent to the computer.

Here are their travel routers filtered to just those that support WiFi 6 and 7: https://store-us.gl-inet.com/collections/travel-routers?filt...


Why not put all of that into a skill file? The context overhead from an MCP connection is significantly higher.

You're right actually. Exa's MCP server is stateless, just a REST wrapper. A skill + CLI would do the same job with way less context cost. Someone already built that (https://github.com/tobalsan/exa).

I've had good results from creating a command-line bash utility (and associated skill) that wraps and supplies credentials opaquely to a cli tool.

Same here. It's not airtight, the agent could technically read the wrapper or env vars, but in practice it doesn't bother. Good enough for most setups.

Stockholder AI sentiment hasn't yet incorporated consumer AI sentiment, it seems.

> lose money

I think that's largely a myth. Anthropic charges something like 10x their internal $/MTok for external resellers.


TI-83 Basic was the first programming language I really felt like I had mastered. For a while in my first CS college class I was writing code in TI basic and translating it to C++. Drugwars and Bowling were the two really impressive games written in TI-Basic.

But discovering z80 assembly was like magic. It was incredibly exciting to go to my dad's office at the university where he worked (where computers had 2 T1 internet lines) to download and try assembly games when they first burst on the scene (I was in 8th grade). Bill Nagel blew my mind with Turbo Breakout and Snake, and later AShell, Penguins, and grayscale Mario... but the best executed and most replayable games I think were Sqrxz and ZTetris on the TI-86 by Jimmy Mardell. Honorable mention to Galaxian and Falldown. I once downloaded the z80 assembly source for a game, printed it to about an inch of paper, and carried it around for weeks trying to understand it...

It was also really cool for some reason (and would often brick the calculator until you took the batteries out) to type random hex pairs into a program and execute it as assembly. "C063" run as assembly - syntax was the random looking Send(9PrgmA where PrgmA is where you typed the hex code - on a TI-83 would scroll tons of random text in an infinite loop.

Does anyone remember the TI website wars? TI Files (later TI Philes) was "so much more awesome" than "the lowly weak ticalc.org"... but look which one is still around :-)


I'm amazed ticalc.org is still alive and kicking. So much nostalgia. Joltima was what convinced me to learn assembly. So far ahead of its time on the TI-86. Full featured RPG with turn-based combat on a graphing calc. Glad the history is still accessible online.


Did you build anything in z80 assembly?


> Bill Nagel

now there's a name that inspired awe in my 12 year old mind.

i didn't know at first how he was able to make those incredible games, only understanding TI Basic myself. mindblowing stuff.


Yes, I checked ticalc.org regularly in hopes of seeing an update to the Zelda: Link’s Awakening port proof of concept demo.


That’s funny I use to do the same in high school 15 years ago


It was Andreas Ess for me. PlaneJump.

That opened my eyes to the world of Assembly, which in turn turned me on to the demoscene, and off I went into a truly magical subculture!


Any demos online that you worked on?


A lot of those old calculators can be had on eBay for less than $30 shipped. I've been building up a small collection - 73, 81, 82, 83, 85, 86, 89


I am enjoying my 92+


I have an 89 but have never owned a 92. Do you think it's worth it to mess around with Fargo and the various games?


I used Webmin[0] back in the day, I wonder how more recent server web UIs like Cockpit stack up.

[0] https://webmin.com/


I looked at this and said "They've made Webmin again."


with all the potential vulnerabilities of 400 npm dependencies (I'm guessing. It's probably higher)


We don't need security. We already know your age and who you are. /s

It's neutered and not as full featured, but not bad in a pinch. All of these web admin tools are hacks that call out to shell scripts and whatnot. It requires a lot of conditional behavior and/or vertical integration. "Linux" has no consistent API for control, so its all duct tape. Webmin is the same, tbh (swap perl for whatever cockpit is written in)

45Drives uses cockpit as the UI layer of their "Houston" operating system. https://www.45drives.com/community/articles/New-Operating-Sy...


Cockpit tends to be less ad-hoc than others ime. Often it'll use dbus on the backend.

It's also socket activated, which is nice.


> All of these web admin tools are hacks that call out to shell scripts and whatnot

I love the smell of RCEs in the morning.


Yes, agreed. Though speed limits higher than 75 are not something I will ever support.*

* Unless we're talking about removing a speed limit altogether and regulating unsafe driving using other criteria.


Autonomous vehicles following proper signalling before lane changes can be safe at arbitrary speeds (see Autobahns working at all). Humans, we should limit passing speed to roughly ~5 mph delta between adjacent lanes and leave it at that.

Humans with adequate following distance in the entire lane can probably manage 10 mph delta. I routinely travel dozens of miles very safely at ~80 with the flow of traffic (including the cops), and been stressed out at 55 in the carpool lane through stop and go traffic in the right-hand lanes due to on ramps/offramps.


What happens at 76mph?


Same thing that happens at 77mph :)

I think 75 is memorable and roughly in the region where the tradeoff between increased kinetic energy and decreased time to arrival per additional unit of velocity becomes untenable.


> the tradeoff between increased kinetic energy and decreased time to arrival per additional unit of velocity becomes untenable

Sounds like a warning page out of the back of a 94 Geo Metro owner's manual.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: