I'm not suggesting great work didn't go into this, but I was able to build a very crude version of this on GPT 3.5. It was evident then that the real power of these models isn't in chat, but in a sandboxed environment where they can recursively iterate on solutions and feedback from their sandboxed environment. I was able to feed mine small applications with bugs and have it comb through and find the bugs, write and apply solutions, write tests for solutions, etc.
Adding features was too hard to implement in my limited spare time, but was clearly possible. I would have needed some form of test running for UIs or CLIs, and I wasn't prepared to go that deep on a project I wasn't going to get much out of.
It was crude and overly specific to what I was trying to get it to do, but it worked well enough to convince me that someone smarter than me could make a capable and truly useful version of it that could actually impact the industry meaningfully.
Adding features was too hard to implement in my limited spare time, but was clearly possible. I would have needed some form of test running for UIs or CLIs, and I wasn't prepared to go that deep on a project I wasn't going to get much out of.
It was crude and overly specific to what I was trying to get it to do, but it worked well enough to convince me that someone smarter than me could make a capable and truly useful version of it that could actually impact the industry meaningfully.