Hacker Newsnew | past | comments | ask | show | jobs | submit | more nialv7's commentslogin

pretty easy to understand - you pay for o3, whereas GPT-4o is free with a usage cap so they want to keep you engaged and lure you in.


I wouldn't be surprised if Andrew Pinski was just wrong. It's anecdotal but my impression of him isn't very good.


> amateur australian chemist

I mean, he has a PhD...


> he has a PhD

> "the judge concluded that Lidden had mental health issues and displayed no malicious intent"

Cause and effect.


It's a kernel in the library operating system sense [0]. A kernel doesn't have to run in a privileged context.

[0]: https://en.wikipedia.org/wiki/Operating_system#Library


I hadn't heard the term library operating system before, but what is described at that link is nothing like your project.

The "kernel" referred to at that link is a full actual kernel, with networking, proceed scheduling, device drivers and hardware access, etc. It means the full implementation of those features. The reason it can be linked into the application's process as a library is because it's taking about a situation where there's only a single application running on the whole machine (presumably a virtual machine). The kernel and application then run together in a single memory space, with the kernel code (maybe the whole thing?) running in kernel mode.

You have implemented the user mode API that makes system calls to the (Windows) kernel. That is not a kernel in any sense. It is confusing and wrong to call it a kernel.


Isn't that what user-mode Linux is?


User mode Linux does actually run something like a real kernel in user space. OP is just providing a POSIX API with no kernel implementation whatsoever. It just forwards those calls to the Window's API. The library OS concept linked to above is different again – that is an actual kernel running in kernel mode but as part of a unikernel (kernel code and application code linked together into a single process).

Still pretty cool but overshadowed by the claim that they've written a whole kernel.


I see. Thanks!


Cygwin is not wine. Wine lets you run existing windows executables unmodified on Linux. Cygwin does not let you run ELFs on windows.


Good point, using wine as the interpreter for PE executables is another cool part of the WINE project.

But it is the case that you can build a Windows application from source on Linux and link it with libwine, in the same way you can build a Unix program from source on Windows and link it with cygwin.dll -- I seem to remember some time ago CorelDRAW (or at least Corel Photo-Paint) was made available for Linux that way, officially supported by Corel.


> we uncover that RL-trained models excel at low k (e.g., pass@1) but are consistently outperformed by base models at high k (e.g., pass@256).

This is a weak argument. I think I get what we are trying to say, but let's take this to the extreme, say pass@10^10^100. Just like a group of monkeys could write Shakespeare if given enough time, a complete random model could probably outperform an RL-trained model at pass@10^10^100. Would we then say the random model can reason too?

Of course the correct reasoning trace will be in the base model's distribution, just like any other well-formed, coherent paragraph. Kind of makes me think, maybe sampling efficiency _is_ intelligence?


If this was just the effect you mention you would not expect the base model to surpass the RL model though. Plus their k are much smaller than that.

I think it's a very interesting and meaningful study.


The authors of the paper address this argument in the QA section.


When the article starts with an AI generated image that adds nothing to the explanation, it tends to make me suspicious if the article itself was written by an AI as well...


I think GP is actually asking whether we can determine if one enters a steady state, i.e. tape no longer changes.


> Works great in Chrome and Slack, and not at all on Android devices.

Hurmm it does work for me in Chrome on Android.


and back then in 2017, AI hasn't really been productized yet, people behind Transformer were researchers, publishing their results were the norm.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: