Hacker Newsnew | past | comments | ask | show | jobs | submit | unignorant's commentslogin

I really enjoyed this article but the claim of no literary fiction making the Publishers Weekly yearly top 10 lists since 2001 isn't really true:

https://en.wikipedia.org/wiki/Publishers_Weekly_list_of_best...

https://en.wikipedia.org/wiki/Publishers_Weekly_list_of_best...

It is true that there isn't that much literary stuff that breaks through, and the stuff that does is usually somewhat crossover (e.g., All the Light We Cannot See in 2015 or Song of Achilles in 2021) but it exists. These two books are shelved under literary codes (though also historical). Song of Achilles in particular is beautifully written and a personal favorite of mine, at least among books published in recent years.

Then there are other works like Little Fires Everywhere and The Midnight Library that I might not consider super literary but nonetheless are also often considered so by book shops or libraries (e.g., https://lightsailed.com/catalog/book/the-midnight-library-a-... the lit fic code is FIC019000).

I was really surprised that Ferrante's Neapolitan series, the best example (I would have thought) of recent work with both high literary acclaim and popular appeal, did not actually make the top 10 list for any year.


Yeah, and looking through the lists makes one suspect that there's a problem of incommensurate measurements... There's a lot of 'very hungry caterpillar' in the recent lists, but I'm unsure whether children's books were even in the running in the 1960's. Or else there's been a revolution in buying books for children since the 60's, which, honestly, I wouldn't be sad about...


yeah, it seems likely the underlying task here (one reasoning step away) was: replace as many fp32 operations as possible in this kernel with fp16. i'm not sure exactly how challenging a port like that is, but intuitively seems a bit less impressive

maybe this intuition is wrong but would be great for the work to address it explicitly if so!


Only seems to have done that in a couple places, like the MatMul. The softmax kernel (https://github.com/ScalingIntelligence/good-kernels/blob/mai...) seem to be entirely bog-standard, and the layernorm kernels are only slightly more interesting.


I looked at the softmax kernel and the cast that it does from a float* to a float4* is extremely brittle -- it's trivial to break by offsetting the input slightly.

Very likely a kernel for a standard library could not employ such a trick that relies on alignment of input pointers. Certainly not without a fallback.


I do a lot of ML work too and recently gave NixOS a try. It's actually not too hard to just use conda/miniconda/micromamba to manage python environments as you would on any other linux system with just a few lines of configuration. Pretty much just add micromamba to your configuration.nix plus a few lines of config for nix-ld. Many other python/ML projects are setup to use docker, and that's another easy option.

I don't have the time or desire to switch all my python/ML work to more conventional Nix, and haven't really had any issues so far.


This technique doesn't actually use RL at all! There’s no policy-gradient training, value function, or self-play RL loop like in AlphaZero/AlphaTensor/AlphaDev.

As far as I can read, the weights of the LLM are not modified. They do some kind of candidate selection via evolutionary algorithms for the LLM prompt, which the LLM then remixes. This process then iterates like a typical evolutionary algorithm.


Thanks for sharing this! I occasionally use google translate and/or GPT4 for similar purposes, but your tool makes the workflow a bit simpler.

I've found creative writing in a target language is great for learning.


Happy to share, thanks!

I also use GPT-4 for explaining the meaning of sentences in more detail (as in JimDabell’s comment). Often my questions are like “how would a native speaker say this colloquially” - I’ve found it really valuable to be able to have a back-and-forth on why something works the way it does


I sort of similarly used LLMs and speech synthesis tools to make a prototype that could generate short (<10min) podcasts in Mandarin on any topic I specified. Being interesting is less important in a language learning context, though it's notable that I haven't used the tool much and prefer listening to Mandarin audiobooks and real human podcasts, perhaps because they are more interesting.


Interesting, has that lanuage learning podcast been useful for you?


I spent a summer at the Santa Fe Institute in 2010 and can say that McCarthy certainly spent a lot of time there. I'm not sure he was a copyeditor, exactly, but SFI is a nice place to hang out for any sort of creative work -- beautiful building and landscape, and very open and collaborative atmosphere.

The institute is on the top of a big hill and I'll always remember how he gave me a lift one day as I was walking up.


I've heard good things about Pilot: https://pilot.com


The immune system is extremely complicated, but very broadly: CD8+ T-cells (also known as killer T-cells) kill infected cells directly, whereas CD4+ T-cells (also known as helper T-cells) release signals that guide many aspects of immune response, including activating CD8+ cells


Among other things, we use it to provide a payment solution for merchants with less friction and a lower fee structure:

-- https://nash.io/payments/

-- https://blog.nash.io/nash-link-why-merchants-will-embrace-si...


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: