Hacker Newsnew | past | comments | ask | show | jobs | submit | hopfenspergerj's commentslogin

What on earth are these questions? They don’t resemble any real use of an llm for work.


No it is not horrible to work with. Even the “standard” settings are quite lax.


I think our respective comment votes disagree.


If you’re meta and you have to defend the AI by admitting “it’s not really intelligent and everything it says is bullshit”, that’s not a position of strength.


Particularly when it's likely that despite the AI bullshitting in the absence of data on its creators, it's also incidentally true that the "black queer momma" persona is a few lines of behavioural stereotypes composed by white male programmers without any particular experience of black queer mommas.


I dont think that necessarily applies when you could easily make a training set from some actual black American people’s writings on the internet or book, or even an individual that self identified that way and train on all their writings around the internet, and result in those same stereotypes when you ask an AI to create such a profile

You dont need a black American engineer or product manager to say “I approve this message” or “halt! my representation is more important here and this product would never fly” as they are just not person the data set was trained on, even if you asked an AI to just create the byline on the profile for such a character

its weirder, and more racially insensitive, for people to be vicariously offended on behalf of the group and still not understand the group. In this case, the engineer or product manager or other decision maker wouldnt have the same background as the person that would call themselves “momma”, let alone it not mattering at all, if you can regurgitate that from a training set


I mean, sure, some programmers with very little experience of queer black mommas could, hypothetically, be so good at curating information from the internet and carrying out training exercises that they created a persona that convincingly represents a queer black momma. Do we think this is what happened in this instance?

In which, ironically, the bot called it a "glaring omission"


the bot is echoing sentiments of comment sections it was trained on and had no idea of its origins

its acting aware and sensitive but only has information about the tech sector as a whole

my critique is about how the standards being applied are dumb all the way down. the standards are not actually that enlightened even if there was more representation congruent with the race/identity being caricatured as representation. nullifying the whole criticism.

the training set is the only thing thats important with a language model. and its just a symptom of dead internet theory, as even the persona’s byline was probably generated by a different language model.


well, yeah, I acknowledged the bot has no idea of its actual origins in my first post. the point is that at some point some actual product manager thought that creating this persona (probably a generic training set plus a few lines of human-authored stereotypes as prompt) and releasing it to the public as an authentic, representative personality was a good idea. Unlike the product manager, the bot's bullshitting was context aware enough to express the sentiment that this was a bit embarrassing


If the intent is to make a recognizable caricature and apply labels to it (cough stereotype), you don't have someone draw themselves. And it's really looking like stereotyping is their intent.


Ruff can automatically upgrade all of the issues you mentioned to match your target minimum python version.


Good to know, thanks!


Are the pandas people considering this as the default string type? Seems like it would be a slam dunk.


That is something I’d like to see but I don’t want to wade into the already very complicated discussion around arrow strings in pandas. If a Pandas developer wanted to take this on I think that would make things easier since there’s so much complexity around strings in Pandas.

That said there is a branch that gets most of the way there: https://github.com/pandas-dev/pandas/pull/58578. The remaining challenges are mostly around getting consensus around how to introduce this change.

If NumPy had StringDType in 2019 instead of 2024 I think Pandas might have had an easier time. Sadly the timing didn’t quite work out.


I'm looking at the very first example, and I'm a little confused. The function `home()` displays a list of messages, but they aren't passed into `home()`. Instead, `messages` is basically a global variable, and some other functions can append messages to it. Then I went looking at some more examples, and I see this pattern repeated. Is this how you're supposed to build webapps with this package? How does it isolate the list of messages for different users?


Which example? I see global vars in a couple of the demos. The game state for the Game of Life makes sense, since it is intended to be shared. The `messages` list in the Chatbot demo is definitely NOT how you'd build a multi-user application, that's mainly showing the styling aspect.

In general, you'd have an actual database and make it so users can only see their own data! See https://github.com/AnswerDotAI/fasthtml/blob/main/examples/a... which adds a filter to queries and DDL statements to ensure that the user can only see/edit their own todos.


You can obviously compensate by changing other taxes to be more progressive, this is such a silly argument.


Sure but that doesn't happen. Start by doing that, not by implementing the regressive part first.


A nuance is perhaps lost, I'm not saying don't do it, merely that simplistic messaging feeds into the binary crisis.


This is an example of bad code.


Maybe you can expand upon that. Now we have no way of knowing why you think it's bad code.


Quick critique: module contains functions with many parameters, many branches, deep nesting, and multiple return points.


those nested if - for - for - if loops are horrendously difficult to understand.

take the fn starting at line 387. they comment why they do certain imports, but this function is comparatively underdocumented. it's not easy to wrap my head around the control flow. some bits are nested about 6 levels too deep for comfort, there are too many positions from which it can return or raise, and the function is about 3x too long

really difficult to grok what is happening here.


It’s not horrible, but I found a few odd things, like f-strings w/o params, long cli options with underscores, non-pythonic if == 0, etc.


Also the main god function is incredibly long and nested, as others mentioned. Nested --> long lines --> black making a mess.


Ensemble of any number GPT 3.5 agents is less accurate than one call to GPT-4.


It's funny because GPT-4 is actually a pile of 3.5s. You just need to set it up correctly.


I guess it's the difference between an ensemble and a mixture of experts, i.e. aggregating outputs from (a) model(s) trained on the same data vs different data (GPT-4). Though GPT-4 presumably does not aggregate, but it routes.


> GPT-4 is actually a pile of 3.5s

I understand the intension and reference you're making. I bet the implementation of GPT-4 is probably something along those lines. However, spreading speculation in definitive language like that when the truth is unknown is dishonest, wouldn't you agree?


Sure, I could it put it less definitively, but realistically, what else can it be? The transformer won't change much and all of the models, at the core use it. It's a closely guarded secret because it's easy to replicate.


That was me on Reddit. I emailed him asking about some of his work on invariant means in the 1970s. He said “I had no taste then”, and told me what a waste of time it was.

At that point, I decided to go into data science instead of trying to get a post doc…


How do you feel about that decision now? I left math after my PhD and often find myself feeling nostalgic about it.


Always something to look forward to in [early] retirement to keep the mind active.


Consider applying for YC's Winter 2026 batch! Applications are open till Nov 10

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: