Hacker News new | past | comments | ask | show | jobs | submit | more lqcfcjx's comments login

This is a great idea. Just a quick feedback: when I search for "llm", I got quite random results.


Well the results are indeed "quite random"... A good test case to improve the spam filtering with :)

It seems to work much better with capital letters for some reason: https://aboutideasnow.com/?q=LLM


After reading thru all the comments, honestly I still don't get the point of this system. What is potential practical value or applications of this model?


This might be a stupid question - even if there's no quality degradation from 10M context, will it be extremely slow in reference?


> I think that safe romantic and adult roleplay with a premium feel will be essential components towards the success of AI companionship in the long-run

I totally agree with this direction. I think it's a wrong decision for some companies to try to make it friendly to a wider age range, but in fact, actions / adult roleplay is what users really need. Who needs a virtual partner that only does safe chatting?

But honestly speaking, you need to make your service more performant. Responses are way too slow, so it's quite hard for me to get a sense of its quality.

I'm also curious about what underlying models are you using? Do you train / fine tune your own llms or it's on top of oai?


Thanks for the feedback!

Frankly, because of resource capacities, the site reserves faster speeds for supporters (we're community supported given the financial complexity of adult content) but I've been seriously thinking lately about how that might be affecting the first user experience. Currently toying with some changes.

The site uses a fine-tune of llama-2-13b, specifically pygmalion-2-13b. I've found it to be particularly engaging and creative in its storytelling, but as you can guess by the size, it has some tradeoffs. I'm currently experimenting with new models, would love to hear suggestions.


This is very impressive. I know in general people are iffy about research benchmark. How does it work to evaluate text-to-video types of use cases? I want to have some intuition on how much this is better than other systems like pika quantatively.


This is a great product! Simple and elegant. I really appreciate your effort to build and share this.

I think people often want to build something big and even complex, but such a simple tool is already very useful for a large audience.


I really have mixed feeling about this. On one hand, having long term memory seems an obviously necessary feature, which can potentially unlock a wide variety of use cases - companionship, more convenience and hopefully provide more personalized responses. Sometimes I find it too inconvenient to share full context (e.g. I won't share my entire social relationship before asking advice about how to communicate with my manager).

However, I wonder to what degree this is a strategic move to build the moat by increasing switch cost. Pi is a great example with memory, but I often find this feature boring as 90% of my tasks are transactional. In fact, in many cases I want AI to surprise me with creative ideas I would never come up with. I would purposely make my prompt vague to get different perspectives.

With that being said, I think being able to switch between these 2 mode with temporary chat is a good middle ground so long as it's easy to toggle. But I'll play with it for a while and see if temporary chat becomes my default.


I'm very impressed by the recent AI progress on making models smaller and more efficient. I just have the feeling that every week there's something big on this space (like what we saw previously from ollama, llava, mixtral...). Apparently the space for on-device models are not fully discovered yet. Very excited to see future products on that direction.


> I'm very impressed by the recent AI progress on making models smaller and more efficient.

That's an odd comment to place in a thread about an image generation model that is bigger than SDXL. Yes, it works in a smaller latent space, yes its faster in the hardware configuration they've used, but its not smaller.


my bad, I misread the post and got the impression that it's a way smaller model. Thanks for correcting me.


+1. I used to be a big fan of bullet journal and keep all my todo list / planning in my notebook. Sometimes I'm not sure if this really has more practical value or it just makes me FEEL better.

But now I'm close to 100% on note taking apps like bear or ios default notes.


This looks very interesting as majority of the world is moving towards bigger "do everything for you" LLMs. Just few preliminary questions after I glanced over the repo and blogs:

- there are also many small models that try to "do everything" like phi, mistral, etc., do you find llmware have better quality and performance? - how does the rag relate features compare with other tools like llamaindex?


Join us for AI Startup School this June 16-17 in San Francisco!

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: