Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

For float and int8, 1024 does indeed outperform 2048. However, for binary, 2048 outperforms 1024.

For our general-purpose embedding model (voyage-3-large), embedding vectors with 2048 dimensions outperform 1024 across the board: https://blog.voyageai.com/2025/01/07/voyage-3-large/



Do you have any insights into this? Perhaps scaling degradation with Matryoshka learning?




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: