From the announcement
“As of now, we have mined 1,580 PySpark tests from the Spark codebase, among which 838 (53.0%) are successful on Sail. We have also mined 2,230 Spark SQL statements or expressions, among which 1,396 (62.6%) can be parsed by Sail”
Kinda early to call this a drop in replacement with those numbers no?
But, with enough parity this project could be a dream for anybody dealing with spark’s dreadful performance. Kudos to the team
The next paragraph explains that: "When looking at the test coverage numbers alone, Sail’s capability may seem limited. But we have found that there is a long tail of failed tests due to formatting discrepancies, edge cases, and less-used SQL functions, which we will continue tackling in future releases."
I am with you that it is still very very early. I'll personally keep an eye on the project.
I'll keep an eye on it too, but for a query engine formatting compliance and edge cases tend to be almost all of the work. It's easy to implement SELECT x FROM y WHERE z.
Yeah but the website literally says “zero code changes”. It’s the long tail that’s dangerous since most people don’t understand it as well as a the core functions
Bit off topic; we are looking for something like this but with a facility for untrusted users to run sandboxed code instead of trusted code. All that I found (but I am relatively new to this field) are hacky and, worse, slow solutions.
Our users (who are perhaps untrusted) write the compute for this and so, they cannot just do whatever they want. But it still has to be performant so kicking off containers so far is not great, however even they we couldn't find done.
It is refreshing to see multiple projects with arrow/datafusion trying to bank on existing and user friendly spark's API instead of reinventing the API all over again.
There is likes of comet and blaze that replace execution backend of spark with datafusion and then you have single process alternatives like sail trying to settle in "not so big data" category.
I am watching evolution of projects powered by datafusion and compatible with spark with keen eye. Early days but quite exciting.
This looks interesting, but the docs are really lacking, to the point where it is barely understandable.
I see some potential wins on it, such as it being a Rust-based, Spark-compatible and better suited for single processor environments, but they are just not explained or developed enough.
> The mission of Sail is to unify stream processing, batch processing, and compute-intensive (AI) workloads. Currently, Sail features a drop-in replacement for Spark SQL and the Spark DataFrame API in single-process settings.
Kinda early to call this a drop in replacement with those numbers no?
But, with enough parity this project could be a dream for anybody dealing with spark’s dreadful performance. Kudos to the team