-
-
Notifications
You must be signed in to change notification settings - Fork 18.6k
Shared Benchmark suite for Pandas-like projects #19988
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
I'd like to see this, probably under the re organization: I like declarative files for describing the benchmarks and libraries (benchmarkees?).
Possibly benchmarks that are appropriate for one library, but wildly inappropriate for another? I'm not sure.
If everyone's contributing it'll cancel out, right? :)
I can start something after the pandas 0.23 release in a few weeks. |
FYI: https://github.com/mm-mansour/Fast-Pandas @mm-mansour nice work on that. Do you have any interest in expanding those benchmarks to alternative implementations (ray, dask, spark)? Some of these could complicate the setup code a decent amount, so "no" is a perfectly fine answer. Would you mind including a license in your repo, so that we can re-use pieces with credit back to the original (and because that's good OSS practice)? |
Any works still going on this, i am exploring ray and i found out that a lot of discussion going around benchmarks of several frameworks (Ray, dask, spark, etc...)? |
Not that I'm aware of.
…On Sun, May 6, 2018 at 3:30 AM, ziedbouf ***@***.***> wrote:
Any works still going on this, i am exploring ray and i found out that a
lot of discussion going around benchmarks of several frameworks (Ray, dask,
spark, etc...)?
—
You are receiving this because you are subscribed to this thread.
Reply to this email directly, view it on GitHub
<#19988 (comment)>,
or mute the thread
<https://github.com/notifications/unsubscribe-auth/ABQHIhsl4kMnTl8JtkoIuJXh2UyYayJUks5tvrSngaJpZM4SbpqR>
.
|
It would be valuable to have a benchmark suite for Pandas-like projects. This would help users reasonably compare the performance tradeoffs of different implementations and help developers identify possible performance issues.
There are, I think, a few axes that such a benchmark suite might engage:
Additionally, there are a few projects that I think might benefit from such an endeavor
Some operational questions:
asv_bench/benchmarks
.Anyway, those are some thoughts. Please let me know if this is out of scope for this issue tracker.
The text was updated successfully, but these errors were encountered: