▲ | dev_l1x_be 5 days ago | |||||||||||||
Not sure what you mean by this. The table concept is the same age as computers. Here is a table, do something with it -> this is the high level df api. All the functions make sense, what is hard to read, write or debug here? I have used Polars to process 600M of xml files (with a bit of a hack) and the polars part of the code is readable with minimal comments. Polars has a better api than pandas, at least the intent is easier to understand. (lazyness, yay) | ||||||||||||||
▲ | phailhaus 5 days ago | parent [-] | |||||||||||||
The problem with the dataframe API is that whenever you want to change a small part of your logic, you usually have to rethink and rewrite the whole solution. It is too difficult to write reusable code. Too many functions that try to do too many things with a million kwargs that each have their own nuances. This is because these libraries tend to favor fewer keystrokes over composable design. So the easy stuff is easy and makes for pretty docs, but the hard stuff is obnoxious to reason through. This article explains it pretty well: https://dynomight.net/numpy/ | ||||||||||||||
|