-
Notifications
You must be signed in to change notification settings - Fork 244
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[FEA] custom kernel for date/timestamp formatting/parsing #10032
Comments
One related issue: #10083 Java API gets: |
Some notes on parsing dates from JSON, based on #9975 Depending on the Spark version, there can be different code paths depending on whether a dateFormat is specified or not. Some of the differences that we need to be able to handle are:
|
When cast string to timestamp, only Spark31x supports special values you mentioned, Spark 320 and 320+ do not support special values. |
Are we just not going to support the special values in Spark 3.1 and document it? or are we going to do special post processing to fix them up? |
I suggest do special post processing: |
Is your feature request related to a problem? Please describe.
Spark uses java for date/timestamp parsing and formatting. We have been using a CUDF kernel that uses formats that are compatible with python/C++. But the java formats are very different, so we have to do a mapping. But there are java formats that are not ambiguous until they are mapped into the format the cudf supports. We really should just write our own kernel that tries to do what Spark/Java does directly.
The text was updated successfully, but these errors were encountered: