-
-
Notifications
You must be signed in to change notification settings - Fork 18.8k
Description
After #55901, we now do inference of the best resolution, and so allow to create non-nanosecond data by default (instead of raising for out of bounds data).
To be clear, it is a very nice improvement to stop raising those OutOfBounds errors while the timestamp would perfectly fit in another resolution. But I do think we could maybe reconsider the exact logic of how to determine the resolution.
With the latest changes you get the following:
>>> pd.to_datetime(["2024-03-22 11:43:01"]).dtype
dtype('<M8[s]')
>>> pd.to_datetime(["2024-03-22 11:43:01.002"]).dtype
dtype('<M8[ms]')
>>> pd.to_datetime(["2024-03-22 11:43:01.002003"]).dtype
dtype('<M8[us]')
The resulting dtype instance depends on the exact input value (not type). I do think this has some downsides:
- The result dtype becomes very data dependent (while in general we want to avoid value dependent behavior)
- You can very easily get multiple datetime dtypes in a workflow, causing more casting (to different unit) than necessary
The fact that pandas by default truncates the string repr of datetimes (i.e. we don't show the subsecond parts if they are all zero, regardless of the actual resolution), in contrast to numpy, also means that round-tripping through a text representation (eg CSV) will very often lead to a change in dtype.
As a potential alternative, we could also decide to have a fixed default resolution (e.g. microseconds), and then the logic for inferring the resolution could be: try to use the default resolution, and only if that does not work (either out of bounds or too much precision, i.e. nanoseconds present), use the inferred resolution from the data.
That still gives some values dependent behaviour, but I think this would make it a lot less common to see. And using a resolution like microseconds is sufficient for by far most use cases (in terms of bounds it supports: [290301 BC, 294241 AD])
Activity
jorisvandenbossche commentedon Jun 17, 2024
cc @pandas-dev/pandas-core
WillAyd commentedon Jun 17, 2024
This sounds reasonable and I think could help simplify the implementation
bashtage commentedon Jun 17, 2024
I think this would be an improvement. Seems like a good idea that anyone working with human times (say down to second precision) with a range of the modern era would get the same basis for the timestamp.
jorisvandenbossche commentedon Jun 17, 2024
I don't think it will simplify things generally, because we still need the current inference logic when the default unit does not fit, but from looking a bit into it, I also don't think it should make the code much more complex.
Pranav-Wadhwa commentedon Jul 19, 2024
take
Pranav-Wadhwa commentedon Jul 19, 2024
Based on discussions, I will update
to_datetime
to always use nanoseconds in the given scenarios.Pranav-Wadhwa commentedon Aug 2, 2024
@jorisvandenbossche would updating the
to_datetime
function to accept a defaultunit='ns'
be a feasible solution for this? Or are there cases where it wouldn't make sense to default to nanoseconds?WillAyd commentedon Aug 3, 2024
@Pranav-Wadhwa nanoseconds is what we used previously, so I don't think we want to go back to that. The OP suggests microseconds as a default resolution, although I'm not sure its as simple as changing the to_datetime signature either.
Before diving into the details I think should get some more agreement from the pandas core team. @jbrockmendel is our datetime guru so let's see if he has any thoughts first
jbrockmendel commentedon Aug 4, 2024
I’m fine with OP suggestion as long as we are internally consistent, I.e. Timestamp constructor
15 remaining items