Interestingly, dependency resolution is not the only NP hard problem uv tries to solve. During development, it also became clear that we needed some way to simplify PEP 508 marker expressions and ask questions like, "are these marker expressions disjoint?"
uv 0.3 introduces a cross platform lock file: https://docs.astral.sh/uv/concepts/projects/#lockfile
More precise details on the compatibility of uv pip
with pip
are documented here: https://docs.astral.sh/uv/pip/compatibility/
You should absolutely not need to handle ISO 8601 and RFC 3339 manually. They are supported via the Display
and FromStr
trait implementations on every main type in Jiff (Span
, Zoned
, Timestamp
, civil::DateTime
, civil::Date
and civil::Time
). It's technically an implementation of a mixture of ISO 8601, RFC 3339 and RFC 9557, but the grammar is specified precisely by Temporal. See: https://docs.rs/jiff/latest/jiff/fmt/temporal/index.html
OK, I've beefed that example up a little bit: https://github.com/BurntSushi/jiff/commit/08dfdde204c739e38147faf70b648e2d417d1c2e
I think the comparison is a bit more muddied now and probably worse overall to be honest. Maybe removing DateTime<Local>
is the right thing to do. I'll think on it.
It's a subtle comparison to make... Probably most people don't even realize that they're doing the wrong thing.
The original name I wanted was gigawatt or some variation there of. :)
Is the cache invalidated if system tzdata is updated?
Yes, although at present, there is a TTL. So an update may take "time" to propagate. jiff::tz::db().reset()
will force the cache to be invalidated. I expect the cache invalidation logic to get tweaked as we get real experience with it.
And what effect does the answer have on the example from “Jiff supports detecting time zone offset conflicts” if both zoned datetimes used the system timezone which got updated between 1. opening 2. parsing the two zoned datetimes.
It's hard to know precisely what you mean. But once you get a jiff::tz::TimeZone
, that value is immutable: https://docs.rs/jiff/latest/jiff/tz/struct.TimeZone.html#a-timezone-is-immutable
New updates to tzdb are only observed when you do a tzdb lookup.
In this section, wouldn’t be more realistic for chrono users to use timezone info around the wire instead of on the wire, rather than using Local+FixedOffset?
That's kinda my point. How do they do that? And does it work with chrono-tz
and tzfile
? And what happens if tzdb updates lead to a serialized datetime with an incorrect offset in a future update of tzdb? There are all sorts of points of failure here that Jiff will handle for you by virtue of tighter integration with tzdb as a first class concept.
How are you doing a date/time library without platform dependencies like libc
or windows-sys
? Are you rolling your own bindings in order to get the local time zone? (Or perhaps you aren't doing that at all.)
Disclosure: I'm the author of the memchr
crate.
You mention the memchr
crate, but you don't seem to have benchmarked it. Instead, you benchmarked the needle
crate (last updated 7 years ago). Can you explain a bit more about your methodology?
The memchr
crate in particular doesn't just use Rabin-Karp. It also uses Two-Way. And SIMD (with support for x86-64, aarch64 and wasm32).
Both Perl and Python use backtracking regex engines and are thus susceptible to similar problems as discussed in the OP.
Cross-posting from reddit:
The PR has more details, but here are a few ad hoc benchmarks using ripgrep on my M2 mac mini while searching a 5.5GB file.
This one is just a case insensitive search. A case insensitive regex expands to something like (ignoring Unicode) [Ss][Hh][Ee][Rr]...
, which means that it has multiple literal prefixes. In fact, you can enumerate them! As long as the set is small enough, this is something that the new SIMD acceleration on aarch64
can handle (and has done for a long time on x86-64
):
$ time rg-before-teddy-aarch64 -i -c 'Sherlock Holmes' OpenSubtitles2018.half.en
3055
real 8.208
user 7.731
sys 0.467
maxmem 5600 MB
faults 191
$ time rg-after-teddy-aarch64 -i -c 'Sherlock Holmes' OpenSubtitles2018.half.en
3055
real 1.137
user 0.695
sys 0.430
maxmem 5904 MB
faults 203
And of course, using multiple literals explicitly also uses this optimization:
$ time rg-before-teddy-aarch64 -c 'Sherlock Holmes|John Watson|Irene Adler|Inspector Lestrade|Professor Moriarty' OpenSubtitles2018.half.en
3804
real 9.055
user 8.580
sys 0.474
maxmem 4912 MB
faults 11
$ time rg-after-teddy-aarch64 -c 'Sherlock Holmes|John Watson|Irene Adler|Inspector Lestrade|Professor Moriarty' OpenSubtitles2018.half.en
3804
real 1.121
user 0.697
sys 0.422
maxmem 4832 MB
faults 11
And it doesn't just work for prefixes, it also works for inner literals too:
$ time rg-before-teddy-aarch64 -c '\w+\s+(Sherlock Holmes|John Watson|Irene Adler|Inspector Lestrade|Professor Moriarty)\s+\w+' OpenSubtitles2018.half.en
773
real 9.065
user 8.586
sys 0.477
maxmem 6384 MB
faults 11
$ time rg-after-teddy-aarch64 -c '\w+\s+(Sherlock Holmes|John Watson|Irene Adler|Inspector Lestrade|Professor Moriarty)\s+\w+' OpenSubtitles2018.half.en
773
real 1.124
user 0.702
sys 0.421
maxmem 6784 MB
faults 11
If you're curious about how the SIMD stuff works, you can read my description of Teddy here. I ported this algorithm out of the Hyperscan project several years ago, and it has been one of the killer ingredients for making ripgrep fast in a lot of common cases. But it only worked on x86-64
. With the rise and popularity of aarch64
and Apple silicon, I was motivated to port it over. I just recently finished analogous work for the memchr
crate as well.
Shortly after we resigned, the top-level team leads, project directors to the Foundation, core team members and the new mods got together to form an interim leadership cohort. Sometimes called the "leadership chat." That then evolved into the Leadership Council by way of an RFC on goverance.
I'm on the uv team. I am quite partial to this approach as well. Alas, it's difficult culturally to pull this off in a pre-existing ecosystem. And in the case of Python at least, it's not totally clear to me that it would avoid the need for solving NP hard problems. See my other comment in this thread about simplifying PEP 508 marker expressions.
Other than avoiding needing a SAT solver to resolve dependencies, the other thing I like about Go's approach is that it makes it very difficult to "lie" about the dependencies you support. In a maximal environment, it's very easy to "depend" on
foo 1.0
but where you actually needfoo 1.1
without issues appearing immediately.