Ok, I see more what you mean. Presumably such information which is
large and not frequently requested (like never after the first time)
won't be cached very long by any reasonable caching strategy.
A minor optimization might be to skip caching the results of applying
the SPACEJUMP method. For that matter, POST and CHECKIN/OUT and
pretty much any method other than GET probably shouldn't be cached,
and naturally documents requiring authorization to access shouldn't be
cached (unless the authorization is global to anyone with access to
that cache, maybe.) That solves much of the problem. There are still
domains (most any search in which arbitrary keywords may be specified)
in which hits are less likely; a smart cache might not cache (or
expire faster) any URL which is a search (i.e. has a ? in it.) Hits
can still happen, though; for instance, my CS Tech Report index got a
whole lot of queries for the sample searches specified in the Mosaic
demo document, and caching those could have been a win for some
locations.
- Marc
-- Marc VanHeyningen mvanheyn@cs.indiana.edu MIME, RIPEM & HTTP spoken here