| Commit message (Collapse) | Author | Age | Files | Lines |
| |
|
|
|
|
| |
* moved source directories into `/src`
* updated bundle.lua for new `src` path
* moved some docs, icon
* removed old test trees
|
| |
|
|
|
|
|
|
|
|
|
|
| |
- Feature: Zen server endpoint `prj/{project}/oplog/{log}/chunks` to post multiple attachments in one request.
- Feature: Zen server endpoint `prj/{project}/oplog/{log}/save` to save an oplog container. Accepts `CbObject` containing a compressed oplog and attachment references organized in blocks.
- Feature: Zen server endpoint `prj/{project}/oplog/{log}/load` to request an oplog container. Responds with an `CbObject` containing a compressed oplog and attachment references organized in blocks.
- Feature: Zen server endpoint `{project}/oplog/{log}/rpc` to initiate an import to or export from an external location and other operations. Use either JSon or CbPackage as payload.
- CbObject/JSon RPC format for `import` and `export` methods:
- CbObject RPC format for `getchunks` method, returns CbPackage with the found chunks, if all chunks are found the number of attachments matches number of chunks requested.
- Feature: Zen server `{project}/oplog/{log}/{hash}` now accepts `HttpVerb::kPost` as well as `HttpVerb::kGet`.
- Feature: Zen command line tool `oplog-export` to export an oplog to an external target using the zenserver oplog export endpoint.
- Feature: Zen command line tool `oplog-import` to import an oplog from an external source using the zenserver oplog import endpoint.
|
| |
|
|
| |
should be replaced with a proper implementation later
|
| |
|
|
|
|
| |
UE side (#208)
removed all use of IoHash::FromBLAKE3() caused by interactions with CompressedBuffer APIs
|
| |
|
|
|
|
|
|
|
|
|
|
|
| |
* Use direct file read and direct buffer allocation for small IoBuffer materalization
* Reduce range of materialized data in CompositeBuffer reading
CompressedBuffer header reading often only need a small part and not the whole file
* reduce lock contention in IoBuffer::Materialize
* Reduce parsing of compressed headers
Validate header type at decompression
* faster CreateDirectories - start from leaf going up and recurse back
* optimized BufferHeader::IsValid
* Add ValidateCompressedHeader to use when we don't need the actual compressed data
Validate that we always get compressed data in CidStore::AddChunk
* changelog
|
| |
|
| |
* Add `import-project` and `export-project` command line parsing
|
| |
|
|
| |
instances (#180)
|
| |
|
|
| |
(#176)
|
| |
|
|
|
| |
memory copy (#174)
When reading upstream, fall back to old rpc response to handle older instances.
|
| |
|
|
|
|
| |
* Pass along endpoint info for each upstream request
* Add more timing details in log
* more log details for single item upstream fetch
* DISABLEDQUERY over SKIP
|
| | |
|
| | |
|
| | |
|
| |
|
|
| |
* only try to parse payload hash if header exists
* changelog
|
| |
|
|
| |
* demote a number of ZEN_ERROR to ZEN_WARN
* changelog
|
| |
|
|
|
| |
* Added CloudCacheSession::GetInlineBlob to properly get CacheValues from Horde
Issue #UE-162151
* validate uncompressed binary from Horde "application/x-jupiter-inline" response
|
| |
|
|
| |
* Implement proper GetCacheValues upstream
* changelog
|
| |
|
| |
* remove legacy derived data interface for Jupiter
|
| |
|
| |
* Fix ZenStructuredCacheClient lifetime issues
|
| |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
- Bumped ZEN_SCHEMA_VERSION
- CasStore no longer a public API, it is hidden behind CidStore
- Moved cas.h from public header folder
- CidStore no longer maps from Cid -> Cas, we store entries in Cas under RawHash
- CasStore now decompresses data to validate content (matching against RawHash)
- CasChunkSet renames to HashKeySet and put in separate header/cpp file
- Disabled "Chunk" command for now as it relied on CAS being exposed as a service
- Changed CAS http service to Cid http server
- Moved "Run" command completely inside ZEN_WITH_EXEC_SERVICES define
- Removed "cas.basic" test
- Uncommented ".exec.basic" test and added return-skip at start of test
- Moved ScrubContext to separate header file
- Renamed CasGC to GcManager
- Cleaned up configuration passing in cas store classes
- Removed CAS stuff from GcContext and clarified naming in class
- Remove migration code
|
| | |
|
| |
|
|
| |
request fails without outright error
|
| | |
|
| |
|
|
| |
instances are deployed with version that support si
|
| | |
|
| |
|
|
|
| |
RPC requests now has a Namespace field under Params instead of one Namespace per cache key
Fall back to legacy upstream HTTP URI format if default namespace is requested
|
| | |
|
| |
|
|
|
| |
DdcNamespace -> DefaultDdcNamespace
BlobStoreNamespace -> DefaultBlobStoreNamespace
|
| | |
|
| | |
|
| | |
|
| | |
|
| | |
|
| |
|
|
|
| |
Shard up g_MappingLock in IoBufferExtendedCore::Materialize() to reduce contention during high load
Don't queue upstream cache records if we don't have any upstreams
|
| |
|
|
| |
Remove force upload, not necessary.
Add a few more timepoints
|
| | |
|
| | |
|
| | |
|
| | |
|
| | |
|
| |
|
|
|
|
| |
writing to zen and horde upstreams. Return failure from HandleGetCacheRecord if the requested type does not match the cachetype. (#55)
* Fix bug with getting values PUT to Jupiter as CompressedBinary. When getting CompressedBinary records from Jupiter, they are expected to now be a record with a reference to the compact binary. This has to be accounted for when performing upstream GETs.
* HandleGetCacheRecord: avoid crashing on invalid type, and avoid sending back data that doesn't match the AcceptType.
|
| | |
|
| |\
| |
| | |
Change Value propagation to Zen or Jupiter
|
| | | |
|
| | |
| |
| |
| | |
This change ensures we retain the right content type of kCompressedBinary when propagating values from Zen->UpstreamZen. This is done via an RPC that posts a CbPackage. Furthermore when propagating from Zen->Jupiter, it composes its own referencing CbObject for them instead of sending a octet content type and Jupiter defining the referencing CbObject. When fetching Values from Jupiter, this new composed CbObject is still interpreted correctly by Zen.
|
| |/
|
| |
Remove the backwards compatibility for the Zen CachePolicy changes now that there has been enough time for all internal users of the old protocol to update.
|
| | |
|
| |\ |
|
| | | |
|
| | | |
|