| Commit message (Collapse) | Author | Age | Files | Lines |
| |
|
| |
* fix validation during stream decompress of zero-size rawsize
|
| |
|
|
| |
exception (#672)
|
| |
|
|
|
| |
- Improvement: Deeper validation of data when scrub is activated (cas/cache/project)
- Improvement: Enabled more multi threading when running scrub operations
- Improvement: Added means to force a scrub operation at startup with a new release using ZEN_DATA_FORCE_SCRUB_VERSION variable in xmake.lua
|
| |
|
| |
* fix source file offset in DecompressToStream implementations
|
| |
|
|
| |
with RawSize = 0 if the offset was out of bounds for the value. (#666)
|
| |
|
|
|
| |
- Improvement: Use fixed size block chunking for know encrypted/compressed file types
- Improvement: Skip trying to compress chunks that are sourced from files that are known to be encrypted/compressed
- Improvement: Add global open file cache for written files increasing throughput during download by reducing overhead of open/close of file by 80%
|
| |
|
| |
* add proper streaming to none compressor type
|
| |
|
|
|
|
|
|
|
| |
* added ValidateStatistics and improved post upload summary
* improved download statistics
* smoother stats update when compressing
* better feedback during stream compresss/decompress
* don't capture TotalPartWriteCount by reference
* disk stats cleanup
* multi-test-download overall timer
|
| |
|
|
|
|
| |
- Improvement: Scavenge .zen temp folders for existing data (downloaded, decompressed or written) from previous failed run
- Improvement: Faster abort during stream compression
- Improvement: Try to move downloaded blobs with rename if possible avoiding an extra disk write
- Improvement: Only clean temp folders on successful or cancelled build - keep it if download fails
|
| |
|
|
| |
- Improvement: Validate hash of decompressed data inline with streaming decompression
- Improvement: Do streaming compression of large blobs to improve memory and I/O performance
|
| |
|
|
| |
* clean up latency parameters and slow down rate updates
* add DecompressToStream
|
| |
|
|
|
|
|
|
|
|
|
|
|
|
| |
* add modification tick to filesystem traversal
* add ShowDetails option to ProgressBar
* log callstack if we terminate process
* handle chunking if MaxSize > 1MB
* BasicFile write helpers and WriteToTempFile simplifications
* bugfix for CompositeBuffer::IterateRange when using DecompressToComposite for actually comrpessed data
revert of earlier optimization
* faster compress/decompress for large disk-based files
* enable progress feedback in IoHash::HashBuffer
* add payload validation in HttpClient::Get
* fix range requests (range is including end byte)
* remove BuildPartId for blob/block related operations in builds api
|
| |
|
|
| |
* fix partial cache miss logging
* Handle zero size partial requests safer
|
| |
|
| |
* refactor so we don't have to re-read data from source to get block sizes
|
| |
|
|
|
| |
* Add CompressedBuffer::GetRange that references source data rather than make a memory copy
* Use Compressed.CopyRange in project store GetChunkRange
* docs for CompressedBuffer::CopyRange and CompressedBuffer::GetRange
|
| |
|
|
|
| |
- Bugfix: Install Ctrl+C handler earlier when doing `zen oplog-export` and `zen oplog-export` to properly cancel jobs
- Improvement: Add ability to block a set of CAS entries from GC in project store
- Improvement: Large attachments and loose files are now split into smaller chunks and stored in blocks during oplog export
|
| |
|
| |
- Bugfix: Allow attachments that contains a raw size of zero
|
| |
|
| |
Added CompressedBufferReader support from UE. This provides some streaming decompression support which can be employed to reduce memory and other resource usage.
|
| |
|
|
| |
changed win64 to be more in line with other platforms wrt xmake config
|
|
|
* moved source directories into `/src`
* updated bundle.lua for new `src` path
* moved some docs, icon
* removed old test trees
|