mirror of
https://gitea.com/actions/cache.git
synced 2024-11-22 20:32:28 +08:00
Add more debugging
This commit is contained in:
parent
994e3b75fc
commit
2cbd952179
2
dist/restore/index.js
vendored
2
dist/restore/index.js
vendored
|
@ -1625,12 +1625,14 @@ function saveCache(cacheId, archivePath) {
|
||||||
const stream = fs.createReadStream(archivePath);
|
const stream = fs.createReadStream(archivePath);
|
||||||
let streamIsClosed = false;
|
let streamIsClosed = false;
|
||||||
stream.on("close", () => {
|
stream.on("close", () => {
|
||||||
|
core.debug("Stream is closed");
|
||||||
streamIsClosed = true;
|
streamIsClosed = true;
|
||||||
});
|
});
|
||||||
const resourceUrl = getCacheApiUrl() + cacheId.toString();
|
const resourceUrl = getCacheApiUrl() + cacheId.toString();
|
||||||
const uploads = [];
|
const uploads = [];
|
||||||
let offset = 0;
|
let offset = 0;
|
||||||
while (!streamIsClosed) {
|
while (!streamIsClosed) {
|
||||||
|
core.debug(`Offset: ${offset}`);
|
||||||
const chunk = stream.read(MAX_CHUNK_SIZE);
|
const chunk = stream.read(MAX_CHUNK_SIZE);
|
||||||
uploads.push(uploadChunk(restClient, resourceUrl, chunk, offset));
|
uploads.push(uploadChunk(restClient, resourceUrl, chunk, offset));
|
||||||
offset += MAX_CHUNK_SIZE;
|
offset += MAX_CHUNK_SIZE;
|
||||||
|
|
2
dist/save/index.js
vendored
2
dist/save/index.js
vendored
|
@ -1625,12 +1625,14 @@ function saveCache(cacheId, archivePath) {
|
||||||
const stream = fs.createReadStream(archivePath);
|
const stream = fs.createReadStream(archivePath);
|
||||||
let streamIsClosed = false;
|
let streamIsClosed = false;
|
||||||
stream.on("close", () => {
|
stream.on("close", () => {
|
||||||
|
core.debug("Stream is closed");
|
||||||
streamIsClosed = true;
|
streamIsClosed = true;
|
||||||
});
|
});
|
||||||
const resourceUrl = getCacheApiUrl() + cacheId.toString();
|
const resourceUrl = getCacheApiUrl() + cacheId.toString();
|
||||||
const uploads = [];
|
const uploads = [];
|
||||||
let offset = 0;
|
let offset = 0;
|
||||||
while (!streamIsClosed) {
|
while (!streamIsClosed) {
|
||||||
|
core.debug(`Offset: ${offset}`);
|
||||||
const chunk = stream.read(MAX_CHUNK_SIZE);
|
const chunk = stream.read(MAX_CHUNK_SIZE);
|
||||||
uploads.push(uploadChunk(restClient, resourceUrl, chunk, offset));
|
uploads.push(uploadChunk(restClient, resourceUrl, chunk, offset));
|
||||||
offset += MAX_CHUNK_SIZE;
|
offset += MAX_CHUNK_SIZE;
|
||||||
|
|
|
@ -186,6 +186,7 @@ export async function saveCache(
|
||||||
const stream = fs.createReadStream(archivePath);
|
const stream = fs.createReadStream(archivePath);
|
||||||
let streamIsClosed = false;
|
let streamIsClosed = false;
|
||||||
stream.on("close", () => {
|
stream.on("close", () => {
|
||||||
|
core.debug("Stream is closed");
|
||||||
streamIsClosed = true;
|
streamIsClosed = true;
|
||||||
});
|
});
|
||||||
|
|
||||||
|
@ -193,6 +194,7 @@ export async function saveCache(
|
||||||
const uploads: Promise<IRestResponse<void>>[] = [];
|
const uploads: Promise<IRestResponse<void>>[] = [];
|
||||||
let offset = 0;
|
let offset = 0;
|
||||||
while (!streamIsClosed) {
|
while (!streamIsClosed) {
|
||||||
|
core.debug(`Offset: ${offset}`);
|
||||||
const chunk: Buffer = stream.read(MAX_CHUNK_SIZE);
|
const chunk: Buffer = stream.read(MAX_CHUNK_SIZE);
|
||||||
uploads.push(uploadChunk(restClient, resourceUrl, chunk, offset));
|
uploads.push(uploadChunk(restClient, resourceUrl, chunk, offset));
|
||||||
offset += MAX_CHUNK_SIZE;
|
offset += MAX_CHUNK_SIZE;
|
||||||
|
|
Loading…
Reference in New Issue
Block a user