Compare commits

..

20 Commits

Author SHA1 Message Date
standardci ed5cfd86db chore(release): publish new version
- @standardnotes/home-server@1.15.66
 - @standardnotes/revisions-server@1.35.5
 - @standardnotes/syncing-server@1.98.4
2023-09-19 11:04:14 +00:00
Karol Sójko a1a3e9f586 fix: add checking for secondary items logs 2023-09-19 12:45:08 +02:00
Karol Sójko a40b17b141 fix: logs for removing already existing content and paging through diff of the content 2023-09-19 12:43:09 +02:00
standardci 18181ed9df chore(release): publish new version
- @standardnotes/home-server@1.15.65
 - @standardnotes/revisions-server@1.35.4
 - @standardnotes/syncing-server@1.98.3
2023-09-19 08:46:37 +00:00
Karol Sójko 0ae028db73 fix: logs formatting during transition for better readability 2023-09-19 10:27:21 +02:00
standardci 79971be672 chore(release): publish new version
- @standardnotes/home-server@1.15.64
 - @standardnotes/revisions-server@1.35.3
 - @standardnotes/syncing-server@1.98.2
2023-09-19 08:18:11 +00:00
Karol Sójko e4fcd738c3 fix(syncing-server): paging through already existing items 2023-09-19 09:59:03 +02:00
standardci 6827e5e218 chore(release): publish new version
- @standardnotes/auth-server@1.143.6
 - @standardnotes/home-server@1.15.63
2023-09-19 06:54:31 +00:00
Karol Sójko b91cd7e232 fix(auth): bump version 2023-09-19 08:36:33 +02:00
Karol Sójko 735d89cdaa fix: building docker image 2023-09-19 08:36:13 +02:00
standardci e8db412bc2 chore(release): publish new version
- @standardnotes/auth-server@1.143.4
 - @standardnotes/home-server@1.15.62
2023-09-19 06:20:59 +00:00
Karol Sójko bf855bb26e fix(auth): add debug logs to transition 2023-09-19 08:02:27 +02:00
Karol Sójko 41cd377145 fix: missing workspace name input param 2023-09-18 14:34:26 +02:00
Karol Sójko d798864caf fix: strip workflows of bundling 2023-09-18 14:18:17 +02:00
standardci cef1998f0e chore(release): publish new version
- @standardnotes/analytics@2.26.16
 - @standardnotes/api-gateway@1.74.14
 - @standardnotes/auth-server@1.143.3
 - @standardnotes/event-store@1.11.44
 - @standardnotes/files-server@1.22.23
 - @standardnotes/home-server@1.15.61
 - @standardnotes/revisions-server@1.35.2
 - @standardnotes/scheduler-server@1.20.48
 - @standardnotes/syncing-server@1.98.1
 - @standardnotes/websockets-server@1.10.45
2023-09-18 11:49:29 +00:00
Karol Sójko c2cbf44594 chore: upgrade node version (#835)
* chore: upgrade node version

* chore: upgrade yarn

* fix: yarn setup:

* remove bundle plugin

* fix bundling

* fix env path for files
2023-09-18 13:32:49 +02:00
standardci de2e167582 chore(release): publish new version
- @standardnotes/analytics@2.26.15
 - @standardnotes/api-gateway@1.74.13
 - @standardnotes/auth-server@1.143.2
 - @standardnotes/domain-core@1.30.0
 - @standardnotes/event-store@1.11.43
 - @standardnotes/files-server@1.22.22
 - @standardnotes/home-server@1.15.60
 - @standardnotes/revisions-server@1.35.1
 - @standardnotes/scheduler-server@1.20.47
 - @standardnotes/settings@1.21.33
 - @standardnotes/syncing-server@1.98.0
 - @standardnotes/websockets-server@1.10.44
2023-09-18 08:44:10 +00:00
Karol Sójko 547a79e231 feat: add publishing notifications for users when a user is added to vault (#834) 2023-09-18 10:27:24 +02:00
standardci 5a3afb3b17 chore(release): publish new version
- @standardnotes/auth-server@1.143.1
 - @standardnotes/home-server@1.15.59
2023-09-15 15:42:29 +00:00
Karol Sójko 66ef4be656 fix(auth): retrieving transition status 2023-09-15 17:23:49 +02:00
2848 changed files with 4420 additions and 5933 deletions
+3 -12
View File
@@ -33,17 +33,12 @@ jobs:
with:
python-version: '3.11'
- name: Create Bundle Dir
id: bundle-dir
run: echo "temp_dir=$(mktemp -d -t ${{ inputs.service_name }}-${{ github.sha }}-XXXXXXX)" >> $GITHUB_OUTPUT
- name: Cache build
id: cache-build
uses: actions/cache@v3
with:
path: |
packages/**/dist
${{ steps.bundle-dir.outputs.temp_dir }}
key: ${{ runner.os }}-${{ inputs.service_name }}-build-${{ github.sha }}
- name: Set up Node
@@ -57,11 +52,7 @@ jobs:
- name: Build
if: steps.cache-build.outputs.cache-hit != 'true'
run: yarn build ${{ inputs.package_path }}
- name: Bundle
if: steps.cache-build.outputs.cache-hit != 'true'
run: yarn workspace ${{ inputs.workspace_name }} bundle --no-compress --output-directory ${{ steps.bundle-dir.outputs.temp_dir }}
run: yarn build
- name: Login to Docker Hub
uses: docker/login-action@v2
@@ -93,8 +84,8 @@ jobs:
uses: docker/build-push-action@v4
with:
builder: ${{ steps.buildx.outputs.name }}
context: ${{ steps.bundle-dir.outputs.temp_dir }}
file: ${{ steps.bundle-dir.outputs.temp_dir }}/${{ inputs.package_path }}/Dockerfile
context: .
file: ${{ inputs.package_path }}/Dockerfile
platforms: linux/amd64,linux/arm64
push: true
tags: |
+2
View File
@@ -14,6 +14,8 @@ newrelic_agent.log
!.yarn/sdks
!.yarn/versions
yarn.build-error.log
packages/files/uploads/*
!packages/files/uploads/.gitkeep
+1 -1
View File
@@ -1 +1 @@
20.2.0
20.6.1
Generated
+1606 -2402
View File
File diff suppressed because one or more lines are too long
+39 -59
View File
@@ -1,5 +1,5 @@
import { URL as URL$1, fileURLToPath, pathToFileURL } from 'url';
import fs from 'fs';
import { URL as URL$1, fileURLToPath, pathToFileURL } from 'url';
import path from 'path';
import moduleExports, { Module } from 'module';
import { createHash } from 'crypto';
@@ -89,7 +89,6 @@ async function copyPromise(destinationFs, destination, sourceFs, source, opts) {
}));
}
async function copyImpl(prelayout, postlayout, destinationFs, destination, sourceFs, source, opts) {
var _a, _b, _c;
const destinationStat = opts.didParentExist ? await maybeLStat(destinationFs, destination) : null;
const sourceStat = await sourceFs.lstatPromise(source);
const { atime, mtime } = opts.stableTime ? { atime: defaultTime, mtime: defaultTime } : sourceStat;
@@ -115,8 +114,8 @@ async function copyImpl(prelayout, postlayout, destinationFs, destination, sourc
throw new Error(`Unsupported file type (${sourceStat.mode})`);
}
}
if (((_a = opts.linkStrategy) == null ? void 0 : _a.type) !== `HardlinkFromIndex` || !sourceStat.isFile()) {
if (updated || ((_b = destinationStat == null ? void 0 : destinationStat.mtime) == null ? void 0 : _b.getTime()) !== mtime.getTime() || ((_c = destinationStat == null ? void 0 : destinationStat.atime) == null ? void 0 : _c.getTime()) !== atime.getTime()) {
if (opts.linkStrategy?.type !== `HardlinkFromIndex` || !sourceStat.isFile()) {
if (updated || destinationStat?.mtime?.getTime() !== mtime.getTime() || destinationStat?.atime?.getTime() !== atime.getTime()) {
postlayout.push(() => destinationFs.lutimesPromise(destination, atime, mtime));
updated = true;
}
@@ -186,7 +185,7 @@ async function copyFileViaIndex(prelayout, postlayout, destinationFs, destinatio
let indexStat = await maybeLStat(destinationFs, indexPath);
if (destinationStat) {
const isDestinationHardlinkedFromIndex = indexStat && destinationStat.dev === indexStat.dev && destinationStat.ino === indexStat.ino;
const isIndexModified = (indexStat == null ? void 0 : indexStat.mtimeMs) !== defaultTimeMs;
const isIndexModified = indexStat?.mtimeMs !== defaultTimeMs;
if (isDestinationHardlinkedFromIndex) {
if (isIndexModified && linkStrategy.autoRepair) {
atomicBehavior = 0 /* Lock */;
@@ -256,8 +255,7 @@ async function copyFileDirect(prelayout, postlayout, destinationFs, destination,
return true;
}
async function copyFile(prelayout, postlayout, destinationFs, destination, destinationStat, sourceFs, source, sourceStat, opts) {
var _a;
if (((_a = opts.linkStrategy) == null ? void 0 : _a.type) === `HardlinkFromIndex`) {
if (opts.linkStrategy?.type === `HardlinkFromIndex`) {
return copyFileViaIndex(prelayout, postlayout, destinationFs, destination, destinationStat, sourceFs, source, sourceStat, opts, opts.linkStrategy);
} else {
return copyFileDirect(prelayout, postlayout, destinationFs, destination, destinationStat, sourceFs, source, sourceStat, opts);
@@ -387,7 +385,7 @@ class FakeFS {
throw error;
}
}
createdDirectory ?? (createdDirectory = subPath);
createdDirectory ??= subPath;
if (chmod != null)
await this.chmodPromise(subPath, chmod);
if (utimes != null) {
@@ -418,7 +416,7 @@ class FakeFS {
throw error;
}
}
createdDirectory ?? (createdDirectory = subPath);
createdDirectory ??= subPath;
if (chmod != null)
this.chmodSync(subPath, chmod);
if (utimes != null) {
@@ -613,12 +611,14 @@ class FakeFS {
throw error;
}
}
async writeJsonPromise(p, data) {
return await this.writeFilePromise(p, `${JSON.stringify(data, null, 2)}
async writeJsonPromise(p, data, { compact = false } = {}) {
const space = compact ? 0 : 2;
return await this.writeFilePromise(p, `${JSON.stringify(data, null, space)}
`);
}
writeJsonSync(p, data) {
return this.writeFileSync(p, `${JSON.stringify(data, null, 2)}
writeJsonSync(p, data, { compact = false } = {}) {
const space = compact ? 0 : 2;
return this.writeFileSync(p, `${JSON.stringify(data, null, space)}
`);
}
async preserveTimePromise(p, cb) {
@@ -852,7 +852,7 @@ class ProxiedFS extends FakeFS {
readFileSync(p, encoding) {
return this.baseFs.readFileSync(this.fsMapToBase(p), encoding);
}
async readdirPromise(p, opts) {
readdirPromise(p, opts) {
return this.baseFs.readdirPromise(this.mapToBase(p), opts);
}
readdirSync(p, opts) {
@@ -932,12 +932,24 @@ class NodeFS extends BasePortableFakeFS {
this.realFs.opendir(npath.fromPortablePath(p), this.makeCallback(resolve, reject));
}
}).then((dir) => {
return Object.defineProperty(dir, `path`, { value: p, configurable: true, writable: true });
const dirWithFixedPath = dir;
Object.defineProperty(dirWithFixedPath, `path`, {
value: p,
configurable: true,
writable: true
});
return dirWithFixedPath;
});
}
opendirSync(p, opts) {
const dir = typeof opts !== `undefined` ? this.realFs.opendirSync(npath.fromPortablePath(p), opts) : this.realFs.opendirSync(npath.fromPortablePath(p));
return Object.defineProperty(dir, `path`, { value: p, configurable: true, writable: true });
const dirWithFixedPath = dir;
Object.defineProperty(dirWithFixedPath, `path`, {
value: p,
configurable: true,
writable: true
});
return dirWithFixedPath;
}
async readPromise(fd, buffer, offset = 0, length = 0, position = -1) {
return await new Promise((resolve, reject) => {
@@ -1215,16 +1227,16 @@ class NodeFS extends BasePortableFakeFS {
}
async readdirPromise(p, opts) {
return await new Promise((resolve, reject) => {
if (opts == null ? void 0 : opts.withFileTypes) {
this.realFs.readdir(npath.fromPortablePath(p), { withFileTypes: true }, this.makeCallback(resolve, reject));
if (opts) {
this.realFs.readdir(npath.fromPortablePath(p), opts, this.makeCallback(resolve, reject));
} else {
this.realFs.readdir(npath.fromPortablePath(p), this.makeCallback((value) => resolve(value), reject));
}
});
}
readdirSync(p, opts) {
if (opts == null ? void 0 : opts.withFileTypes) {
return this.realFs.readdirSync(npath.fromPortablePath(p), { withFileTypes: true });
if (opts) {
return this.realFs.readdirSync(npath.fromPortablePath(p), opts);
} else {
return this.realFs.readdirSync(npath.fromPortablePath(p));
}
@@ -1359,9 +1371,6 @@ class VirtualFS extends ProxiedFS {
}
const [major, minor] = process.versions.node.split(`.`).map((value) => parseInt(value, 10));
const HAS_CONSOLIDATED_HOOKS = major > 16 || major === 16 && minor >= 12;
const HAS_UNFLAGGED_JSON_MODULES = major > 17 || major === 17 && minor >= 5 || major === 16 && minor >= 15;
const HAS_JSON_IMPORT_ASSERTION_REQUIREMENT = major > 17 || major === 17 && minor >= 1 || major === 16 && minor > 14;
const WATCH_MODE_MESSAGE_USES_ARRAYS = major > 19 || major === 19 && minor >= 2 || major === 18 && minor >= 13;
const HAS_LAZY_LOADED_TRANSLATORS = major > 19 || major === 19 && minor >= 3;
@@ -1427,11 +1436,7 @@ function getFileFormat(filepath) {
);
}
case `.json`: {
if (HAS_UNFLAGGED_JSON_MODULES)
return `json`;
throw new Error(
`Unknown file extension ".json" for ${filepath}`
);
return `json`;
}
case `.js`: {
const pkg = readPackageScope(filepath);
@@ -1452,38 +1457,15 @@ function getFileFormat(filepath) {
}
}
async function getFormat$1(resolved, context, defaultGetFormat) {
const url = tryParseURL(resolved);
if ((url == null ? void 0 : url.protocol) !== `file:`)
return defaultGetFormat(resolved, context, defaultGetFormat);
const format = getFileFormat(fileURLToPath(url));
if (format) {
return {
format
};
}
return defaultGetFormat(resolved, context, defaultGetFormat);
}
async function getSource$1(urlString, context, defaultGetSource) {
const url = tryParseURL(urlString);
if ((url == null ? void 0 : url.protocol) !== `file:`)
return defaultGetSource(urlString, context, defaultGetSource);
return {
source: await fs.promises.readFile(fileURLToPath(url), `utf8`)
};
}
async function load$1(urlString, context, nextLoad) {
var _a;
const url = tryParseURL(urlString);
if ((url == null ? void 0 : url.protocol) !== `file:`)
if (url?.protocol !== `file:`)
return nextLoad(urlString, context, nextLoad);
const filePath = fileURLToPath(url);
const format = getFileFormat(filePath);
if (!format)
return nextLoad(urlString, context, nextLoad);
if (HAS_JSON_IMPORT_ASSERTION_REQUIREMENT && format === `json` && ((_a = context.importAssertions) == null ? void 0 : _a.type) !== `json`) {
if (format === `json` && context.importAssertions?.type !== `json`) {
const err = new TypeError(`[ERR_IMPORT_ASSERTION_TYPE_MISSING]: Module "${urlString}" needs an import assertion of type "json"`);
err.code = `ERR_IMPORT_ASSERTION_TYPE_MISSING`;
throw err;
@@ -1500,7 +1482,7 @@ async function load$1(urlString, context, nextLoad) {
}
return {
format,
source: await fs.promises.readFile(filePath, `utf8`),
source: format === `commonjs` ? void 0 : await fs.promises.readFile(filePath, `utf8`),
shortCircuit: true
};
}
@@ -1991,7 +1973,7 @@ async function resolve$1(originalSpecifier, context, nextResolve) {
specifier = fileURLToPath(url);
}
const { parentURL, conditions = [] } = context;
const issuer = parentURL ? fileURLToPath(parentURL) : process.cwd();
const issuer = parentURL && tryParseURL(parentURL)?.protocol === `file:` ? fileURLToPath(parentURL) : process.cwd();
const pnpapi = findPnpApi(issuer) ?? (url ? findPnpApi(specifier) : null);
if (!pnpapi)
return nextResolve(originalSpecifier, context, nextResolve);
@@ -2068,8 +2050,6 @@ if (!HAS_LAZY_LOADED_TRANSLATORS) {
}
const resolve = resolve$1;
const getFormat = HAS_CONSOLIDATED_HOOKS ? void 0 : getFormat$1;
const getSource = HAS_CONSOLIDATED_HOOKS ? void 0 : getSource$1;
const load = HAS_CONSOLIDATED_HOOKS ? load$1 : void 0;
const load = load$1;
export { getFormat, getSource, load, resolve };
export { load, resolve };
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file not shown.

Some files were not shown because too many files have changed in this diff Show More