Compare commits
9 Commits
developmen
...
ft/tests_s
Author | SHA1 | Date |
---|---|---|
Antonin Coulibaly | c75435ec8c | |
Antonin Coulibaly | 8ade0f71e9 | |
Antonin Coulibaly | 3268fa1cc7 | |
David Pineau | 6987e9031a | |
David Pineau | 42a8dcc968 | |
David Pineau | 812d2bec67 | |
David Pineau | f0391eb5d1 | |
David Pineau | 93291d55e4 | |
David Pineau | bb63fc629c |
|
@ -1,31 +0,0 @@
|
|||
name: Tests
|
||||
|
||||
on:
|
||||
push:
|
||||
branches-ignore:
|
||||
- development/**
|
||||
- q/*/**
|
||||
|
||||
jobs:
|
||||
tests:
|
||||
runs-on: ubuntu-latest
|
||||
steps:
|
||||
- name: Checkokut
|
||||
uses: actions/checkout@v4
|
||||
- name: Install deps
|
||||
run: sudo apt-get update -q
|
||||
- uses: actions/setup-node@v4
|
||||
with:
|
||||
node-version: '16'
|
||||
- name: Install Yarn
|
||||
run: npm install -g yarn
|
||||
- name: install dependencies
|
||||
run: yarn install --frozen-lockfile
|
||||
- name: run lint
|
||||
run: echo "linter is disabled temporarily ()" || yarn run --silent lint -- --max-warnings 0
|
||||
- name: run lint_md
|
||||
run: yarn --silent lint_md
|
||||
- name: run test
|
||||
run: yarn test
|
||||
- name: run coverage
|
||||
run: yarn coverage
|
11
README.md
11
README.md
|
@ -1,5 +1,6 @@
|
|||
# WereLogs
|
||||
|
||||
[![CircleCI][badgepub]](https://circleci.com/gh/scality/werelogs)
|
||||
[![Scality CI][badgepriv]](http://ci.ironmann.io/gh/scality/werelogs)
|
||||
|
||||
This repository provides a NodeJS Library that aims to be an efficient logging
|
||||
|
@ -19,7 +20,7 @@ https://github.com/scality/Guidelines/blob/master/CONTRIBUTING.md).
|
|||
In order to install WereLogs, you can use NPM with github's HTTP url, and save
|
||||
it in your own package.json:
|
||||
|
||||
```sh
|
||||
```
|
||||
$> npm i --save scality/werelogs
|
||||
```
|
||||
|
||||
|
@ -157,3 +158,11 @@ In order to find out the known issues, it is advised to take a look at the
|
|||
[project's github page](http://github.com/scality/werelogs). There, you should
|
||||
be able to find the issues, tagged with the releases they are impacting,
|
||||
whether they're open or closed.
|
||||
|
||||
## Contributing
|
||||
|
||||
The contributing rules for this project are defined in the associated
|
||||
CONTRIBUTING.md file.
|
||||
|
||||
[badgepub]: https://circleci.com/gh/scality/werelogs.svg?style=svg
|
||||
[badgepriv]: http://ci.ironmann.io/gh/scality/werelogs.svg?style=svg&circle-token=a946e81ad65b99814403b5e57f017d9ecbe93f0a
|
||||
|
|
|
@ -0,0 +1,23 @@
|
|||
general:
|
||||
branches:
|
||||
ignore:
|
||||
- /^ultron\/.*/ # Ignore ultron/* branches
|
||||
artifacts:
|
||||
- coverage/
|
||||
- doc/
|
||||
|
||||
machine:
|
||||
node:
|
||||
version: 6.9.5
|
||||
test:
|
||||
override:
|
||||
- npm run lint_md
|
||||
- npm run lint -- --max-warnings 0
|
||||
- npm run gendoc
|
||||
- npm run coverage
|
||||
# running ft_test packs werelogs and installs it + deps into
|
||||
# tests/functional. Pack is like publishing werelogs in a local tgz
|
||||
# archive that can be installed.
|
||||
# This step shall ensure that no issue is encountered when installing
|
||||
# the package, and allows to functionally test werelogs.
|
||||
- npm run ft_test
|
|
@ -1,65 +1,45 @@
|
|||
interface WerelogsConfigOptions {
|
||||
level?: 'trace' | 'debug' | 'info' | 'warn' | 'error' | 'fatal';
|
||||
dump?: 'trace' | 'debug' | 'info' | 'warn' | 'error' | 'fatal';
|
||||
streams?: object[];
|
||||
}
|
||||
interface config {}
|
||||
|
||||
declare class WerelogsConfig {
|
||||
constructor(config?: WerelogsConfigOptions);
|
||||
reset(): WerelogsConfig;
|
||||
update(config: WerelogsConfig): WerelogsConfig;
|
||||
logger: any;
|
||||
level: 'trace' | 'debug' | 'info' | 'warn' | 'error' | 'fatal';
|
||||
dump: 'trace' | 'debug' | 'info' | 'warn' | 'error' | 'fatal';
|
||||
end: 'trace' | 'debug' | 'info' | 'warn' | 'error' | 'fatal';
|
||||
}
|
||||
|
||||
interface LogDictionary {
|
||||
interface LogDictionnary {
|
||||
httpMethod?: string;
|
||||
httpURL?: string;
|
||||
[field: string]: any;
|
||||
}
|
||||
|
||||
declare module 'werelogs' {
|
||||
export class RequestLogger {
|
||||
constructor(
|
||||
logger: any,
|
||||
logLevel: string,
|
||||
dumpThreshold: string,
|
||||
endLevel: string,
|
||||
uids?: string|Array<string>
|
||||
);
|
||||
getUids(): Array<string>;
|
||||
getSerializedUids(): string;
|
||||
addDefaultFields(fields: LogDictionary): LogDictionary;
|
||||
trace(msg: string, data?: LogDictionary): void;
|
||||
debug(msg: string, data?: LogDictionary): void;
|
||||
info(msg: string, data?: LogDictionary): void;
|
||||
warn(msg: string, data?: LogDictionary): void;
|
||||
error(msg: string, data?: LogDictionary): void;
|
||||
fatal(msg: string, data?: LogDictionary): void;
|
||||
end(msg: string, data?: LogDictionary): void;
|
||||
errorEnd(msg: string, data?:LogDictionary): void;
|
||||
}
|
||||
declare class RequestLogger {
|
||||
constructor(
|
||||
logger: any,
|
||||
logLevel: string,
|
||||
dumpThreshold: string,
|
||||
endLevel: string,
|
||||
uids?: string|Array<string>
|
||||
);
|
||||
getUids(): Array<string>;
|
||||
getSerializedUids(): string;
|
||||
addDefaultFields(fields: LogDictionnary): LogDictionnary;
|
||||
trace(msg: string, data?: LogDictionnary): void;
|
||||
debug(msg: string, data?: LogDictionnary): void;
|
||||
info(msg: string, data?: LogDictionnary): void;
|
||||
warn(msg: string, data?: LogDictionnary): void;
|
||||
error(msg: string, data?: LogDictionnary): void;
|
||||
fatal(msg: string, data?: LogDictionnary): void;
|
||||
end(msg: string, data?: LogDictionnary): void;
|
||||
}
|
||||
|
||||
declare module 'werelogs' {
|
||||
export class Logger {
|
||||
name: string;
|
||||
constructor(name: string);
|
||||
|
||||
constructor(name: string, config?: config);
|
||||
setLevel(levelName: string): void;
|
||||
setDumpLevelThreshold(levelName: string): void;
|
||||
newRequestLogger(uids?: string|Array<string>): RequestLogger;
|
||||
newRequestLoggerFromSerializedUids(uids: string): RequestLogger;
|
||||
trace(msg: string, data?: LogDictionary): void;
|
||||
debug(msg: string, data?: LogDictionary): void;
|
||||
info(msg: string, data?: LogDictionary): void;
|
||||
warn(msg: string, data?: LogDictionary): void;
|
||||
error(msg: string, data?: LogDictionary): void;
|
||||
fatal(msg: string, data?: LogDictionary): void;
|
||||
}
|
||||
|
||||
export function configure(config: WerelogsConfigOptions): void;
|
||||
|
||||
export class API {
|
||||
constructor(config: WerelogsConfigOptions);
|
||||
reconfigure(config: WerelogsConfigOptions): void;
|
||||
Logger: Logger;
|
||||
trace(msg: string, data?: LogDictionnary): void;
|
||||
debug(msg: string, data?: LogDictionnary): void;
|
||||
info(msg: string, data?: LogDictionnary): void;
|
||||
warn(msg: string, data?: LogDictionnary): void;
|
||||
error(msg: string, data?: LogDictionnary): void;
|
||||
fatal(msg: string, data?: LogDictionnary): void;
|
||||
}
|
||||
}
|
||||
|
|
52
index.js
52
index.js
|
@ -1,51 +1 @@
|
|||
const API = require('./lib/api.js');
|
||||
const stderrUtils = require('./lib/stderrUtils');
|
||||
|
||||
/*
|
||||
* For convenience purposes, we provide an already instanciated API; so that
|
||||
* old uses of the imported Logger class can be kept as-is. For quick logging,
|
||||
* this also provides a hassle-free way to log using werelogs.
|
||||
*/
|
||||
const werelogs = new API();
|
||||
|
||||
module.exports = {
|
||||
Logger: werelogs.Logger,
|
||||
configure: werelogs.reconfigure.bind(werelogs),
|
||||
Werelogs: API,
|
||||
/**
|
||||
* Timestamp logs going to stderr
|
||||
*
|
||||
* @example <caption>Simplest usage</caption>
|
||||
* ```
|
||||
* const { stderrUtils } = require('werelogs');
|
||||
* stderrUtils.catchAndTimestampStderr();
|
||||
* ```
|
||||
*
|
||||
* @example <caption>Manage process exit</caption>
|
||||
* ```
|
||||
* const { stderrUtils } = require('werelogs');
|
||||
* // set exitCode to null to keep process running on uncaughtException
|
||||
* stderrUtils.catchAndTimestampStderr(undefined, null);
|
||||
* // application init
|
||||
* process.on('uncaughtException', (err) => {
|
||||
* // custom handling, close connections, files
|
||||
* this.worker.kill(); // or process.exit(1);
|
||||
* });
|
||||
* // Note you could use prependListener to execute your callback first
|
||||
* // and then let stderrUtils exit the process.
|
||||
* ```
|
||||
*
|
||||
* @example <caption>Custom listener</caption>
|
||||
* ```
|
||||
* const { stderrUtils } = require('werelogs');
|
||||
* stderrUtils.catchAndTimestampWarning();
|
||||
* // application init
|
||||
* process.on('uncaughtException', (err, origin) => {
|
||||
* stderrUtils.printErrorWithTimestamp(err, origin);
|
||||
* // close and stop everything
|
||||
* process.exit(1);
|
||||
* });
|
||||
* ```
|
||||
*/
|
||||
stderrUtils,
|
||||
};
|
||||
module.exports = { Logger: require('./lib/Logger.js') };
|
||||
|
|
|
@ -1,5 +1,4 @@
|
|||
|
||||
// eslint-disable-line strict
|
||||
'use strict'; // eslint-disable-line strict
|
||||
|
||||
const LogLevel = require('./LogLevel.js');
|
||||
const SimpleLogger = require('./SimpleLogger.js');
|
||||
|
@ -15,31 +14,10 @@ class Config {
|
|||
* This is the default constructor of the Config Object, and the only way
|
||||
* to instanciate it (with default parameters).
|
||||
*
|
||||
* @param {object} conf - A configuration object for werelogs.
|
||||
* @param {string} conf.level - The string name of the logging level
|
||||
* ('trace', 'debug', 'info', 'warn',
|
||||
* 'error' and 'fatal' in order of
|
||||
* importance.)
|
||||
* @param {string} conf.dump - The string name of the log dumping
|
||||
* level ('trace', 'debug', 'info',
|
||||
* 'warn', 'error' and 'fatal' in order
|
||||
* of importance.)
|
||||
* @param {object[]} conf.streams - The array of streams into which to
|
||||
* log. This is an Array of objects
|
||||
* which have a field named 'stream',
|
||||
* which is writeable.
|
||||
*
|
||||
* @returns {undefined}
|
||||
*/
|
||||
constructor(conf) {
|
||||
this.logLevel = 'info';
|
||||
this.dumpThreshold = 'error';
|
||||
this.endLevel = 'info';
|
||||
this.streams = [{ level: 'trace', stream: process.stdout }];
|
||||
this.simpleLogger = new SimpleLogger('werelogs', this.streams);
|
||||
if (conf) {
|
||||
this.update(conf);
|
||||
}
|
||||
constructor() {
|
||||
this.reset();
|
||||
}
|
||||
|
||||
/**
|
||||
|
@ -72,9 +50,11 @@ class Config {
|
|||
* 'error' and 'fatal' in order of
|
||||
* importance.)
|
||||
* @param {object[]} config.streams - The array of streams into which to
|
||||
* log. This is an Array of objects
|
||||
* which have a field named 'stream',
|
||||
* which is writeable.
|
||||
* log. Their configuration is directly
|
||||
* related to the expected bunyan
|
||||
* streams array, for compatibility
|
||||
* purposes (except that the 'level'
|
||||
* field is not accounted for)
|
||||
*
|
||||
* @see [Bunyan's documentation]{@link
|
||||
* https://github.com/trentm/node-bunyan/blob/master/README.md#streams} for
|
||||
|
@ -89,43 +69,29 @@ class Config {
|
|||
|
||||
const checkedConfig = config || {};
|
||||
|
||||
if (Object.prototype.hasOwnProperty.call(checkedConfig, 'level')) {
|
||||
if (checkedConfig.hasOwnProperty('level')) {
|
||||
LogLevel.throwIfInvalid(checkedConfig.level);
|
||||
}
|
||||
if (Object.prototype.hasOwnProperty.call(checkedConfig, 'dump')) {
|
||||
LogLevel.throwIfInvalid(checkedConfig.dump);
|
||||
}
|
||||
// for check log level vs. log dump level
|
||||
const newLogLevel = checkedConfig.level || this.logLevel;
|
||||
const newLogDumpLevel = checkedConfig.dump || this.dumpThreshold;
|
||||
if (newLogDumpLevel
|
||||
&& !LogLevel.shouldLog(newLogDumpLevel, newLogLevel)) {
|
||||
throw new Error(
|
||||
'Logging level should be at most logging dump level',
|
||||
);
|
||||
}
|
||||
|
||||
if (Object.prototype.hasOwnProperty.call(checkedConfig, 'level')) {
|
||||
this.logLevel = checkedConfig.level;
|
||||
}
|
||||
|
||||
if (Object.prototype.hasOwnProperty.call(checkedConfig, 'dump')) {
|
||||
if (checkedConfig.hasOwnProperty('dump')) {
|
||||
LogLevel.throwIfInvalid(checkedConfig.dump);
|
||||
this.dumpThreshold = checkedConfig.dump;
|
||||
}
|
||||
|
||||
if (Object.prototype.hasOwnProperty.call(checkedConfig, 'end')) {
|
||||
if (checkedConfig.hasOwnProperty('end')) {
|
||||
LogLevel.throwIfInvalid(checkedConfig.end);
|
||||
this.endLevel = checkedConfig.end;
|
||||
}
|
||||
|
||||
if (Object.prototype.hasOwnProperty.call(checkedConfig, 'streams')) {
|
||||
if (checkedConfig.hasOwnProperty('streams')) {
|
||||
if (!Array.isArray(checkedConfig.streams)) {
|
||||
throw new TypeError('WereLogs config.streams must be an Array '
|
||||
+ 'of Writeable Streams.');
|
||||
throw new Error('WereLogs config.streams must be an Array ' +
|
||||
'of Writeable Streams.');
|
||||
}
|
||||
if (!checkedConfig.streams.length) {
|
||||
throw new Error('Werelogs config.streams must contain at '
|
||||
+ 'least one stream.');
|
||||
throw new Error('Werelogs config.streams must contain at ' +
|
||||
'least one stream.');
|
||||
}
|
||||
this.streams = checkedConfig.streams.map(stream => {
|
||||
stream.level = 'trace'; // eslint-disable-line no-param-reassign
|
||||
|
@ -188,4 +154,4 @@ class Config {
|
|||
}
|
||||
}
|
||||
|
||||
module.exports = Config;
|
||||
module.exports = new Config();
|
||||
|
|
|
@ -0,0 +1,247 @@
|
|||
'use strict'; // eslint-disable-line strict
|
||||
|
||||
const Utils = require('./Utils.js');
|
||||
const objectCopy = Utils.objectCopy;
|
||||
|
||||
/**
|
||||
* This class manages the internal `default fields` for logging classes, with a
|
||||
* notion of parent/child relationships and hierarchy.
|
||||
* A Child element will inherit the fields from the Parent element, and
|
||||
* complete (or even override part of it) with its own fields. For instance,
|
||||
* a child can redefine (and thus override in the final result) a field that is
|
||||
* already defined by its parent.
|
||||
*
|
||||
* This class shall be used embedded within loggers, to provide the `Default
|
||||
* fields` and hierarchy logics.
|
||||
*
|
||||
* @private @property {DefaultFields} parent - The parent node in the
|
||||
* hierarchy
|
||||
* @private @property {DefaultFields[]} children - The list of children nodes
|
||||
* in the hierarchy
|
||||
* @private @property {Object} fields - The dictionary of fields
|
||||
* defined for this node
|
||||
* @private @property {Object} parentFields - The dictionary of fields
|
||||
* inherited by this node from
|
||||
* its parent node
|
||||
* @private @property {Object} precomputedFields - The prepared dictionary of
|
||||
* aggregated fields containing
|
||||
* the parent's fields
|
||||
* overriden with the node's
|
||||
* own fields. This is used as
|
||||
* a trick to reduce the
|
||||
* computational need.
|
||||
*
|
||||
* @mixin
|
||||
*/
|
||||
class DefaultFields {
|
||||
constructor() {
|
||||
this.parent = null;
|
||||
this.children = [];
|
||||
this.fields = {};
|
||||
this.parentFields = {};
|
||||
this.precomputedFields = {};
|
||||
}
|
||||
|
||||
/**
|
||||
* This function allows to manually destroy a node, so that all references
|
||||
* to parents or children are released. This helps in avoiding the creation
|
||||
* of memory leaks due to our internal cyclic references that cannot be
|
||||
* automatically released due to the mechanics of GC in Node.JS, and its
|
||||
* shitty prevention of meaningful OOP designs.
|
||||
*
|
||||
* All hail << The Coming of the Great White Handkerchief >> that will
|
||||
* wipe this misery out of our lives !
|
||||
*
|
||||
* @returns {undefined}
|
||||
*/
|
||||
_dfDestroy() {
|
||||
this.children.forEach(child => {
|
||||
child.setParent(null);
|
||||
});
|
||||
this.setParent(null);
|
||||
this.children = [];
|
||||
}
|
||||
|
||||
/**
|
||||
* Add a node to the list of children of the current node, effectively
|
||||
* making it one of the node's children.
|
||||
*
|
||||
* @param {DefaultFields} child - The node to add as a child of the current
|
||||
* node
|
||||
*
|
||||
* @return {undefined}
|
||||
*/
|
||||
_dfRegisterChild(child) {
|
||||
this.children.push(child);
|
||||
this._dfNotifyChild(child);
|
||||
}
|
||||
|
||||
/**
|
||||
* Remove a node from the list of children of the current node, effectively
|
||||
* cutting off the relationship between the two.
|
||||
*
|
||||
* @param {DefaultFields} toRemove - The node to remove from the list of
|
||||
* children of the current node.
|
||||
*
|
||||
* @return {undefined}
|
||||
*/
|
||||
_dfUnregisterChild(toRemove) {
|
||||
this.children = this.children.filter(child => child !== toRemove);
|
||||
}
|
||||
|
||||
/**
|
||||
* Utility function to notify one child node of an update of the node's
|
||||
* precomputed fields.
|
||||
*
|
||||
* @param {DefaultField} child - The child node to notify
|
||||
*
|
||||
* @return {undefined}
|
||||
*/
|
||||
_dfNotifyChild(child) {
|
||||
child._dfSetParentFields(this.precomputedFields);
|
||||
}
|
||||
|
||||
/**
|
||||
* Utility function to notify every children node of an update of the
|
||||
* node's precomputed fields.
|
||||
*
|
||||
* @return {undefined}
|
||||
*/
|
||||
_dfNotifyChildren() {
|
||||
this.children.forEach(child => this._dfNotifyChild(child));
|
||||
}
|
||||
|
||||
/**
|
||||
* This function allows changing the Parent node of the current node,
|
||||
* consequently changing the resulting aggregation of the hierarchy of
|
||||
* fields. This can be used for a temporary switch of parent node.
|
||||
*
|
||||
* @param {DefaultFields} parent - The new parent node to set for the
|
||||
* current node
|
||||
*
|
||||
* @return {DefaultFields|null} The previous parent
|
||||
*/
|
||||
setParent(parent) {
|
||||
const oldParent = this.parent;
|
||||
if (parent === oldParent) {
|
||||
return oldParent;
|
||||
}
|
||||
if (oldParent) {
|
||||
oldParent._dfUnregisterChild(this);
|
||||
}
|
||||
this.parent = parent || null;
|
||||
if (this.parent) {
|
||||
this.parent._dfRegisterChild(this);
|
||||
} else {
|
||||
this._dfSetParentFields();
|
||||
}
|
||||
return oldParent;
|
||||
}
|
||||
|
||||
/**
|
||||
* Internal function to partially recompute the precomputedFields through
|
||||
* inclusion of the newly defined fields into the precomputed ones. This
|
||||
* inclusion may override some already defined fields.
|
||||
*
|
||||
* This function shall always be the last one called when updating the
|
||||
* internal fields, as it also triggers the update of the children nodes.
|
||||
*
|
||||
* @param {Object} newFields - The dictionary of newFields to include into
|
||||
* the precomputedFields
|
||||
*
|
||||
* @return {undefined}
|
||||
*/
|
||||
_dfAugmentPrecomputedFields() {
|
||||
objectCopy(this.precomputedFields, this.fields);
|
||||
this._dfNotifyChildren();
|
||||
}
|
||||
|
||||
/**
|
||||
* Internal function to update the fields provided by the parent node in
|
||||
* the DefaultFields hierarchy. It serves as a notification hook to refresh
|
||||
* the precomputed fields depending on the parent node's fields.
|
||||
* Two situations may lead to calling this function:
|
||||
* 1. The parent node's updated its preComputed fields and notified its
|
||||
* children, including the current node
|
||||
* 2. The node reset its own fields, and we must re-compute to remove
|
||||
* obsolete fields previously provided by the current node.
|
||||
*
|
||||
* @param {Object} parentFields - the precomputedFields from the parent node
|
||||
*
|
||||
* @return {undefined}
|
||||
*/
|
||||
_dfSetParentFields(parentFields) {
|
||||
this.parentFields = parentFields || {};
|
||||
this.precomputedFields = objectCopy({}, this.parentFields);
|
||||
this._dfAugmentPrecomputedFields();
|
||||
}
|
||||
|
||||
/**
|
||||
* This function allows to reset the fields managed by the DefaultFields
|
||||
* instance. It automatically triggers the re-computation of the
|
||||
* precomputed fields, cascading through the node and its children.
|
||||
*
|
||||
* /!\ This function may lead to an important use of the computational
|
||||
* resources if over-used.
|
||||
*
|
||||
* @return {undefined}
|
||||
*/
|
||||
resetDefaultFields() {
|
||||
const oldFields = this.fields;
|
||||
this.fields = {};
|
||||
// Automatically triggers the recomputation of precomputedFields
|
||||
this._dfSetParentFields(this.parentFields);
|
||||
return oldFields;
|
||||
}
|
||||
|
||||
/**
|
||||
* This function allows the user to remove one or more items from the
|
||||
* defaultFields's dict.
|
||||
*
|
||||
* @param {String[]} fields - List of the names of the fields to be removed
|
||||
* from the internal dictionary of default
|
||||
* fields
|
||||
*
|
||||
* @return {Object} The previous set of default fields
|
||||
*/
|
||||
removeDefaultFields(fields) {
|
||||
const toRemove = {};
|
||||
fields.forEach(key => {
|
||||
toRemove[key] = undefined;
|
||||
});
|
||||
return this.addDefaultFields(toRemove);
|
||||
}
|
||||
|
||||
/**
|
||||
* This function allows the user to add default fields to include into all
|
||||
* JSON log entries generated through this request logger. As this function
|
||||
* attempt not to modify the provided fields object, it copies the field
|
||||
* into a new object for safe keeping.
|
||||
*
|
||||
* @param {Object} fields The dictionnary of additional fields to include
|
||||
* by default for this instance of the
|
||||
* RequestLogger.
|
||||
*
|
||||
* @return {Object} The previous set of default fields (can be
|
||||
* safely ignored).
|
||||
*/
|
||||
addDefaultFields(fields) {
|
||||
const oldFields = this.fields;
|
||||
this.fields = objectCopy({}, this.fields, fields);
|
||||
this._dfAugmentPrecomputedFields(fields);
|
||||
return oldFields;
|
||||
}
|
||||
|
||||
/**
|
||||
* This function returns the node's precomputed fields, that includes all
|
||||
* of its hierarchy's fields and its own. This is intended to retrieve the
|
||||
* final form of the fields managed by the object.
|
||||
*
|
||||
* @return {Object} The precomputed fields to be added to a log entry
|
||||
*/
|
||||
_dfGetFields() {
|
||||
return this.precomputedFields;
|
||||
}
|
||||
}
|
||||
|
||||
module.exports = DefaultFields;
|
|
@ -1,5 +1,4 @@
|
|||
|
||||
// eslint-disable-line strict
|
||||
'use strict'; // eslint-disable-line strict
|
||||
|
||||
const logLevels = [
|
||||
'trace',
|
||||
|
@ -19,15 +18,10 @@ const logLevels = [
|
|||
*
|
||||
* @throw {Error} A human-readable message that tells which log
|
||||
* levels are supported.
|
||||
* @throw {TypeError} A human-readable message indicating that the
|
||||
* provided logging level was not a string
|
||||
*
|
||||
* @returns {undefined}
|
||||
*/
|
||||
function throwIfInvalid(level) {
|
||||
if (typeof level !== 'string') {
|
||||
throw new TypeError('Logging level should be a string');
|
||||
}
|
||||
if (logLevels.indexOf(level) === -1) {
|
||||
throw new RangeError(`Invalid logging level: ${level} is neither`
|
||||
+ ` ${logLevels.join(', ')}.`);
|
||||
|
|
|
@ -1,32 +1,64 @@
|
|||
|
||||
// eslint-disable-line strict
|
||||
'use strict'; // eslint-disable-line strict
|
||||
|
||||
const LogLevel = require('./LogLevel.js');
|
||||
const RequestLogger = require('./RequestLogger.js');
|
||||
const { unserializeUids } = require('./Utils.js');
|
||||
const Utils = require('./Utils.js');
|
||||
const Config = require('./Config.js');
|
||||
const DefaultFields = require('./DefaultFields.js');
|
||||
|
||||
const unserializeUids = Utils.unserializeUids;
|
||||
const objectCopy = Utils.objectCopy;
|
||||
|
||||
class Logger extends DefaultFields {
|
||||
|
||||
class Logger {
|
||||
/**
|
||||
* This is the constructor of the Logger class. It takes optional
|
||||
* configuration parameters, that allow to modify its behavior.
|
||||
*
|
||||
* @param {Werelogs.Config} config - An instanciated Werelogs Config object
|
||||
* @param {string|Object} fields - Deprecated: {string}: The name of the
|
||||
* Logger that will be included in the log
|
||||
* entries
|
||||
* Advised: {Object} A set of fields that
|
||||
* will be used as the default fields for
|
||||
* this Logger, and will be included in
|
||||
* all later log entries.
|
||||
*
|
||||
* @param {string} name - The name of the Logger. It can be found later on
|
||||
* in the log entries.
|
||||
* @param {object} config - A configuration object for werelogs.
|
||||
* @param {string} config.level - The string name of the logging level
|
||||
* ('trace', 'debug', 'info', 'warn',
|
||||
* 'error' and 'fatal' in order of
|
||||
* importance.)
|
||||
* @param {string} config.dump - The string name of the log dumping level
|
||||
* ('trace', 'debug', 'info', 'warn',
|
||||
* 'error' and 'fatal' in order of
|
||||
* importance.)
|
||||
*
|
||||
* @returns {undefined}
|
||||
*/
|
||||
constructor(config, name) {
|
||||
if (!(config instanceof Config)) {
|
||||
throw new TypeError('Invalid parameter Type for "config".');
|
||||
constructor(fields, config) {
|
||||
super();
|
||||
/* TODO XXX FIXME Remove starting at version 8.0 FIXME XXX TODO
|
||||
* vvvvvvvvvvvvvvvvvvvvvvvv */
|
||||
if (typeof fields === 'string') {
|
||||
this.addDefaultFields({ name: fields });
|
||||
} else if (typeof fields === 'object') {
|
||||
/* ^^^^^^^^^^^^^^^^^^^^^^^
|
||||
* TODO XXX FIXME Remove starting at version 8.0 FIXME XXX TODO */
|
||||
this.addDefaultFields(fields);
|
||||
/* TODO XXX FIXME Remove starting at version 8.0 FIXME XXX TODO
|
||||
* vvvvvvvvvvvvvvvvvvvvvvvv */
|
||||
}
|
||||
if (!(typeof name === 'string' || name instanceof String)) {
|
||||
throw new TypeError('Invalid parameter Type for "name".');
|
||||
}
|
||||
this.config = config;
|
||||
this.name = name;
|
||||
/* ^^^^^^^^^^^^^^^^^^^^^^^
|
||||
* TODO XXX FIXME Remove starting at version 8.0 FIXME XXX TODO */
|
||||
Config.update(config);
|
||||
}
|
||||
|
||||
setLevel(levelName) {
|
||||
Config.update({ level: levelName });
|
||||
}
|
||||
|
||||
setDumpThreshold(levelName) {
|
||||
Config.update({ dump: levelName });
|
||||
}
|
||||
|
||||
/**
|
||||
|
@ -38,10 +70,10 @@ class Logger {
|
|||
* @returns {RequestLogger} A Valid Request Logger
|
||||
*/
|
||||
newRequestLogger(uids) {
|
||||
const rLog = new RequestLogger(this.config.logger,
|
||||
this.config.level, this.config.dump,
|
||||
this.config.end, uids);
|
||||
rLog.addDefaultFields({ name: this.name });
|
||||
const rLog = new RequestLogger(Config.logger,
|
||||
Config.level, Config.dump, Config.end,
|
||||
uids);
|
||||
rLog.setParent(this);
|
||||
return rLog;
|
||||
}
|
||||
|
||||
|
@ -54,18 +86,16 @@ class Logger {
|
|||
* @returns {RequestLogger} A Valid Request Logger
|
||||
*/
|
||||
newRequestLoggerFromSerializedUids(serializedUids) {
|
||||
const rLog = new RequestLogger(this.config.logger,
|
||||
this.config.level, this.config.dump,
|
||||
this.config.end,
|
||||
unserializeUids(serializedUids));
|
||||
rLog.addDefaultFields({ name: this.name });
|
||||
const rLog = new RequestLogger(Config.logger,
|
||||
Config.level, Config.dump, Config.end,
|
||||
unserializeUids(serializedUids));
|
||||
rLog.setParent(this);
|
||||
return rLog;
|
||||
}
|
||||
|
||||
_doLog(levelName, msg, data) {
|
||||
const sLogger = this.config.logger;
|
||||
const finalData = { name: this.name, time: Date.now() };
|
||||
if (!LogLevel.shouldLog(levelName, this.config.level)) {
|
||||
const sLogger = Config.logger;
|
||||
if (!LogLevel.shouldLog(levelName, Config.level)) {
|
||||
return;
|
||||
}
|
||||
if (data !== undefined && typeof data !== 'object') {
|
||||
|
@ -74,10 +104,12 @@ class Logger {
|
|||
callparams: [msg, data],
|
||||
},
|
||||
'Werelogs API was mis-used.'
|
||||
+ ' This development error should be fixed ASAP.',
|
||||
);
|
||||
+ ' This development error should be fixed ASAP.');
|
||||
return;
|
||||
}
|
||||
const finalData = objectCopy({},
|
||||
this._dfGetFields(),
|
||||
{ time: Date.now() });
|
||||
if (data) {
|
||||
Object.keys(data).forEach(k => {
|
||||
finalData[k] = data[k];
|
||||
|
|
|
@ -1,49 +1,54 @@
|
|||
'use strict'; // eslint-disable-line strict
|
||||
|
||||
// eslint-disable-line strict
|
||||
const assert = require('assert');
|
||||
|
||||
const LogLevel = require('./LogLevel.js');
|
||||
const Utils = require('./Utils.js');
|
||||
|
||||
const { serializeUids, generateUid, objectCopy } = Utils;
|
||||
const DefaultFields = require('./DefaultFields.js');
|
||||
const serializeUids = Utils.serializeUids;
|
||||
const generateUid = Utils.generateUid;
|
||||
const objectCopy = Utils.objectCopy;
|
||||
|
||||
function ensureUidValidity(uid) {
|
||||
if (uid.indexOf(':') !== -1) {
|
||||
throw new Error(`RequestLogger UID "${uid}" contains an illegal `
|
||||
+ 'character: \':\'.');
|
||||
throw new Error(`RequestLogger UID "${uid}" contains an illegal ` +
|
||||
'character: \':\'.');
|
||||
}
|
||||
return uid;
|
||||
}
|
||||
|
||||
class EndLogger {
|
||||
/*
|
||||
* @mixes DefaultFields
|
||||
*/
|
||||
class EndLogger extends DefaultFields {
|
||||
constructor(reqLogger) {
|
||||
super();
|
||||
this.logger = reqLogger;
|
||||
this.fields = {};
|
||||
this.setParent(reqLogger);
|
||||
}
|
||||
|
||||
augmentedLog(level, msg, data) {
|
||||
// We can alter current instance, as it won't be usable after this
|
||||
// call.
|
||||
this.fields = objectCopy(this.fields, data || {});
|
||||
return this.logger.log(level, msg, this.fields, true);
|
||||
}
|
||||
|
||||
/**
|
||||
* This function allows the user to add default fields to include into all
|
||||
* JSON log entries generated through this request logger. As this function
|
||||
* attempt not to modify the provided fields object, it copies the field
|
||||
* into a new object for safe keeping.
|
||||
*
|
||||
* @param {object} fields The dictionnary of additional fields to include
|
||||
* by default for this instance of the
|
||||
* RequestLogger.
|
||||
*
|
||||
* @returns {object} The previous set of default fields (can be
|
||||
* safely ignored).
|
||||
*/
|
||||
addDefaultFields(fields) {
|
||||
const oldFields = this.fields;
|
||||
this.fields = objectCopy({}, this.fields, fields);
|
||||
return oldFields;
|
||||
assert.strictEqual(this.logger.elapsedTime, null, 'The logger\'s'
|
||||
+ 'end() wrapper should not be called more than'
|
||||
+ ' once.');
|
||||
const fields = objectCopy(this.precomputedFields, data || {});
|
||||
const res = this.logger.log(level, msg, fields, true);
|
||||
/* XXX NOTE
|
||||
*
|
||||
* This is our "manual" memory management.
|
||||
* Since child and parent `DefaultFields` are actually refering to each
|
||||
* other, the GC cannot release the child logger.
|
||||
* Also, we have no means to automatically unregister the child from
|
||||
* its parent, nor to have the parent notify the child of its update
|
||||
* without keeping those references. So here we go, manual memory
|
||||
* management in Garbage-collected languages !
|
||||
*
|
||||
* All hail << The Coming of the Great White Handkerchief >> that will
|
||||
* wipe this misery out of our lives !
|
||||
*
|
||||
* XXX NOTE */
|
||||
this._dfDestroy();
|
||||
return res;
|
||||
}
|
||||
|
||||
/**
|
||||
|
@ -137,7 +142,8 @@ class EndLogger {
|
|||
* the global log level; and is used to track the log events for one given
|
||||
* request.
|
||||
*/
|
||||
class RequestLogger {
|
||||
class RequestLogger extends DefaultFields {
|
||||
|
||||
/**
|
||||
* Constructor of the WereLogs Request Logger.
|
||||
* This function takes a logger instance, a logging level, and a last
|
||||
|
@ -173,7 +179,8 @@ class RequestLogger {
|
|||
* @returns {undefined}
|
||||
*/
|
||||
constructor(logger, logLevel, dumpThreshold, endLevel, uids) {
|
||||
let uidList;
|
||||
super();
|
||||
let uidList = undefined;
|
||||
|
||||
if (!LogLevel.shouldLog(dumpThreshold, logLevel)) {
|
||||
throw new Error('Logging Dump level should be equal or'
|
||||
|
@ -183,13 +190,12 @@ class RequestLogger {
|
|||
if (uids !== undefined && Array.isArray(uids)) {
|
||||
uidList = uids.map(uid => ensureUidValidity(uid));
|
||||
uidList.push(generateUid());
|
||||
} else if (uids !== undefined && typeof uids === 'string') {
|
||||
} else if (uids !== undefined && typeof(uids) === 'string') {
|
||||
uidList = [ensureUidValidity(uids)];
|
||||
}
|
||||
this.uids = uidList || [generateUid()];
|
||||
|
||||
this.entries = [];
|
||||
this.fields = {};
|
||||
this.logLevel = logLevel;
|
||||
this.dumpThreshold = dumpThreshold;
|
||||
this.endLevel = endLevel;
|
||||
|
@ -229,25 +235,6 @@ class RequestLogger {
|
|||
return serializeUids(this.uids);
|
||||
}
|
||||
|
||||
/**
|
||||
* This function allows the user to add default fields to include into all
|
||||
* JSON log entries generated through this request logger. As this function
|
||||
* attempt not to modify the provided fields object, it copies the field
|
||||
* into a new object for safe keeping.
|
||||
*
|
||||
* @param {object} fields The dictionnary of additional fields to include
|
||||
* by default for this instance of the
|
||||
* RequestLogger.
|
||||
*
|
||||
* @returns {object} The previous set of default fields (can be
|
||||
* safely ignored).
|
||||
*/
|
||||
addDefaultFields(fields) {
|
||||
const oldFields = this.fields;
|
||||
this.fields = objectCopy({}, this.fields, fields);
|
||||
return oldFields;
|
||||
}
|
||||
|
||||
/**
|
||||
* Logging function to write a trace-level log entry.
|
||||
*
|
||||
|
@ -365,6 +352,8 @@ class RequestLogger {
|
|||
if (msg === undefined && data === undefined) {
|
||||
return this.endLogger;
|
||||
}
|
||||
assert.strictEqual(this.elapsedTime, null, 'The "end()" logging method '
|
||||
+ 'should not be called more than once.');
|
||||
return this.log(this.endLevel, msg, data, true);
|
||||
}
|
||||
|
||||
|
@ -381,6 +370,8 @@ class RequestLogger {
|
|||
* @returns {undefined}
|
||||
*/
|
||||
errorEnd(msg, data) {
|
||||
assert.strictEqual(this.elapsedTime, null, 'The "end()" logging method '
|
||||
+ 'should not be called more than once.');
|
||||
return this.log('error', msg, data, true);
|
||||
}
|
||||
|
||||
|
@ -422,11 +413,10 @@ class RequestLogger {
|
|||
+ ' This development error should be fixed ASAP.',
|
||||
{
|
||||
callparams: [msg, logFields],
|
||||
},
|
||||
);
|
||||
});
|
||||
return;
|
||||
}
|
||||
const fields = objectCopy({}, this.fields, logFields || {});
|
||||
const fields = objectCopy({}, this._dfGetFields(), logFields || {});
|
||||
const endFlag = isEnd || false;
|
||||
|
||||
/*
|
||||
|
@ -440,12 +430,6 @@ class RequestLogger {
|
|||
// eslint-disable-next-line camelcase
|
||||
fields.req_id = serializeUids(this.uids);
|
||||
if (endFlag) {
|
||||
if (this.elapsedTime !== null) {
|
||||
// reset elapsedTime to avoid an infinite recursion
|
||||
// while logging the error
|
||||
this.elapsedTime = null;
|
||||
this.error('RequestLogger.end() has been called more than once');
|
||||
}
|
||||
this.elapsedTime = process.hrtime(this.startTime);
|
||||
// eslint-disable-next-line camelcase
|
||||
fields.elapsed_ms = this.elapsedTime[0] * 1000
|
||||
|
@ -466,6 +450,10 @@ class RequestLogger {
|
|||
} else if (LogLevel.shouldLog(level, this.logLevel)) {
|
||||
this.doLogIO(logEntry);
|
||||
}
|
||||
|
||||
if (isEnd) {
|
||||
this.setParent(null);
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
|
|
|
@ -1,17 +1,7 @@
|
|||
|
||||
// eslint-disable-line strict
|
||||
'use strict'; // eslint-disable-line strict
|
||||
|
||||
const os = require('os');
|
||||
const safeJSONStringify = require('safe-json-stringify');
|
||||
const fastJSONStringify = require('fast-safe-stringify')
|
||||
|
||||
function errorStackReplacer(key, value) {
|
||||
if (value instanceof Error) {
|
||||
return value.stack;
|
||||
}
|
||||
return value;
|
||||
}
|
||||
|
||||
/*
|
||||
* This function safely stringifies JSON. If an exception occcurs (due to
|
||||
* circular references, exceptions thrown from object getters etc.), the module
|
||||
|
@ -23,13 +13,12 @@ function errorStackReplacer(key, value) {
|
|||
function safeStringify(obj) {
|
||||
let str;
|
||||
try {
|
||||
// Try to stringify the object (fast version)
|
||||
str = fastJSONStringify(obj, errorStackReplacer);
|
||||
str = JSON.stringify(obj);
|
||||
} catch (e) {
|
||||
// fallback to remove circular object references or other exceptions
|
||||
// eslint-disable-next-line no-param-reassign
|
||||
obj.unsafeJSON = true;
|
||||
return safeJSONStringify(obj, errorStackReplacer);
|
||||
return safeJSONStringify(obj);
|
||||
}
|
||||
return str;
|
||||
}
|
||||
|
@ -45,11 +34,11 @@ function isWriteableStream(s) {
|
|||
class SimpleLogger {
|
||||
constructor(name, streams) {
|
||||
this.name = name;
|
||||
this.streams = [{ level: 'trace', stream: process.stdout }];
|
||||
this.streams = [{ level: 'info', stream: process.stdout }];
|
||||
if (streams) {
|
||||
if (!Array.isArray(streams)) {
|
||||
throw new Error('Invalid streams. streams must be an array'
|
||||
+ ' list of writeable streams');
|
||||
throw new Error('Invalid streams. streams must be an array' +
|
||||
' list of writeable streams');
|
||||
}
|
||||
/*
|
||||
* This is for backwards compatibility. current config in projects
|
||||
|
@ -79,9 +68,8 @@ class SimpleLogger {
|
|||
logFields.hostname = this.hostname;
|
||||
logFields.pid = process.pid;
|
||||
|
||||
const safeString = safeStringify(logFields);
|
||||
this.streams.forEach(s => s.stream
|
||||
.write(`${safeString}\n`));
|
||||
.write(`${safeStringify(logFields)}\n`));
|
||||
}
|
||||
|
||||
info(fields, message) {
|
||||
|
|
|
@ -1,5 +1,4 @@
|
|||
|
||||
// eslint-disable-line strict
|
||||
'use strict'; // eslint-disable-line strict
|
||||
|
||||
/**
|
||||
* @constant
|
||||
|
|
72
lib/api.js
72
lib/api.js
|
@ -1,72 +0,0 @@
|
|||
|
||||
// eslint-disable-line strict
|
||||
|
||||
const Config = require('./Config.js');
|
||||
const Logger = require('./Logger.js');
|
||||
|
||||
class API {
|
||||
/**
|
||||
* This is the constructor of the Logger class. It takes optional
|
||||
* configuration parameters, that allow to modify its behavior.
|
||||
*
|
||||
* @param {object} config - A configuration object for werelogs.
|
||||
* @param {string} config.level - The name of the logging level ('trace',
|
||||
* 'debug', 'info', 'warn', 'error' and
|
||||
* 'fatal' in order of importance.)
|
||||
* @param {string} config.dump - The name of the log dumping level
|
||||
* ('trace', 'debug', 'info', 'warn',
|
||||
* 'error' and 'fatal' in order of
|
||||
* importance.)
|
||||
* @param {object[]} config.streams - The streams into which to log. This
|
||||
* is an Array of objects which have a
|
||||
* field named 'stream', which is
|
||||
* writeable.
|
||||
*/
|
||||
constructor(config) {
|
||||
this.config = new Config(config);
|
||||
this.preboundLogger = Logger.bind(null, this.config);
|
||||
}
|
||||
|
||||
/**
|
||||
* This is a thunk function that allows reconfiguring the streams and log
|
||||
* levels of all Logger and future RequestLogger objects. Note that
|
||||
* existing RequestLogger will live their lifespan retaining the old
|
||||
* configuration.
|
||||
* If the provided configuration is erroneous, the function may throw
|
||||
* exceptions depending on the detected configuration error. Please see the
|
||||
* Config class's documentation about that.
|
||||
*
|
||||
* @throws {TypeError} - One of the provided arguments is not of the
|
||||
* expected type
|
||||
* @throws {RangeError} - The logging level provided is not part of the
|
||||
* supported logging levels
|
||||
* @throws {Error} - A human-readable message providing details about
|
||||
* a logic error due to the input parameters
|
||||
* provided.
|
||||
*
|
||||
* @param {object} config - A configuration object for werelogs.
|
||||
* @param {string} config.level - The name of the logging level ('trace',
|
||||
* 'debug', 'info', 'warn', 'error' and
|
||||
* 'fatal' in order of importance.)
|
||||
* @param {string} config.dump - The name of the log dumping level
|
||||
* ('trace', 'debug', 'info', 'warn',
|
||||
* 'error' and 'fatal' in order of
|
||||
* importance.)
|
||||
* @param {object[]} config.streams - The streams into which to log. This
|
||||
* is an Array of objects which have a
|
||||
* field named 'stream', which is
|
||||
* writeable.
|
||||
*
|
||||
* @returns {undefined}
|
||||
*
|
||||
*/
|
||||
reconfigure(config) {
|
||||
this.config.update(config);
|
||||
}
|
||||
|
||||
get Logger() {
|
||||
return this.preboundLogger;
|
||||
}
|
||||
}
|
||||
|
||||
module.exports = API;
|
|
@ -1,106 +0,0 @@
|
|||
/**
|
||||
* @returns {string} a timestamp in ISO format YYYY-MM-DDThh:mm:ss.sssZ
|
||||
*/
|
||||
const defaultTimestamp = () => new Date().toISOString();
|
||||
|
||||
/**
|
||||
* Prints on stderr a timestamp, the origin and the error
|
||||
*
|
||||
* If no other instructions are needed on uncaughtException,
|
||||
* consider using `catchAndTimestampStderr` directly.
|
||||
*
|
||||
* @example
|
||||
* process.on('uncaughtException', (err, origin) => {
|
||||
* printErrorWithTimestamp(err, origin);
|
||||
* // server.close();
|
||||
* // file.close();
|
||||
* process.nextTick(() => process.exit(1));
|
||||
* });
|
||||
* // Don't forget to timestamp warning
|
||||
* catchAndTimestampWarning();
|
||||
* @param {Error} err see process event uncaughtException
|
||||
* @param {uncaughtException|unhandledRejection} origin see process event
|
||||
* @param {string} [date=`defaultTimestamp()`] Date to print
|
||||
* @returns {boolean} see process.stderr.write
|
||||
*/
|
||||
function printErrorWithTimestamp(
|
||||
err, origin, date = defaultTimestamp(),
|
||||
) {
|
||||
return process.stderr.write(`${date}: ${origin}:\n${err.stack}\n`);
|
||||
}
|
||||
|
||||
/**
|
||||
* Prefer using `catchAndTimestampStderr` instead of this function.
|
||||
*
|
||||
* Adds listener for uncaughtException to print with timestamp.
|
||||
*
|
||||
* If you want to manage the end of the process, you can set exitCode to null.
|
||||
* Or use `printErrorWithTimestamp` in your own uncaughtException listener.
|
||||
*
|
||||
* @param {Function} [dateFct=`defaultTimestamp`] Fct returning a formatted date
|
||||
* @param {*} [exitCode=1] On uncaughtException, if not null, `process.exit`
|
||||
* will be called with this value
|
||||
* @returns {undefined}
|
||||
*/
|
||||
function catchAndTimestampUncaughtException(
|
||||
dateFct = defaultTimestamp, exitCode = 1,
|
||||
) {
|
||||
process.on('uncaughtException', (err, origin) => {
|
||||
printErrorWithTimestamp(err, origin, dateFct());
|
||||
if (exitCode !== null) {
|
||||
process.nextTick(() => process.exit(exitCode));
|
||||
}
|
||||
});
|
||||
}
|
||||
|
||||
/**
|
||||
* Forces the use of `--trace-warnings` and adds a date in warning.detail
|
||||
* The warning will be printed by the default `onWarning`
|
||||
*
|
||||
* @param {string} [dateFct=`defaultTimestamp`] Fct returning a formatted date
|
||||
* @returns {undefined}
|
||||
*/
|
||||
function catchAndTimestampWarning(dateFct = defaultTimestamp) {
|
||||
process.traceProcessWarnings = true;
|
||||
// must be executed first, before the default `onWarning`
|
||||
process.prependListener('warning', warning => {
|
||||
if (warning.detail) {
|
||||
// eslint-disable-next-line no-param-reassign
|
||||
warning.detail += `\nAbove Warning Date: ${dateFct()}`;
|
||||
} else {
|
||||
// eslint-disable-next-line no-param-reassign
|
||||
warning.detail = `Above Warning Date: ${dateFct()}`;
|
||||
}
|
||||
});
|
||||
}
|
||||
|
||||
/**
|
||||
* Adds listener for uncaughtException and warning to print them with timestamp.
|
||||
*
|
||||
* If you want to manage the end of the process, you can set exitCode to null.
|
||||
* Or use `printErrorWithTimestamp` in your own uncaughtException listener.
|
||||
*
|
||||
* @example
|
||||
* const { stderrUtils } = require('werelogs');
|
||||
* // first instruction in your index.js or entrypoint
|
||||
* stderrUtils.catchAndTimestampStderr();
|
||||
*
|
||||
* @param {Function} [dateFct=`defaultTimestamp`] Fct returning a formatted date
|
||||
* @param {*} [exitCode=1] On uncaughtException, if not null, `process.exit`
|
||||
* will be called with this value
|
||||
* @returns {undefined}
|
||||
*/
|
||||
function catchAndTimestampStderr(
|
||||
dateFct = defaultTimestamp, exitCode = 1,
|
||||
) {
|
||||
catchAndTimestampUncaughtException(dateFct, exitCode);
|
||||
catchAndTimestampWarning(dateFct);
|
||||
}
|
||||
|
||||
module.exports = {
|
||||
defaultTimestamp,
|
||||
printErrorWithTimestamp,
|
||||
catchAndTimestampUncaughtException,
|
||||
catchAndTimestampWarning,
|
||||
catchAndTimestampStderr,
|
||||
};
|
33
package.json
33
package.json
|
@ -1,18 +1,18 @@
|
|||
{
|
||||
"name": "werelogs",
|
||||
"engines": {
|
||||
"node": ">=10"
|
||||
"node": "6.9.5"
|
||||
},
|
||||
"version": "8.1.5",
|
||||
"version": "7.0.0",
|
||||
"description": "An efficient raw JSON logging library aimed at micro-services architectures.",
|
||||
"main": "index.js",
|
||||
"scripts": {
|
||||
"gendoc": "jsdoc $(git ls-files 'lib/*.js') -d doc",
|
||||
"lint": "eslint $(git ls-files '*.js')",
|
||||
"lint_md": "markdownlint $(git ls-files '*.md')",
|
||||
"lint_md": "mdlint $(git ls-files '*.md')",
|
||||
"test": "mocha tests/unit/",
|
||||
"ft_test": "(npm pack && cp werelogs-*.tgz tests/functional && cd tests/functional && cp -R ../../node_modules/ node_modules/ && npm install werelogs-*.tgz && ./node_modules/.bin/mocha . multi-modules/ && rm -rf tests/functional/node_modules tests/functional/werelogs-*.tgz tests/functional/*lock*)",
|
||||
"coverage": "nyc ./node_modules/.bin/_mocha tests/unit"
|
||||
"ft_test": "rm -rf tests/functional/node_modules && npm pack && cp -R node_modules werelogs-*.tgz tests/functional && cd tests/functional && npm install werelogs-*.tgz && ./node_modules/.bin/mocha . multi-modules/ && cd -",
|
||||
"coverage": "istanbul cover ./node_modules/.bin/_mocha tests/unit"
|
||||
},
|
||||
"repository": {
|
||||
"type": "git",
|
||||
|
@ -33,20 +33,17 @@
|
|||
},
|
||||
"homepage": "https://github.com/scality/werelogs#readme",
|
||||
"dependencies": {
|
||||
"fast-safe-stringify": "^2.1.1",
|
||||
"safe-json-stringify": "^1.2.0"
|
||||
"safe-json-stringify": "^1.0.3"
|
||||
},
|
||||
"devDependencies": {
|
||||
"eslint": "^7.32.0",
|
||||
"eslint-config-airbnb": "^18.2.1",
|
||||
"eslint-config-scality": "git+https://git.yourcmc.ru/vitalif/zenko-eslint-config-scality.git",
|
||||
"eslint-plugin-import": "^2.22.1",
|
||||
"eslint-plugin-jsx-a11y": "^6.4.1",
|
||||
"eslint-plugin-react": "^7.26.0",
|
||||
"eslint-plugin-react-hooks": "^4.2.0",
|
||||
"jsdoc": "^3.4.3",
|
||||
"markdownlint-cli": "^0.27.1",
|
||||
"mocha": ">=3.1.2",
|
||||
"nyc": "^15.1.0"
|
||||
"eslint-plugin-react": "^4.3.0",
|
||||
"eslint": "^2.4.0",
|
||||
"eslint-config-airbnb": "^6.0.0",
|
||||
"eslint-config-scality": "scality/Guidelines",
|
||||
"istanbul": "^1.0.0-alpha",
|
||||
"istanbul-api": "==1.0.0-alpha.9",
|
||||
"jsdoc": "^3.4.0",
|
||||
"mdlint": "^0.1.0",
|
||||
"mocha": "^3.2.0"
|
||||
}
|
||||
}
|
||||
|
|
|
@ -1,11 +1,11 @@
|
|||
|
||||
// eslint-disable-line strict
|
||||
'use strict'; // eslint-disable-line strict
|
||||
|
||||
const assert = require('assert');
|
||||
|
||||
const LogLevel = require('../lib/LogLevel.js');
|
||||
|
||||
class DummyLogger {
|
||||
|
||||
constructor() {
|
||||
this.ops = [];
|
||||
this.counts = {
|
||||
|
@ -57,8 +57,8 @@ function computeBehavior(filterLevel, logLevel, testLevel) {
|
|||
|
||||
return {
|
||||
value,
|
||||
msg: `Expected ${logLevel} to be called ${value} times with `
|
||||
+ `filter level ${filterLevel}.`,
|
||||
msg: `Expected ${logLevel} to be called ${value} times with ` +
|
||||
`filter level ${filterLevel}.`,
|
||||
};
|
||||
}
|
||||
|
||||
|
@ -107,10 +107,9 @@ function loggingMisuseGenerator(test, createLogger) {
|
|||
logger.info.apply(logger, test.args);
|
||||
},
|
||||
Error,
|
||||
`Werelogs should not throw with ${test.desc}`,
|
||||
);
|
||||
`Werelogs should not throw with ${test.desc}`);
|
||||
assert(dummyLogger.ops[0][0], 'fatal',
|
||||
'Expected the Module Logger to have logged a fatal message.');
|
||||
'Expected the Module Logger to have logged a fatal message.');
|
||||
done();
|
||||
};
|
||||
}
|
||||
|
|
|
@ -1,12 +1,10 @@
|
|||
|
||||
// eslint-disable-line strict
|
||||
'use strict'; // eslint-disable-line strict
|
||||
|
||||
const assert = require('assert');
|
||||
const { PassThrough } = require('stream');
|
||||
const PassThrough = require('stream').PassThrough;
|
||||
const pass = new PassThrough;
|
||||
|
||||
const pass = new PassThrough();
|
||||
|
||||
const werelogs = require('werelogs'); // eslint-disable-line
|
||||
const Logger = require('werelogs').Logger;
|
||||
|
||||
// With PassThrough, SimpleLogger can use it as Writeable stream and all the
|
||||
// data being written can be read into a variable
|
||||
|
@ -17,23 +15,23 @@ pass.on('data', data => {
|
|||
logBuffer.records.push(data.toString());
|
||||
});
|
||||
|
||||
werelogs.configure({
|
||||
level: 'info',
|
||||
dump: 'error',
|
||||
streams: [{
|
||||
stream: pass,
|
||||
type: 'raw',
|
||||
}],
|
||||
});
|
||||
|
||||
function createModuleLogger() {
|
||||
return new werelogs.Logger('FT-test');
|
||||
function createModuleLogger(fields) {
|
||||
const defaultFields = fields || 'FT-test';
|
||||
return new Logger(defaultFields, {
|
||||
level: 'info',
|
||||
dump: 'error',
|
||||
streams: [{
|
||||
stream: pass,
|
||||
type: 'raw',
|
||||
}],
|
||||
});
|
||||
}
|
||||
|
||||
|
||||
function checkFields(fields) {
|
||||
const record = JSON.parse(logBuffer.records[0].trim());
|
||||
Object.keys(fields).forEach(k => {
|
||||
if (Object.prototype.hasOwnProperty.call(fields, k)) {
|
||||
if (fields.hasOwnProperty(k)) {
|
||||
assert.deepStrictEqual(record[k], fields[k]);
|
||||
}
|
||||
});
|
||||
|
@ -52,7 +50,7 @@ describe('Werelogs is usable as a dependency', () => {
|
|||
assert.doesNotThrow(
|
||||
createModuleLogger,
|
||||
Error,
|
||||
'Werelogs threw an exception trying to create a ModuleLogger.',
|
||||
'Werelogs threw an exception trying to create a ModuleLogger.'
|
||||
);
|
||||
done();
|
||||
});
|
||||
|
@ -68,11 +66,8 @@ describe('Werelogs is usable as a dependency', () => {
|
|||
it('Should be able to log a message and additional fields', done => {
|
||||
const logger = createModuleLogger();
|
||||
const msg = 'This is a message with added fields';
|
||||
const fields = {
|
||||
errorCode: 9,
|
||||
description: 'TestError',
|
||||
options: { dump: false },
|
||||
};
|
||||
const fields = { errorCode: 9, description: 'TestError',
|
||||
options: { dump: false } };
|
||||
logger.info(msg, fields);
|
||||
assert.strictEqual(parseLogEntry().message, msg);
|
||||
checkFields(fields);
|
||||
|
@ -88,7 +83,7 @@ describe('Werelogs is usable as a dependency', () => {
|
|||
assert.doesNotThrow(
|
||||
() => createModuleLogger().newRequestLogger(),
|
||||
Error,
|
||||
'Werelogs threw an exception trying to create a ModuleLogger.',
|
||||
'Werelogs threw an exception trying to create a ModuleLogger.'
|
||||
);
|
||||
done();
|
||||
});
|
||||
|
@ -104,15 +99,42 @@ describe('Werelogs is usable as a dependency', () => {
|
|||
it('Should be able to log a message and additional fields', done => {
|
||||
const logger = createModuleLogger().newRequestLogger();
|
||||
const msg = 'This is a message with added fields';
|
||||
const fields = {
|
||||
errorCode: 9,
|
||||
description: 'TestError',
|
||||
options: { dump: false },
|
||||
};
|
||||
const fields = { errorCode: 9, description: 'TestError',
|
||||
options: { dump: false } };
|
||||
logger.info(msg, fields);
|
||||
assert.strictEqual(parseLogEntry().message, msg);
|
||||
checkFields(fields);
|
||||
done();
|
||||
});
|
||||
|
||||
it('Should not log a removed field', done => {
|
||||
const logger = createModuleLogger().newRequestLogger();
|
||||
const msg = 'This is a message with no fields(removed)';
|
||||
const fields = { errorCode: 0, description: 'TestNotFailing' };
|
||||
logger.addDefaultFields(fields);
|
||||
logger.removeDefaultFields(['errorCode', 'description']);
|
||||
logger.info(msg);
|
||||
assert.strictEqual(parseLogEntry().message, msg);
|
||||
assert(!parseLogEntry().hasOwnProperty('errorCode'));
|
||||
assert(!parseLogEntry().hasOwnProperty('description'));
|
||||
done();
|
||||
});
|
||||
|
||||
it('Should include the parent Loggers default fields', done => {
|
||||
const mFields = {
|
||||
name: 'TestModule',
|
||||
submodule: 'functional',
|
||||
};
|
||||
const logger = createModuleLogger(mFields);
|
||||
const rLog = logger.newRequestLogger();
|
||||
const msg =
|
||||
"This is a message including the module's default fields";
|
||||
rLog.info(msg);
|
||||
assert.strictEqual(parseLogEntry().message, msg);
|
||||
assert.deepStrictEqual(parseLogEntry().name, mFields.name);
|
||||
assert.deepStrictEqual(parseLogEntry().submodule,
|
||||
mFields.submodule);
|
||||
done();
|
||||
});
|
||||
});
|
||||
});
|
||||
|
|
|
@ -1,14 +1,13 @@
|
|||
const assert = require('assert');
|
||||
const { PassThrough } = require('stream');
|
||||
const PassThrough = require('stream').PassThrough;
|
||||
|
||||
const Werelogs = require('werelogs'); // eslint-disable-line
|
||||
const Werelogs = require('werelogs').Logger;
|
||||
const modules = [
|
||||
require('./module1.js'),
|
||||
require('./module2.js'),
|
||||
require('./module3.js'),
|
||||
];
|
||||
|
||||
const pass = new PassThrough();
|
||||
const pass = new PassThrough;
|
||||
|
||||
const logBuffer = {
|
||||
records: [],
|
||||
|
@ -17,10 +16,10 @@ pass.on('data', data => {
|
|||
logBuffer.records.push(JSON.parse(data.toString().trim()));
|
||||
});
|
||||
|
||||
describe('Config is shared and unique within one API', () => {
|
||||
it('should find all log entries in the RingBuffer with the right '
|
||||
+ 'module name', done => {
|
||||
Werelogs.configure({
|
||||
describe('Config is shared and unique within one program', () => {
|
||||
it('should find all log entries in the RingBuffer with the right ' +
|
||||
'module name', done => {
|
||||
const log = new Werelogs('test-index', {
|
||||
level: 'debug',
|
||||
dump: 'fatal',
|
||||
streams: [{
|
||||
|
@ -28,7 +27,6 @@ describe('Config is shared and unique within one API', () => {
|
|||
stream: pass,
|
||||
}],
|
||||
});
|
||||
const log = new Werelogs.Logger('test-index');
|
||||
modules.forEach(mod => { mod(); });
|
||||
log.warn('Logging as warn');
|
||||
const rLog = log.newRequestLogger();
|
||||
|
|
|
@ -1,4 +1,4 @@
|
|||
const Werelogs = require('werelogs').Logger; // eslint-disable-line
|
||||
const Werelogs = require('werelogs').Logger;
|
||||
|
||||
const log = new Werelogs('test-mod1');
|
||||
|
||||
|
|
|
@ -1,4 +1,4 @@
|
|||
const Werelogs = require('werelogs').Logger; // eslint-disable-line
|
||||
const Werelogs = require('werelogs').Logger;
|
||||
|
||||
const log = new Werelogs('test-mod2');
|
||||
|
||||
|
|
|
@ -1,4 +1,4 @@
|
|||
const Werelogs = require('werelogs').Logger; // eslint-disable-line
|
||||
const Werelogs = require('werelogs').Logger;
|
||||
|
||||
const log = new Werelogs('test-mod3');
|
||||
|
||||
|
|
|
@ -5,54 +5,51 @@ const assert = require('assert');
|
|||
const Config = require('../../lib/Config.js');
|
||||
|
||||
describe('Config', () => {
|
||||
const config = new Config();
|
||||
|
||||
beforeEach(() => {
|
||||
config.reset();
|
||||
Config.reset();
|
||||
});
|
||||
|
||||
it('should work with default configuration', done => {
|
||||
assert.doesNotThrow(
|
||||
() => {
|
||||
config.logger.info('test message');
|
||||
Config.logger.info('test message');
|
||||
},
|
||||
Error,
|
||||
);
|
||||
Error);
|
||||
done();
|
||||
});
|
||||
|
||||
it('log level should be updateable', done => {
|
||||
config.update({ level: 'debug' });
|
||||
assert.strictEqual(config.level, 'debug', 'Expected config\'s log level to be updated.');
|
||||
Config.update({ level: 'debug' });
|
||||
assert.strictEqual(Config.level, 'debug', 'Expected Config\'s log level to be updated.');
|
||||
done();
|
||||
});
|
||||
|
||||
it('dump threshold should be updateable', done => {
|
||||
const origDump = config.dump;
|
||||
assert.notStrictEqual(origDump, 'warn', 'Expected original config.dump to differ from value to update.');
|
||||
config.update({ dump: 'warn' });
|
||||
assert.strictEqual(config.dump, 'warn', 'Expected config\'s dump threshold to be updated.');
|
||||
const origDump = Config.dump;
|
||||
assert.notStrictEqual(origDump, 'warn', 'Expected original Config.dump to differ from value to update.');
|
||||
Config.update({ dump: 'warn' });
|
||||
assert.strictEqual(Config.dump, 'warn', 'Expected Config\'s dump threshold to be updated.');
|
||||
done();
|
||||
});
|
||||
|
||||
it('end logging level should be updateable', done => {
|
||||
const origEnd = config.end;
|
||||
assert.notStrictEqual(origEnd, 'trace', 'Expected original config.end to differ from value to update.');
|
||||
config.update({ end: 'trace' });
|
||||
assert.strictEqual(config.end, 'trace', 'Expected config\'s end log level to be updated.');
|
||||
const origEnd = Config.end;
|
||||
assert.notStrictEqual(origEnd, 'trace', 'Expected original Config.end to differ from value to update.');
|
||||
Config.update({ end: 'trace' });
|
||||
assert.strictEqual(Config.end, 'trace', 'Expected Config\'s end log level to be updated.');
|
||||
done();
|
||||
});
|
||||
|
||||
it('should not be modified by an empty config object', done => {
|
||||
const origLevel = config.level;
|
||||
const origDump = config.dump;
|
||||
const origLogger = config.logger;
|
||||
const origStreams = config.streams;
|
||||
config.update({});
|
||||
assert.deepStrictEqual(origLevel, config.level, 'Expected logging level not to have changed.');
|
||||
assert.deepStrictEqual(origDump, config.dump, 'Expected dump threshold not to have changed.');
|
||||
assert.strictEqual(origLogger, config.logger, 'Expected logger not to have changed.');
|
||||
assert.deepStrictEqual(origStreams, config.streams, 'Expected streams not to have changed.');
|
||||
const origLevel = Config.level;
|
||||
const origDump = Config.dump;
|
||||
const origLogger = Config.logger;
|
||||
const origStreams = Config.streams;
|
||||
Config.update({});
|
||||
assert.deepStrictEqual(origLevel, Config.level, 'Expected logging level not to have changed.');
|
||||
assert.deepStrictEqual(origDump, Config.dump, 'Expected dump threshold not to have changed.');
|
||||
assert.strictEqual(origLogger, Config.logger, 'Expected logger not to have changed.');
|
||||
assert.deepStrictEqual(origStreams, Config.streams, 'Expected streams not to have changed.');
|
||||
done();
|
||||
});
|
||||
});
|
||||
|
|
|
@ -0,0 +1,515 @@
|
|||
'use strict'; // eslint-disable-line strict
|
||||
|
||||
const assert = require('assert');
|
||||
|
||||
const DefaultFields = require('../../lib/DefaultFields.js');
|
||||
|
||||
describe('class DefaultFields', () => {
|
||||
describe('Basic Fields logic', () => {
|
||||
it('Can get the resulting fields', done => {
|
||||
const df = new DefaultFields();
|
||||
const fields = df._dfGetFields();
|
||||
assert(fields !== null && typeof fields === 'object');
|
||||
done();
|
||||
});
|
||||
|
||||
it('No fields are set by default', done => {
|
||||
const df = new DefaultFields();
|
||||
assert.deepStrictEqual(df._dfGetFields(), {});
|
||||
done();
|
||||
});
|
||||
|
||||
it('Validate defensive coding prevents undefined parentFields',
|
||||
done => {
|
||||
const df = new DefaultFields();
|
||||
df._dfSetParentFields();
|
||||
assert.deepStrictEqual(df._dfGetFields(), {});
|
||||
done();
|
||||
});
|
||||
|
||||
describe('With fields', () => {
|
||||
let df = null;
|
||||
const fields = {
|
||||
name: 'Testing',
|
||||
testing: true,
|
||||
count: 4,
|
||||
data: 'doggy',
|
||||
};
|
||||
|
||||
beforeEach(done => {
|
||||
df = new DefaultFields();
|
||||
df.addDefaultFields(fields);
|
||||
done();
|
||||
});
|
||||
|
||||
it('Can add a dictionary of new fields', done => {
|
||||
assert.deepStrictEqual(df._dfGetFields(), fields);
|
||||
done();
|
||||
});
|
||||
|
||||
it('Can remove a list of invalid fields', done => {
|
||||
df.removeDefaultFields(['invalid', 'notthere']);
|
||||
done();
|
||||
});
|
||||
|
||||
it('Can remove a list of fields', done => {
|
||||
df.removeDefaultFields(['data', 'count']);
|
||||
assert.strictEqual(df._dfGetFields().data, undefined);
|
||||
assert.strictEqual(df._dfGetFields().count, undefined);
|
||||
assert.strictEqual(df._dfGetFields().name, 'Testing');
|
||||
assert.strictEqual(df._dfGetFields().testing, true);
|
||||
done();
|
||||
});
|
||||
|
||||
|
||||
it('Can reset the fields', done => {
|
||||
df.resetDefaultFields();
|
||||
assert.deepStrictEqual(df._dfGetFields(), {});
|
||||
done();
|
||||
});
|
||||
});
|
||||
});
|
||||
|
||||
describe('Basic Parent-Child logic', () => {
|
||||
let parentNode = null;
|
||||
|
||||
beforeEach(done => {
|
||||
parentNode = new DefaultFields();
|
||||
done();
|
||||
});
|
||||
|
||||
function linkedChild(pNode) {
|
||||
const childNode = new DefaultFields();
|
||||
childNode.setParent(pNode);
|
||||
return childNode;
|
||||
}
|
||||
|
||||
it('Can set a parent to a node', done => {
|
||||
const childNode = linkedChild(parentNode);
|
||||
assert.strictEqual(childNode.parent, parentNode);
|
||||
assert.notStrictEqual(parentNode.children.indexOf(childNode), -1);
|
||||
done();
|
||||
});
|
||||
|
||||
it('Can remove a child from a node', done => {
|
||||
const childNode = linkedChild(parentNode);
|
||||
parentNode._dfUnregisterChild(childNode);
|
||||
assert.strictEqual(parentNode.children.indexOf(childNode), -1);
|
||||
done();
|
||||
});
|
||||
|
||||
it('Can reset the parent of a node (w/ undefined)', done => {
|
||||
const childNode = linkedChild(parentNode);
|
||||
childNode.setParent(undefined);
|
||||
assert.strictEqual(childNode.parent, null);
|
||||
assert.strictEqual(parentNode.children.indexOf(childNode), -1);
|
||||
done();
|
||||
});
|
||||
|
||||
it('Can reset the parent of a node (w/ null)', done => {
|
||||
const childNode = linkedChild(parentNode);
|
||||
childNode.setParent(null);
|
||||
assert.strictEqual(childNode.parent, null);
|
||||
assert.strictEqual(parentNode.children.indexOf(childNode), -1);
|
||||
done();
|
||||
});
|
||||
});
|
||||
|
||||
describe('Single parent-child relationship', () => {
|
||||
let parentNode = null;
|
||||
let childNode = null;
|
||||
|
||||
beforeEach(done => {
|
||||
parentNode = new DefaultFields();
|
||||
childNode = new DefaultFields();
|
||||
childNode.setParent(parentNode);
|
||||
done();
|
||||
});
|
||||
|
||||
it('Child can define its own fields', done => {
|
||||
const fields = {
|
||||
child: true,
|
||||
parent: false,
|
||||
test: 1,
|
||||
};
|
||||
childNode.addDefaultFields(fields);
|
||||
assert.deepStrictEqual(childNode._dfGetFields(), fields);
|
||||
done();
|
||||
});
|
||||
|
||||
it('Parent can define its own fields', done => {
|
||||
const fields = {
|
||||
child: false,
|
||||
parent: true,
|
||||
test: 2,
|
||||
};
|
||||
parentNode.addDefaultFields(fields);
|
||||
assert.deepStrictEqual(parentNode._dfGetFields(), fields);
|
||||
done();
|
||||
});
|
||||
|
||||
it('Child inherits parents fields', done => {
|
||||
const fields = {
|
||||
child: true,
|
||||
parent: false,
|
||||
test: 3,
|
||||
};
|
||||
parentNode.addDefaultFields(fields);
|
||||
assert.deepStrictEqual(childNode._dfGetFields(), fields);
|
||||
done();
|
||||
});
|
||||
|
||||
it('Child inherits successive parent field updates', done => {
|
||||
const pFields1 = {
|
||||
parent: true,
|
||||
test: 4,
|
||||
};
|
||||
const rFields1 = {
|
||||
parent: true,
|
||||
test: 4,
|
||||
};
|
||||
const pFields2 = { child: false };
|
||||
const rFields2 = {
|
||||
parent: true,
|
||||
test: 4,
|
||||
child: false,
|
||||
};
|
||||
const pFields3 = {
|
||||
data: 'pouet',
|
||||
};
|
||||
const rFields3 = {
|
||||
parent: true,
|
||||
test: 4,
|
||||
child: false,
|
||||
data: 'pouet',
|
||||
};
|
||||
parentNode.addDefaultFields(pFields1);
|
||||
assert.deepStrictEqual(childNode._dfGetFields(), rFields1);
|
||||
parentNode.addDefaultFields(pFields2);
|
||||
assert.deepStrictEqual(childNode._dfGetFields(), rFields2);
|
||||
parentNode.addDefaultFields(pFields3);
|
||||
assert.deepStrictEqual(childNode._dfGetFields(), rFields3);
|
||||
done();
|
||||
});
|
||||
|
||||
it('Child inherits reset parent fields', done => {
|
||||
const pFields = {
|
||||
parent: true,
|
||||
test: 5,
|
||||
};
|
||||
parentNode.addDefaultFields(pFields);
|
||||
assert.deepStrictEqual(childNode._dfGetFields(), pFields);
|
||||
parentNode.resetDefaultFields();
|
||||
assert.deepStrictEqual(childNode._dfGetFields(), {});
|
||||
done();
|
||||
});
|
||||
|
||||
it('Child mixes parent and own fields', done => {
|
||||
const pFields = { parent: true };
|
||||
const cFields = {
|
||||
child: true,
|
||||
test: 6,
|
||||
};
|
||||
const rFields = {
|
||||
parent: true,
|
||||
child: true,
|
||||
test: 6,
|
||||
};
|
||||
parentNode.addDefaultFields(pFields);
|
||||
childNode.addDefaultFields(cFields);
|
||||
assert.deepStrictEqual(childNode._dfGetFields(), rFields);
|
||||
done();
|
||||
});
|
||||
|
||||
it('Child overrides conflicting parent fields', done => {
|
||||
const pFields = {
|
||||
parent: true,
|
||||
child: false,
|
||||
test: 0,
|
||||
};
|
||||
const cFields = {
|
||||
child: true,
|
||||
test: 7,
|
||||
};
|
||||
const rFields = {
|
||||
parent: true,
|
||||
child: true,
|
||||
test: 7,
|
||||
};
|
||||
parentNode.addDefaultFields(pFields);
|
||||
childNode.addDefaultFields(cFields);
|
||||
assert.deepStrictEqual(childNode._dfGetFields(), rFields);
|
||||
done();
|
||||
});
|
||||
});
|
||||
|
||||
describe('Multiple-level parent-child relationship', () => {
|
||||
let ggpNode = null;
|
||||
let gpNode = null;
|
||||
let pNode = null;
|
||||
let childNode = null;
|
||||
|
||||
beforeEach(done => {
|
||||
ggpNode = new DefaultFields();
|
||||
gpNode = new DefaultFields();
|
||||
gpNode.setParent(ggpNode);
|
||||
pNode = new DefaultFields();
|
||||
pNode.setParent(gpNode);
|
||||
childNode = new DefaultFields();
|
||||
childNode.setParent(pNode);
|
||||
done();
|
||||
});
|
||||
|
||||
it('Child inherits from whole hierarchy (newer -> older)',
|
||||
done => {
|
||||
const ggpFields = { great: true };
|
||||
const gpFields = { grand: true };
|
||||
const pFields = { parent: true };
|
||||
const rFields = {
|
||||
great: true,
|
||||
grand: true,
|
||||
parent: true,
|
||||
};
|
||||
pNode.addDefaultFields(pFields);
|
||||
gpNode.addDefaultFields(gpFields);
|
||||
ggpNode.addDefaultFields(ggpFields);
|
||||
assert.deepStrictEqual(childNode._dfGetFields(), rFields);
|
||||
done();
|
||||
});
|
||||
|
||||
it('Child inherits from whole hierarchy (older -> newer)',
|
||||
done => {
|
||||
const ggpFields = { great: true };
|
||||
const gpFields = { grand: true };
|
||||
const pFields = { parent: true };
|
||||
const rFields = {
|
||||
great: true,
|
||||
grand: true,
|
||||
parent: true,
|
||||
};
|
||||
ggpNode.addDefaultFields(ggpFields);
|
||||
gpNode.addDefaultFields(gpFields);
|
||||
pNode.addDefaultFields(pFields);
|
||||
assert.deepStrictEqual(childNode._dfGetFields(), rFields);
|
||||
done();
|
||||
});
|
||||
|
||||
it('Nodes inherit in-hierarchy fields reset', done => {
|
||||
const ggpFields = { great: true };
|
||||
const gpFields = { grand: true };
|
||||
const pFields = { parent: true };
|
||||
const rFields = {
|
||||
great: true,
|
||||
// grand: true, // Part 'reset'
|
||||
parent: true,
|
||||
};
|
||||
ggpNode.addDefaultFields(ggpFields);
|
||||
gpNode.addDefaultFields(gpFields);
|
||||
pNode.addDefaultFields(pFields);
|
||||
gpNode.resetDefaultFields();
|
||||
assert.deepStrictEqual(childNode._dfGetFields(), rFields);
|
||||
done();
|
||||
});
|
||||
|
||||
it('Field overriding is cascading through generations (newer -> older)',
|
||||
done => {
|
||||
const ggpFields = { generation: 0 };
|
||||
const gpFields = { generation: 1 };
|
||||
const pFields = { generation: 2 };
|
||||
const cFields = { generation: 3 };
|
||||
childNode.addDefaultFields(cFields);
|
||||
pNode.addDefaultFields(pFields);
|
||||
gpNode.addDefaultFields(gpFields);
|
||||
ggpNode.addDefaultFields(ggpFields);
|
||||
assert.deepStrictEqual(childNode._dfGetFields(), cFields);
|
||||
assert.deepStrictEqual(pNode._dfGetFields(), pFields);
|
||||
assert.deepStrictEqual(gpNode._dfGetFields(), gpFields);
|
||||
assert.deepStrictEqual(ggpNode._dfGetFields(), ggpFields);
|
||||
done();
|
||||
});
|
||||
|
||||
it('Field overriding is cascading through generations (older -> newer)',
|
||||
done => {
|
||||
const ggpFields = { generation: 0 };
|
||||
const gpFields = { generation: 1 };
|
||||
const pFields = { generation: 2 };
|
||||
const cFields = { generation: 3 };
|
||||
ggpNode.addDefaultFields(ggpFields);
|
||||
gpNode.addDefaultFields(gpFields);
|
||||
pNode.addDefaultFields(pFields);
|
||||
childNode.addDefaultFields(cFields);
|
||||
assert.deepStrictEqual(childNode._dfGetFields(), cFields);
|
||||
assert.deepStrictEqual(pNode._dfGetFields(), pFields);
|
||||
assert.deepStrictEqual(gpNode._dfGetFields(), gpFields);
|
||||
assert.deepStrictEqual(ggpNode._dfGetFields(), ggpFields);
|
||||
done();
|
||||
});
|
||||
|
||||
it('Destroying intermediate level breaks relationships', done => {
|
||||
const ggpFields = { ggp: 1 };
|
||||
const gpFields = { gp: 1 };
|
||||
const pFields = { p: 1 };
|
||||
const cFields = { c: 1 };
|
||||
ggpNode.addDefaultFields(ggpFields);
|
||||
gpNode.addDefaultFields(gpFields);
|
||||
pNode.addDefaultFields(pFields);
|
||||
childNode.addDefaultFields(cFields);
|
||||
pNode._dfDestroy();
|
||||
assert.strictEqual(gpNode.children.indexOf(pNode), -1);
|
||||
assert.strictEqual(pNode.parent, null);
|
||||
assert.strictEqual(pNode.children.indexOf(childNode), -1);
|
||||
assert.strictEqual(childNode.parent, null);
|
||||
assert.deepStrictEqual(pNode._dfGetFields(), pFields);
|
||||
assert.deepStrictEqual(childNode._dfGetFields(), cFields);
|
||||
done();
|
||||
});
|
||||
|
||||
it('Destroying intermediate level(2) breaks relationships', done => {
|
||||
const ggpFields = { ggp: 1 };
|
||||
const gpFields = { gp: 1 };
|
||||
const pFields = { p: 1 };
|
||||
const cFields = { c: 1 };
|
||||
const rCFields = { p: 1, c: 1 };
|
||||
ggpNode.addDefaultFields(ggpFields);
|
||||
gpNode.addDefaultFields(gpFields);
|
||||
pNode.addDefaultFields(pFields);
|
||||
childNode.addDefaultFields(cFields);
|
||||
gpNode._dfDestroy();
|
||||
assert.strictEqual(ggpNode.children.indexOf(gpNode), -1);
|
||||
assert.strictEqual(gpNode.parent, null);
|
||||
assert.strictEqual(gpNode.children.indexOf(pNode), -1);
|
||||
assert.strictEqual(pNode.parent, null);
|
||||
assert.deepStrictEqual(gpNode._dfGetFields(), gpFields);
|
||||
assert.deepStrictEqual(childNode._dfGetFields(), rCFields);
|
||||
done();
|
||||
});
|
||||
});
|
||||
|
||||
describe('Topology changes', () => {
|
||||
let ggpNode1 = null;
|
||||
let ggpNode2 = null;
|
||||
let gpNode1 = null;
|
||||
let gpNode2 = null;
|
||||
let pNode1 = null;
|
||||
let pNode2 = null;
|
||||
let cNode = null;
|
||||
|
||||
const ggp1Fields = { ggp1: true, generation: 0 };
|
||||
const ggp2Fields = { ggp2: true, generation: 0 };
|
||||
const gp1Fields = { gp1: true, generation: 1 };
|
||||
const gp2Fields = { gp2: true, generation: 1 };
|
||||
const p1Fields = { p1: true, generation: 2 };
|
||||
const p2Fields = { p2: true, generation: 2 };
|
||||
const cFields = { c: true, generation: 3 };
|
||||
|
||||
const startFields = {
|
||||
ggp1: true,
|
||||
gp1: true,
|
||||
p1: true,
|
||||
c: true,
|
||||
generation: 3,
|
||||
};
|
||||
|
||||
beforeEach(done => {
|
||||
ggpNode1 = new DefaultFields();
|
||||
gpNode1 = new DefaultFields();
|
||||
gpNode1.setParent(ggpNode1);
|
||||
pNode1 = new DefaultFields();
|
||||
pNode1.setParent(gpNode1);
|
||||
|
||||
ggpNode2 = new DefaultFields();
|
||||
gpNode2 = new DefaultFields();
|
||||
gpNode2.setParent(ggpNode2);
|
||||
pNode2 = new DefaultFields();
|
||||
pNode2.setParent(gpNode2);
|
||||
|
||||
cNode = new DefaultFields();
|
||||
cNode.setParent(pNode1);
|
||||
|
||||
ggpNode1.addDefaultFields(ggp1Fields);
|
||||
ggpNode2.addDefaultFields(ggp2Fields);
|
||||
gpNode1.addDefaultFields(gp1Fields);
|
||||
gpNode2.addDefaultFields(gp2Fields);
|
||||
pNode1.addDefaultFields(p1Fields);
|
||||
pNode2.addDefaultFields(p2Fields);
|
||||
cNode.addDefaultFields(cFields);
|
||||
|
||||
done();
|
||||
});
|
||||
|
||||
it('Nodes are updated accordingly when a parent changes '
|
||||
+ '(change whole upper hierarchy)',
|
||||
done => {
|
||||
const rFields = {
|
||||
ggp2: true,
|
||||
gp2: true,
|
||||
p2: true,
|
||||
c: true,
|
||||
generation: 3,
|
||||
};
|
||||
assert.deepStrictEqual(cNode._dfGetFields(), startFields);
|
||||
cNode.setParent(pNode2);
|
||||
assert.deepStrictEqual(cNode._dfGetFields(), rFields);
|
||||
done();
|
||||
});
|
||||
|
||||
it('Nodes are updated accordingly when a parent changes '
|
||||
+ '(change part of the upper hierarchy)',
|
||||
done => {
|
||||
const rFields = {
|
||||
ggp2: true,
|
||||
gp2: true,
|
||||
p1: true,
|
||||
c: true,
|
||||
generation: 3,
|
||||
};
|
||||
assert.deepStrictEqual(cNode._dfGetFields(), startFields);
|
||||
pNode1.setParent(gpNode2);
|
||||
assert.deepStrictEqual(cNode._dfGetFields(), rFields);
|
||||
done();
|
||||
});
|
||||
|
||||
it('Nodes are updated accordingly when a parent changes '
|
||||
+ '(shortcut hierarchy)',
|
||||
done => {
|
||||
const rFields = {
|
||||
ggp2: true,
|
||||
c: true,
|
||||
generation: 3,
|
||||
};
|
||||
assert.deepStrictEqual(cNode._dfGetFields(), startFields);
|
||||
cNode.setParent(ggpNode2);
|
||||
assert.deepStrictEqual(cNode._dfGetFields(), rFields);
|
||||
done();
|
||||
});
|
||||
|
||||
it('Nodes are updated accordingly when a parent is unset '
|
||||
+ '(remove whole hierarchy)',
|
||||
done => {
|
||||
const rFields = {
|
||||
c: true,
|
||||
generation: 3,
|
||||
};
|
||||
assert.deepStrictEqual(cNode._dfGetFields(), startFields);
|
||||
cNode.setParent(null);
|
||||
assert.deepStrictEqual(cNode._dfGetFields(), rFields);
|
||||
done();
|
||||
});
|
||||
|
||||
it('Nodes are updated accordingly when a parent is unset '
|
||||
+ '(remove upper hierarchy)',
|
||||
done => {
|
||||
const rFields = {
|
||||
gp1: true,
|
||||
p1: true,
|
||||
c: true,
|
||||
generation: 3,
|
||||
};
|
||||
assert.deepStrictEqual(cNode._dfGetFields(), startFields);
|
||||
gpNode1.setParent(null);
|
||||
assert.deepStrictEqual(cNode._dfGetFields(), rFields);
|
||||
done();
|
||||
});
|
||||
});
|
||||
});
|
||||
|
|
@ -1,5 +1,4 @@
|
|||
|
||||
// eslint-disable-line strict
|
||||
'use strict'; // eslint-disable-line strict
|
||||
|
||||
const assert = require('assert');
|
||||
|
||||
|
@ -13,8 +12,7 @@ function generateValidThrowTest(level) {
|
|||
},
|
||||
Error,
|
||||
'Expected level to be valid and '
|
||||
+ 'the function not to throw an Error.',
|
||||
);
|
||||
+ 'the function not to throw an Error.');
|
||||
done();
|
||||
};
|
||||
}
|
||||
|
@ -28,28 +26,27 @@ describe('LogLevel', () => {
|
|||
},
|
||||
RangeError,
|
||||
'Expected function to throw an Error instance due to '
|
||||
+ 'invalid log level.',
|
||||
);
|
||||
+ 'invalid log level.');
|
||||
done();
|
||||
});
|
||||
|
||||
it('should not throw on "trace" level',
|
||||
generateValidThrowTest('trace'));
|
||||
generateValidThrowTest('trace'));
|
||||
|
||||
it('should not throw on "debug" level',
|
||||
generateValidThrowTest('debug'));
|
||||
generateValidThrowTest('debug'));
|
||||
|
||||
it('should not throw on "info" level',
|
||||
generateValidThrowTest('info'));
|
||||
generateValidThrowTest('info'));
|
||||
|
||||
it('should not throw on "warn" level',
|
||||
generateValidThrowTest('warn'));
|
||||
generateValidThrowTest('warn'));
|
||||
|
||||
it('should not throw on "error" level',
|
||||
generateValidThrowTest('error'));
|
||||
generateValidThrowTest('error'));
|
||||
|
||||
it('should not throw on "fatal" level',
|
||||
generateValidThrowTest('fatal'));
|
||||
generateValidThrowTest('fatal'));
|
||||
});
|
||||
|
||||
describe('shouldLog(level, floor)', () => {
|
||||
|
@ -57,8 +54,7 @@ describe('LogLevel', () => {
|
|||
assert.strictEqual(
|
||||
LogLevel.shouldLog('trace', 'trace'),
|
||||
true,
|
||||
'Expected trace floor to allow logging trace level.',
|
||||
);
|
||||
'Expected trace floor to allow logging trace level.');
|
||||
done();
|
||||
});
|
||||
|
||||
|
@ -66,8 +62,7 @@ describe('LogLevel', () => {
|
|||
assert.strictEqual(
|
||||
LogLevel.shouldLog('debug', 'debug'),
|
||||
true,
|
||||
'Expected debug floor to allow logging debug level.',
|
||||
);
|
||||
'Expected debug floor to allow logging debug level.');
|
||||
done();
|
||||
});
|
||||
|
||||
|
@ -75,8 +70,7 @@ describe('LogLevel', () => {
|
|||
assert.strictEqual(
|
||||
LogLevel.shouldLog('info', 'info'),
|
||||
true,
|
||||
'Expected info floor to allow logging info level.',
|
||||
);
|
||||
'Expected info floor to allow logging info level.');
|
||||
done();
|
||||
});
|
||||
|
||||
|
@ -84,8 +78,7 @@ describe('LogLevel', () => {
|
|||
assert.strictEqual(
|
||||
LogLevel.shouldLog('warn', 'warn'),
|
||||
true,
|
||||
'Expected warn floor to allow logging warn level.',
|
||||
);
|
||||
'Expected warn floor to allow logging warn level.');
|
||||
done();
|
||||
});
|
||||
|
||||
|
@ -93,8 +86,7 @@ describe('LogLevel', () => {
|
|||
assert.strictEqual(
|
||||
LogLevel.shouldLog('error', 'error'),
|
||||
true,
|
||||
'Expected error floor to allow logging error level.',
|
||||
);
|
||||
'Expected error floor to allow logging error level.');
|
||||
done();
|
||||
});
|
||||
|
||||
|
@ -102,8 +94,7 @@ describe('LogLevel', () => {
|
|||
assert.strictEqual(
|
||||
LogLevel.shouldLog('fatal', 'fatal'),
|
||||
true,
|
||||
'Expected fatal floor to allow logging fatal level.',
|
||||
);
|
||||
'Expected fatal floor to allow logging fatal level.');
|
||||
done();
|
||||
});
|
||||
});
|
||||
|
|
|
@ -1,15 +1,15 @@
|
|||
|
||||
// eslint-disable-line strict
|
||||
'use strict'; // eslint-disable-line strict
|
||||
|
||||
const assert = require('assert');
|
||||
|
||||
const { genericFilterGenerator, loggingMisuseGenerator, DummyLogger } = require('../Utils');
|
||||
const Utils = require('../Utils.js');
|
||||
const genericFilterGenerator = Utils.genericFilterGenerator;
|
||||
const loggingMisuseGenerator = Utils.loggingMisuseGenerator;
|
||||
const DummyLogger = Utils.DummyLogger;
|
||||
|
||||
const Config = require('../../lib/Config.js');
|
||||
const RequestLogger = require('../../lib/RequestLogger.js');
|
||||
const Logger = require('../../lib/Logger.js');
|
||||
|
||||
const config = new Config();
|
||||
const Logger = require('../../index.js').Logger;
|
||||
|
||||
/*
|
||||
* This function is a thunk-function calling the Utils' filterGenerator with
|
||||
|
@ -18,14 +18,18 @@ const config = new Config();
|
|||
*/
|
||||
function filterGenerator(logLevel, callLevel) {
|
||||
function createModuleLogger(dummyLogger, filterLevel) {
|
||||
const logger = new Logger('TestModuleLogger',
|
||||
{
|
||||
level: filterLevel,
|
||||
dump: 'fatal',
|
||||
});
|
||||
/*
|
||||
* Here, patch the config by setting a specifically designed dummyLogger
|
||||
* Here, patch the Config by setting a specificly designed dummyLogger
|
||||
* for testing purposes that will help us collect runtime data.
|
||||
*/
|
||||
const testConfig = new Config({ level: filterLevel, dump: 'fatal' });
|
||||
testConfig.simpleLogger = dummyLogger;
|
||||
Config.simpleLogger = dummyLogger;
|
||||
|
||||
return new Logger(testConfig, 'TestModuleLogger');
|
||||
return logger;
|
||||
}
|
||||
|
||||
return genericFilterGenerator(logLevel, callLevel, createModuleLogger);
|
||||
|
@ -33,92 +37,146 @@ function filterGenerator(logLevel, callLevel) {
|
|||
|
||||
function checkFields(src, result) {
|
||||
Object.keys(src).forEach(k => {
|
||||
if (Object.prototype.hasOwnProperty.call(src, k)) {
|
||||
if (src.hasOwnProperty(k)) {
|
||||
assert.deepStrictEqual(result[k], src[k]);
|
||||
}
|
||||
});
|
||||
assert.ok(Object.prototype.hasOwnProperty.call(result, 'time'));
|
||||
assert.ok(result.hasOwnProperty('time'));
|
||||
// Time field should be current give or take 1s
|
||||
assert.ok((Date.now() - result.time) < 1000);
|
||||
}
|
||||
|
||||
|
||||
describe('Logger is usable:', () => {
|
||||
describe('WereLogs Logger is usable:', () => {
|
||||
beforeEach(() => {
|
||||
config.reset();
|
||||
Config.reset();
|
||||
});
|
||||
|
||||
it('Cannot be instanciated without parameters', done => {
|
||||
assert.throws(
|
||||
() => new Logger(),
|
||||
TypeError,
|
||||
'Logger Instanciation should not succeed without parameter.',
|
||||
);
|
||||
done();
|
||||
});
|
||||
|
||||
it('Cannot be instanciated with only a config', done => {
|
||||
assert.throws(
|
||||
() => new Logger(config),
|
||||
TypeError,
|
||||
'Logger Instanciation should not be succeed without a name.',
|
||||
);
|
||||
done();
|
||||
});
|
||||
|
||||
it('Cannot be instanciated with a bad config type', done => {
|
||||
assert.throws(
|
||||
() => new Logger({ level: 'info' }, 'WereLogsTest'),
|
||||
TypeError,
|
||||
'Logger Instanciation should not succeed with a bad config type.',
|
||||
);
|
||||
done();
|
||||
});
|
||||
|
||||
it('Cannot be instanciated with only a name', done => {
|
||||
assert.throws(
|
||||
it('Can be instanciated with only a name', done => {
|
||||
assert.doesNotThrow(
|
||||
() => new Logger('WereLogsTest'),
|
||||
TypeError,
|
||||
'Logger Instanciation should not succeed with only a name.',
|
||||
);
|
||||
Error,
|
||||
'WereLogs Instanciation should not throw any kind of error.');
|
||||
done();
|
||||
});
|
||||
|
||||
it('Cannot be instanciated with invalid log level', done => {
|
||||
assert.throws(
|
||||
() => new Logger('test', { level: 'invalidlevel' }),
|
||||
RangeError,
|
||||
// eslint-disable-next-line max-len
|
||||
'WereLogs should not be instanciable without the proper logging levels.');
|
||||
done();
|
||||
});
|
||||
|
||||
it('Cannot be instanciated with invalid dump threshold level', done => {
|
||||
assert.throws(
|
||||
() => new Logger('test', { level: 'trace', dump: 'invalidlevel' }),
|
||||
RangeError,
|
||||
// eslint-disable-next-line max-len
|
||||
'WereLogs should not be instanciable without the proper dumping threshold levels.');
|
||||
|
||||
done();
|
||||
});
|
||||
|
||||
it('Cannot be instanciated with a non-Array in config.streams', done => {
|
||||
assert.throws(
|
||||
() => new Logger('test', { streams: process.stdout }),
|
||||
Error,
|
||||
// eslint-disable-next-line max-len
|
||||
'Werelogs should not be instanciable with a stream option that is not an array.');
|
||||
done();
|
||||
});
|
||||
|
||||
it('Cannot be instanciated with an empty Array in config.streams', done => {
|
||||
assert.throws(
|
||||
() => new Logger('test', { streams: [] }),
|
||||
Error,
|
||||
// eslint-disable-next-line max-len
|
||||
'Werelogs should not be instanciable with an empty array for the streams option.');
|
||||
done();
|
||||
});
|
||||
|
||||
it('Cannot set logging level to invalid level at runtime', done => {
|
||||
const logger = new Logger('test');
|
||||
assert.throws(
|
||||
() => {
|
||||
logger.setLevel('invalidLevel');
|
||||
},
|
||||
RangeError,
|
||||
// eslint-disable-next-line max-len
|
||||
'WereLogs should not be able to set log level to an invalid level.');
|
||||
done();
|
||||
});
|
||||
|
||||
it('Can set logging level at runtime', done => {
|
||||
const logger = new Logger('test');
|
||||
assert.doesNotThrow(
|
||||
() => {
|
||||
logger.setLevel('fatal');
|
||||
},
|
||||
RangeError,
|
||||
'WereLogs should be able to set log level at runtime.');
|
||||
done();
|
||||
});
|
||||
|
||||
it('Cannot set dump threshold to invalid level at runtime', done => {
|
||||
const logger = new Logger('test');
|
||||
assert.throws(
|
||||
() => {
|
||||
logger.setDumpThreshold('invalidLevel');
|
||||
},
|
||||
RangeError,
|
||||
// eslint-disable-next-line max-len
|
||||
'WereLogs should not be able to set dump threshold to an invalid level.');
|
||||
done();
|
||||
});
|
||||
|
||||
it('Can set dump threshold at runtime', done => {
|
||||
const logger = new Logger('test');
|
||||
assert.doesNotThrow(
|
||||
() => {
|
||||
logger.setDumpThreshold('fatal');
|
||||
},
|
||||
RangeError,
|
||||
'WereLogs should be able to set dump threshold at runtime.');
|
||||
done();
|
||||
});
|
||||
|
||||
it('Can create Per-Request Loggers', done => {
|
||||
const logger = new Logger(config, 'test');
|
||||
const logger = new Logger('test');
|
||||
assert.doesNotThrow(
|
||||
() => {
|
||||
logger.newRequestLogger();
|
||||
},
|
||||
Error,
|
||||
'Werelogs should not throw when creating a request logger.',
|
||||
);
|
||||
'Werelogs should not throw when creating a request logger.');
|
||||
const reqLogger = logger.newRequestLogger();
|
||||
assert(reqLogger instanceof RequestLogger, 'RequestLogger');
|
||||
done();
|
||||
});
|
||||
|
||||
it('Can create Per-Request Loggers from a Serialized UID Array', done => {
|
||||
const logger = new Logger(config, 'test');
|
||||
const logger = new Logger('test');
|
||||
assert.doesNotThrow(
|
||||
() => {
|
||||
logger.newRequestLogger();
|
||||
},
|
||||
Error,
|
||||
// eslint-disable-next-line max-len
|
||||
'Werelogs should not throw when creating a request logger from a Serialized UID Array.',
|
||||
);
|
||||
'Werelogs should not throw when creating a request logger from a Serialized UID Array.');
|
||||
const reqLogger = logger.newRequestLoggerFromSerializedUids(
|
||||
'OneUID:SecondUID:TestUID:YouWinUID',
|
||||
);
|
||||
'OneUID:SecondUID:TestUID:YouWinUID');
|
||||
assert(reqLogger instanceof RequestLogger, 'RequestLogger');
|
||||
assert.deepStrictEqual(reqLogger.getUids().slice(0, -1),
|
||||
['OneUID', 'SecondUID', 'TestUID', 'YouWinUID']);
|
||||
['OneUID', 'SecondUID', 'TestUID', 'YouWinUID']);
|
||||
done();
|
||||
});
|
||||
|
||||
it('Uses the additional fields as expected', done => {
|
||||
const dummyLogger = new DummyLogger();
|
||||
config.simpleLogger = dummyLogger;
|
||||
const logger = new Logger(config, 'test');
|
||||
const logger = new Logger('test');
|
||||
Config.simpleLogger = dummyLogger;
|
||||
const fields = {
|
||||
ip: '127.0.0.1',
|
||||
method: 'GET',
|
||||
|
@ -134,27 +192,27 @@ describe('Logger is usable:', () => {
|
|||
describe('Does not crash and logs a fatal message when mis-using its logging API', () => {
|
||||
const testValues = [
|
||||
{ desc: 'a string as second argument', args: ['test', 'second-param-string'] },
|
||||
{ desc: 'a function as second argument', args: ['test', () => { }] }, // eslint-disable-line arrow-body-style
|
||||
{ desc: 'a function as second argument', args: ['test', () => { return; }] }, // eslint-disable-line arrow-body-style
|
||||
{ desc: 'a Number as second argument', args: ['test', 1] },
|
||||
{ desc: 'more than 2 arguments', args: ['test', 2, 3, 4] },
|
||||
];
|
||||
/* eslint-enable max-len */
|
||||
function createMisusableLogger(dummyLogger) {
|
||||
config.simpleLogger = dummyLogger;
|
||||
const logger = new Logger(config, 'test');
|
||||
const logger = new Logger('test');
|
||||
Config.simpleLogger = dummyLogger;
|
||||
return logger;
|
||||
}
|
||||
|
||||
for (let i = 0; i < testValues.length; ++i) {
|
||||
const test = testValues[i];
|
||||
it(`Does not crash with ${test.desc}`,
|
||||
loggingMisuseGenerator(test, createMisusableLogger));
|
||||
loggingMisuseGenerator(test, createMisusableLogger));
|
||||
}
|
||||
});
|
||||
});
|
||||
|
||||
/* eslint-disable no-multi-spaces, max-len */
|
||||
describe('Logger can log as specified by the log level', () => {
|
||||
describe('Werelogs Module-level Logger can log as specified by the log level', () => {
|
||||
it('Trace level does not filter trace level out', filterGenerator('trace', 'trace'));
|
||||
it('Trace level does not filter debug level out', filterGenerator('trace', 'debug'));
|
||||
it('Trace level does not filter info level out', filterGenerator('trace', 'info'));
|
||||
|
@ -198,3 +256,85 @@ describe('Logger can log as specified by the log level', () => {
|
|||
it('Fatal level does not filter fatal level out', filterGenerator('fatal', 'fatal'));
|
||||
});
|
||||
/* eslint-enable no-multi-spaces, max-len */
|
||||
|
||||
describe('Werelogs Module-level Logger provide the DefaultFields logic', () => {
|
||||
const loggerConfig = {
|
||||
level: 'info',
|
||||
dump: 'fatal',
|
||||
end: 'info',
|
||||
};
|
||||
|
||||
it('should not modify the object passed as a parameter', done => {
|
||||
const defaultFields = {
|
||||
name: 'TestDefaultFields1',
|
||||
data: 0,
|
||||
};
|
||||
const add1 = {
|
||||
attr1: 0,
|
||||
};
|
||||
const add2 = {
|
||||
attr2: 'string',
|
||||
};
|
||||
const dummyLogger = new DummyLogger();
|
||||
Config.simpleLogger = dummyLogger;
|
||||
const logger = new Logger(defaultFields, loggerConfig);
|
||||
logger.addDefaultFields(add1);
|
||||
logger.addDefaultFields(add2);
|
||||
assert.deepStrictEqual(add1, { attr1: 0 });
|
||||
assert.deepStrictEqual(add2, { attr2: 'string' });
|
||||
done();
|
||||
});
|
||||
|
||||
it('should add one added default field to the log entries', done => {
|
||||
const defaultFields = {
|
||||
name: 'TestDefaultFields2',
|
||||
data: 0,
|
||||
};
|
||||
const clientInfo = {
|
||||
clientIP: '127.0.0.1',
|
||||
};
|
||||
const dummyLogger = new DummyLogger();
|
||||
Config.simpleLogger = dummyLogger;
|
||||
const logger = new Logger(defaultFields, loggerConfig);
|
||||
logger.addDefaultFields(clientInfo);
|
||||
logger.info('test message');
|
||||
assert.strictEqual(dummyLogger.ops[0][1][0].clientIP,
|
||||
clientInfo.clientIP);
|
||||
done();
|
||||
});
|
||||
|
||||
it('should add multiple added default fields to the log entries',
|
||||
done => {
|
||||
const defaultFields = {
|
||||
name: 'TestDefaultFields3',
|
||||
data: 0,
|
||||
};
|
||||
const clientInfo = {
|
||||
clientIP: '127.0.0.1',
|
||||
clientPort: '1337',
|
||||
};
|
||||
const requestInfo = {
|
||||
object: '/tata/self.txt',
|
||||
creator: 'Joddy',
|
||||
};
|
||||
const dummyLogger = new DummyLogger();
|
||||
Config.simpleLogger = dummyLogger;
|
||||
const logger = new Logger(defaultFields, loggerConfig);
|
||||
logger.addDefaultFields(clientInfo);
|
||||
logger.addDefaultFields(requestInfo);
|
||||
logger.info('test message');
|
||||
assert.strictEqual(dummyLogger.ops[0][1][0].clientIP,
|
||||
clientInfo.clientIP);
|
||||
assert.strictEqual(dummyLogger.ops[0][1][0].clientPort,
|
||||
clientInfo.clientPort);
|
||||
assert.strictEqual(dummyLogger.ops[0][1][0].object,
|
||||
requestInfo.object);
|
||||
assert.strictEqual(dummyLogger.ops[0][1][0].creator,
|
||||
requestInfo.creator);
|
||||
assert.strictEqual(dummyLogger.ops[0][1][0].name,
|
||||
defaultFields.name);
|
||||
assert.strictEqual(dummyLogger.ops[0][1][0].data,
|
||||
defaultFields.data);
|
||||
done();
|
||||
});
|
||||
});
|
||||
|
|
|
@ -1,9 +1,11 @@
|
|||
|
||||
// eslint-disable-line strict
|
||||
'use strict'; // eslint-disable-line strict
|
||||
|
||||
const assert = require('assert');
|
||||
|
||||
const { DummyLogger, genericFilterGenerator, loggingMisuseGenerator } = require('../Utils.js');
|
||||
const Utils = require('../Utils.js');
|
||||
const DummyLogger = Utils.DummyLogger;
|
||||
const genericFilterGenerator = Utils.genericFilterGenerator;
|
||||
const loggingMisuseGenerator = Utils.loggingMisuseGenerator;
|
||||
|
||||
const RequestLogger = require('../../lib/RequestLogger.js');
|
||||
|
||||
|
@ -22,7 +24,7 @@ function filterGenerator(logLevel, callLevel) {
|
|||
|
||||
|
||||
function runLoggingDumpTest(commandHistory, expectedHistory, expectedCounts,
|
||||
done) {
|
||||
done) {
|
||||
const dummyLogger = new DummyLogger();
|
||||
const reqLogger = new RequestLogger(dummyLogger, 'trace', 'error', 'info');
|
||||
|
||||
|
@ -44,9 +46,9 @@ function runLoggingDumpTest(commandHistory, expectedHistory, expectedCounts,
|
|||
|
||||
expectedHistory.every((val, index) => {
|
||||
assert.strictEqual(dummyLogger.ops[index][0], val[0],
|
||||
'Expected log entry levels to match.');
|
||||
'Expected log entry levels to match.');
|
||||
assert.strictEqual(dummyLogger.ops[index][1][1], val[1],
|
||||
'Expected log entry values to match.');
|
||||
'Expected log entry values to match.');
|
||||
return true;
|
||||
});
|
||||
assert.deepEqual(dummyLogger.counts, expectedCounts);
|
||||
|
@ -60,8 +62,7 @@ describe('RequestLogger', () => {
|
|||
assert.throws(
|
||||
() => new RequestLogger(undefined, 'fatal', 'debug', 'info'),
|
||||
Error,
|
||||
'Dump level "debug" should not be valid with logging level "fatal".',
|
||||
);
|
||||
'Dump level "debug" should not be valid with logging level "fatal".');
|
||||
done();
|
||||
});
|
||||
|
||||
|
@ -69,8 +70,7 @@ describe('RequestLogger', () => {
|
|||
assert.doesNotThrow(
|
||||
() => new RequestLogger(undefined, 'debug', 'fatal', 'info'),
|
||||
Error,
|
||||
'Dump level "fatal" should be valid with logging level "debug".',
|
||||
);
|
||||
'Dump level "fatal" should be valid with logging level "debug".');
|
||||
done();
|
||||
});
|
||||
});
|
||||
|
@ -98,8 +98,7 @@ describe('RequestLogger', () => {
|
|||
assert.throws(
|
||||
() => new RequestLogger(undefined, 'debug', 'fatal', 'info', 'pouet:tata'),
|
||||
Error,
|
||||
'UID string "pouet:tata" should be rejected by the RequestLogger constructor.',
|
||||
);
|
||||
'UID string "pouet:tata" should be rejected by the RequestLogger constructor.');
|
||||
done();
|
||||
});
|
||||
|
||||
|
@ -117,8 +116,7 @@ describe('RequestLogger', () => {
|
|||
assert.throws(
|
||||
() => new RequestLogger(undefined, 'debug', 'fatal', 'info', ['OneUID', 'SecondUID', 'Test:DashUID']),
|
||||
Error,
|
||||
'UID string "Test:DashUID" should be rejected by the RequestLogger constructor.',
|
||||
);
|
||||
'UID string "Test:DashUID" should be rejected by the RequestLogger constructor.');
|
||||
done();
|
||||
});
|
||||
});
|
||||
|
@ -188,7 +186,7 @@ describe('RequestLogger', () => {
|
|||
describe('Does not crash when mis-using its logging API', () => {
|
||||
const testValues = [
|
||||
{ desc: 'a string as second argument', args: ['test', 'second-param-string'] },
|
||||
{ desc: 'a function as second argument', args: ['test', function f() { }] },
|
||||
{ desc: 'a function as second argument', args: ['test', function f() { return; }] },
|
||||
{ desc: 'a Number as second argument', args: ['test', 1] },
|
||||
{ desc: 'more than 2 arguments', args: ['test', 2, 3, 4] },
|
||||
];
|
||||
|
@ -199,7 +197,7 @@ describe('RequestLogger', () => {
|
|||
for (let i = 0; i < testValues.length; ++i) {
|
||||
const test = testValues[i];
|
||||
it(`Does not crash with ${test.desc}`,
|
||||
loggingMisuseGenerator(test, createMisusableRequestLogger));
|
||||
loggingMisuseGenerator(test, createMisusableRequestLogger));
|
||||
}
|
||||
});
|
||||
|
||||
|
@ -250,35 +248,35 @@ describe('RequestLogger', () => {
|
|||
|
||||
describe('Logging API regression testing', () => {
|
||||
it('Should not alter the input fields when not actually logging',
|
||||
done => {
|
||||
const dummyLogger = new DummyLogger();
|
||||
const reqLogger = new RequestLogger(dummyLogger,
|
||||
'info', 'fatal', 'info');
|
||||
const refFields = { hits: 45, count: 32 };
|
||||
const usedFields = { ...refFields };
|
||||
reqLogger.debug('test', usedFields);
|
||||
assert.deepStrictEqual(usedFields, refFields);
|
||||
done();
|
||||
});
|
||||
done => {
|
||||
const dummyLogger = new DummyLogger();
|
||||
const reqLogger = new RequestLogger(dummyLogger,
|
||||
'info', 'fatal', 'info');
|
||||
const refFields = { hits: 45, count: 32 };
|
||||
const usedFields = Object.assign({}, refFields);
|
||||
reqLogger.debug('test', usedFields);
|
||||
assert.deepStrictEqual(usedFields, refFields);
|
||||
done();
|
||||
});
|
||||
|
||||
it('Should not alter the input fields when actually logging',
|
||||
done => {
|
||||
const dummyLogger = new DummyLogger();
|
||||
const reqLogger = new RequestLogger(dummyLogger,
|
||||
'info', 'fatal', 'info');
|
||||
const refFields = { hits: 45, count: 32 };
|
||||
const usedFields = { ...refFields };
|
||||
reqLogger.info('test', usedFields);
|
||||
assert.deepStrictEqual(usedFields, refFields);
|
||||
done();
|
||||
});
|
||||
done => {
|
||||
const dummyLogger = new DummyLogger();
|
||||
const reqLogger = new RequestLogger(dummyLogger,
|
||||
'info', 'fatal', 'info');
|
||||
const refFields = { hits: 45, count: 32 };
|
||||
const usedFields = Object.assign({}, refFields);
|
||||
reqLogger.info('test', usedFields);
|
||||
assert.deepStrictEqual(usedFields, refFields);
|
||||
done();
|
||||
});
|
||||
|
||||
it('Should not alter the input fields when dumping', done => {
|
||||
const dummyLogger = new DummyLogger();
|
||||
const reqLogger = new RequestLogger(dummyLogger,
|
||||
'info', 'fatal', 'info');
|
||||
'info', 'fatal', 'info');
|
||||
const refFields = { hits: 45, count: 32 };
|
||||
const usedFields = { ...refFields };
|
||||
const usedFields = Object.assign({}, refFields);
|
||||
reqLogger.error('test', usedFields);
|
||||
assert.deepStrictEqual(usedFields, refFields);
|
||||
done();
|
||||
|
@ -295,7 +293,7 @@ describe('RequestLogger', () => {
|
|||
};
|
||||
const dummyLogger = new DummyLogger();
|
||||
const reqLogger = new RequestLogger(dummyLogger,
|
||||
'info', 'fatal', 'info');
|
||||
'info', 'fatal', 'info');
|
||||
reqLogger.addDefaultFields(add1);
|
||||
reqLogger.addDefaultFields(add2);
|
||||
assert.deepStrictEqual(add1, { attr1: 0 });
|
||||
|
@ -309,70 +307,71 @@ describe('RequestLogger', () => {
|
|||
};
|
||||
const dummyLogger = new DummyLogger();
|
||||
const reqLogger = new RequestLogger(dummyLogger,
|
||||
'info', 'fatal', 'info');
|
||||
'info', 'fatal', 'info');
|
||||
reqLogger.addDefaultFields(clientInfo);
|
||||
reqLogger.info('test message');
|
||||
assert.strictEqual(clientInfo.clientIP,
|
||||
dummyLogger.ops[0][1][0].clientIP);
|
||||
dummyLogger.ops[0][1][0].clientIP);
|
||||
done();
|
||||
});
|
||||
|
||||
it('should add multiple added default fields to the log entries',
|
||||
done => {
|
||||
const clientInfo = {
|
||||
clientIP: '127.0.0.1',
|
||||
clientPort: '1337',
|
||||
};
|
||||
const requestInfo = {
|
||||
object: '/tata/self.txt',
|
||||
creator: 'Joddy',
|
||||
};
|
||||
const dummyLogger = new DummyLogger();
|
||||
const reqLogger = new RequestLogger(dummyLogger,
|
||||
'info', 'fatal', 'info');
|
||||
reqLogger.addDefaultFields(clientInfo);
|
||||
reqLogger.addDefaultFields(requestInfo);
|
||||
reqLogger.info('test message');
|
||||
assert.strictEqual(clientInfo.clientIP,
|
||||
dummyLogger.ops[0][1][0].clientIP);
|
||||
assert.strictEqual(clientInfo.clientPort,
|
||||
dummyLogger.ops[0][1][0].clientPort);
|
||||
assert.strictEqual(requestInfo.object,
|
||||
dummyLogger.ops[0][1][0].object);
|
||||
assert.strictEqual(requestInfo.creator,
|
||||
dummyLogger.ops[0][1][0].creator);
|
||||
done();
|
||||
});
|
||||
done => {
|
||||
const clientInfo = {
|
||||
clientIP: '127.0.0.1',
|
||||
clientPort: '1337',
|
||||
};
|
||||
const requestInfo = {
|
||||
object: '/tata/self.txt',
|
||||
creator: 'Joddy',
|
||||
};
|
||||
const dummyLogger = new DummyLogger();
|
||||
const reqLogger = new RequestLogger(dummyLogger,
|
||||
'info', 'fatal', 'info');
|
||||
reqLogger.addDefaultFields(clientInfo);
|
||||
reqLogger.addDefaultFields(requestInfo);
|
||||
reqLogger.info('test message');
|
||||
assert.strictEqual(clientInfo.clientIP,
|
||||
dummyLogger.ops[0][1][0].clientIP);
|
||||
assert.strictEqual(clientInfo.clientPort,
|
||||
dummyLogger.ops[0][1][0].clientPort);
|
||||
assert.strictEqual(requestInfo.object,
|
||||
dummyLogger.ops[0][1][0].object);
|
||||
assert.strictEqual(requestInfo.creator,
|
||||
dummyLogger.ops[0][1][0].creator);
|
||||
done();
|
||||
});
|
||||
});
|
||||
|
||||
describe('Automatic Elapsed Time computation', () => {
|
||||
describe('Deprecated API:', () => {
|
||||
it('should include an "elapsed_ms" field in the last log entry',
|
||||
done => {
|
||||
const dummyLogger = new DummyLogger();
|
||||
const reqLogger = new RequestLogger(dummyLogger,
|
||||
'info', 'fatal', 'info');
|
||||
reqLogger.end('Last message');
|
||||
assert.strictEqual(dummyLogger.ops[0][1][1], 'Last message');
|
||||
assert.notStrictEqual(dummyLogger.ops[0][1][0].elapsed_ms,
|
||||
undefined);
|
||||
assert.strictEqual(typeof dummyLogger.ops[0][1][0]
|
||||
.elapsed_ms, 'number');
|
||||
done();
|
||||
});
|
||||
done => {
|
||||
const dummyLogger = new DummyLogger();
|
||||
const reqLogger = new RequestLogger(dummyLogger,
|
||||
'info', 'fatal', 'info');
|
||||
reqLogger.end('Last message');
|
||||
assert.strictEqual(dummyLogger.ops[0][1][1], 'Last message');
|
||||
assert.notStrictEqual(dummyLogger.ops[0][1][0].elapsed_ms,
|
||||
undefined);
|
||||
assert.strictEqual(typeof(dummyLogger.ops[0][1][0]
|
||||
.elapsed_ms),
|
||||
'number');
|
||||
done();
|
||||
});
|
||||
|
||||
// eslint-disable-next-line max-len
|
||||
it('should include an "elapsed_ms" field in the last log entry and be error level', () => {
|
||||
const dummyLogger = new DummyLogger();
|
||||
const reqLogger = new RequestLogger(dummyLogger,
|
||||
'info', 'fatal', 'info');
|
||||
'info', 'fatal', 'info');
|
||||
reqLogger.errorEnd('Last message failed');
|
||||
assert.strictEqual(dummyLogger.ops[0][1][1],
|
||||
'Last message failed');
|
||||
'Last message failed');
|
||||
assert.notStrictEqual(dummyLogger.ops[0][1][0].elapsed_ms,
|
||||
undefined);
|
||||
assert.strictEqual(typeof dummyLogger.ops[0][1][0].elapsed_ms,
|
||||
'number');
|
||||
undefined);
|
||||
assert.strictEqual(typeof(dummyLogger.ops[0][1][0].elapsed_ms),
|
||||
'number');
|
||||
assert.strictEqual(dummyLogger.ops[0][0], 'error');
|
||||
});
|
||||
});
|
||||
|
@ -393,7 +392,7 @@ describe('RequestLogger', () => {
|
|||
endLogging[level](reqLogger.end())('Last message');
|
||||
assert.strictEqual(dummyLogger.ops[0][1][1], 'Last message');
|
||||
assert.notStrictEqual(dummyLogger.ops[0][1][0].elapsed_ms, undefined);
|
||||
assert.strictEqual(typeof dummyLogger.ops[0][1][0].elapsed_ms, 'number');
|
||||
assert.strictEqual(typeof(dummyLogger.ops[0][1][0].elapsed_ms), 'number');
|
||||
assert.strictEqual(dummyLogger.ops[0][0], level);
|
||||
done();
|
||||
});
|
||||
|
@ -408,87 +407,54 @@ describe('RequestLogger', () => {
|
|||
// end() logging
|
||||
reqLogger.end().error('Test Augmented END', { endValue: 42 });
|
||||
assert.strictEqual(dummyLogger.ops[0][1][1], 'Test Augmented END');
|
||||
assert.strictEqual(typeof dummyLogger.ops[0][1][0].elapsed_ms,
|
||||
'number');
|
||||
assert.strictEqual(typeof(dummyLogger.ops[0][1][0].elapsed_ms),
|
||||
'number');
|
||||
assert.strictEqual(dummyLogger.ops[0][1][0].endFlag, true);
|
||||
assert.strictEqual(dummyLogger.ops[0][1][0].endValue, 42);
|
||||
done();
|
||||
});
|
||||
|
||||
it('should log an error in addition to request logs when end() called more than once',
|
||||
done => {
|
||||
const dummyLogger = new DummyLogger();
|
||||
const reqLogger = new RequestLogger(dummyLogger, 'trace', 'fatal');
|
||||
reqLogger.end().info('after first call to end()');
|
||||
reqLogger.end().debug('after second call to end()');
|
||||
assert.strictEqual(dummyLogger.ops.length, 3);
|
||||
assert.strictEqual(dummyLogger.ops[0][0], 'info');
|
||||
assert.strictEqual(dummyLogger.ops[0][1][1], 'after first call to end()');
|
||||
assert.strictEqual(dummyLogger.ops[1][0], 'error');
|
||||
assert.strictEqual(dummyLogger.ops[2][0], 'debug');
|
||||
assert.strictEqual(dummyLogger.ops[2][1][1], 'after second call to end()');
|
||||
done();
|
||||
});
|
||||
});
|
||||
|
||||
describe('Log History dumped when logging floor level reached', () => {
|
||||
it('Dumping duplicates log entries', done => {
|
||||
const commandHistory = ['info', 'error'];
|
||||
const expectedHistory = [['info', 0], ['info', 0], ['error', 1]];
|
||||
const expectedCounts = {
|
||||
trace: 0,
|
||||
debug: 0,
|
||||
info: 2,
|
||||
warn: 0,
|
||||
error: 1,
|
||||
fatal: 0,
|
||||
};
|
||||
const expectedCounts = { trace: 0, debug: 0, info: 2, warn: 0,
|
||||
error: 1, fatal: 0 };
|
||||
|
||||
runLoggingDumpTest(commandHistory, expectedHistory, expectedCounts,
|
||||
done);
|
||||
done);
|
||||
done();
|
||||
});
|
||||
|
||||
it('Dumping Keeps logging history order', done => {
|
||||
const commandHistory = ['trace', 'info', 'debug', 'error'];
|
||||
const expectedHistory = [['trace', 0], ['info', 1], ['debug', 2],
|
||||
['trace', 0], ['info', 1], ['debug', 2],
|
||||
['error', 3]];
|
||||
const expectedCounts = {
|
||||
trace: 2,
|
||||
debug: 2,
|
||||
info: 2,
|
||||
warn: 0,
|
||||
error: 1,
|
||||
fatal: 0,
|
||||
};
|
||||
['trace', 0], ['info', 1], ['debug', 2],
|
||||
['error', 3]];
|
||||
const expectedCounts = { trace: 2, debug: 2, info: 2, warn: 0,
|
||||
error: 1, fatal: 0 };
|
||||
|
||||
runLoggingDumpTest(commandHistory, expectedHistory, expectedCounts,
|
||||
done);
|
||||
done);
|
||||
done();
|
||||
});
|
||||
|
||||
it('Dumping multiple times does not re-dump already-dumped entries',
|
||||
done => {
|
||||
const commandHistory = ['trace', 'info', 'debug', 'error',
|
||||
'warn', 'debug', 'fatal'];
|
||||
const expectedHistory = [['trace', 0], ['info', 1], ['debug', 2],
|
||||
['trace', 0], ['info', 1], ['debug', 2],
|
||||
['error', 3], ['warn', 4], ['debug', 5],
|
||||
['warn', 4], ['debug', 5],
|
||||
['fatal', 6]];
|
||||
const expectedCounts = {
|
||||
trace: 2,
|
||||
debug: 4,
|
||||
info: 2,
|
||||
warn: 2,
|
||||
error: 1,
|
||||
fatal: 1,
|
||||
};
|
||||
done => {
|
||||
const commandHistory = ['trace', 'info', 'debug', 'error',
|
||||
'warn', 'debug', 'fatal'];
|
||||
const expectedHistory = [['trace', 0], ['info', 1], ['debug', 2],
|
||||
['trace', 0], ['info', 1], ['debug', 2],
|
||||
['error', 3], ['warn', 4], ['debug', 5],
|
||||
['warn', 4], ['debug', 5],
|
||||
['fatal', 6]];
|
||||
const expectedCounts = { trace: 2, debug: 4, info: 2, warn: 2,
|
||||
error: 1, fatal: 1 };
|
||||
|
||||
runLoggingDumpTest(commandHistory, expectedHistory,
|
||||
expectedCounts, done);
|
||||
done();
|
||||
});
|
||||
runLoggingDumpTest(commandHistory, expectedHistory,
|
||||
expectedCounts, done);
|
||||
done();
|
||||
});
|
||||
});
|
||||
});
|
||||
|
|
|
@ -1,17 +1,17 @@
|
|||
|
||||
// eslint-disable-line strict
|
||||
'use strict'; // eslint-disable-line strict
|
||||
|
||||
const assert = require('assert');
|
||||
|
||||
const {
|
||||
generateUid, serializeUids, unserializeUids, objectCopy,
|
||||
} = require('../../lib/Utils');
|
||||
const Utils = require('../../lib/Utils.js');
|
||||
const generateUid = Utils.generateUid;
|
||||
const serializeUids = Utils.serializeUids;
|
||||
const unserializeUids = Utils.unserializeUids;
|
||||
const objectCopy = Utils.objectCopy;
|
||||
|
||||
describe('Utils: generateUid', () => {
|
||||
it('generates a string-typed ID', done => {
|
||||
const uid = generateUid();
|
||||
assert.strictEqual(typeof uid, 'string',
|
||||
`The generated ID is not a String (${typeof uid})`);
|
||||
assert.strictEqual(typeof(uid), 'string',
|
||||
`The generated ID is not a String (${typeof(uid)})`);
|
||||
done();
|
||||
});
|
||||
it('generate roughly unique IDs', done => {
|
||||
|
@ -24,8 +24,8 @@ describe('Utils: generateUid', () => {
|
|||
}
|
||||
Object.keys(generated).every(uid => {
|
||||
assert.strictEqual(generated[uid], 1,
|
||||
`Uid ${uid} was generated ${generated[uid]} `
|
||||
+ 'times: It is not even remotely unique.');
|
||||
`Uid ${uid} was generated ${generated[uid]} ` +
|
||||
'times: It is not even remotely unique.');
|
||||
return {};
|
||||
});
|
||||
done();
|
||||
|
@ -37,7 +37,7 @@ describe('Utils: serializeUids', () => {
|
|||
const uidList = ['FirstUID', 'SecondUID', 'ThirdUID'];
|
||||
const serializedUIDs = serializeUids(uidList);
|
||||
assert.strictEqual(serializedUIDs, 'FirstUID:SecondUID:ThirdUID',
|
||||
'Serialized UID List should match expected value.');
|
||||
'Serialized UID List should match expected value.');
|
||||
done();
|
||||
});
|
||||
|
||||
|
@ -54,12 +54,8 @@ describe('Utils: objectCopy', () => {
|
|||
it('copies all the properties from source to target object', done => {
|
||||
const target = { foo: 'bar' };
|
||||
const source = { id: 1, name: 'demo', value: { a: 1, b: 2, c: 3 } };
|
||||
const result = {
|
||||
foo: 'bar',
|
||||
id: 1,
|
||||
name: 'demo',
|
||||
value: { a: 1, b: 2, c: 3 },
|
||||
};
|
||||
const result = { foo: 'bar', id: 1, name: 'demo',
|
||||
value: { a: 1, b: 2, c: 3 } };
|
||||
objectCopy(target, source);
|
||||
assert.deepStrictEqual(target, result,
|
||||
'target should have the same properties as source');
|
||||
|
@ -67,32 +63,21 @@ describe('Utils: objectCopy', () => {
|
|||
});
|
||||
|
||||
it('copies all the properties from multiple sources to target object',
|
||||
done => {
|
||||
const target = { foo: 'bar' };
|
||||
const source1 = {
|
||||
id: 1,
|
||||
name: 'demo1',
|
||||
value: { a: 1, b: 2, c: 3 },
|
||||
};
|
||||
// eslint-disable-next-line camelcase
|
||||
const source2 = {
|
||||
req_id: 2,
|
||||
method: 'test',
|
||||
err: { code: 'error', msg: 'test' },
|
||||
};
|
||||
const result = {
|
||||
foo: 'bar',
|
||||
id: 1,
|
||||
name: 'demo1',
|
||||
value: { a: 1, b: 2, c: 3 },
|
||||
// eslint-disable-next-line camelcase
|
||||
req_id: 2,
|
||||
method: 'test',
|
||||
err: { code: 'error', msg: 'test' },
|
||||
};
|
||||
objectCopy(target, source1, source2);
|
||||
assert.deepStrictEqual(target, result,
|
||||
'target should have the same properties as source');
|
||||
done();
|
||||
});
|
||||
done => {
|
||||
const target = { foo: 'bar' };
|
||||
const source1 = { id: 1, name: 'demo1',
|
||||
value: { a: 1, b: 2, c: 3 } };
|
||||
// eslint-disable-next-line camelcase
|
||||
const source2 = { req_id: 2, method: 'test',
|
||||
err: { code: 'error', msg: 'test' } };
|
||||
const result = { foo: 'bar', id: 1, name: 'demo1',
|
||||
value: { a: 1, b: 2, c: 3 },
|
||||
// eslint-disable-next-line camelcase
|
||||
req_id: 2, method: 'test',
|
||||
err: { code: 'error', msg: 'test' } };
|
||||
objectCopy(target, source1, source2);
|
||||
assert.deepStrictEqual(target, result,
|
||||
'target should have the same properties as source');
|
||||
done();
|
||||
});
|
||||
});
|
||||
|
|
|
@ -1,17 +0,0 @@
|
|||
#!/usr/bin/env node
|
||||
// Convert string args into primitive value
|
||||
const fromStr = (str, primitive) => (str === `${primitive}` ? primitive : str);
|
||||
const date = fromStr(process.argv[2], undefined);
|
||||
const exitCode = fromStr(fromStr(process.argv[3], null), undefined);
|
||||
|
||||
const { stderrUtils } = require('../../../../index');
|
||||
|
||||
stderrUtils.catchAndTimestampStderr(
|
||||
date ? () => date : undefined,
|
||||
exitCode,
|
||||
);
|
||||
|
||||
process.emitWarning('TestWarningMessage');
|
||||
// This will print warning after printing error before exit
|
||||
throw new Error('TestingError');
|
||||
|
|
@ -1,23 +0,0 @@
|
|||
#!/usr/bin/env node
|
||||
// Convert string args into primitive value
|
||||
const fromStr = (str, primitive) => (str === `${primitive}` ? primitive : str);
|
||||
const date = fromStr(process.argv[2], undefined);
|
||||
const exitCode = fromStr(fromStr(process.argv[3], null), undefined);
|
||||
const promise = fromStr(process.argv[4], true);
|
||||
|
||||
const { stderrUtils } = require('../../../../index');
|
||||
|
||||
stderrUtils.catchAndTimestampUncaughtException(
|
||||
date ? () => date : undefined,
|
||||
exitCode,
|
||||
);
|
||||
|
||||
// Executed if process does not exit, process is in undefined behavior (bad)
|
||||
// eslint-disable-next-line no-console
|
||||
setTimeout(() => console.log('EXECUTED AFTER UNCAUGHT EXCEPTION'), 1);
|
||||
|
||||
if (promise === true) {
|
||||
Promise.reject();
|
||||
} else {
|
||||
throw new Error('TestingError');
|
||||
}
|
|
@ -1,38 +0,0 @@
|
|||
#!/usr/bin/env node
|
||||
// Convert string args into primitive value
|
||||
const fromStr = (str, primitive) => (str === `${primitive}` ? primitive : str);
|
||||
const date = fromStr(process.argv[2], undefined);
|
||||
const name = fromStr(process.argv[3], undefined);
|
||||
const code = fromStr(process.argv[4], undefined);
|
||||
const detail = fromStr(process.argv[5], undefined);
|
||||
|
||||
const { stderrUtils } = require('../../../../index');
|
||||
|
||||
stderrUtils.catchAndTimestampWarning(
|
||||
date ? () => date : undefined,
|
||||
);
|
||||
|
||||
const warning = new Error('TestWarningMessage');
|
||||
|
||||
if (name) warning.name = name;
|
||||
if (code) warning.code = code;
|
||||
if (detail) warning.detail = detail;
|
||||
|
||||
process.emitWarning(warning);
|
||||
|
||||
/*
|
||||
Examples:
|
||||
|
||||
(node:203831) Error: TestWarningMessage
|
||||
at Object.<anonymous> (catchWarning.js:15:17)
|
||||
...
|
||||
at node:internal/main/run_main_module:22:47
|
||||
Above Warning Date: 2024-06-26T16:32:55.505Z
|
||||
|
||||
(node:205151) [TEST01] CUSTOM: TestWarningMessage
|
||||
at Object.<anonymous> (catchWarning.js:15:17)
|
||||
...
|
||||
at node:internal/main/run_main_module:22:47
|
||||
Some additional detail
|
||||
Above Warning Date: Tue, 31 Dec 2024 10:20:30 GMT
|
||||
*/
|
|
@ -1,309 +0,0 @@
|
|||
const assert = require('assert');
|
||||
const { execFile } = require('child_process');
|
||||
|
||||
const stderrUtils = require('../../lib/stderrUtils');
|
||||
|
||||
/** Simple regex for ISO YYYY-MM-DDThh:mm:ss.sssZ */
|
||||
// eslint-disable-next-line max-len
|
||||
const defaultDateRegex = /\d{4}-[01]\d-[0-3]\dT[0-2]\d:[0-5]\d:[0-5]\d\.\d+(?:[+-][0-2]\d:[0-5]\d|Z)/;
|
||||
|
||||
// eslint-disable-next-line valid-jsdoc
|
||||
/** another format: Tue, 31 Dec 2024 10:20:30 GMT */
|
||||
const customDate = () => new Date('2024-12-31T10:20:30.444Z').toUTCString();
|
||||
|
||||
describe('stderrUtils', () => {
|
||||
const errStackRegex = /Error: TestingError\n(?:.*\sat\s.*\n)+/;
|
||||
|
||||
describe('defaultTimestamp', () => {
|
||||
it('should match ISO format', () => {
|
||||
assert.match(stderrUtils.defaultTimestamp(), defaultDateRegex);
|
||||
});
|
||||
});
|
||||
|
||||
describe('printErrorWithTimestamp', () => {
|
||||
let stderrText;
|
||||
const originalStderrWrite = process.stderr.write;
|
||||
const mockedStderrWrite = text => { stderrText = text; return true; };
|
||||
const err = new Error('TestingError');
|
||||
const origin = 'uncaughtException';
|
||||
|
||||
beforeEach(() => {
|
||||
stderrText = undefined;
|
||||
process.stderr.write = mockedStderrWrite;
|
||||
});
|
||||
|
||||
afterEach(() => {
|
||||
process.stderr.write = originalStderrWrite;
|
||||
stderrText = undefined;
|
||||
});
|
||||
|
||||
it(
|
||||
'should write to stderr with current date, origin and stacktrace',
|
||||
() => {
|
||||
const written = stderrUtils
|
||||
.printErrorWithTimestamp(err, origin);
|
||||
|
||||
assert.strictEqual(written, true);
|
||||
const [firstLine, errStack] = stderrText.split(':\n');
|
||||
const [errDate, errOrigin] = firstLine.split(': ');
|
||||
|
||||
assert.match(errDate, defaultDateRegex);
|
||||
assert.strictEqual(errOrigin, origin);
|
||||
assert.strictEqual(errStack, `${err.stack}\n`);
|
||||
},
|
||||
);
|
||||
|
||||
it(
|
||||
'should write to stderr with custom date, origin and stacktrace',
|
||||
() => {
|
||||
const written = stderrUtils
|
||||
.printErrorWithTimestamp(err, origin, customDate());
|
||||
|
||||
assert.strictEqual(written, true);
|
||||
const [firstLine, errStack] = stderrText.split(':\n');
|
||||
const [errDate, errOrigin] = firstLine.split(': ');
|
||||
|
||||
assert.strictEqual(errDate, customDate());
|
||||
assert.strictEqual(errOrigin, origin);
|
||||
assert.strictEqual(errStack, `${err.stack}\n`);
|
||||
},
|
||||
);
|
||||
});
|
||||
|
||||
const execOptions = {
|
||||
cwd: __dirname,
|
||||
// Subprocess should always stop alone
|
||||
// But just in case, kill subprocess after 500ms.
|
||||
// Leave enough time for `nyc` that runs slower.
|
||||
timeout: 500,
|
||||
};
|
||||
|
||||
// Execute in another process to notice the process exit
|
||||
// Therefore, looks more like a functional test
|
||||
const timeoutHint = (ms, retries) =>
|
||||
`Test fixture process timed out after ${ms}ms with ${retries} retries.\n` +
|
||||
'Due to nyc coverage first run slowing down process.\nIncrease execOptions.timeout to fix';
|
||||
|
||||
describe('catchAndTimestampUncaughtException', () => {
|
||||
[
|
||||
{ desc: 'with default date' },
|
||||
{ desc: 'with custom date', date: customDate() },
|
||||
{ desc: 'with custom exitCode 42', exitCode: 42 },
|
||||
{ desc: 'without exit on uncaught exception', exitCode: null },
|
||||
{ desc: 'for unhandled promise', promise: true },
|
||||
].forEach(({
|
||||
desc, date, exitCode, promise,
|
||||
}) => describe(desc, () => {
|
||||
/** for before all hook that doesn't support this.retries */
|
||||
let retries = 4;
|
||||
let err;
|
||||
let stdout;
|
||||
let stderr;
|
||||
let errStack;
|
||||
let errDate;
|
||||
let errOrigin;
|
||||
|
||||
before('run process catchUncaughtException', function beforeAllHook(done) {
|
||||
execFile(
|
||||
'./fixtures/stderrUtils/catchUncaughtException.js',
|
||||
[`${date}`, `${exitCode}`, `${promise}`],
|
||||
execOptions,
|
||||
(subErr, subStdout, subStderr) => {
|
||||
if (subErr?.killed) {
|
||||
retries--;
|
||||
if (retries <= 0) {
|
||||
assert.fail(timeoutHint(execOptions.timeout, retries));
|
||||
}
|
||||
execOptions.timeout *= 2;
|
||||
return beforeAllHook(done);
|
||||
}
|
||||
err = subErr;
|
||||
stdout = subStdout;
|
||||
stderr = subStderr;
|
||||
let firstLine;
|
||||
[firstLine, errStack] = stderr.split(':\n');
|
||||
[errDate, errOrigin] = firstLine.split(': ');
|
||||
done();
|
||||
},
|
||||
);
|
||||
});
|
||||
|
||||
if (exitCode === null) {
|
||||
it('should not be an error (or timeout)',
|
||||
() => assert.ifError(err));
|
||||
it('should have stdout (printed after uncaught exception)',
|
||||
() => assert.match(stdout,
|
||||
/^.*EXECUTED AFTER UNCAUGHT EXCEPTION(?:.|\n)*$/));
|
||||
} else {
|
||||
it('should be an error',
|
||||
() => assert.ok(err));
|
||||
it(`should have exitCode ${exitCode || 1}`,
|
||||
() => assert.strictEqual(err.code, exitCode || 1));
|
||||
it('should have empty stdout',
|
||||
() => assert.strictEqual(stdout, ''));
|
||||
}
|
||||
|
||||
it('should have stderr',
|
||||
() => assert.ok(stderr));
|
||||
it('should have date in stderr first line',
|
||||
() => (date
|
||||
? assert.strictEqual(errDate, date)
|
||||
: assert.match(errDate, defaultDateRegex)));
|
||||
|
||||
it('should have origin in stderr first line',
|
||||
() => (promise === true
|
||||
? assert.strictEqual(errOrigin, 'unhandledRejection')
|
||||
: assert.strictEqual(errOrigin, 'uncaughtException')));
|
||||
|
||||
if (!promise) {
|
||||
it('should have stack trace on stderr',
|
||||
() => assert.match(errStack, errStackRegex));
|
||||
}
|
||||
}));
|
||||
});
|
||||
|
||||
describe('catchAndTimestampWarning (also tests node onWarning)', () => {
|
||||
[
|
||||
{ desc: 'with default date' },
|
||||
{ desc: 'with custom date', date: customDate() },
|
||||
{ desc: 'with deprecation warning', name: 'DeprecationWarning' },
|
||||
{
|
||||
desc: 'with custom warning',
|
||||
name: 'CUSTOM',
|
||||
code: 'TEST01',
|
||||
detail: 'Some additional detail',
|
||||
},
|
||||
].forEach(({
|
||||
desc, date, name, code, detail,
|
||||
}) => describe(desc, () => {
|
||||
/** for before all hook that doesn't support this.retries */
|
||||
let retries = 4;
|
||||
let err;
|
||||
let stdout;
|
||||
let stderr;
|
||||
|
||||
before('run process catchWarning', function beforeAllHook(done) {
|
||||
execFile(
|
||||
'./fixtures/stderrUtils/catchWarning.js',
|
||||
[`${date}`, `${name}`, `${code}`, `${detail}`],
|
||||
execOptions,
|
||||
(subErr, subStdout, subStderr) => {
|
||||
if (subErr?.killed) {
|
||||
retries--;
|
||||
if (retries <= 0) {
|
||||
assert.fail(timeoutHint(execOptions.timeout, retries));
|
||||
}
|
||||
execOptions.timeout *= 2;
|
||||
return beforeAllHook(done);
|
||||
}
|
||||
err = subErr;
|
||||
stdout = subStdout;
|
||||
stderr = subStderr;
|
||||
done();
|
||||
},
|
||||
);
|
||||
});
|
||||
|
||||
it('should not be an error (or timeout)',
|
||||
() => assert.ifError(err));
|
||||
it('should have empty stdout',
|
||||
() => assert.strictEqual(stdout, ''));
|
||||
it('should have stderr',
|
||||
() => assert.ok(stderr));
|
||||
it('should have message on stderr first line, then stack trace',
|
||||
() => assert.match(stderr,
|
||||
/^.*TestWarningMessage\n(?:\s+at\s.*\n)+/));
|
||||
|
||||
if (code) {
|
||||
it('should have code on stderr first line',
|
||||
() => assert.match(stderr, new RegExp(`^.*[${code}]`)));
|
||||
}
|
||||
|
||||
if (name) {
|
||||
it('should have name on stderr first line',
|
||||
() => assert.match(stderr, new RegExp(`^.*${name}:`)));
|
||||
}
|
||||
|
||||
if (detail) {
|
||||
it('should have detail on stderr',
|
||||
() => assert.match(stderr, new RegExp(`.*${detail}.*`)));
|
||||
}
|
||||
|
||||
it(`should have ${date ? 'custom' : 'default'} date on stderr`,
|
||||
() => assert.match(stderr, new RegExp(
|
||||
`\nAbove Warning Date: ${
|
||||
date || defaultDateRegex.source}\n`,
|
||||
)));
|
||||
}));
|
||||
});
|
||||
|
||||
describe('catchAndTimestampStderr', () => {
|
||||
[
|
||||
{ desc: 'with default date' },
|
||||
{ desc: 'with custom date', date: customDate() },
|
||||
{ desc: 'with exit code', exitCode: 42 },
|
||||
|
||||
].forEach(({
|
||||
desc, date, exitCode,
|
||||
}) => describe(desc, () => {
|
||||
/** for before all hook that doesn't support this.retries */
|
||||
let retries = 4;
|
||||
let err;
|
||||
let stdout;
|
||||
let stderr;
|
||||
|
||||
before('run process catchStderr', function beforeAllHook(done) {
|
||||
execFile(
|
||||
'./fixtures/stderrUtils/catchStderr.js',
|
||||
[`${date}`, `${exitCode}`],
|
||||
execOptions,
|
||||
(subErr, subStdout, subStderr) => {
|
||||
if (subErr?.killed) {
|
||||
retries--;
|
||||
if (retries <= 0) {
|
||||
assert.fail(timeoutHint(execOptions.timeout, retries));
|
||||
}
|
||||
execOptions.timeout *= 2;
|
||||
return beforeAllHook(done);
|
||||
}
|
||||
err = subErr;
|
||||
stdout = subStdout;
|
||||
stderr = subStderr;
|
||||
done();
|
||||
},
|
||||
);
|
||||
});
|
||||
|
||||
it('should be an error',
|
||||
() => assert.ok(err));
|
||||
it(`should have exitCode ${exitCode || 1}`,
|
||||
() => assert.strictEqual(err.code, exitCode || 1));
|
||||
it('should have empty stdout',
|
||||
() => assert.strictEqual(stdout, ''));
|
||||
|
||||
it('should have stderr',
|
||||
() => assert.ok(stderr));
|
||||
|
||||
// 2024-06-26T15:04:55.364Z: uncaughtException:
|
||||
// Error: TestingError
|
||||
// at Object.<anonymous> (catchStderr.js:16:7)
|
||||
// at node:internal/main/run_main_module:22:47
|
||||
it('should have error date, origin and stacktrace in stderr',
|
||||
() => assert.match(stderr,
|
||||
new RegExp(`${date || defaultDateRegex.source
|
||||
}: uncaughtException:\n${errStackRegex.source}`)));
|
||||
|
||||
// (node:171245) Warning: TestWarningMessage
|
||||
// at Object.<anonymous> (catchStderr.js:14:9)
|
||||
// at node:internal/main/run_main_module:22:47
|
||||
// Above Warning Date: 2024-06-26T15:04:55.365Z
|
||||
it('should have warning with stacktrace in stderr', () => {
|
||||
const trace = 'Warning: TestWarningMessage\n(?:\\s+at\\s.*\n)+';
|
||||
const detail = `(?:.|\n)*?(?<=\n)Above Warning Date: ${
|
||||
date || defaultDateRegex.source}\n`;
|
||||
assert.match(stderr,
|
||||
new RegExp(`${trace}${detail}`));
|
||||
});
|
||||
}));
|
||||
});
|
||||
});
|
Loading…
Reference in New Issue