1 |
|
2 |
|
3 |
|
4 |
|
5 |
|
6 |
|
7 |
|
8 |
|
9 |
|
10 |
|
11 |
|
12 |
|
13 |
|
14 | 'use strict';
|
15 |
|
16 | const async = require('async');
|
17 | const stream = require('stream');
|
18 | const error = require('./error.js');
|
19 | const debug = require('debug')('couchbackup:writer');
|
20 |
|
21 | module.exports = function(db, bufferSize, parallelism, ee) {
|
22 | const writer = new stream.Transform({ objectMode: true });
|
23 | let buffer = [];
|
24 | let written = 0;
|
25 | let linenumber = 0;
|
26 |
|
27 |
|
28 |
|
29 |
|
30 |
|
31 | const q = async.queue(function(payload, cb) {
|
32 |
|
33 | if (payload.docs && payload.docs[0] && payload.docs[0]._rev) {
|
34 | payload.new_edits = false;
|
35 | debug('Using new_edits false mode.');
|
36 | }
|
37 |
|
38 | if (!didError) {
|
39 | db.service.postBulkDocs({
|
40 | db: db.db,
|
41 | bulkDocs: payload
|
42 | }).then(response => {
|
43 | if (!response.result || (payload.new_edits === false && response.result.length > 0)) {
|
44 | throw new Error(`Error writing batch with new_edits:${payload.new_edits !== false}` +
|
45 | ` and ${response.result ? response.result.length : 'unavailable'} items`);
|
46 | }
|
47 | written += payload.docs.length;
|
48 | writer.emit('restored', { documents: payload.docs.length, total: written });
|
49 | cb();
|
50 | }).catch(err => {
|
51 | err = error.convertResponseError(err);
|
52 | debug(`Error writing docs ${err.name} ${err.message}`);
|
53 | cb(err, payload);
|
54 | });
|
55 | }
|
56 | }, parallelism);
|
57 |
|
58 | let didError = false;
|
59 |
|
60 |
|
61 | function processBuffer(flush, callback) {
|
62 | function taskCallback(err, payload) {
|
63 | if (err && !didError) {
|
64 | debug(`Queue task failed with error ${err.name}`);
|
65 | didError = true;
|
66 | q.kill();
|
67 | writer.emit('error', err);
|
68 | }
|
69 | }
|
70 |
|
71 | if (flush || buffer.length >= bufferSize) {
|
72 |
|
73 |
|
74 | do {
|
75 |
|
76 | const toSend = buffer.splice(0, bufferSize);
|
77 |
|
78 |
|
79 | debug(`Adding ${toSend.length} to the write queue.`);
|
80 | q.push({ docs: toSend }, taskCallback);
|
81 | } while (buffer.length >= bufferSize);
|
82 |
|
83 |
|
84 | if (flush && buffer.length > 0) {
|
85 | debug(`Adding remaining ${buffer.length} to the write queue.`);
|
86 | q.push({ docs: buffer }, taskCallback);
|
87 | }
|
88 |
|
89 |
|
90 | async.until(
|
91 |
|
92 |
|
93 | function(callback) {
|
94 |
|
95 | if (didError) {
|
96 | return callback(null, true);
|
97 | }
|
98 | if (flush) {
|
99 | callback(null, q.idle() && q.length() === 0);
|
100 | } else {
|
101 | callback(null, q.length() <= parallelism * 2);
|
102 | }
|
103 | },
|
104 | function(cb) {
|
105 | setTimeout(cb, 20);
|
106 | },
|
107 |
|
108 | function() {
|
109 | if (flush && !didError) {
|
110 | writer.emit('finished', { total: written });
|
111 | }
|
112 |
|
113 | callback();
|
114 | });
|
115 | } else {
|
116 | callback();
|
117 | }
|
118 | }
|
119 |
|
120 |
|
121 | writer._transform = function(obj, encoding, done) {
|
122 |
|
123 |
|
124 |
|
125 | linenumber++;
|
126 | if (!didError && obj !== '') {
|
127 |
|
128 | try {
|
129 | const arr = JSON.parse(obj);
|
130 |
|
131 |
|
132 | if (typeof arr === 'object' && arr.length > 0) {
|
133 |
|
134 | buffer = buffer.concat(arr);
|
135 |
|
136 |
|
137 |
|
138 |
|
139 |
|
140 |
|
141 | this.pause();
|
142 |
|
143 |
|
144 | processBuffer(false, done);
|
145 | } else {
|
146 | ee.emit('error', new error.BackupError('BackupFileJsonError', `Error on line ${linenumber} of backup file - not an array`));
|
147 | done();
|
148 | }
|
149 | } catch (e) {
|
150 | ee.emit('error', new error.BackupError('BackupFileJsonError', `Error on line ${linenumber} of backup file - cannot parse as JSON`));
|
151 |
|
152 | done();
|
153 | }
|
154 | } else {
|
155 | done();
|
156 | }
|
157 | };
|
158 |
|
159 |
|
160 | writer._flush = function(done) {
|
161 | processBuffer(true, done);
|
162 | };
|
163 | return writer;
|
164 | };
|