This commit is contained in:
Sai1919
2016-11-10 23:24:55 +05:30
parent 4649d3749c
commit 66104578bf

View File

@@ -1034,7 +1034,6 @@ describe('Tests', function () {
describe('performance testing', function () { describe('performance testing', function () {
it('should properly parse more than 500 MB of file.', function (done) { it('should properly parse more than 500 MB of file.', function (done) {
var firstChunk = fs.readFileSync('./test/TestFiles/MB_and_GB_size_files/firstChunk.xml')
var parser = new ParserFactory({resourcePath: '/items/item'}) var parser = new ParserFactory({resourcePath: '/items/item'})
// var wsStream = fs.createWriteStream('./test/TestFiles/MB_and_GB_size_files/MBFile.xml') // var wsStream = fs.createWriteStream('./test/TestFiles/MB_and_GB_size_files/MBFile.xml')
// var rsStream = fs.createReadStream('./test/TestFiles/MB_and_GB_size_files/MBFile.xml') // var rsStream = fs.createReadStream('./test/TestFiles/MB_and_GB_size_files/MBFile.xml')
@@ -1046,6 +1045,7 @@ describe('Tests', function () {
xmlStream._read = function noop () {} xmlStream._read = function noop () {}
var dataChunk var dataChunk
this.timeout(900000) this.timeout(900000)
var firstChunk = fs.readFileSync('./test/TestFiles/MB_and_GB_size_files/firstChunk.xml')
xmlStream.push(firstChunk) xmlStream.push(firstChunk)
for (var i = 0; i < 2200; i++) { for (var i = 0; i < 2200; i++) {
dataChunk = fs.readFileSync('./test/TestFiles/MB_and_GB_size_files/repetitiveChunk.xml') dataChunk = fs.readFileSync('./test/TestFiles/MB_and_GB_size_files/repetitiveChunk.xml')
@@ -1079,7 +1079,6 @@ describe('Tests', function () {
}) })
it('should properly parse more than 1 GB of file.', function (done) { it('should properly parse more than 1 GB of file.', function (done) {
var firstChunk = fs.readFileSync('./test/TestFiles/MB_and_GB_size_files/firstChunk.xml')
var parser = new ParserFactory({resourcePath: '/items/item'}) var parser = new ParserFactory({resourcePath: '/items/item'})
// var wsStream = fs.createWriteStream('./test/TestFiles/MB_and_GB_size_files/MBFile.xml') // var wsStream = fs.createWriteStream('./test/TestFiles/MB_and_GB_size_files/MBFile.xml')
// var rsStream = fs.createReadStream('./test/TestFiles/MB_and_GB_size_files/MBFile.xml') // var rsStream = fs.createReadStream('./test/TestFiles/MB_and_GB_size_files/MBFile.xml')
@@ -1091,6 +1090,7 @@ describe('Tests', function () {
xmlStream._read = function noop () {} xmlStream._read = function noop () {}
var dataChunk var dataChunk
this.timeout(900000) this.timeout(900000)
var firstChunk = fs.readFileSync('./test/TestFiles/MB_and_GB_size_files/firstChunk.xml')
xmlStream.push(firstChunk) xmlStream.push(firstChunk)
for (var i = 0; i < 4400; i++) { for (var i = 0; i < 4400; i++) {
dataChunk = fs.readFileSync('./test/TestFiles/MB_and_GB_size_files/repetitiveChunk.xml') dataChunk = fs.readFileSync('./test/TestFiles/MB_and_GB_size_files/repetitiveChunk.xml')