@@ -63,7 +63,7 @@ const exists = async (p) => {
63
63
}
64
64
} ) ( )
65
65
. then ( ( x ) => + x || 0 , ( e ) => {
66
- e && process . stderr . write ( `${ e && e . stack || e } \n` ) ;
66
+ e && process . stderr . write ( `${ e ? .stack || e } \n` ) ;
67
67
return process . exitCode || 1 ;
68
68
} ) . then ( ( code ) => process . exit ( code ) ) ;
69
69
@@ -141,7 +141,7 @@ function validateReaderIntegration(jsonData, arrowBuffer) {
141
141
for ( const [ jsonRecordBatch , binaryRecordBatch ] of zip ( jsonReader , binaryReader ) ) {
142
142
compareTableIsh ( jsonRecordBatch , binaryRecordBatch ) ;
143
143
}
144
- } catch ( e ) { throw new Error ( `${ msg } : fail \n ${ e && e . stack || e } ` ) ; }
144
+ } catch ( e ) { throw new Error ( `${ msg } : fail \n ${ e ? .stack || e } ` ) ; }
145
145
process . stdout . write ( `${ msg } : pass\n` ) ;
146
146
}
147
147
@@ -151,7 +151,7 @@ function validateTableFromBuffersIntegration(jsonData, arrowBuffer) {
151
151
const jsonTable = Table . from ( jsonData ) ;
152
152
const binaryTable = Table . from ( arrowBuffer ) ;
153
153
compareTableIsh ( jsonTable , binaryTable ) ;
154
- } catch ( e ) { throw new Error ( `${ msg } : fail \n ${ e && e . stack || e } ` ) ; }
154
+ } catch ( e ) { throw new Error ( `${ msg } : fail \n ${ e ? .stack || e } ` ) ; }
155
155
process . stdout . write ( `${ msg } : pass\n` ) ;
156
156
}
157
157
@@ -164,7 +164,7 @@ function validateTableToBuffersIntegration(srcFormat, arrowFormat) {
164
164
const srcTable = Table . from ( srcFormat === `json` ? jsonData : arrowBuffer ) ;
165
165
const dstTable = Table . from ( srcTable . serialize ( `binary` , arrowFormat === `stream` ) ) ;
166
166
compareTableIsh ( dstTable , refTable ) ;
167
- } catch ( e ) { throw new Error ( `${ msg } : fail \n ${ e && e . stack || e } ` ) ; }
167
+ } catch ( e ) { throw new Error ( `${ msg } : fail \n ${ e ? .stack || e } ` ) ; }
168
168
process . stdout . write ( `${ msg } : pass\n` ) ;
169
169
} ;
170
170
}
0 commit comments