From 9a1d36a76ab8a70d3283941e0f1084d1369cf556 Mon Sep 17 00:00:00 2001 From: Connor Manning Date: Tue, 26 Mar 2024 10:03:09 -0500 Subject: [PATCH] v0.0.6 --- package.json | 4 ++-- src/copc/hierarchy.test.ts | 4 ++-- src/las/point-data.test.ts | 36 +++++++++++++++--------------------- 3 files changed, 19 insertions(+), 25 deletions(-) diff --git a/package.json b/package.json index 95351b2..cc40247 100644 --- a/package.json +++ b/package.json @@ -1,7 +1,7 @@ { "name": "copc", "license": "MIT", - "version": "0.0.5", + "version": "0.0.6", "main": "lib/index.js", "scripts": { "build": "tsc --project tsconfig.production.json && tsc-alias", @@ -31,6 +31,6 @@ } }, "browser": { - "fs": false + "fs": false } } diff --git a/src/copc/hierarchy.test.ts b/src/copc/hierarchy.test.ts index 1cb675f..bc83af6 100644 --- a/src/copc/hierarchy.test.ts +++ b/src/copc/hierarchy.test.ts @@ -91,7 +91,7 @@ test('parse invalid', () => { const b = pack({ key, pointCount: 1, pointDataOffset, pointDataLength }) expect(() => Hierarchy.parse(Buffer.concat([b, Buffer.alloc(1)]))).toThrow( - /length/i + /length/i, ) - expect(() => Hierarchy.parse(b.slice(0, -1))).toThrow(/length/i) + expect(() => Hierarchy.parse(b.subarray(0, -1))).toThrow(/length/i) }) diff --git a/src/las/point-data.test.ts b/src/las/point-data.test.ts index ac0fa64..847f2b9 100644 --- a/src/las/point-data.test.ts +++ b/src/las/point-data.test.ts @@ -12,25 +12,23 @@ const chunkTableAdjustment = 8 test('decompress chunk', async () => { const buffer = await fs.readFile(join(Test.dirname, 'data/ellipsoid-1.4.laz')) const header = Las.Header.parse(buffer) - const compressed = buffer.slice(header.pointDataOffset + chunkTableAdjustment) + const compressed = buffer.subarray( + header.pointDataOffset + chunkTableAdjustment, + ) const pointdata = await decompressChunk(compressed, header) - const getter: Getter = async (begin, end) => buffer.slice(begin, end) + const getter: Getter = async (begin, end) => buffer.subarray(begin, end) const vlrs = await Las.Vlr.walk(getter, header) const ebVlr = Las.Vlr.find(vlrs, 'LASF_Spec', 4) const eb = ebVlr && Las.ExtraBytes.parse(await Las.Vlr.fetch(getter, ebVlr)) const view = Las.View.create(pointdata, header, eb) - const getters = [ - 'X', - 'Y', - 'Z', - 'Intensity', - 'InvertedIntensity', - ].map(view.getter) + const getters = ['X', 'Y', 'Z', 'Intensity', 'InvertedIntensity'].map( + view.getter, + ) { - const [x, y, z, i, ii] = getters.map(f => f(0)) + const [x, y, z, i, ii] = getters.map((f) => f(0)) expect(x >= header.min[0]) expect(y >= header.min[1]) expect(z >= header.min[2]) @@ -40,7 +38,7 @@ test('decompress chunk', async () => { expect(i).not.toEqual(ii) } { - const [x, y, z, i, ii] = getters.map(f => f(1)) + const [x, y, z, i, ii] = getters.map((f) => f(1)) expect(x >= header.min[0]) expect(y >= header.min[1]) expect(z >= header.min[2]) @@ -56,22 +54,18 @@ test('decompress file', async () => { const header = Las.Header.parse(buffer) const pointdata = await decompressFile(buffer) - const getter: Getter = async (begin, end) => buffer.slice(begin, end) + const getter: Getter = async (begin, end) => buffer.subarray(begin, end) const vlrs = await Las.Vlr.walk(getter, header) const ebVlr = Las.Vlr.find(vlrs, 'LASF_Spec', 4) const eb = ebVlr && Las.ExtraBytes.parse(await Las.Vlr.fetch(getter, ebVlr)) const view = Las.View.create(pointdata, header, eb) - const getters = [ - 'X', - 'Y', - 'Z', - 'Intensity', - 'InvertedIntensity', - ].map(view.getter) + const getters = ['X', 'Y', 'Z', 'Intensity', 'InvertedIntensity'].map( + view.getter, + ) { - const [x, y, z, i, ii] = getters.map(f => f(0)) + const [x, y, z, i, ii] = getters.map((f) => f(0)) expect(x >= header.min[0]) expect(y >= header.min[1]) expect(z >= header.min[2]) @@ -81,7 +75,7 @@ test('decompress file', async () => { expect(i).not.toEqual(ii) } { - const [x, y, z, i, ii] = getters.map(f => f(1)) + const [x, y, z, i, ii] = getters.map((f) => f(1)) expect(x >= header.min[0]) expect(y >= header.min[1]) expect(z >= header.min[2])