mirror of
https://github.com/Drop-OSS/droplet.git
synced 2025-11-12 15:52:47 +10:00
Compare commits
5 Commits
| Author | SHA1 | Date | |
|---|---|---|---|
| bb678b4b3a | |||
| cc94798962 | |||
| 7811818a72 | |||
| b6910e717b | |||
| 45a26c7156 |
2
.gitignore
vendored
2
.gitignore
vendored
@ -9,7 +9,7 @@ npm-debug.log*
|
||||
yarn-debug.log*
|
||||
yarn-error.log*
|
||||
lerna-debug.log*
|
||||
.test
|
||||
.test*
|
||||
.tsimp
|
||||
|
||||
# Diagnostic reports (https://nodejs.org/api/report.html)
|
||||
|
||||
@ -9,11 +9,12 @@ crate-type = ["cdylib"]
|
||||
|
||||
[dependencies]
|
||||
# Default enable napi4 feature, see https://nodejs.org/api/n-api.html#node-api-version-matrix
|
||||
napi = { version = "2.12.2", default-features = false, features = [
|
||||
napi = { version = "3.0.0-alpha.33", default-features = false, features = [
|
||||
"napi4",
|
||||
"async",
|
||||
"web_stream",
|
||||
] }
|
||||
napi-derive = "2.12.2"
|
||||
napi-derive = "3.0.0-alpha.33"
|
||||
hex = "0.4.3"
|
||||
serde_json = "1.0.128"
|
||||
md5 = "0.7.0"
|
||||
@ -21,6 +22,8 @@ time-macros = "0.2.22"
|
||||
time = "0.3.41"
|
||||
webpki = "0.22.4"
|
||||
ring = "0.17.14"
|
||||
tokio = { version = "1.45.1", features = ["fs"] }
|
||||
tokio-util = { version = "0.7.15", features = ["codec"] }
|
||||
|
||||
[dependencies.x509-parser]
|
||||
version = "0.17.0"
|
||||
|
||||
46
__test__/utils.spec.mjs
Normal file
46
__test__/utils.spec.mjs
Normal file
@ -0,0 +1,46 @@
|
||||
import test from "ava";
|
||||
import fs from "node:fs";
|
||||
import path from "path";
|
||||
|
||||
import droplet from "../index.js";
|
||||
|
||||
test("check alt thread util", async (t) => {
|
||||
let endtime1, endtime2;
|
||||
|
||||
droplet.callAltThreadFunc(async () => {
|
||||
await new Promise((r) => setTimeout(r, 100));
|
||||
endtime1 = Date.now();
|
||||
});
|
||||
|
||||
await new Promise((r) => setTimeout(r, 500));
|
||||
endtime2 = Date.now();
|
||||
|
||||
const difference = endtime2 - endtime1;
|
||||
if (difference >= 600) {
|
||||
t.fail("likely isn't multithreaded, difference: " + difference);
|
||||
}
|
||||
|
||||
t.pass();
|
||||
});
|
||||
|
||||
test("read file", async (t) => {
|
||||
const dirName = "./.test2";
|
||||
if (fs.existsSync(dirName)) fs.rmSync(dirName, { recursive: true });
|
||||
fs.mkdirSync(dirName, { recursive: true });
|
||||
|
||||
const testString = "g'day what's up my koala bros\n".repeat(10000);
|
||||
|
||||
fs.writeFileSync("./.test2/TESTFILE", testString);
|
||||
|
||||
const stream = droplet.readFile("./.test2", "TESTFILE");
|
||||
|
||||
let finalString = "";
|
||||
|
||||
for await (const chunk of stream) {
|
||||
// Do something with each 'chunk'
|
||||
finalString += String.fromCharCode.apply(null, chunk);
|
||||
}
|
||||
|
||||
t.assert(finalString == testString, "file strings don't match");
|
||||
fs.rmSync(dirName, { recursive: true });
|
||||
});
|
||||
20
index.d.ts
vendored
20
index.d.ts
vendored
@ -3,12 +3,14 @@
|
||||
|
||||
/* auto-generated by NAPI-RS */
|
||||
|
||||
export declare function hasBackendForPath(path: string): boolean
|
||||
export declare function listFiles(path: string): Array<string>
|
||||
export declare function callAltThreadFunc(callback: (...args: any[]) => any): void
|
||||
export declare function generateManifest(dir: string, progress: (...args: any[]) => any, log: (...args: any[]) => any, callback: (...args: any[]) => any): void
|
||||
export declare function generateRootCa(): Array<string>
|
||||
export declare function generateClientCertificate(clientId: string, clientName: string, rootCa: string, rootCaPrivate: string): Array<string>
|
||||
export declare function verifyClientCertificate(clientCert: string, rootCa: string): boolean
|
||||
export declare function signNonce(privateKey: string, nonce: string): string
|
||||
export declare function verifyNonce(publicCert: string, nonce: string, signature: string): boolean
|
||||
function hasBackendForPath(path: string): boolean
|
||||
function listFiles(path: string): Array<string>
|
||||
function readFile(path: string, subPath: string): ReadableStream<Buffer> | null
|
||||
function callAltThreadFunc(tsfn: ((err: Error | null, ) => any)): void
|
||||
function generateManifest(dir: string, progressSfn: ((err: Error | null, arg: number) => any), logSfn: ((err: Error | null, arg: string) => any), callbackSfn: ((err: Error | null, arg: string) => any)): void
|
||||
function generateRootCa(): Array<string>
|
||||
function generateClientCertificate(clientId: string, clientName: string, rootCa: string, rootCaPrivate: string): Array<string>
|
||||
function verifyClientCertificate(clientCert: string, rootCa: string): boolean
|
||||
function signNonce(privateKey: string, nonce: string): string
|
||||
function verifyNonce(publicCert: string, nonce: string, signature: string): boolean
|
||||
undefinedundefined
|
||||
|
||||
4
index.js
4
index.js
@ -310,10 +310,11 @@ if (!nativeBinding) {
|
||||
throw new Error(`Failed to load native binding`)
|
||||
}
|
||||
|
||||
const { hasBackendForPath, listFiles, callAltThreadFunc, generateManifest, generateRootCa, generateClientCertificate, verifyClientCertificate, signNonce, verifyNonce } = nativeBinding
|
||||
const { hasBackendForPath, listFiles, readFile, callAltThreadFunc, generateManifest, generateRootCa, generateClientCertificate, verifyClientCertificate, signNonce, verifyNonce, } = nativeBinding
|
||||
|
||||
module.exports.hasBackendForPath = hasBackendForPath
|
||||
module.exports.listFiles = listFiles
|
||||
module.exports.readFile = readFile
|
||||
module.exports.callAltThreadFunc = callAltThreadFunc
|
||||
module.exports.generateManifest = generateManifest
|
||||
module.exports.generateRootCa = generateRootCa
|
||||
@ -321,3 +322,4 @@ module.exports.generateClientCertificate = generateClientCertificate
|
||||
module.exports.verifyClientCertificate = verifyClientCertificate
|
||||
module.exports.signNonce = signNonce
|
||||
module.exports.verifyNonce = verifyNonce
|
||||
module.exports.undefined = undefined
|
||||
|
||||
@ -1,6 +1,6 @@
|
||||
{
|
||||
"name": "@drop-oss/droplet",
|
||||
"version": "1.1.2",
|
||||
"version": "1.2.0",
|
||||
"main": "index.js",
|
||||
"types": "index.d.ts",
|
||||
"napi": {
|
||||
@ -21,7 +21,7 @@
|
||||
},
|
||||
"license": "MIT",
|
||||
"devDependencies": {
|
||||
"@napi-rs/cli": "^2.18.4",
|
||||
"@napi-rs/cli": "2.18.4",
|
||||
"@types/node": "^22.13.10",
|
||||
"ava": "^6.2.0"
|
||||
},
|
||||
|
||||
@ -2,12 +2,19 @@
|
||||
use std::os::unix::fs::PermissionsExt;
|
||||
use std::{
|
||||
fs::{self, metadata, File},
|
||||
io::BufReader,
|
||||
io::{self, BufReader, ErrorKind, Read},
|
||||
path::{Path, PathBuf},
|
||||
task::Poll,
|
||||
};
|
||||
|
||||
const CHUNK_SIZE: usize = 1024 * 1024 * 64;
|
||||
|
||||
use napi::{
|
||||
bindgen_prelude::*,
|
||||
tokio_stream::{Stream, StreamExt},
|
||||
};
|
||||
use tokio_util::{
|
||||
bytes::BytesMut,
|
||||
codec::{BytesCodec, FramedRead},
|
||||
};
|
||||
|
||||
fn _list_files(vec: &mut Vec<PathBuf>, path: &Path) {
|
||||
if metadata(path).unwrap().is_dir() {
|
||||
@ -30,7 +37,7 @@ pub struct VersionFile {
|
||||
|
||||
pub trait VersionBackend: 'static {
|
||||
fn list_files(&self, path: &Path) -> Vec<VersionFile>;
|
||||
fn reader(&self, file: &VersionFile) -> BufReader<File>;
|
||||
fn reader(&self, file: &VersionFile) -> Option<File>;
|
||||
}
|
||||
|
||||
pub struct PathVersionBackend {
|
||||
@ -70,10 +77,10 @@ impl VersionBackend for PathVersionBackend {
|
||||
results
|
||||
}
|
||||
|
||||
fn reader(&self, file: &VersionFile) -> BufReader<File> {
|
||||
let file = File::open(self.base_dir.join(file.relative_filename.clone())).unwrap();
|
||||
let reader = BufReader::with_capacity(CHUNK_SIZE, file);
|
||||
return reader;
|
||||
fn reader(&self, file: &VersionFile) -> Option<File> {
|
||||
let file = File::open(self.base_dir.join(file.relative_filename.clone())).ok()?;
|
||||
|
||||
return Some(file);
|
||||
}
|
||||
}
|
||||
|
||||
@ -85,7 +92,7 @@ impl VersionBackend for ArchiveVersionBackend {
|
||||
todo!()
|
||||
}
|
||||
|
||||
fn reader(&self, file: &VersionFile) -> BufReader<File> {
|
||||
fn reader(&self, file: &VersionFile) -> Option<File> {
|
||||
todo!()
|
||||
}
|
||||
}
|
||||
@ -120,4 +127,39 @@ pub fn list_files(path: String) -> Vec<String> {
|
||||
let backend = create_backend_for_path(path).unwrap();
|
||||
let files = backend.list_files(path);
|
||||
files.into_iter().map(|e| e.relative_filename).collect()
|
||||
}
|
||||
}
|
||||
|
||||
#[napi]
|
||||
pub fn read_file(
|
||||
path: String,
|
||||
sub_path: String,
|
||||
env: &Env,
|
||||
) -> Option<ReadableStream<'static, BufferSlice<'static>>> {
|
||||
let path = Path::new(&path);
|
||||
let backend = create_backend_for_path(path).unwrap();
|
||||
let version_file = VersionFile {
|
||||
relative_filename: sub_path,
|
||||
permission: 0, // Shouldn't matter
|
||||
};
|
||||
// Use `?` operator for cleaner error propagation from `Option`
|
||||
let reader = backend.reader(&version_file)?;
|
||||
|
||||
// Convert std::fs::File to tokio::fs::File for async operations
|
||||
let reader = tokio::fs::File::from_std(reader);
|
||||
|
||||
// Create a FramedRead stream with BytesCodec for chunking
|
||||
|
||||
let stream = FramedRead::new(reader, BytesCodec::new())
|
||||
// Use StreamExt::map to transform each Result item
|
||||
.map(|result_item| {
|
||||
result_item
|
||||
// Apply Result::map to transform Ok(BytesMut) to Ok(Vec<u8>)
|
||||
.map(|bytes| bytes.to_vec())
|
||||
// Apply Result::map_err to transform Err(std::io::Error) to Err(napi::Error)
|
||||
.map_err(|e| napi::Error::from(e)) // napi::Error implements From<tokio::io::Error>
|
||||
});
|
||||
// Create the napi-rs ReadableStream from the tokio_stream::Stream
|
||||
// The unwrap() here means if stream creation fails, it will panic.
|
||||
// For a production system, consider returning Result<Option<...>> and handling this.
|
||||
Some(ReadableStream::create_with_stream_bytes(env, stream).unwrap())
|
||||
}
|
||||
|
||||
@ -1,23 +1,22 @@
|
||||
use std::{
|
||||
collections::HashMap,
|
||||
fs::File,
|
||||
io::{BufRead, BufReader},
|
||||
path::Path,
|
||||
thread,
|
||||
collections::HashMap, fs::File, io::{BufRead, BufReader}, path::Path, rc::Rc, sync::Arc, thread
|
||||
};
|
||||
|
||||
#[cfg(unix)]
|
||||
use std::os::unix::fs::PermissionsExt;
|
||||
|
||||
use napi::{
|
||||
threadsafe_function::{ErrorStrategy, ThreadsafeFunction, ThreadsafeFunctionCallMode},
|
||||
Error, JsFunction,
|
||||
bindgen_prelude::Function,
|
||||
threadsafe_function::{ThreadsafeFunction, ThreadsafeFunctionCallMode},
|
||||
Env, Error, Result,
|
||||
};
|
||||
use serde_json::json;
|
||||
use uuid::Uuid;
|
||||
|
||||
use crate::file_utils::create_backend_for_path;
|
||||
|
||||
const CHUNK_SIZE: usize = 1024 * 1024 * 64;
|
||||
|
||||
#[derive(serde::Serialize)]
|
||||
struct ChunkData {
|
||||
permissions: u32,
|
||||
@ -27,14 +26,10 @@ struct ChunkData {
|
||||
}
|
||||
|
||||
#[napi]
|
||||
pub fn call_alt_thread_func(callback: JsFunction) -> Result<(), Error> {
|
||||
let tsfn: ThreadsafeFunction<u32, ErrorStrategy::CalleeHandled> = callback
|
||||
.create_threadsafe_function(0, |ctx| {
|
||||
ctx.env.create_uint32(ctx.value + 1).map(|v| vec![v])
|
||||
})?;
|
||||
let tsfn = tsfn.clone();
|
||||
pub fn call_alt_thread_func(tsfn: Arc<ThreadsafeFunction<()>>) -> Result<(), String> {
|
||||
let tsfn_cloned = tsfn.clone();
|
||||
thread::spawn(move || {
|
||||
tsfn.call(Ok(0), ThreadsafeFunctionCallMode::NonBlocking);
|
||||
tsfn_cloned.call(Ok(()), ThreadsafeFunctionCallMode::Blocking);
|
||||
});
|
||||
Ok(())
|
||||
}
|
||||
@ -42,24 +37,10 @@ pub fn call_alt_thread_func(callback: JsFunction) -> Result<(), Error> {
|
||||
#[napi]
|
||||
pub fn generate_manifest(
|
||||
dir: String,
|
||||
progress: JsFunction,
|
||||
log: JsFunction,
|
||||
callback: JsFunction,
|
||||
) -> Result<(), Error> {
|
||||
let progress_sfn: ThreadsafeFunction<i32, ErrorStrategy::CalleeHandled> = progress
|
||||
.create_threadsafe_function(0, |ctx| ctx.env.create_int32(ctx.value).map(|v| vec![v]))
|
||||
.unwrap();
|
||||
let log_sfn: ThreadsafeFunction<String, ErrorStrategy::CalleeHandled> = log
|
||||
.create_threadsafe_function(0, |ctx| {
|
||||
ctx.env.create_string_from_std(ctx.value).map(|v| vec![v])
|
||||
})
|
||||
.unwrap();
|
||||
let callback_sfn: ThreadsafeFunction<String, ErrorStrategy::CalleeHandled> = callback
|
||||
.create_threadsafe_function(0, |ctx| {
|
||||
ctx.env.create_string_from_std(ctx.value).map(|v| vec![v])
|
||||
})
|
||||
.unwrap();
|
||||
|
||||
progress_sfn: ThreadsafeFunction<i32>,
|
||||
log_sfn: ThreadsafeFunction<String>,
|
||||
callback_sfn: ThreadsafeFunction<String>,
|
||||
) -> Result<(), String> {
|
||||
thread::spawn(move || {
|
||||
let base_dir = Path::new(&dir);
|
||||
let backend = create_backend_for_path(base_dir).unwrap();
|
||||
@ -72,7 +53,8 @@ pub fn generate_manifest(
|
||||
let mut i: i32 = 0;
|
||||
|
||||
for version_file in files {
|
||||
let mut reader = backend.reader(&version_file);
|
||||
let mut raw_reader= backend.reader(&version_file).unwrap();
|
||||
let mut reader = BufReader::with_capacity(CHUNK_SIZE, raw_reader);
|
||||
|
||||
let mut chunk_data = ChunkData {
|
||||
permissions: version_file.permission,
|
||||
@ -101,8 +83,7 @@ pub fn generate_manifest(
|
||||
|
||||
let log_str = format!(
|
||||
"Processed chunk {} for {}",
|
||||
chunk_index,
|
||||
&version_file.relative_filename
|
||||
chunk_index, &version_file.relative_filename
|
||||
);
|
||||
log_sfn.call(Ok(log_str), ThreadsafeFunctionCallMode::Blocking);
|
||||
|
||||
|
||||
@ -9,7 +9,7 @@ __metadata:
|
||||
version: 0.0.0-use.local
|
||||
resolution: "@drop-oss/droplet@workspace:."
|
||||
dependencies:
|
||||
"@napi-rs/cli": "npm:^2.18.4"
|
||||
"@napi-rs/cli": "npm:2.18.4"
|
||||
"@types/node": "npm:^22.13.10"
|
||||
ava: "npm:^6.2.0"
|
||||
languageName: unknown
|
||||
@ -55,7 +55,7 @@ __metadata:
|
||||
languageName: node
|
||||
linkType: hard
|
||||
|
||||
"@napi-rs/cli@npm:^2.18.4":
|
||||
"@napi-rs/cli@npm:2.18.4":
|
||||
version: 2.18.4
|
||||
resolution: "@napi-rs/cli@npm:2.18.4"
|
||||
bin:
|
||||
|
||||
Reference in New Issue
Block a user