summaryrefslogtreecommitdiff
diff options
context:
space:
mode:
authorLuca Casonato <hello@lcas.dev>2024-06-24 11:47:12 +0200
committerGitHub <noreply@github.com>2024-06-24 11:47:12 +0200
commit1e8a6b94b1dcf98a2ae4de97b3e98e7b3e4e8f7f (patch)
tree89ae2bc343dea6bf17ca9d512ea80b51540347ca
parentff535061077d2b67e20154a7dfefe8ca92502c5a (diff)
fix(ext/node): rewrite crypto.Hash (#24302)
Changes in this PR: - Added new fixed size hash algorithms (blake2b512, blake2s256, sha512-224, sha512-256, sha3-224, sha3-256, sha3-384, sha3-512, sm3) - Added variable size hash algorithms (the concept), with the algorithms shake128 and shake256 - Use cppgc instead of resources for the hasher - Enable Node's crypto.Hash tests and fix found bugs
-rw-r--r--Cargo.lock40
-rw-r--r--ext/node/Cargo.toml3
-rw-r--r--ext/node/ops/crypto/digest.rs290
-rw-r--r--ext/node/ops/crypto/mod.rs73
-rw-r--r--ext/node/polyfills/internal/crypto/hash.ts232
-rw-r--r--tests/integration/node_unit_tests.rs1
-rw-r--r--tests/node_compat/config.jsonc8
-rw-r--r--tests/node_compat/runner/TODO.md1
-rw-r--r--tests/node_compat/test.ts16
-rw-r--r--tests/node_compat/test/fixtures/sample.pngbin0 -> 9416 bytes
-rw-r--r--tests/node_compat/test/parallel/test-crypto-hash.js285
-rw-r--r--tests/unit_node/crypto/crypto_hash_test.ts26
-rw-r--r--tests/unit_node/crypto/crypto_misc_test.ts18
13 files changed, 760 insertions, 233 deletions
diff --git a/Cargo.lock b/Cargo.lock
index d35c101cc..c65d3314d 100644
--- a/Cargo.lock
+++ b/Cargo.lock
@@ -481,6 +481,15 @@ dependencies = [
]
[[package]]
+name = "blake2"
+version = "0.10.6"
+source = "registry+https://github.com/rust-lang/crates.io-index"
+checksum = "46502ad458c9a52b69d4d4d32775c788b7a1b85e8bc9d482d92250fc0e3f8efe"
+dependencies = [
+ "digest",
+]
+
+[[package]]
name = "block"
version = "0.1.6"
source = "registry+https://github.com/rust-lang/crates.io-index"
@@ -1653,6 +1662,7 @@ dependencies = [
"aead-gcm-stream",
"aes",
"async-trait",
+ "blake2",
"brotli",
"bytes",
"cbc",
@@ -1706,8 +1716,10 @@ dependencies = [
"serde",
"sha1",
"sha2",
+ "sha3",
"signature",
"simd-json",
+ "sm3",
"spki",
"tokio",
"url",
@@ -3667,6 +3679,15 @@ dependencies = [
]
[[package]]
+name = "keccak"
+version = "0.1.5"
+source = "registry+https://github.com/rust-lang/crates.io-index"
+checksum = "ecc2af9a1119c51f12a14607e783cb977bde58bc069ff0c3da1095e635d70654"
+dependencies = [
+ "cpufeatures",
+]
+
+[[package]]
name = "khronos-egl"
version = "6.0.0"
source = "registry+https://github.com/rust-lang/crates.io-index"
@@ -5796,6 +5817,16 @@ dependencies = [
]
[[package]]
+name = "sha3"
+version = "0.10.8"
+source = "registry+https://github.com/rust-lang/crates.io-index"
+checksum = "75872d278a8f37ef87fa0ddbda7802605cb18344497949862c0d4dcb291eba60"
+dependencies = [
+ "digest",
+ "keccak",
+]
+
+[[package]]
name = "shell-escape"
version = "0.1.5"
source = "registry+https://github.com/rust-lang/crates.io-index"
@@ -5911,6 +5942,15 @@ dependencies = [
]
[[package]]
+name = "sm3"
+version = "0.4.2"
+source = "registry+https://github.com/rust-lang/crates.io-index"
+checksum = "ebb9a3b702d0a7e33bc4d85a14456633d2b165c2ad839c5fd9a8417c1ab15860"
+dependencies = [
+ "digest",
+]
+
+[[package]]
name = "smallvec"
version = "1.13.2"
source = "registry+https://github.com/rust-lang/crates.io-index"
diff --git a/ext/node/Cargo.toml b/ext/node/Cargo.toml
index 2a6fd758d..8b5895bc7 100644
--- a/ext/node/Cargo.toml
+++ b/ext/node/Cargo.toml
@@ -17,6 +17,7 @@ path = "lib.rs"
aead-gcm-stream = "0.1"
aes.workspace = true
async-trait.workspace = true
+blake2 = "0.10.6"
brotli.workspace = true
bytes.workspace = true
cbc.workspace = true
@@ -70,8 +71,10 @@ sec1 = "0.7"
serde = "1.0.149"
sha1.workspace = true
sha2.workspace = true
+sha3 = "0.10.8"
signature.workspace = true
simd-json = "0.13.4"
+sm3 = "0.4.2"
spki.workspace = true
tokio.workspace = true
url.workspace = true
diff --git a/ext/node/ops/crypto/digest.rs b/ext/node/ops/crypto/digest.rs
index 25bb0ab98..588ea7425 100644
--- a/ext/node/ops/crypto/digest.rs
+++ b/ext/node/ops/crypto/digest.rs
@@ -1,107 +1,293 @@
// Copyright 2018-2024 the Deno authors. All rights reserved. MIT license.
-use deno_core::error::type_error;
+use deno_core::error::generic_error;
use deno_core::error::AnyError;
-use deno_core::Resource;
+use deno_core::GcResource;
use digest::Digest;
use digest::DynDigest;
-use std::borrow::Cow;
+use digest::ExtendableOutput;
+use digest::Update;
use std::cell::RefCell;
use std::rc::Rc;
-pub enum Hash {
- Md4(Box<md4::Md4>),
- Md5(Box<md5::Md5>),
- Ripemd160(Box<ripemd::Ripemd160>),
- Sha1(Box<sha1::Sha1>),
- Sha224(Box<sha2::Sha224>),
- Sha256(Box<sha2::Sha256>),
- Sha384(Box<sha2::Sha384>),
- Sha512(Box<sha2::Sha512>),
+pub struct Hasher {
+ pub hash: Rc<RefCell<Option<Hash>>>,
}
-pub struct Context {
- pub hash: Rc<RefCell<Hash>>,
-}
+impl GcResource for Hasher {}
+
+impl Hasher {
+ pub fn new(
+ algorithm: &str,
+ output_length: Option<usize>,
+ ) -> Result<Self, AnyError> {
+ let hash = Hash::new(algorithm, output_length)?;
-impl Context {
- pub fn new(algorithm: &str) -> Result<Self, AnyError> {
Ok(Self {
- hash: Rc::new(RefCell::new(Hash::new(algorithm)?)),
+ hash: Rc::new(RefCell::new(Some(hash))),
})
}
- pub fn update(&self, data: &[u8]) {
- self.hash.borrow_mut().update(data);
+ pub fn update(&self, data: &[u8]) -> bool {
+ if let Some(hash) = self.hash.borrow_mut().as_mut() {
+ hash.update(data);
+ true
+ } else {
+ false
+ }
}
- pub fn digest(self) -> Result<Box<[u8]>, AnyError> {
- let hash = Rc::try_unwrap(self.hash)
- .map_err(|_| type_error("Hash context is already in use"))?;
-
- let hash = hash.into_inner();
- Ok(hash.digest_and_drop())
+ pub fn digest(&self) -> Option<Box<[u8]>> {
+ let hash = self.hash.borrow_mut().take()?;
+ Some(hash.digest_and_drop())
}
-}
-impl Clone for Context {
- fn clone(&self) -> Self {
- Self {
- hash: Rc::new(RefCell::new(self.hash.borrow().clone())),
- }
+ pub fn clone_inner(
+ &self,
+ output_length: Option<usize>,
+ ) -> Result<Option<Self>, AnyError> {
+ let hash = self.hash.borrow();
+ let Some(hash) = hash.as_ref() else {
+ return Ok(None);
+ };
+ let hash = hash.clone_hash(output_length)?;
+ Ok(Some(Self {
+ hash: Rc::new(RefCell::new(Some(hash))),
+ }))
}
}
-impl Resource for Context {
- fn name(&self) -> Cow<str> {
- "cryptoDigest".into()
- }
+pub enum Hash {
+ Blake2b512(Box<blake2::Blake2b512>),
+ Blake2s256(Box<blake2::Blake2s256>),
+
+ Md4(Box<md4::Md4>),
+ Md5(Box<md5::Md5>),
+
+ Ripemd160(Box<ripemd::Ripemd160>),
+
+ Sha1(Box<sha1::Sha1>),
+
+ Sha224(Box<sha2::Sha224>),
+ Sha256(Box<sha2::Sha256>),
+ Sha384(Box<sha2::Sha384>),
+ Sha512(Box<sha2::Sha512>),
+ Sha512_224(Box<sha2::Sha512_224>),
+ Sha512_256(Box<sha2::Sha512_256>),
+
+ Sha3_224(Box<sha3::Sha3_224>),
+ Sha3_256(Box<sha3::Sha3_256>),
+ Sha3_384(Box<sha3::Sha3_384>),
+ Sha3_512(Box<sha3::Sha3_512>),
+
+ Sm3(Box<sm3::Sm3>),
+
+ Shake128(Box<sha3::Shake128>, /* output_length: */ Option<usize>),
+ Shake256(Box<sha3::Shake256>, /* output_length: */ Option<usize>),
}
use Hash::*;
impl Hash {
- pub fn new(algorithm_name: &str) -> Result<Self, AnyError> {
- Ok(match algorithm_name {
+ pub fn new(
+ algorithm_name: &str,
+ output_length: Option<usize>,
+ ) -> Result<Self, AnyError> {
+ match algorithm_name {
+ "shake128" => return Ok(Shake128(Default::default(), output_length)),
+ "shake256" => return Ok(Shake256(Default::default(), output_length)),
+ _ => {}
+ }
+
+ let algorithm = match algorithm_name {
+ "blake2b512" => Blake2b512(Default::default()),
+ "blake2s256" => Blake2s256(Default::default()),
+
"md4" => Md4(Default::default()),
"md5" => Md5(Default::default()),
+
"ripemd160" => Ripemd160(Default::default()),
+
"sha1" => Sha1(Default::default()),
"sha224" => Sha224(Default::default()),
"sha256" => Sha256(Default::default()),
"sha384" => Sha384(Default::default()),
"sha512" => Sha512(Default::default()),
- _ => return Err(type_error("unsupported algorithm")),
- })
+ "sha512-224" => Sha512_224(Default::default()),
+ "sha512-256" => Sha512_256(Default::default()),
+
+ "sha3-224" => Sha3_224(Default::default()),
+ "sha3-256" => Sha3_256(Default::default()),
+ "sha3-384" => Sha3_384(Default::default()),
+ "sha3-512" => Sha3_512(Default::default()),
+
+ "sm3" => Sm3(Default::default()),
+
+ _ => {
+ return Err(generic_error(format!(
+ "Digest method not supported: {algorithm_name}"
+ )))
+ }
+ };
+ if let Some(length) = output_length {
+ if length != algorithm.output_length() {
+ return Err(generic_error(
+ "Output length mismatch for non-extendable algorithm",
+ ));
+ }
+ }
+ Ok(algorithm)
+ }
+
+ pub fn output_length(&self) -> usize {
+ match self {
+ Blake2b512(context) => context.output_size(),
+ Blake2s256(context) => context.output_size(),
+
+ Md4(context) => context.output_size(),
+ Md5(context) => context.output_size(),
+
+ Ripemd160(context) => context.output_size(),
+
+ Sha1(context) => context.output_size(),
+ Sha224(context) => context.output_size(),
+ Sha256(context) => context.output_size(),
+ Sha384(context) => context.output_size(),
+ Sha512(context) => context.output_size(),
+ Sha512_224(context) => context.output_size(),
+ Sha512_256(context) => context.output_size(),
+
+ Sha3_224(context) => context.output_size(),
+ Sha3_256(context) => context.output_size(),
+ Sha3_384(context) => context.output_size(),
+ Sha3_512(context) => context.output_size(),
+
+ Sm3(context) => context.output_size(),
+
+ Shake128(_, _) => unreachable!(
+ "output_length() should not be called on extendable algorithms"
+ ),
+ Shake256(_, _) => unreachable!(
+ "output_length() should not be called on extendable algorithms"
+ ),
+ }
}
pub fn update(&mut self, data: &[u8]) {
match self {
+ Blake2b512(context) => Digest::update(&mut **context, data),
+ Blake2s256(context) => Digest::update(&mut **context, data),
+
Md4(context) => Digest::update(&mut **context, data),
Md5(context) => Digest::update(&mut **context, data),
+
Ripemd160(context) => Digest::update(&mut **context, data),
+
Sha1(context) => Digest::update(&mut **context, data),
Sha224(context) => Digest::update(&mut **context, data),
Sha256(context) => Digest::update(&mut **context, data),
Sha384(context) => Digest::update(&mut **context, data),
Sha512(context) => Digest::update(&mut **context, data),
+ Sha512_224(context) => Digest::update(&mut **context, data),
+ Sha512_256(context) => Digest::update(&mut **context, data),
+
+ Sha3_224(context) => Digest::update(&mut **context, data),
+ Sha3_256(context) => Digest::update(&mut **context, data),
+ Sha3_384(context) => Digest::update(&mut **context, data),
+ Sha3_512(context) => Digest::update(&mut **context, data),
+
+ Sm3(context) => Digest::update(&mut **context, data),
+
+ Shake128(context, _) => Update::update(&mut **context, data),
+ Shake256(context, _) => Update::update(&mut **context, data),
};
}
pub fn digest_and_drop(self) -> Box<[u8]> {
match self {
+ Blake2b512(context) => context.finalize(),
+ Blake2s256(context) => context.finalize(),
+
Md4(context) => context.finalize(),
Md5(context) => context.finalize(),
+
Ripemd160(context) => context.finalize(),
+
Sha1(context) => context.finalize(),
Sha224(context) => context.finalize(),
Sha256(context) => context.finalize(),
Sha384(context) => context.finalize(),
Sha512(context) => context.finalize(),
+ Sha512_224(context) => context.finalize(),
+ Sha512_256(context) => context.finalize(),
+
+ Sha3_224(context) => context.finalize(),
+ Sha3_256(context) => context.finalize(),
+ Sha3_384(context) => context.finalize(),
+ Sha3_512(context) => context.finalize(),
+
+ Sm3(context) => context.finalize(),
+
+ // The default output lengths align with Node.js
+ Shake128(context, output_length) => {
+ context.finalize_boxed(output_length.unwrap_or(16))
+ }
+ Shake256(context, output_length) => {
+ context.finalize_boxed(output_length.unwrap_or(32))
+ }
+ }
+ }
+
+ pub fn clone_hash(
+ &self,
+ output_length: Option<usize>,
+ ) -> Result<Self, AnyError> {
+ let hash = match self {
+ Shake128(context, _) => {
+ return Ok(Shake128(context.clone(), output_length))
+ }
+ Shake256(context, _) => {
+ return Ok(Shake256(context.clone(), output_length))
+ }
+
+ Blake2b512(context) => Blake2b512(context.clone()),
+ Blake2s256(context) => Blake2s256(context.clone()),
+
+ Md4(context) => Md4(context.clone()),
+ Md5(context) => Md5(context.clone()),
+
+ Ripemd160(context) => Ripemd160(context.clone()),
+
+ Sha1(context) => Sha1(context.clone()),
+ Sha224(context) => Sha224(context.clone()),
+ Sha256(context) => Sha256(context.clone()),
+ Sha384(context) => Sha384(context.clone()),
+ Sha512(context) => Sha512(context.clone()),
+ Sha512_224(context) => Sha512_224(context.clone()),
+ Sha512_256(context) => Sha512_256(context.clone()),
+
+ Sha3_224(context) => Sha3_224(context.clone()),
+ Sha3_256(context) => Sha3_256(context.clone()),
+ Sha3_384(context) => Sha3_384(context.clone()),
+ Sha3_512(context) => Sha3_512(context.clone()),
+
+ Sm3(context) => Sm3(context.clone()),
+ };
+
+ if let Some(length) = output_length {
+ if length != hash.output_length() {
+ return Err(generic_error(
+ "Output length mismatch for non-extendable algorithm",
+ ));
+ }
}
+
+ Ok(hash)
}
pub fn get_hashes() -> Vec<&'static str> {
vec![
+ "blake2s256",
+ "blake2b512",
"md4",
"md5",
"ripemd160",
@@ -110,21 +296,15 @@ impl Hash {
"sha256",
"sha384",
"sha512",
+ "sha512-224",
+ "sha512-256",
+ "sha3-224",
+ "sha3-256",
+ "sha3-384",
+ "sha3-512",
+ "shake128",
+ "shake256",
+ "sm3",
]
}
}
-
-impl Clone for Hash {
- fn clone(&self) -> Self {
- match self {
- Md4(_) => Md4(Default::default()),
- Md5(_) => Md5(Default::default()),
- Ripemd160(_) => Ripemd160(Default::default()),
- Sha1(_) => Sha1(Default::default()),
- Sha224(_) => Sha224(Default::default()),
- Sha256(_) => Sha256(Default::default()),
- Sha384(_) => Sha384(Default::default()),
- Sha512(_) => Sha512(Default::default()),
- }
- }
-}
diff --git a/ext/node/ops/crypto/mod.rs b/ext/node/ops/crypto/mod.rs
index 666ce8409..8ded3420a 100644
--- a/ext/node/ops/crypto/mod.rs
+++ b/ext/node/ops/crypto/mod.rs
@@ -7,7 +7,6 @@ use deno_core::serde_v8::BigInt as V8BigInt;
use deno_core::unsync::spawn_blocking;
use deno_core::JsBuffer;
use deno_core::OpState;
-use deno_core::ResourceId;
use deno_core::StringOrBuffer;
use deno_core::ToJsBuffer;
use elliptic_curve::sec1::ToEncodedPoint;
@@ -96,18 +95,13 @@ pub fn op_node_check_prime_bytes_async(
})
}
-#[op2(fast)]
-#[smi]
+#[op2]
+#[cppgc]
pub fn op_node_create_hash(
- state: &mut OpState,
#[string] algorithm: &str,
-) -> u32 {
- state
- .resource_table
- .add(match digest::Context::new(algorithm) {
- Ok(context) => context,
- Err(_) => return 0,
- })
+ output_length: Option<u32>,
+) -> Result<digest::Hasher, AnyError> {
+ digest::Hasher::new(algorithm, output_length.map(|l| l as usize))
}
#[op2]
@@ -118,65 +112,44 @@ pub fn op_node_get_hashes() -> Vec<&'static str> {
#[op2(fast)]
pub fn op_node_hash_update(
- state: &mut OpState,
- #[smi] rid: u32,
+ #[cppgc] hasher: &digest::Hasher,
#[buffer] data: &[u8],
) -> bool {
- let context = match state.resource_table.get::<digest::Context>(rid) {
- Ok(context) => context,
- _ => return false,
- };
- context.update(data);
- true
+ hasher.update(data)
}
#[op2(fast)]
pub fn op_node_hash_update_str(
- state: &mut OpState,
- #[smi] rid: u32,
+ #[cppgc] hasher: &digest::Hasher,
#[string] data: &str,
) -> bool {
- let context = match state.resource_table.get::<digest::Context>(rid) {
- Ok(context) => context,
- _ => return false,
- };
- context.update(data.as_bytes());
- true
+ hasher.update(data.as_bytes())
}
#[op2]
-#[serde]
+#[buffer]
pub fn op_node_hash_digest(
- state: &mut OpState,
- #[smi] rid: ResourceId,
-) -> Result<ToJsBuffer, AnyError> {
- let context = state.resource_table.take::<digest::Context>(rid)?;
- let context = Rc::try_unwrap(context)
- .map_err(|_| type_error("Hash context is already in use"))?;
- Ok(context.digest()?.into())
+ #[cppgc] hasher: &digest::Hasher,
+) -> Option<Box<[u8]>> {
+ hasher.digest()
}
#[op2]
#[string]
pub fn op_node_hash_digest_hex(
- state: &mut OpState,
- #[smi] rid: ResourceId,
-) -> Result<String, AnyError> {
- let context = state.resource_table.take::<digest::Context>(rid)?;
- let context = Rc::try_unwrap(context)
- .map_err(|_| type_error("Hash context is already in use"))?;
- let digest = context.digest()?;
- Ok(faster_hex::hex_string(&digest))
+ #[cppgc] hasher: &digest::Hasher,
+) -> Option<String> {
+ let digest = hasher.digest()?;
+ Some(faster_hex::hex_string(&digest))
}
-#[op2(fast)]
-#[smi]
+#[op2]
+#[cppgc]
pub fn op_node_hash_clone(
- state: &mut OpState,
- #[smi] rid: ResourceId,
-) -> Result<ResourceId, AnyError> {
- let context = state.resource_table.get::<digest::Context>(rid)?;
- Ok(state.resource_table.add(context.as_ref().clone()))
+ #[cppgc] hasher: &digest::Hasher,
+ output_length: Option<u32>,
+) -> Result<Option<digest::Hasher>, AnyError> {
+ hasher.clone_inner(output_length.map(|l| l as usize))
}
#[op2]
diff --git a/ext/node/polyfills/internal/crypto/hash.ts b/ext/node/polyfills/internal/crypto/hash.ts
index a1d61f953..2e040be25 100644
--- a/ext/node/polyfills/internal/crypto/hash.ts
+++ b/ext/node/polyfills/internal/crypto/hash.ts
@@ -13,8 +13,8 @@ import {
op_node_hash_update,
op_node_hash_update_str,
} from "ext:core/ops";
+import { primordials } from "ext:core/mod.js";
-import { TextEncoder } from "ext:deno_web/08_text_encoding.js";
import { Buffer } from "node:buffer";
import { Transform } from "node:stream";
import {
@@ -22,7 +22,11 @@ import {
forgivingBase64UrlEncode as encodeToBase64Url,
} from "ext:deno_web/00_infra.js";
import type { TransformOptions } from "ext:deno_node/_stream.d.ts";
-import { validateString } from "ext:deno_node/internal/validators.mjs";
+import {
+ validateEncoding,
+ validateString,
+ validateUint32,
+} from "ext:deno_node/internal/validators.mjs";
import type {
BinaryToTextEncoding,
Encoding,
@@ -32,119 +36,148 @@ import {
KeyObject,
prepareSecretKey,
} from "ext:deno_node/internal/crypto/keys.ts";
+import {
+ ERR_CRYPTO_HASH_FINALIZED,
+ ERR_INVALID_ARG_TYPE,
+ NodeError,
+} from "ext:deno_node/internal/errors.ts";
+import LazyTransform from "ext:deno_node/internal/streams/lazy_transform.mjs";
+import {
+ getDefaultEncoding,
+ toBuf,
+} from "ext:deno_node/internal/crypto/util.ts";
+import { isArrayBufferView } from "ext:deno_node/internal/util/types.ts";
+
+const { ReflectApply, ObjectSetPrototypeOf } = primordials;
-// TODO(@littledivy): Use Result<T, E> instead of boolean when
-// https://bugs.chromium.org/p/v8/issues/detail?id=13600 is fixed.
function unwrapErr(ok: boolean) {
- if (!ok) {
- throw new Error("Context is not initialized");
- }
+ if (!ok) throw new ERR_CRYPTO_HASH_FINALIZED();
}
-const coerceToBytes = (data: string | BufferSource): Uint8Array => {
- if (data instanceof Uint8Array) {
- return data;
- } else if (typeof data === "string") {
- // This assumes UTF-8, which may not be correct.
- return new TextEncoder().encode(data);
- } else if (ArrayBuffer.isView(data)) {
- return new Uint8Array(data.buffer, data.byteOffset, data.byteLength);
- } else if (data instanceof ArrayBuffer) {
- return new Uint8Array(data);
- } else {
- throw new TypeError("expected data to be string | BufferSource");
- }
-};
+declare const __hasher: unique symbol;
+type Hasher = { __hasher: typeof __hasher };
-/**
- * The Hash class is a utility for creating hash digests of data. It can be used in one of two ways:
- *
- * - As a stream that is both readable and writable, where data is written to produce a computed hash digest on the readable side, or
- * - Using the hash.update() and hash.digest() methods to produce the computed hash.
- *
- * The crypto.createHash() method is used to create Hash instances. Hash objects are not to be created directly using the new keyword.
- */
-export class Hash extends Transform {
- #context: number;
+const kHandle = Symbol("kHandle");
- constructor(
- algorithm: string | number,
- _opts?: TransformOptions,
- ) {
- super({
- transform(chunk: string, _encoding: string, callback: () => void) {
- op_node_hash_update(context, coerceToBytes(chunk));
- callback();
- },
- flush(callback: () => void) {
- this.push(this.digest(undefined));
- callback();
- },
- });
+export function Hash(
+ this: Hash,
+ algorithm: string | Hasher,
+ options?: { outputLength?: number },
+): Hash {
+ if (!(this instanceof Hash)) {
+ return new Hash(algorithm, options);
+ }
+ if (!(typeof algorithm === "object")) {
+ validateString(algorithm, "algorithm");
+ }
+ const xofLen = typeof options === "object" && options !== null
+ ? options.outputLength
+ : undefined;
+ if (xofLen !== undefined) {
+ validateUint32(xofLen, "options.outputLength");
+ }
- if (typeof algorithm === "string") {
- this.#context = op_node_create_hash(
- algorithm.toLowerCase(),
+ try {
+ this[kHandle] = typeof algorithm === "object"
+ ? op_node_hash_clone(algorithm, xofLen)
+ : op_node_create_hash(algorithm.toLowerCase(), xofLen);
+ } catch (err) {
+ // TODO(lucacasonato): don't do this
+ if (err.message === "Output length mismatch for non-extendable algorithm") {
+ throw new NodeError(
+ "ERR_OSSL_EVP_NOT_XOF_OR_INVALID_LENGTH",
+ "Invalid XOF digest length",
);
- if (this.#context === 0) {
- throw new TypeError(`Unknown hash algorithm: ${algorithm}`);
- }
} else {
- this.#context = algorithm;
+ throw err;
}
+ }
+
+ if (this[kHandle] === null) throw new ERR_CRYPTO_HASH_FINALIZED();
+
+ ReflectApply(LazyTransform, this, [options]);
+}
- const context = this.#context;
+interface Hash {
+ [kHandle]: object;
+}
+
+ObjectSetPrototypeOf(Hash.prototype, LazyTransform.prototype);
+ObjectSetPrototypeOf(Hash, LazyTransform);
+
+Hash.prototype.copy = function copy(options?: { outputLength: number }) {
+ return new Hash(this[kHandle], options);
+};
+
+Hash.prototype._transform = function _transform(
+ chunk: string | Buffer,
+ encoding: Encoding | "buffer",
+ callback: () => void,
+) {
+ this.update(chunk, encoding);
+ callback();
+};
+
+Hash.prototype._flush = function _flush(callback: () => void) {
+ this.push(this.digest());
+ callback();
+};
+
+Hash.prototype.update = function update(
+ data: string | Buffer,
+ encoding: Encoding | "buffer",
+) {
+ encoding = encoding || getDefaultEncoding();
+
+ if (typeof data === "string") {
+ validateEncoding(data, encoding);
+ } else if (!isArrayBufferView(data)) {
+ throw new ERR_INVALID_ARG_TYPE(
+ "data",
+ ["string", "Buffer", "TypedArray", "DataView"],
+ data,
+ );
}
- copy(): Hash {
- return new Hash(op_node_hash_clone(this.#context));
+ if (
+ typeof data === "string" && (encoding === "utf8" || encoding === "buffer")
+ ) {
+ unwrapErr(op_node_hash_update_str(this[kHandle], data));
+ } else {
+ unwrapErr(op_node_hash_update(this[kHandle], toBuf(data, encoding)));
}
- /**
- * Updates the hash content with the given data.
- */
- update(data: string | ArrayBuffer, _encoding?: string): this {
- if (typeof data === "string") {
- unwrapErr(op_node_hash_update_str(this.#context, data));
- } else {
- unwrapErr(op_node_hash_update(this.#context, coerceToBytes(data)));
- }
+ return this;
+};
- return this;
- }
+Hash.prototype.digest = function digest(outputEncoding: Encoding | "buffer") {
+ outputEncoding = outputEncoding || getDefaultEncoding();
+ outputEncoding = `${outputEncoding}`;
- /**
- * Calculates the digest of all of the data.
- *
- * If encoding is provided a string will be returned; otherwise a Buffer is returned.
- *
- * Supported encodings are currently 'hex', 'binary', 'base64', 'base64url'.
- */
- digest(encoding?: string): Buffer | string {
- if (encoding === "hex") {
- return op_node_hash_digest_hex(this.#context);
- }
+ if (outputEncoding === "hex") {
+ const result = op_node_hash_digest_hex(this[kHandle]);
+ if (result === null) throw new ERR_CRYPTO_HASH_FINALIZED();
+ return result;
+ }
- const digest = op_node_hash_digest(this.#context);
- if (encoding === undefined) {
+ const digest = op_node_hash_digest(this[kHandle]);
+ if (digest === null) throw new ERR_CRYPTO_HASH_FINALIZED();
+
+ // TODO(@littedivy): Fast paths for below encodings.
+ switch (outputEncoding) {
+ case "binary":
+ return String.fromCharCode(...digest);
+ case "base64":
+ return encodeToBase64(digest);
+ case "base64url":
+ return encodeToBase64Url(digest);
+ case undefined:
+ case "buffer":
return Buffer.from(digest);
- }
-
- // TODO(@littedivy): Fast paths for below encodings.
- switch (encoding) {
- case "binary":
- return String.fromCharCode(...digest);
- case "base64":
- return encodeToBase64(digest);
- case "base64url":
- return encodeToBase64Url(digest);
- case "buffer":
- return Buffer.from(digest);
- default:
- return Buffer.from(digest).toString(encoding);
- }
+ default:
+ return Buffer.from(digest).toString(outputEncoding);
}
-}
+};
export function Hmac(
hmac: string,
@@ -171,7 +204,7 @@ class HmacImpl extends Transform {
super({
transform(chunk: string, encoding: string, callback: () => void) {
// deno-lint-ignore no-explicit-any
- self.update(coerceToBytes(chunk), encoding as any);
+ self.update(Buffer.from(chunk), encoding as any);
callback();
},
flush(callback: () => void) {
@@ -219,9 +252,10 @@ class HmacImpl extends Transform {
digest(encoding?: BinaryToTextEncoding): Buffer | string {
const result = this.#hash.digest();
- return new Hash(this.#algorithm).update(this.#opad).update(result).digest(
- encoding,
- );
+ return new Hash(this.#algorithm).update(this.#opad).update(result)
+ .digest(
+ encoding,
+ );
}
update(data: string | ArrayBuffer, inputEncoding?: Encoding): this {
diff --git a/tests/integration/node_unit_tests.rs b/tests/integration/node_unit_tests.rs
index b067f3121..d0b6d1fbd 100644
--- a/tests/integration/node_unit_tests.rs
+++ b/tests/integration/node_unit_tests.rs
@@ -61,6 +61,7 @@ util::unit_test_factory!(
crypto_cipher_gcm_test = crypto / crypto_cipher_gcm_test,
crypto_hash_test = crypto / crypto_hash_test,
crypto_key_test = crypto / crypto_key_test,
+ crypto_misc_test = crypto / crypto_misc_test,
crypto_sign_test = crypto / crypto_sign_test,
events_test,
dgram_test,
diff --git a/tests/node_compat/config.jsonc b/tests/node_compat/config.jsonc
index 0a661c0a9..83d91eac1 100644
--- a/tests/node_compat/config.jsonc
+++ b/tests/node_compat/config.jsonc
@@ -132,16 +132,17 @@
"tmpdir.js"
],
"fixtures": [
- "GH-1899-output.js",
"a.js",
- "child-process-spawn-node.js",
"child_process_should_emit_error.js",
+ "child-process-spawn-node.js",
"echo.js",
"elipses.txt",
"empty.txt",
"exit.js",
+ "GH-1899-output.js",
"loop.js",
"print-chars.js",
+ "sample.png",
"x.txt"
],
"fixtures/keys": ["agent1-cert.pem", "agent1-key.pem", "ca1-cert.pem"],
@@ -253,6 +254,7 @@
"test-console-tty-colors.js",
"test-crypto-dh-shared.js",
"test-crypto-dh.js",
+ "test-crypto-hash.js",
"test-crypto-hkdf.js",
"test-crypto-hmac.js",
"test-crypto-prime.js",
@@ -701,8 +703,8 @@
"test-zlib-zero-windowBits.js"
],
"pseudo-tty": [
- "console-dumb-tty.js",
"console_colors.js",
+ "console-dumb-tty.js",
"no_dropped_stdio.js",
"no_interleaved_stdio.js",
"test-tty-color-support-warning-2.js",
diff --git a/tests/node_compat/runner/TODO.md b/tests/node_compat/runner/TODO.md
index d5dd2fc7b..24618dc88 100644
--- a/tests/node_compat/runner/TODO.md
+++ b/tests/node_compat/runner/TODO.md
@@ -464,7 +464,6 @@ NOTE: This file should not be manually edited. Please edit `tests/node_compat/co
- [parallel/test-crypto-from-binary.js](https://github.com/nodejs/node/tree/v18.12.1/test/parallel/test-crypto-from-binary.js)
- [parallel/test-crypto-getcipherinfo.js](https://github.com/nodejs/node/tree/v18.12.1/test/parallel/test-crypto-getcipherinfo.js)
- [parallel/test-crypto-hash-stream-pipe.js](https://github.com/nodejs/node/tree/v18.12.1/test/parallel/test-crypto-hash-stream-pipe.js)
-- [parallel/test-crypto-hash.js](https://github.com/nodejs/node/tree/v18.12.1/test/parallel/test-crypto-hash.js)
- [parallel/test-crypto-key-objects-messageport.js](https://github.com/nodejs/node/tree/v18.12.1/test/parallel/test-crypto-key-objects-messageport.js)
- [parallel/test-crypto-key-objects.js](https://github.com/nodejs/node/tree/v18.12.1/test/parallel/test-crypto-key-objects.js)
- [parallel/test-crypto-keygen-async-dsa-key-object.js](https://github.com/nodejs/node/tree/v18.12.1/test/parallel/test-crypto-keygen-async-dsa-key-object.js)
diff --git a/tests/node_compat/test.ts b/tests/node_compat/test.ts
index 6f15f2d0b..939fdf52a 100644
--- a/tests/node_compat/test.ts
+++ b/tests/node_compat/test.ts
@@ -16,7 +16,7 @@
import { magenta } from "@std/fmt/colors.ts";
import { pooledMap } from "@std/async/pool.ts";
import { dirname, fromFileUrl, join } from "@std/path/mod.ts";
-import { fail } from "@std/assert/mod.ts";
+import { assertEquals, fail } from "@std/assert/mod.ts";
import {
config,
getPathsFromTestSuites,
@@ -169,12 +169,14 @@ Deno.test("Node.js compatibility", async (t) => {
function checkConfigTestFilesOrder(testFileLists: Array<string[]>) {
for (const testFileList of testFileLists) {
const sortedTestList = JSON.parse(JSON.stringify(testFileList));
- sortedTestList.sort();
- if (JSON.stringify(testFileList) !== JSON.stringify(sortedTestList)) {
- throw new Error(
- `File names in \`config.json\` are not correct order.`,
- );
- }
+ sortedTestList.sort((a: string, b: string) =>
+ a.toLowerCase().localeCompare(b.toLowerCase())
+ );
+ assertEquals(
+ testFileList,
+ sortedTestList,
+ "File names in `config.json` are not correct order.",
+ );
}
}
diff --git a/tests/node_compat/test/fixtures/sample.png b/tests/node_compat/test/fixtures/sample.png
new file mode 100644
index 000000000..258622019
--- /dev/null
+++ b/tests/node_compat/test/fixtures/sample.png
Binary files differ
diff --git a/tests/node_compat/test/parallel/test-crypto-hash.js b/tests/node_compat/test/parallel/test-crypto-hash.js
new file mode 100644
index 000000000..18c57dab7
--- /dev/null
+++ b/tests/node_compat/test/parallel/test-crypto-hash.js
@@ -0,0 +1,285 @@
+// deno-fmt-ignore-file
+// deno-lint-ignore-file
+
+// Copyright Joyent and Node contributors. All rights reserved. MIT license.
+// Taken from Node 18.12.1
+// This file is automatically generated by `tests/node_compat/runner/setup.ts`. Do not modify this file manually.
+
+'use strict';
+const common = require('../common');
+if (!common.hasCrypto)
+ common.skip('missing crypto');
+
+const assert = require('assert');
+const crypto = require('crypto');
+const fs = require('fs');
+
+const fixtures = require('../common/fixtures');
+
+let cryptoType;
+let digest;
+
+// Test hashing
+const a1 = crypto.createHash('sha1').update('Test123').digest('hex');
+const a2 = crypto.createHash('sha256').update('Test123').digest('base64');
+const a3 = crypto.createHash('sha512').update('Test123').digest(); // buffer
+const a4 = crypto.createHash('sha1').update('Test123').digest('buffer');
+
+// stream interface
+let a5 = crypto.createHash('sha512');
+a5.end('Test123');
+a5 = a5.read();
+
+let a6 = crypto.createHash('sha512');
+a6.write('Te');
+a6.write('st');
+a6.write('123');
+a6.end();
+a6 = a6.read();
+
+let a7 = crypto.createHash('sha512');
+a7.end();
+a7 = a7.read();
+
+let a8 = crypto.createHash('sha512');
+a8.write('');
+a8.end();
+a8 = a8.read();
+
+if (!common.hasFipsCrypto) {
+ cryptoType = 'md5';
+ digest = 'latin1';
+ const a0 = crypto.createHash(cryptoType).update('Test123').digest(digest);
+ assert.strictEqual(
+ a0,
+ 'h\u00ea\u00cb\u0097\u00d8o\fF!\u00fa+\u000e\u0017\u00ca\u00bd\u008c',
+ `${cryptoType} with ${digest} digest failed to evaluate to expected hash`
+ );
+}
+cryptoType = 'md5';
+digest = 'hex';
+assert.strictEqual(
+ a1,
+ '8308651804facb7b9af8ffc53a33a22d6a1c8ac2',
+ `${cryptoType} with ${digest} digest failed to evaluate to expected hash`);
+cryptoType = 'sha256';
+digest = 'base64';
+assert.strictEqual(
+ a2,
+ '2bX1jws4GYKTlxhloUB09Z66PoJZW+y+hq5R8dnx9l4=',
+ `${cryptoType} with ${digest} digest failed to evaluate to expected hash`);
+cryptoType = 'sha512';
+digest = 'latin1';
+assert.deepStrictEqual(
+ a3,
+ Buffer.from(
+ '\u00c1(4\u00f1\u0003\u001fd\u0097!O\'\u00d4C/&Qz\u00d4' +
+ '\u0094\u0015l\u00b8\u008dQ+\u00db\u001d\u00c4\u00b5}\u00b2' +
+ '\u00d6\u0092\u00a3\u00df\u00a2i\u00a1\u009b\n\n*\u000f' +
+ '\u00d7\u00d6\u00a2\u00a8\u0085\u00e3<\u0083\u009c\u0093' +
+ '\u00c2\u0006\u00da0\u00a1\u00879(G\u00ed\'',
+ 'latin1'),
+ `${cryptoType} with ${digest} digest failed to evaluate to expected hash`);
+cryptoType = 'sha1';
+digest = 'hex';
+assert.deepStrictEqual(
+ a4,
+ Buffer.from('8308651804facb7b9af8ffc53a33a22d6a1c8ac2', 'hex'),
+ `${cryptoType} with ${digest} digest failed to evaluate to expected hash`
+);
+
+// Stream interface should produce the same result.
+assert.deepStrictEqual(a5, a3);
+assert.deepStrictEqual(a6, a3);
+assert.notStrictEqual(a7, undefined);
+assert.notStrictEqual(a8, undefined);
+
+// Test multiple updates to same hash
+const h1 = crypto.createHash('sha1').update('Test123').digest('hex');
+const h2 = crypto.createHash('sha1').update('Test').update('123').digest('hex');
+assert.strictEqual(h1, h2);
+
+// Test hashing for binary files
+const fn = fixtures.path('sample.png');
+const sha1Hash = crypto.createHash('sha1');
+const fileStream = fs.createReadStream(fn);
+fileStream.on('data', function(data) {
+ sha1Hash.update(data);
+});
+fileStream.on('close', common.mustCall(function() {
+ // Test SHA1 of sample.png
+ assert.strictEqual(sha1Hash.digest('hex'),
+ '22723e553129a336ad96e10f6aecdf0f45e4149e');
+}));
+
+// Issue https://github.com/nodejs/node-v0.x-archive/issues/2227: unknown digest
+// method should throw an error.
+assert.throws(function() {
+ crypto.createHash('xyzzy');
+}, /Digest method not supported/);
+
+// Issue https://github.com/nodejs/node/issues/9819: throwing encoding used to
+// segfault.
+assert.throws(
+ () => crypto.createHash('sha256').digest({
+ toString: () => { throw new Error('boom'); },
+ }),
+ {
+ name: 'Error',
+ message: 'boom'
+ });
+
+// Issue https://github.com/nodejs/node/issues/25487: error message for invalid
+// arg type to update method should include all possible types
+assert.throws(
+ () => crypto.createHash('sha256').update(),
+ {
+ code: 'ERR_INVALID_ARG_TYPE',
+ name: 'TypeError',
+ });
+
+// Default UTF-8 encoding
+const hutf8 = crypto.createHash('sha512').update('УТФ-8 text').digest('hex');
+assert.strictEqual(
+ hutf8,
+ '4b21bbd1a68e690a730ddcb5a8bc94ead9879ffe82580767ad7ec6fa8ba2dea6' +
+ '43a821af66afa9a45b6a78c712fecf0e56dc7f43aef4bcfc8eb5b4d8dca6ea5b');
+
+assert.notStrictEqual(
+ hutf8,
+ crypto.createHash('sha512').update('УТФ-8 text', 'latin1').digest('hex'));
+
+const h3 = crypto.createHash('sha256');
+h3.digest();
+
+assert.throws(
+ () => h3.digest(),
+ {
+ code: 'ERR_CRYPTO_HASH_FINALIZED',
+ name: 'Error'
+ });
+
+assert.throws(
+ () => h3.update('foo'),
+ {
+ code: 'ERR_CRYPTO_HASH_FINALIZED',
+ name: 'Error'
+ });
+
+assert.strictEqual(
+ crypto.createHash('sha256').update('test').digest('ucs2'),
+ crypto.createHash('sha256').update('test').digest().toString('ucs2'));
+
+assert.throws(
+ () => crypto.createHash(),
+ {
+ code: 'ERR_INVALID_ARG_TYPE',
+ name: 'TypeError',
+ message: 'The "algorithm" argument must be of type string. ' +
+ 'Received undefined'
+ }
+);
+
+{
+ const Hash = crypto.Hash;
+ const instance = crypto.Hash('sha256');
+ assert(instance instanceof Hash, 'Hash is expected to return a new instance' +
+ ' when called without `new`');
+}
+
+// Test XOF hash functions and the outputLength option.
+{
+ // Default outputLengths.
+ assert.strictEqual(crypto.createHash('shake128').digest('hex'),
+ '7f9c2ba4e88f827d616045507605853e');
+ assert.strictEqual(crypto.createHash('shake128', null).digest('hex'),
+ '7f9c2ba4e88f827d616045507605853e');
+ assert.strictEqual(crypto.createHash('shake256').digest('hex'),
+ '46b9dd2b0ba88d13233b3feb743eeb24' +
+ '3fcd52ea62b81b82b50c27646ed5762f');
+ assert.strictEqual(crypto.createHash('shake256', { outputLength: 0 })
+ .copy() // Default outputLength.
+ .digest('hex'),
+ '46b9dd2b0ba88d13233b3feb743eeb24' +
+ '3fcd52ea62b81b82b50c27646ed5762f');
+
+ // Short outputLengths.
+ assert.strictEqual(crypto.createHash('shake128', { outputLength: 0 })
+ .digest('hex'),
+ '');
+ assert.strictEqual(crypto.createHash('shake128', { outputLength: 5 })
+ .copy({ outputLength: 0 })
+ .digest('hex'),
+ '');
+ assert.strictEqual(crypto.createHash('shake128', { outputLength: 5 })
+ .digest('hex'),
+ '7f9c2ba4e8');
+ assert.strictEqual(crypto.createHash('shake128', { outputLength: 0 })
+ .copy({ outputLength: 5 })
+ .digest('hex'),
+ '7f9c2ba4e8');
+ assert.strictEqual(crypto.createHash('shake128', { outputLength: 15 })
+ .digest('hex'),
+ '7f9c2ba4e88f827d61604550760585');
+ assert.strictEqual(crypto.createHash('shake256', { outputLength: 16 })
+ .digest('hex'),
+ '46b9dd2b0ba88d13233b3feb743eeb24');
+
+ // Large outputLengths.
+ assert.strictEqual(crypto.createHash('shake128', { outputLength: 128 })
+ .digest('hex'),
+ '7f9c2ba4e88f827d616045507605853e' +
+ 'd73b8093f6efbc88eb1a6eacfa66ef26' +
+ '3cb1eea988004b93103cfb0aeefd2a68' +
+ '6e01fa4a58e8a3639ca8a1e3f9ae57e2' +
+ '35b8cc873c23dc62b8d260169afa2f75' +
+ 'ab916a58d974918835d25e6a435085b2' +
+ 'badfd6dfaac359a5efbb7bcc4b59d538' +
+ 'df9a04302e10c8bc1cbf1a0b3a5120ea');
+ const superLongHash = crypto.createHash('shake256', {
+ outputLength: 1024 * 1024
+ }).update('The message is shorter than the hash!')
+ .digest('hex');
+ assert.strictEqual(superLongHash.length, 2 * 1024 * 1024);
+ assert.ok(superLongHash.endsWith('193414035ddba77bf7bba97981e656ec'));
+ assert.ok(superLongHash.startsWith('a2a28dbc49cfd6e5d6ceea3d03e77748'));
+
+ // Non-XOF hash functions should accept valid outputLength options as well.
+ assert.strictEqual(crypto.createHash('sha224', { outputLength: 28 })
+ .digest('hex'),
+ 'd14a028c2a3a2bc9476102bb288234c4' +
+ '15a2b01f828ea62ac5b3e42f');
+
+ // Passing invalid sizes should throw during creation.
+ assert.throws(() => {
+ crypto.createHash('sha256', { outputLength: 28 });
+ }, {
+ code: 'ERR_OSSL_EVP_NOT_XOF_OR_INVALID_LENGTH'
+ });
+
+ for (const outputLength of [null, {}, 'foo', false]) {
+ assert.throws(() => crypto.createHash('sha256', { outputLength }),
+ { code: 'ERR_INVALID_ARG_TYPE' });
+ }
+
+ for (const outputLength of [-1, .5, Infinity, 2 ** 90]) {
+ assert.throws(() => crypto.createHash('sha256', { outputLength }),
+ { code: 'ERR_OUT_OF_RANGE' });
+ }
+}
+
+{
+ const h = crypto.createHash('sha512');
+ h.digest();
+ assert.throws(() => h.copy(), { code: 'ERR_CRYPTO_HASH_FINALIZED' });
+ assert.throws(() => h.digest(), { code: 'ERR_CRYPTO_HASH_FINALIZED' });
+}
+
+{
+ const a = crypto.createHash('sha512').update('abc');
+ const b = a.copy();
+ const c = b.copy().update('def');
+ const d = crypto.createHash('sha512').update('abcdef');
+ assert.strictEqual(a.digest('hex'), b.digest('hex'));
+ assert.strictEqual(c.digest('hex'), d.digest('hex'));
+}
diff --git a/tests/unit_node/crypto/crypto_hash_test.ts b/tests/unit_node/crypto/crypto_hash_test.ts
index 74223067e..96bc1d51b 100644
--- a/tests/unit_node/crypto/crypto_hash_test.ts
+++ b/tests/unit_node/crypto/crypto_hash_test.ts
@@ -1,11 +1,5 @@
// Copyright 2018-2024 the Deno authors. All rights reserved. MIT license.
-import {
- createHash,
- createHmac,
- getHashes,
- randomFillSync,
- randomUUID,
-} from "node:crypto";
+import { createHash, createHmac, getHashes } from "node:crypto";
import { Buffer } from "node:buffer";
import { Readable } from "node:stream";
import { assert, assertEquals } from "@std/assert/mod.ts";
@@ -123,17 +117,13 @@ Deno.test("[node/crypto.getHashes]", () => {
}
});
-Deno.test("[node/crypto.getRandomUUID] works the same way as Web Crypto API", () => {
- assertEquals(randomUUID().length, crypto.randomUUID().length);
- assertEquals(typeof randomUUID(), typeof crypto.randomUUID());
+Deno.test("[node/crypto.hash] supports buffer args", () => {
+ const buffer = Buffer.from("abc");
+ const d = createHash("sha1").update(buffer).digest("hex");
+ assertEquals(d, "a9993e364706816aba3e25717850c26c9cd0d89d");
});
-Deno.test("[node/crypto.randomFillSync] supported arguments", () => {
- const buf = new Uint8Array(10);
-
- assert(randomFillSync(buf));
- assert(randomFillSync(buf, 0));
- // @ts-ignore: arraybuffer arguments are valid.
- assert(randomFillSync(buf.buffer));
- assert(randomFillSync(new DataView(buf.buffer)));
+Deno.test("[node/crypto.hash] does not leak", () => {
+ const hasher = createHash("sha1");
+ hasher.update("abc");
});
diff --git a/tests/unit_node/crypto/crypto_misc_test.ts b/tests/unit_node/crypto/crypto_misc_test.ts
new file mode 100644
index 000000000..8132f2e99
--- /dev/null
+++ b/tests/unit_node/crypto/crypto_misc_test.ts
@@ -0,0 +1,18 @@
+// Copyright 2018-2024 the Deno authors. All rights reserved. MIT license.
+import { randomFillSync, randomUUID } from "node:crypto";
+import { assert, assertEquals } from "../../unit/test_util.ts";
+
+Deno.test("[node/crypto.getRandomUUID] works the same way as Web Crypto API", () => {
+ assertEquals(randomUUID().length, crypto.randomUUID().length);
+ assertEquals(typeof randomUUID(), typeof crypto.randomUUID());
+});
+
+Deno.test("[node/crypto.randomFillSync] supported arguments", () => {
+ const buf = new Uint8Array(10);
+
+ assert(randomFillSync(buf));
+ assert(randomFillSync(buf, 0));
+ // @ts-ignore: arraybuffer arguments are valid.
+ assert(randomFillSync(buf.buffer));
+ assert(randomFillSync(new DataView(buf.buffer)));
+});