From 980ad3fbf9329f5e923e1fa5bea4f821eb897238 Mon Sep 17 00:00:00 2001 From: slvrtrn Date: Tue, 1 Oct 2024 13:57:01 +0200 Subject: [PATCH] Remove fdescribe --- .../__tests__/integration/web_select_streaming.test.ts | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/packages/client-web/__tests__/integration/web_select_streaming.test.ts b/packages/client-web/__tests__/integration/web_select_streaming.test.ts index 1a82642..66cf293 100644 --- a/packages/client-web/__tests__/integration/web_select_streaming.test.ts +++ b/packages/client-web/__tests__/integration/web_select_streaming.test.ts @@ -2,7 +2,7 @@ import type { ClickHouseClient, Row } from '@clickhouse/client-common' import { createTestClient } from '@test/utils' import { genLargeStringsDataset } from '@test/utils/datasets' -fdescribe('[Web] SELECT streaming', () => { +describe('[Web] SELECT streaming', () => { let client: ClickHouseClient> afterEach(async () => { await client.close() @@ -211,7 +211,7 @@ fdescribe('[Web] SELECT streaming', () => { // See https://github.com/ClickHouse/clickhouse-js/issues/171 for more details // Here we generate a large enough dataset to break into multiple chunks while streaming, // effectively testing the implementation of incomplete rows handling - fdescribe('should correctly process multiple chunks', () => { + describe('should correctly process multiple chunks', () => { describe('large amount of rows', () => { it('should work with .json()', async () => { const { table, values } = await genLargeStringsDataset(client, {