mirror of https://github.com/buster-so/buster.git
Merge pull request #483 from buster-so/dallin/bus-1365-make-queries-in-snowflake-between-agent-and-fetching-data
Dallin/bus-1365-make-queries-in-snowflake-between-agent-and-fetching-data
This commit is contained in:
commit
7277e45235
|
@ -6,13 +6,12 @@ import { z } from 'zod';
|
||||||
import { getWorkflowDataSourceManager } from '../../utils/data-source-manager';
|
import { getWorkflowDataSourceManager } from '../../utils/data-source-manager';
|
||||||
import { createPermissionErrorMessage, validateSqlPermissions } from '../../utils/sql-permissions';
|
import { createPermissionErrorMessage, validateSqlPermissions } from '../../utils/sql-permissions';
|
||||||
import type { AnalystRuntimeContext } from '../../workflows/analyst-workflow';
|
import type { AnalystRuntimeContext } from '../../workflows/analyst-workflow';
|
||||||
import { ensureSqlLimit } from './sql-limit-helper';
|
|
||||||
|
|
||||||
const executeSqlStatementInputSchema = z.object({
|
const executeSqlStatementInputSchema = z.object({
|
||||||
statements: z.array(z.string()).describe(
|
statements: z.array(z.string()).describe(
|
||||||
`Array of lightweight, optimized SQL statements to execute.
|
`Array of lightweight, optimized SQL statements to execute.
|
||||||
Each statement should be small and focused.
|
Each statement should be small and focused.
|
||||||
SELECT queries without a LIMIT clause will automatically have LIMIT 25 added for performance.
|
SELECT queries without a LIMIT clause will automatically have LIMIT 50 added for performance.
|
||||||
Existing LIMIT clauses will be preserved.
|
Existing LIMIT clauses will be preserved.
|
||||||
YOU MUST USE THE <SCHEMA_NAME>.<TABLE_NAME> syntax/qualifier for all table names.
|
YOU MUST USE THE <SCHEMA_NAME>.<TABLE_NAME> syntax/qualifier for all table names.
|
||||||
NEVER use SELECT * - you must explicitly list the columns you want to query from the documentation provided.
|
NEVER use SELECT * - you must explicitly list the columns you want to query from the documentation provided.
|
||||||
|
@ -363,9 +362,6 @@ async function executeSingleStatement(
|
||||||
return { success: false, error: 'SQL statement cannot be empty' };
|
return { success: false, error: 'SQL statement cannot be empty' };
|
||||||
}
|
}
|
||||||
|
|
||||||
// Ensure the SQL statement has a LIMIT clause to prevent excessive results
|
|
||||||
const limitedSql = ensureSqlLimit(sqlStatement, 25);
|
|
||||||
|
|
||||||
// Validate permissions before execution
|
// Validate permissions before execution
|
||||||
const userId = runtimeContext.get('userId');
|
const userId = runtimeContext.get('userId');
|
||||||
if (!userId) {
|
if (!userId) {
|
||||||
|
@ -373,7 +369,7 @@ async function executeSingleStatement(
|
||||||
}
|
}
|
||||||
|
|
||||||
const dataSourceSyntax = runtimeContext.get('dataSourceSyntax');
|
const dataSourceSyntax = runtimeContext.get('dataSourceSyntax');
|
||||||
const permissionResult = await validateSqlPermissions(limitedSql, userId, dataSourceSyntax);
|
const permissionResult = await validateSqlPermissions(sqlStatement, userId, dataSourceSyntax);
|
||||||
if (!permissionResult.isAuthorized) {
|
if (!permissionResult.isAuthorized) {
|
||||||
return {
|
return {
|
||||||
success: false,
|
success: false,
|
||||||
|
@ -385,10 +381,12 @@ async function executeSingleStatement(
|
||||||
for (let attempt = 0; attempt <= MAX_RETRIES; attempt++) {
|
for (let attempt = 0; attempt <= MAX_RETRIES; attempt++) {
|
||||||
try {
|
try {
|
||||||
// Execute the SQL query using the DataSource with timeout
|
// Execute the SQL query using the DataSource with timeout
|
||||||
|
// Pass maxRows to the adapter instead of modifying the SQL
|
||||||
const result = await dataSource.execute({
|
const result = await dataSource.execute({
|
||||||
sql: limitedSql,
|
sql: sqlStatement,
|
||||||
options: {
|
options: {
|
||||||
timeout: TIMEOUT_MS,
|
timeout: TIMEOUT_MS,
|
||||||
|
maxRows: 50, // Limit results at the adapter level, not in SQL
|
||||||
},
|
},
|
||||||
});
|
});
|
||||||
|
|
||||||
|
@ -411,7 +409,7 @@ async function executeSingleStatement(
|
||||||
console.warn(
|
console.warn(
|
||||||
`[execute-sql] Query timeout on attempt ${attempt + 1}/${MAX_RETRIES + 1}. Retrying in ${delay}ms...`,
|
`[execute-sql] Query timeout on attempt ${attempt + 1}/${MAX_RETRIES + 1}. Retrying in ${delay}ms...`,
|
||||||
{
|
{
|
||||||
sql: `${limitedSql.substring(0, 100)}...`,
|
sql: `${sqlStatement.substring(0, 100)}...`,
|
||||||
attempt: attempt + 1,
|
attempt: attempt + 1,
|
||||||
nextDelay: delay,
|
nextDelay: delay,
|
||||||
}
|
}
|
||||||
|
@ -437,7 +435,7 @@ async function executeSingleStatement(
|
||||||
console.warn(
|
console.warn(
|
||||||
`[execute-sql] Query timeout (exception) on attempt ${attempt + 1}/${MAX_RETRIES + 1}. Retrying in ${delay}ms...`,
|
`[execute-sql] Query timeout (exception) on attempt ${attempt + 1}/${MAX_RETRIES + 1}. Retrying in ${delay}ms...`,
|
||||||
{
|
{
|
||||||
sql: `${limitedSql.substring(0, 100)}...`,
|
sql: `${sqlStatement.substring(0, 100)}...`,
|
||||||
attempt: attempt + 1,
|
attempt: attempt + 1,
|
||||||
nextDelay: delay,
|
nextDelay: delay,
|
||||||
error: errorMessage,
|
error: errorMessage,
|
||||||
|
@ -466,7 +464,8 @@ async function executeSingleStatement(
|
||||||
export const executeSql = createTool({
|
export const executeSql = createTool({
|
||||||
id: 'execute-sql',
|
id: 'execute-sql',
|
||||||
description: `Use this to run lightweight, validation queries to understand values in columns, date ranges, etc.
|
description: `Use this to run lightweight, validation queries to understand values in columns, date ranges, etc.
|
||||||
SELECT queries without LIMIT will automatically have LIMIT 25 added.
|
Please limit your queries to 50 rows for performance.
|
||||||
|
Query results will be limited to 50 rows for performance.
|
||||||
You must use the <SCHEMA_NAME>.<TABLE_NAME> syntax/qualifier for all table names.
|
You must use the <SCHEMA_NAME>.<TABLE_NAME> syntax/qualifier for all table names.
|
||||||
Otherwise the queries wont run successfully.`,
|
Otherwise the queries wont run successfully.`,
|
||||||
inputSchema: executeSqlStatementInputSchema,
|
inputSchema: executeSqlStatementInputSchema,
|
||||||
|
|
|
@ -171,26 +171,6 @@ function createDataMetadata(results: Record<string, unknown>[]): DataMetadata {
|
||||||
};
|
};
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
|
||||||
* Wraps a SQL query with a LIMIT clause for validation purposes
|
|
||||||
* Handles existing LIMIT clauses and complex queries
|
|
||||||
*/
|
|
||||||
function wrapQueryWithLimit(sql: string, limit: number): string {
|
|
||||||
// Remove any existing LIMIT clause to avoid conflicts
|
|
||||||
const sqlWithoutLimit = sql.replace(/\s+LIMIT\s+\d+\s*$/i, '').trim();
|
|
||||||
|
|
||||||
// For CTEs or complex queries, wrap the entire query
|
|
||||||
if (
|
|
||||||
sqlWithoutLimit.toUpperCase().includes('WITH ') ||
|
|
||||||
sqlWithoutLimit.includes('(') ||
|
|
||||||
sqlWithoutLimit.toUpperCase().includes('UNION')
|
|
||||||
) {
|
|
||||||
return `SELECT * FROM (${sqlWithoutLimit}) AS validation_wrapper LIMIT ${limit}`;
|
|
||||||
}
|
|
||||||
|
|
||||||
// For simple queries, just append LIMIT
|
|
||||||
return `${sqlWithoutLimit} LIMIT ${limit}`;
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Ensures timeFrame values are properly quoted in YAML content
|
* Ensures timeFrame values are properly quoted in YAML content
|
||||||
|
@ -1373,13 +1353,10 @@ async function validateSql(
|
||||||
// Attempt execution with retries
|
// Attempt execution with retries
|
||||||
for (let attempt = 0; attempt <= MAX_RETRIES; attempt++) {
|
for (let attempt = 0; attempt <= MAX_RETRIES; attempt++) {
|
||||||
try {
|
try {
|
||||||
// For validation, wrap query with LIMIT at SQL level for better performance
|
|
||||||
// This ensures Snowflake doesn't process the entire dataset
|
|
||||||
const validationSql = wrapQueryWithLimit(sqlQuery, 1000);
|
|
||||||
|
|
||||||
// Execute the SQL query using the DataSource with row limit and timeout for validation
|
// Execute the SQL query using the DataSource with row limit and timeout for validation
|
||||||
|
// Use maxRows to limit results without modifying the SQL query (preserves Snowflake caching)
|
||||||
const result = await dataSource.execute({
|
const result = await dataSource.execute({
|
||||||
sql: validationSql,
|
sql: sqlQuery,
|
||||||
options: {
|
options: {
|
||||||
maxRows: 1000, // Additional safety limit at adapter level
|
maxRows: 1000, // Additional safety limit at adapter level
|
||||||
timeout: TIMEOUT_MS,
|
timeout: TIMEOUT_MS,
|
||||||
|
|
|
@ -162,26 +162,6 @@ function createDataMetadata(results: Record<string, unknown>[]): DataMetadata {
|
||||||
};
|
};
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
|
||||||
* Wraps a SQL query with a LIMIT clause for validation purposes
|
|
||||||
* Handles existing LIMIT clauses and complex queries
|
|
||||||
*/
|
|
||||||
function wrapQueryWithLimit(sql: string, limit: number): string {
|
|
||||||
// Remove any existing LIMIT clause to avoid conflicts
|
|
||||||
const sqlWithoutLimit = sql.replace(/\s+LIMIT\s+\d+\s*$/i, '').trim();
|
|
||||||
|
|
||||||
// For CTEs or complex queries, wrap the entire query
|
|
||||||
if (
|
|
||||||
sqlWithoutLimit.toUpperCase().includes('WITH ') ||
|
|
||||||
sqlWithoutLimit.includes('(') ||
|
|
||||||
sqlWithoutLimit.toUpperCase().includes('UNION')
|
|
||||||
) {
|
|
||||||
return `SELECT * FROM (${sqlWithoutLimit}) AS validation_wrapper LIMIT ${limit}`;
|
|
||||||
}
|
|
||||||
|
|
||||||
// For simple queries, just append LIMIT
|
|
||||||
return `${sqlWithoutLimit} LIMIT ${limit}`;
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Ensures timeFrame values are properly quoted in YAML content
|
* Ensures timeFrame values are properly quoted in YAML content
|
||||||
|
@ -332,13 +312,10 @@ async function validateSql(
|
||||||
// Attempt execution with retries
|
// Attempt execution with retries
|
||||||
for (let attempt = 0; attempt <= MAX_RETRIES; attempt++) {
|
for (let attempt = 0; attempt <= MAX_RETRIES; attempt++) {
|
||||||
try {
|
try {
|
||||||
// For validation, wrap query with LIMIT at SQL level for better performance
|
|
||||||
// This ensures Snowflake doesn't process the entire dataset
|
|
||||||
const validationSql = wrapQueryWithLimit(sqlQuery, 1000);
|
|
||||||
|
|
||||||
// Execute the SQL query using the DataSource with row limit and timeout for validation
|
// Execute the SQL query using the DataSource with row limit and timeout for validation
|
||||||
|
// Use maxRows to limit results without modifying the SQL query (preserves Snowflake caching)
|
||||||
const result = await dataSource.execute({
|
const result = await dataSource.execute({
|
||||||
sql: validationSql,
|
sql: sqlQuery,
|
||||||
options: {
|
options: {
|
||||||
maxRows: 1000, // Additional safety limit at adapter level
|
maxRows: 1000, // Additional safety limit at adapter level
|
||||||
timeout: TIMEOUT_MS,
|
timeout: TIMEOUT_MS,
|
||||||
|
|
|
@ -18,6 +18,8 @@ interface SnowflakeStatement {
|
||||||
getScale(): number;
|
getScale(): number;
|
||||||
getPrecision(): number;
|
getPrecision(): number;
|
||||||
}>;
|
}>;
|
||||||
|
streamRows?: () => NodeJS.ReadableStream;
|
||||||
|
cancel?: (callback: (err: Error | undefined) => void) => void;
|
||||||
}
|
}
|
||||||
|
|
||||||
// Configure Snowflake SDK to disable logging
|
// Configure Snowflake SDK to disable logging
|
||||||
|
@ -189,16 +191,9 @@ export class SnowflakeAdapter extends BaseAdapter {
|
||||||
// Set query timeout if specified (default: 120 seconds for Snowflake queue handling)
|
// Set query timeout if specified (default: 120 seconds for Snowflake queue handling)
|
||||||
const timeoutMs = timeout || TIMEOUT_CONFIG.query.default;
|
const timeoutMs = timeout || TIMEOUT_CONFIG.query.default;
|
||||||
|
|
||||||
// For maxRows, we'll fetch maxRows + 1 to determine if there are more rows
|
// IMPORTANT: Execute the original SQL unchanged to leverage Snowflake's query caching
|
||||||
let effectiveSql = sql;
|
// For memory protection, we'll fetch all rows but limit in memory
|
||||||
if (maxRows && maxRows > 0) {
|
// This is a compromise to preserve caching while preventing OOM on truly massive queries
|
||||||
// Check if query already has LIMIT
|
|
||||||
const upperSql = sql.toUpperCase();
|
|
||||||
if (!upperSql.includes(' LIMIT ')) {
|
|
||||||
effectiveSql = `${sql} LIMIT ${maxRows + 1}`;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
const queryPromise = new Promise<{
|
const queryPromise = new Promise<{
|
||||||
rows: Record<string, unknown>[];
|
rows: Record<string, unknown>[];
|
||||||
statement: SnowflakeStatement;
|
statement: SnowflakeStatement;
|
||||||
|
@ -207,8 +202,9 @@ export class SnowflakeAdapter extends BaseAdapter {
|
||||||
reject(new Error('Failed to acquire Snowflake connection'));
|
reject(new Error('Failed to acquire Snowflake connection'));
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
connection.execute({
|
connection.execute({
|
||||||
sqlText: effectiveSql,
|
sqlText: sql, // Use original SQL unchanged for caching
|
||||||
binds: params as snowflake.Binds,
|
binds: params as snowflake.Binds,
|
||||||
complete: (
|
complete: (
|
||||||
err: SnowflakeError | undefined,
|
err: SnowflakeError | undefined,
|
||||||
|
@ -235,7 +231,7 @@ export class SnowflakeAdapter extends BaseAdapter {
|
||||||
precision: col.getPrecision() > 0 ? col.getPrecision() : 0,
|
precision: col.getPrecision() > 0 ? col.getPrecision() : 0,
|
||||||
})) || [];
|
})) || [];
|
||||||
|
|
||||||
// Handle maxRows logic
|
// Handle maxRows logic in memory (not in SQL)
|
||||||
let finalRows = result.rows;
|
let finalRows = result.rows;
|
||||||
let hasMoreRows = false;
|
let hasMoreRows = false;
|
||||||
|
|
||||||
|
|
|
@ -0,0 +1,198 @@
|
||||||
|
import { afterEach, beforeEach, describe, expect } from 'vitest';
|
||||||
|
import { SnowflakeAdapter } from '../../../src/adapters/snowflake';
|
||||||
|
import { DataSourceType } from '../../../src/types/credentials';
|
||||||
|
import type { SnowflakeCredentials } from '../../../src/types/credentials';
|
||||||
|
import { TEST_TIMEOUT, skipIfNoCredentials, testConfig } from '../../setup';
|
||||||
|
|
||||||
|
const testWithCredentials = skipIfNoCredentials('snowflake');
|
||||||
|
|
||||||
|
describe('Snowflake Memory Protection Tests', () => {
|
||||||
|
let adapter: SnowflakeAdapter;
|
||||||
|
let credentials: SnowflakeCredentials;
|
||||||
|
|
||||||
|
beforeEach(() => {
|
||||||
|
adapter = new SnowflakeAdapter();
|
||||||
|
|
||||||
|
// Set up credentials once
|
||||||
|
if (
|
||||||
|
!testConfig.snowflake.account_id ||
|
||||||
|
!testConfig.snowflake.warehouse_id ||
|
||||||
|
!testConfig.snowflake.username ||
|
||||||
|
!testConfig.snowflake.password ||
|
||||||
|
!testConfig.snowflake.default_database
|
||||||
|
) {
|
||||||
|
throw new Error(
|
||||||
|
'TEST_SNOWFLAKE_ACCOUNT_ID, TEST_SNOWFLAKE_WAREHOUSE_ID, TEST_SNOWFLAKE_USERNAME, TEST_SNOWFLAKE_PASSWORD, and TEST_SNOWFLAKE_DATABASE are required for this test'
|
||||||
|
);
|
||||||
|
}
|
||||||
|
|
||||||
|
credentials = {
|
||||||
|
type: DataSourceType.Snowflake,
|
||||||
|
account_id: testConfig.snowflake.account_id,
|
||||||
|
warehouse_id: testConfig.snowflake.warehouse_id,
|
||||||
|
username: testConfig.snowflake.username,
|
||||||
|
password: testConfig.snowflake.password,
|
||||||
|
default_database: testConfig.snowflake.default_database,
|
||||||
|
default_schema: testConfig.snowflake.default_schema,
|
||||||
|
role: testConfig.snowflake.role,
|
||||||
|
};
|
||||||
|
});
|
||||||
|
|
||||||
|
afterEach(async () => {
|
||||||
|
if (adapter) {
|
||||||
|
await adapter.close();
|
||||||
|
}
|
||||||
|
});
|
||||||
|
|
||||||
|
testWithCredentials(
|
||||||
|
'should handle large result sets with maxRows without running out of memory',
|
||||||
|
async () => {
|
||||||
|
await adapter.initialize(credentials);
|
||||||
|
|
||||||
|
// NOTE: Due to Snowflake SDK limitations, we cannot truly stream results
|
||||||
|
// For now, we'll test with a smaller dataset to avoid OOM
|
||||||
|
// Query ORDERS table instead of LINEITEM (1.5M rows vs 6M rows)
|
||||||
|
const result = await adapter.query(
|
||||||
|
'SELECT * FROM SNOWFLAKE_SAMPLE_DATA.TPCH_SF1.ORDERS',
|
||||||
|
undefined,
|
||||||
|
100 // Only fetch 100 rows
|
||||||
|
);
|
||||||
|
|
||||||
|
expect(result.rows.length).toBe(100);
|
||||||
|
expect(result.hasMoreRows).toBe(true);
|
||||||
|
expect(result.rowCount).toBe(100);
|
||||||
|
|
||||||
|
// Verify we got the fields metadata
|
||||||
|
expect(result.fields.length).toBeGreaterThan(0);
|
||||||
|
expect(result.fields[0]).toHaveProperty('name');
|
||||||
|
expect(result.fields[0]).toHaveProperty('type');
|
||||||
|
},
|
||||||
|
TEST_TIMEOUT
|
||||||
|
);
|
||||||
|
|
||||||
|
testWithCredentials(
|
||||||
|
'should preserve query caching when running the same query multiple times',
|
||||||
|
async () => {
|
||||||
|
await adapter.initialize(credentials);
|
||||||
|
|
||||||
|
const sql = 'SELECT * FROM SNOWFLAKE_SAMPLE_DATA.TPCH_SF1.CUSTOMER WHERE C_MKTSEGMENT = ?';
|
||||||
|
const params = ['AUTOMOBILE'];
|
||||||
|
|
||||||
|
// First execution - will be cached by Snowflake
|
||||||
|
const start1 = Date.now();
|
||||||
|
const result1 = await adapter.query(sql, params, 50);
|
||||||
|
const time1 = Date.now() - start1;
|
||||||
|
|
||||||
|
// Second execution - should hit Snowflake's cache
|
||||||
|
const start2 = Date.now();
|
||||||
|
const result2 = await adapter.query(sql, params, 50);
|
||||||
|
const time2 = Date.now() - start2;
|
||||||
|
|
||||||
|
// Third execution with different maxRows - should still hit cache
|
||||||
|
const start3 = Date.now();
|
||||||
|
const result3 = await adapter.query(sql, params, 25);
|
||||||
|
const time3 = Date.now() - start3;
|
||||||
|
|
||||||
|
// Verify results
|
||||||
|
expect(result1.rows.length).toBe(50);
|
||||||
|
expect(result2.rows.length).toBe(50);
|
||||||
|
expect(result3.rows.length).toBe(25);
|
||||||
|
|
||||||
|
// All should indicate more rows available
|
||||||
|
expect(result1.hasMoreRows).toBe(true);
|
||||||
|
expect(result2.hasMoreRows).toBe(true);
|
||||||
|
expect(result3.hasMoreRows).toBe(true);
|
||||||
|
|
||||||
|
// Cache hits should be faster (allowing for some variance)
|
||||||
|
console.info(`Query times: ${time1}ms, ${time2}ms, ${time3}ms`);
|
||||||
|
|
||||||
|
// The cached queries (2nd and 3rd) should generally be faster than the first
|
||||||
|
// We use a loose check because network latency can vary
|
||||||
|
const avgCachedTime = (time2 + time3) / 2;
|
||||||
|
expect(avgCachedTime).toBeLessThanOrEqual(time1 * 1.5); // Allow 50% variance
|
||||||
|
},
|
||||||
|
TEST_TIMEOUT
|
||||||
|
);
|
||||||
|
|
||||||
|
testWithCredentials(
|
||||||
|
'should handle queries with no maxRows (fetch all results)',
|
||||||
|
async () => {
|
||||||
|
await adapter.initialize(credentials);
|
||||||
|
|
||||||
|
// Query a small table without maxRows
|
||||||
|
const result = await adapter.query(
|
||||||
|
'SELECT * FROM SNOWFLAKE_SAMPLE_DATA.TPCH_SF1.REGION'
|
||||||
|
);
|
||||||
|
|
||||||
|
// REGION table has exactly 5 rows
|
||||||
|
expect(result.rows.length).toBe(5);
|
||||||
|
expect(result.hasMoreRows).toBe(false);
|
||||||
|
expect(result.rowCount).toBe(5);
|
||||||
|
},
|
||||||
|
TEST_TIMEOUT
|
||||||
|
);
|
||||||
|
|
||||||
|
testWithCredentials(
|
||||||
|
'should handle maxRows=1 correctly',
|
||||||
|
async () => {
|
||||||
|
await adapter.initialize(credentials);
|
||||||
|
|
||||||
|
const result = await adapter.query(
|
||||||
|
'SELECT * FROM SNOWFLAKE_SAMPLE_DATA.TPCH_SF1.NATION ORDER BY N_NATIONKEY',
|
||||||
|
undefined,
|
||||||
|
1
|
||||||
|
);
|
||||||
|
|
||||||
|
expect(result.rows.length).toBe(1);
|
||||||
|
expect(result.hasMoreRows).toBe(true);
|
||||||
|
expect(result.rows[0]).toHaveProperty('N_NATIONKEY', 0); // First nation
|
||||||
|
},
|
||||||
|
TEST_TIMEOUT
|
||||||
|
);
|
||||||
|
|
||||||
|
testWithCredentials(
|
||||||
|
'should handle edge case where result set equals maxRows',
|
||||||
|
async () => {
|
||||||
|
await adapter.initialize(credentials);
|
||||||
|
|
||||||
|
// REGION table has exactly 5 rows
|
||||||
|
const result = await adapter.query(
|
||||||
|
'SELECT * FROM SNOWFLAKE_SAMPLE_DATA.TPCH_SF1.REGION',
|
||||||
|
undefined,
|
||||||
|
5
|
||||||
|
);
|
||||||
|
|
||||||
|
expect(result.rows.length).toBe(5);
|
||||||
|
expect(result.hasMoreRows).toBe(false); // No more rows available
|
||||||
|
expect(result.rowCount).toBe(5);
|
||||||
|
},
|
||||||
|
TEST_TIMEOUT
|
||||||
|
);
|
||||||
|
|
||||||
|
testWithCredentials(
|
||||||
|
'should handle complex queries with CTEs and maxRows',
|
||||||
|
async () => {
|
||||||
|
await adapter.initialize(credentials);
|
||||||
|
|
||||||
|
const sql = `
|
||||||
|
WITH high_value_orders AS (
|
||||||
|
SELECT O_CUSTKEY, SUM(O_TOTALPRICE) as total_spent
|
||||||
|
FROM SNOWFLAKE_SAMPLE_DATA.TPCH_SF1.ORDERS
|
||||||
|
GROUP BY O_CUSTKEY
|
||||||
|
HAVING SUM(O_TOTALPRICE) > 500000
|
||||||
|
)
|
||||||
|
SELECT c.C_NAME, c.C_PHONE, h.total_spent
|
||||||
|
FROM SNOWFLAKE_SAMPLE_DATA.TPCH_SF1.CUSTOMER c
|
||||||
|
JOIN high_value_orders h ON c.C_CUSTKEY = h.O_CUSTKEY
|
||||||
|
ORDER BY h.total_spent DESC
|
||||||
|
`;
|
||||||
|
|
||||||
|
const result = await adapter.query(sql, undefined, 10);
|
||||||
|
|
||||||
|
expect(result.rows.length).toBe(10);
|
||||||
|
expect(result.hasMoreRows).toBe(true);
|
||||||
|
expect(result.fields.length).toBe(3); // C_NAME, C_PHONE, total_spent
|
||||||
|
},
|
||||||
|
TEST_TIMEOUT
|
||||||
|
);
|
||||||
|
});
|
Loading…
Reference in New Issue