- Added logging for server startup with port number in main.rs
- Removed hardcoded GITHUB_WEBHOOK_SECRET from github-webhook-validator tests
- Updated slack-notifier tests to remove BUSTER_URL environment variable
- Refactored FirecrawlService tests to use async creation method and removed API key from environment
- Add mock for @buster/secrets module
- Mock getSecret function to return test values
- Remove outdated BRAINTRUST_KEY env stubbing
- Fixes CI/CD failures due to missing Infisical secrets
Resolved conflicts by:
- Using storage provider approach from staging in cleanup-export-file.ts
- Keeping Opus fallback model from our env var work in sonnet-4.ts
- Merging dependencies in data-source package.json (kept @buster/secrets, added new deps from staging)
- Regenerated pnpm-lock.yaml after resolving package.json conflicts
- Added bucketName to S3 integration responses and handlers to support bucket-specific operations.
- Updated tests to reflect the inclusion of bucketName in integration creation and retrieval processes.
- Improved error handling for fetching bucket names from the vault during S3 integration retrieval.
These changes improve the S3 integration management by allowing for more granular control and visibility of storage buckets.
- Accept deletion of env.d.ts and validate-env.ts (replaced by centralized secrets)
- Remove SLACK_INTEGRATION_ENABLED (now always enabled per staging changes)
- Update Slack OAuth service to use @buster/secrets instead of process.env
- Add BUSTER_URL to server keys
- Fix Snowflake test timeout issues
- Simplified S3 integration handler functions by consolidating imports and improving code readability.
- Updated error handling in the S3 integration process to enhance clarity and maintainability.
- Refactored storage provider functions to utilize a more modular approach, separating concerns for better organization.
- Introduced utility functions for common operations, improving code reuse and reducing duplication.
These changes enhance the overall structure and maintainability of the S3 integration management features.
- Added routes for creating, retrieving, and deleting S3 integrations in the API.
- Introduced handlers for S3 integration operations, including validation of user permissions and storage credentials.
- Updated database schema to support S3 integrations, including a new table and associated queries.
- Integrated storage provider logic to handle S3, R2, and GCS configurations.
- Enhanced error handling and response structures for integration operations.
This commit lays the groundwork for managing storage integrations within the application, allowing users to connect and manage their S3 storage solutions.
- Modified the SSL configuration in both the PostgreSQL adapter and its tests to use { rejectUnauthorized: false } instead of a boolean true value.
- Ensured consistency in handling SSL settings across the adapter and its tests.
- Update all LLM model files to use getSecretSync with AI_KEYS constants
- Migrate OpenAI GPT models (gpt-5, gpt-5-mini, gpt-5-nano) to use secrets system
- Migrate Anthropic models (sonnet-4, haiku-3-5) to use secrets for both Anthropic API and Vertex AI credentials
- Update provider files (openai.ts, anthropic.ts, vertex.ts) to use getSecretSync
- Migrate evaluation workflow to use BRAINTRUST_KEY from secrets
- Implement proper error handling for missing credentials with try-catch blocks
- Test files will be migrated in a separate update
All non-test source files now use centralized secrets management instead of direct process.env access.
- Updated all database adapter tests and implementations to replace the 'database' field with 'default_database' for consistency.
- Ensured backward compatibility in the Redshift adapter by allowing both 'database' and 'default_database' fields.
- Enhanced SQLServer and MySQL adapters to reflect the new credential structure, improving clarity and maintainability.
- Added checks to ensure toolCallId exists and is a valid string before processing tool results.
- Preserves the original part if toolCallId is missing or invalid, improving robustness in message conversion.
- Updated logic to prioritize cache as the source of truth during streaming, with asynchronous database updates for persistence.
- Improved error handling for background database updates, ensuring cache validity even if DB update fails.
- Cleaned up mergeResponseMessages and mergeReasoningMessages functions by removing redundant code.
- Introduced lru-cache version 11.1.0 in package.json and pnpm-lock.yaml.
- Refactored updateMessageEntries to utilize TypeScript-based merge logic with write-through caching.
- Improved error handling for fetching existing message entries and updated database logic to merge new entries with existing ones.
- Removed the creation and handling of raw LLM message entries in both start and finish functions.
- Updated message entry updates to only include reasoning messages, simplifying the logic and improving clarity.
- Added imports for various dashboard, metrics, and report tools to the analyst workflow.
- Updated the `runAnalystWorkflow` function to return a comprehensive output, including tool call results, charts created, and summary statistics.
- Implemented logic to extract tool calls and segment messages by user requests, improving the overall workflow output structure.