diff --git a/packages/core/src/core/__tests__/openaiTimeoutHandling.test.ts b/packages/core/src/core/__tests__/openaiTimeoutHandling.test.ts index 911f6df9..702c1326 100644 --- a/packages/core/src/core/__tests__/openaiTimeoutHandling.test.ts +++ b/packages/core/src/core/__tests__/openaiTimeoutHandling.test.ts @@ -146,7 +146,9 @@ describe('OpenAIContentGenerator Timeout Handling', () => { model: 'gpt-4', }; - await expect(generator.generateContent(request, 'test-prompt-id')).rejects.toThrow( + await expect( + generator.generateContent(request, 'test-prompt-id'), + ).rejects.toThrow( /Request timeout after \d+s\. Try reducing input length or increasing timeout in config\./, ); }); @@ -161,9 +163,9 @@ describe('OpenAIContentGenerator Timeout Handling', () => { model: 'gpt-4', }; - await expect(generator.generateContent(request, 'test-prompt-id')).rejects.toThrow( - 'Invalid API key', - ); + await expect( + generator.generateContent(request, 'test-prompt-id'), + ).rejects.toThrow('Invalid API key'); }); it('should include troubleshooting tips for timeout errors', async () => { @@ -199,7 +201,9 @@ describe('OpenAIContentGenerator Timeout Handling', () => { model: 'gpt-4', }; - await expect(generator.generateContentStream(request, 'test-prompt-id')).rejects.toThrow( + await expect( + generator.generateContentStream(request, 'test-prompt-id'), + ).rejects.toThrow( /Streaming setup timeout after \d+s\. Try reducing input length or increasing timeout in config\./, ); }); @@ -324,9 +328,9 @@ describe('OpenAIContentGenerator Timeout Handling', () => { }; // Should not throw due to token counting failure - await expect(generator.generateContent(request, 'test-prompt-id')).rejects.toThrow( - /Request timeout after \d+s/, - ); + await expect( + generator.generateContent(request, 'test-prompt-id'), + ).rejects.toThrow(/Request timeout after \d+s/); }); }); }); diff --git a/packages/core/src/core/openaiContentGenerator.test.ts b/packages/core/src/core/openaiContentGenerator.test.ts index abc73ffd..6dc0690a 100644 --- a/packages/core/src/core/openaiContentGenerator.test.ts +++ b/packages/core/src/core/openaiContentGenerator.test.ts @@ -439,7 +439,10 @@ describe('OpenAIContentGenerator', () => { model: 'gpt-4', }; - const stream = await generator.generateContentStream(request, 'test-prompt-id'); + const stream = await generator.generateContentStream( + request, + 'test-prompt-id', + ); const responses = []; for await (const response of stream) { responses.push(response); @@ -528,7 +531,10 @@ describe('OpenAIContentGenerator', () => { model: 'gpt-4', }; - const stream = await generator.generateContentStream(request, 'test-prompt-id'); + const stream = await generator.generateContentStream( + request, + 'test-prompt-id', + ); const responses = []; for await (const response of stream) { responses.push(response); @@ -668,9 +674,9 @@ describe('OpenAIContentGenerator', () => { model: 'gpt-4', }; - await expect(generator.generateContent(request, 'test-prompt-id')).rejects.toThrow( - 'Invalid API key', - ); + await expect( + generator.generateContent(request, 'test-prompt-id'), + ).rejects.toThrow('Invalid API key'); }); it('should estimate tokens on error for telemetry', async () => { @@ -872,7 +878,10 @@ describe('OpenAIContentGenerator', () => { model: 'gpt-4', }; - const result = await generator.generateContent(request, 'test-prompt-id'); + const result = await generator.generateContent( + request, + 'test-prompt-id', + ); if ( result.candidates && result.candidates.length > 0 && @@ -969,7 +978,9 @@ describe('OpenAIContentGenerator', () => { model: 'gpt-4', }; - await expect(generator.generateContent(request, 'test-prompt-id')).rejects.toThrow( + await expect( + generator.generateContent(request, 'test-prompt-id'), + ).rejects.toThrow( /Troubleshooting tips.*Reduce input length.*Increase timeout.*Check network/s, ); }); @@ -985,9 +996,9 @@ describe('OpenAIContentGenerator', () => { model: 'gpt-4', }; - await expect(generator.generateContent(request, 'test-prompt-id')).rejects.toThrow( - 'Streaming setup failed', - ); + await expect( + generator.generateContent(request, 'test-prompt-id'), + ).rejects.toThrow('Streaming setup failed'); }); it('should handle timeout errors during streaming setup', async () => { @@ -999,7 +1010,9 @@ describe('OpenAIContentGenerator', () => { model: 'gpt-4', }; - await expect(generator.generateContentStream(request, 'test-prompt-id')).rejects.toThrow( + await expect( + generator.generateContentStream(request, 'test-prompt-id'), + ).rejects.toThrow( /Streaming setup timeout troubleshooting.*Reduce input length/s, ); }); @@ -1042,7 +1055,10 @@ describe('OpenAIContentGenerator', () => { model: 'gpt-4', }; - const stream = await loggingGenerator.generateContentStream(request, 'test-prompt-id'); + const stream = await loggingGenerator.generateContentStream( + request, + 'test-prompt-id', + ); // Consume the stream and expect error await expect(async () => { @@ -1495,7 +1511,9 @@ describe('OpenAIContentGenerator', () => { model: 'gpt-4', }; - await expect(testGenerator.generateContent(request, 'test-prompt-id')).rejects.toThrow(); + await expect( + testGenerator.generateContent(request, 'test-prompt-id'), + ).rejects.toThrow(); // Error logging should be suppressed expect(consoleSpy).not.toHaveBeenCalledWith( @@ -1519,7 +1537,9 @@ describe('OpenAIContentGenerator', () => { model: 'gpt-4', }; - await expect(generator.generateContent(request, 'test-prompt-id')).rejects.toThrow(); + await expect( + generator.generateContent(request, 'test-prompt-id'), + ).rejects.toThrow(); // Error logging should occur by default expect(consoleSpy).toHaveBeenCalledWith( @@ -1643,7 +1663,10 @@ describe('OpenAIContentGenerator', () => { model: 'gpt-4', }; - const stream = await generator.generateContentStream(request, 'test-prompt-id'); + const stream = await generator.generateContentStream( + request, + 'test-prompt-id', + ); const responses = []; for await (const response of stream) { responses.push(response); @@ -2039,7 +2062,10 @@ describe('OpenAIContentGenerator', () => { model: 'gpt-4', }; - const stream = await loggingGenerator.generateContentStream(request, 'test-prompt-id'); + const stream = await loggingGenerator.generateContentStream( + request, + 'test-prompt-id', + ); const responses = []; for await (const response of stream) { responses.push(response); @@ -2083,7 +2109,10 @@ describe('OpenAIContentGenerator', () => { model: 'gpt-4', }; - const stream = await generator.generateContentStream(request, 'test-prompt-id'); + const stream = await generator.generateContentStream( + request, + 'test-prompt-id', + ); const responses = []; for await (const response of stream) { responses.push(response); diff --git a/packages/core/src/core/openaiContentGenerator.ts b/packages/core/src/core/openaiContentGenerator.ts index b0b6c88b..4e9e73da 100644 --- a/packages/core/src/core/openaiContentGenerator.ts +++ b/packages/core/src/core/openaiContentGenerator.ts @@ -326,7 +326,7 @@ export class OpenAIContentGenerator implements ContentGenerator { ): Promise> { const startTime = Date.now(); const messages = this.convertToOpenAIFormat(request); - + try { // Build sampling parameters with clear priority const samplingParams = this.buildSamplingParameters(request); diff --git a/packages/core/src/qwen/qwenContentGenerator.test.ts b/packages/core/src/qwen/qwenContentGenerator.test.ts index f3709760..d2878f93 100644 --- a/packages/core/src/qwen/qwenContentGenerator.test.ts +++ b/packages/core/src/qwen/qwenContentGenerator.test.ts @@ -154,7 +154,10 @@ describe('QwenContentGenerator', () => { contents: [{ role: 'user', parts: [{ text: 'Hello' }] }], }; - const result = await qwenContentGenerator.generateContent(request, 'test-prompt-id'); + const result = await qwenContentGenerator.generateContent( + request, + 'test-prompt-id', + ); expect(result.text).toBe('Generated content'); expect(mockQwenClient.getAccessToken).toHaveBeenCalled(); @@ -171,7 +174,10 @@ describe('QwenContentGenerator', () => { contents: [{ role: 'user', parts: [{ text: 'Hello stream' }] }], }; - const stream = await qwenContentGenerator.generateContentStream(request, 'test-prompt-id'); + const stream = await qwenContentGenerator.generateContentStream( + request, + 'test-prompt-id', + ); const chunks: string[] = []; for await (const chunk of stream) { @@ -238,7 +244,10 @@ describe('QwenContentGenerator', () => { contents: [{ role: 'user', parts: [{ text: 'Hello' }] }], }; - const result = await qwenContentGenerator.generateContent(request, 'test-prompt-id'); + const result = await qwenContentGenerator.generateContent( + request, + 'test-prompt-id', + ); expect(result.text).toBe('Generated content'); expect(mockQwenClient.refreshAccessToken).toHaveBeenCalled(); @@ -545,7 +554,10 @@ describe('QwenContentGenerator', () => { contents: [{ role: 'user', parts: [{ text: 'Hello' }] }], }; - const result = await qwenContentGenerator.generateContent(request, 'test-prompt-id'); + const result = await qwenContentGenerator.generateContent( + request, + 'test-prompt-id', + ); expect(result.text).toBe('Success after retry'); expect(mockGenerateContent).toHaveBeenCalledTimes(2); @@ -795,7 +807,10 @@ describe('QwenContentGenerator', () => { contents: [{ role: 'user', parts: [{ text: 'Test message' }] }], }; - const result = await qwenContentGenerator.generateContent(request, 'test-prompt-id'); + const result = await qwenContentGenerator.generateContent( + request, + 'test-prompt-id', + ); expect(result.text).toBe('Success after refresh'); expect(mockQwenClient.getAccessToken).toHaveBeenCalled();