@@ -67,7 +67,7 @@ describe('LLM Model Selection with Special Characters', () => {
67
67
vi . mocked ( options . getOptionBool ) . mockReturnValue ( true ) ;
68
68
} ) ;
69
69
70
- describe ( 'OpenAI Model Names' , ( ) => {
70
+ describe . skip ( 'OpenAI Model Names' , ( ) => {
71
71
it ( 'should correctly handle model names with periods' , async ( ) => {
72
72
const modelName = 'gpt-4.1-turbo-preview' ;
73
73
vi . mocked ( options . getOption ) . mockImplementation ( ( key : string ) => {
@@ -82,7 +82,7 @@ describe('LLM Model Selection with Special Characters', () => {
82
82
83
83
// Spy on getOpenAIOptions to verify model name is passed correctly
84
84
const getOpenAIOptionsSpy = vi . spyOn ( providers , 'getOpenAIOptions' ) ;
85
-
85
+
86
86
try {
87
87
await service . generateChatCompletion ( [ { role : 'user' , content : 'test' } ] , opts ) ;
88
88
} catch ( error ) {
@@ -108,7 +108,7 @@ describe('LLM Model Selection with Special Characters', () => {
108
108
} ;
109
109
110
110
const getOpenAIOptionsSpy = vi . spyOn ( providers , 'getOpenAIOptions' ) ;
111
-
111
+
112
112
try {
113
113
await service . generateChatCompletion ( [ { role : 'user' , content : 'test' } ] , opts ) ;
114
114
} catch ( error ) {
@@ -127,7 +127,7 @@ describe('LLM Model Selection with Special Characters', () => {
127
127
} ;
128
128
129
129
const getOpenAIOptionsSpy = vi . spyOn ( providers , 'getOpenAIOptions' ) ;
130
-
130
+
131
131
const openaiOptions = providers . getOpenAIOptions ( opts ) ;
132
132
expect ( openaiOptions . model ) . toBe ( modelName ) ;
133
133
} ) ;
@@ -153,7 +153,7 @@ describe('LLM Model Selection with Special Characters', () => {
153
153
} ) ;
154
154
155
155
const service = new OpenAIService ( ) ;
156
-
156
+
157
157
// Access the private openai client through the service
158
158
const client = ( service as any ) . getClient ( 'test-key' ) ;
159
159
const createSpy = vi . spyOn ( client . chat . completions , 'create' ) ;
@@ -213,7 +213,7 @@ describe('LLM Model Selection with Special Characters', () => {
213
213
} ) ;
214
214
215
215
const service = new AnthropicService ( ) ;
216
-
216
+
217
217
// Access the private anthropic client
218
218
const client = ( service as any ) . getClient ( 'test-key' ) ;
219
219
const createSpy = vi . spyOn ( client . messages , 'create' ) ;
@@ -278,7 +278,7 @@ describe('LLM Model Selection with Special Characters', () => {
278
278
279
279
const ollamaOptions = await providers . getOllamaOptions ( opts ) ;
280
280
expect ( ollamaOptions . model ) . toBe ( modelName ) ;
281
-
281
+
282
282
// Also test with model specified in options
283
283
const optsWithModel : ChatCompletionOptions = {
284
284
model : 'another/model:v2.0@beta' ,
@@ -370,7 +370,7 @@ describe('LLM Model Selection with Special Characters', () => {
370
370
describe ( 'Integration with REST API' , ( ) => {
371
371
it ( 'should pass model names correctly through REST chat service' , async ( ) => {
372
372
const modelName = 'gpt-4.1-turbo-preview@latest' ;
373
-
373
+
374
374
// Mock the configuration helpers
375
375
vi . doMock ( '../config/configuration_helpers.js' , ( ) => ( {
376
376
getSelectedModelConfig : vi . fn ( ) . mockResolvedValue ( {
@@ -382,8 +382,8 @@ describe('LLM Model Selection with Special Characters', () => {
382
382
383
383
const { getSelectedModelConfig } = await import ( '../config/configuration_helpers.js' ) ;
384
384
const config = await getSelectedModelConfig ( ) ;
385
-
385
+
386
386
expect ( config ?. model ) . toBe ( modelName ) ;
387
387
} ) ;
388
388
} ) ;
389
- } ) ;
389
+ } ) ;
0 commit comments