@@ -350,7 +350,7 @@ def mock_glm4_completion() -> list:
350
350
351
351
async def test_glm4_astream (mock_glm4_completion : list ) -> None :
352
352
llm_name = "glm-4"
353
- llm = ChatOpenAI (model = llm_name , stream_usage = True )
353
+ llm = ChatOpenAI (model = llm_name , stream_usage = True , streaming = True )
354
354
mock_client = AsyncMock ()
355
355
356
356
async def mock_create (* args : Any , ** kwargs : Any ) -> MockAsyncContextManager :
@@ -375,7 +375,7 @@ async def mock_create(*args: Any, **kwargs: Any) -> MockAsyncContextManager:
375
375
376
376
def test_glm4_stream (mock_glm4_completion : list ) -> None :
377
377
llm_name = "glm-4"
378
- llm = ChatOpenAI (model = llm_name , stream_usage = True )
378
+ llm = ChatOpenAI (model = llm_name , stream_usage = True , streaming = True )
379
379
mock_client = MagicMock ()
380
380
381
381
def mock_create (* args : Any , ** kwargs : Any ) -> MockSyncContextManager :
@@ -431,7 +431,7 @@ def mock_deepseek_completion() -> list[dict]:
431
431
432
432
async def test_deepseek_astream (mock_deepseek_completion : list ) -> None :
433
433
llm_name = "deepseek-chat"
434
- llm = ChatOpenAI (model = llm_name , stream_usage = True )
434
+ llm = ChatOpenAI (model = llm_name , stream_usage = True , streaming = True )
435
435
mock_client = AsyncMock ()
436
436
437
437
async def mock_create (* args : Any , ** kwargs : Any ) -> MockAsyncContextManager :
@@ -455,7 +455,7 @@ async def mock_create(*args: Any, **kwargs: Any) -> MockAsyncContextManager:
455
455
456
456
def test_deepseek_stream (mock_deepseek_completion : list ) -> None :
457
457
llm_name = "deepseek-chat"
458
- llm = ChatOpenAI (model = llm_name , stream_usage = True )
458
+ llm = ChatOpenAI (model = llm_name , stream_usage = True , streaming = True )
459
459
mock_client = MagicMock ()
460
460
461
461
def mock_create (* args : Any , ** kwargs : Any ) -> MockSyncContextManager :
@@ -499,7 +499,7 @@ def mock_openai_completion() -> list[dict]:
499
499
500
500
async def test_openai_astream (mock_openai_completion : list ) -> None :
501
501
llm_name = "gpt-4o"
502
- llm = ChatOpenAI (model = llm_name , stream_usage = True )
502
+ llm = ChatOpenAI (model = llm_name , stream_usage = True , streaming = True )
503
503
mock_client = AsyncMock ()
504
504
505
505
async def mock_create (* args : Any , ** kwargs : Any ) -> MockAsyncContextManager :
@@ -523,7 +523,7 @@ async def mock_create(*args: Any, **kwargs: Any) -> MockAsyncContextManager:
523
523
524
524
def test_openai_stream (mock_openai_completion : list ) -> None :
525
525
llm_name = "gpt-4o"
526
- llm = ChatOpenAI (model = llm_name , stream_usage = True )
526
+ llm = ChatOpenAI (model = llm_name , stream_usage = True , streaming = True )
527
527
mock_client = MagicMock ()
528
528
529
529
def mock_create (* args : Any , ** kwargs : Any ) -> MockSyncContextManager :
0 commit comments