Skip to content

v1.2.0: OpenAI Responses API Integration & Dual-Mode Support

Latest

Choose a tag to compare

@pavelsukhachev pavelsukhachev released this 13 Jul 16:22
· 1 commit to main since this release

πŸš€ Major Release: OpenAI Responses API Integration

✨ New Features

πŸ†• OpenAI Responses API Integration

  • Full support for 2025 Responses API with gpt-4o + image_generation tool
  • Native multimodal understanding with better context awareness
  • Enhanced prompt following and superior text rendering in images
  • Real-time streaming with partial previews via Responses API

πŸ”„ Dual API Architecture

  • Seamless API switching via API_MODE environment variable
  • Default to Responses API (recommended) with fallback to Images API
  • Backward compatibility with existing Images API implementations
  • Unified interface supporting both legacy and modern workflows

πŸ–ΌοΈ Enhanced Image Editing

  • Multi-image input support with proper message formatting
  • Context-aware editing with conversation history integration
  • Previous response ID linking for multi-turn editing sessions
  • Advanced mask support for precise inpainting operations

πŸ§ͺ Comprehensive Test Coverage

  • 99+ tests with full API coverage for both implementations
  • Complete error handling and edge case coverage
  • Real API validation with integration testing
  • Mock-based testing with proper dependency injection

πŸ”§ Technical Improvements

βš™οΈ Enhanced Configuration System

  • API_MODE: Switch between 'responses' (default) and 'images'
  • RESPONSES_MODEL: Model selection for Responses API (default: gpt-4o)
  • Flexible model switching between dedicated and integrated approaches

πŸ—οΈ SOLID Architecture

  • Complete refactoring following dependency injection patterns
  • Single Responsibility principle with focused services
  • Interface-based design for maximum extensibility
  • Test-driven development with comprehensive coverage

πŸ“š Updated Documentation

πŸ“– Comprehensive Guides

  • Dual API documentation with feature comparison table
  • Enhanced environment variables documentation
  • Updated streaming examples for both API modes
  • API mode selection guidance and best practices

πŸ“‹ API Comparison

Feature Responses API (gpt-4o) Images API (gpt-image-1)
Latest Technology βœ… 2025 Responses API ⚠️ Legacy API
Text in Images βœ… Superior βœ… Good
Context Awareness βœ… Excellent ⚠️ Limited
Streaming βœ… Partial previews ⚠️ Final only
Multi-turn βœ… Full support ⚠️ Basic

πŸ”§ Migration Guide

For New Users

  • Default configuration uses Responses API automatically
  • No action required - just set OPENAI_API_KEY

For Existing Users

  • Backward compatible - existing configurations continue to work
  • Optional upgrade: Set API_MODE=responses to use latest features
  • Gradual migration supported with dual-mode architecture

πŸ“¦ Installation

# Clone and install
git clone https://github.com/pavelsukhachev/mcp-server-gpt-image.git
cd mcp-server-gpt-image
npm install

# Configure (uses Responses API by default)
echo "OPENAI_API_KEY=your-api-key-here" > .env

# Run
npm run start:http

πŸš€ What's Next

This release establishes MCP Server GPT Image as the most advanced OpenAI image generation server available, supporting both cutting-edge Responses API and legacy Images API with seamless switching.

Perfect for:

  • 🎨 Advanced image generation workflows
  • πŸ”„ Multi-turn conversation contexts
  • ⚑ Real-time streaming applications
  • πŸ§ͺ Development and production environments

Full Changelog: https://github.com/pavelsukhachev/mcp-server-gpt-image/blob/main/CHANGELOG.md