Skip to main content

Release Note v2025.03

1. Public Inference Service Now Supports API Access

  • Users can now access the public inference service via API, enabling seamless integration into applications and workflows for more convenient inference calls.

2. kTransformer Support Added to the Inference Engine

  • The inference engine now supports kTransformer, offering an ultra-low-cost solution for large model inference. This allows AI models to run efficiently on low-end hardware without relying on expensive high-end GPUs.

3. Supports dynamic modification of inference framework parameters

  • Supports dynamic modification of inference framework parameters, allowing users to customize parameters when creating a dedicated instance to meet specific needs.

4. New Widget Support in Dedicated Instance Sandbox

  • Image-text-to-text: Converts image and text inputs into text, ideal for multimodal tasks.
  • Feature extraction: Supports feature extraction, enhancing model analysis and vectorization tasks.

5. MCP Server Support in Application Space

  • Users can now deploy dedicated MCP Servers directly within the application space.

6. Model Source ID Display Added

  • Model details page now includes a model source ID, improving traceability and transparency of model origins.

7. Enhanced Interaction for Events & Blogs

  • New tags, search, and comments features make it easier to find relevant content and engage with the community.

8. AgenticOps Product Map Released

  • We introduce AgenticOps, an innovative framework for Agent development, deployment, and optimization. Our product suite—CSGHub, StarShip, and AutoHub—is closely integrated with AgenticOps, forming a complete ecosystem for Agent development and operations. A detailed product map is now available on the homepage.

Version List

Click Version List to check all release notes.