ENH: Add MinerU2.5-2509-1.2B for VLLM#4523
Conversation
|
to fix this issue, #4518 |
|
Could you rebase the code? |
|
Could you please take a look at two things?
|
|
You need to modify the JSON file in ModelHub. Directly editing JSON within the code is not permitted. You may need to revert your recent changes and then edit the JSON in ModelHub, which will automatically commit to the current PR. |
|
As you said "transformers_dependencies ; #engine# == \“Transformers\”" ” |
|
验证成功。请帮忙通过。 |
Integrate MinerU VLM model support into Xinference: Core Changes: - xinference/model/llm/llm_family.json: Add mineru-vlm model configuration - xinference/model/llm/vllm/core.py: Add base64 image/PDF handling - xinference/ui/gradio/chat_interface.py: Add PDF upload support in chat UI - xinference/ui/gradio/media_interface.py: Enhance media interface - xinference/model/image/model_spec.json: Update OCR model specifications Features: - VLM chat with multi-modal support (image, video, audio, PDF) - Base64 data URI handling for seamless file uploads - MinerU VLM model ready for serving via vLLM engine
- Fix W293: Remove trailing whitespace from blank lines in media_interface.py - Fix E303: Remove extra blank lines in media_interface.py - Fix F841: Remove unused variable file_ext in media_interface.py - Fix W391: Remove blank line at end of core.py
|
Replaced by #4569 |
Add MinerU2.5-2509-1.2B as a Vision-Language Model to Xinference
Support image upload and interactive analysis in the VLM Chat interface
It is recommended to use MinerU CLI for processing PDF documents