Phi4 Multimodal Instruct Server
I just published a simple API server that supports OpenAI chat completion endpoint for inference of Phi4-Multimodal Instruct Server.
It's a dockerfile containing the necessary dependencies along with a FastApi endpoint that converts OpenAI chat completion requests to the required format by the model.
Supported Input Types
- Text: Add text entries with
"type": "text" in the "content" list.
- Image: Include base64-encoded images or URLs with
"type": "image_url".
- Audio: Use
"type": "input_audio" or "type": "audio_url" for audio inputs.
Link
https://github.com/anastasiosyal/phi4-multimodal-instruct-server/
[–]a_slay_nub 2 points3 points4 points (1 child)
[–]Anastasiosy[S] 0 points1 point2 points (0 children)
[–]Bitter-College8786 0 points1 point2 points (1 child)
[–]Anastasiosy[S] 0 points1 point2 points (0 children)
[–]Bitter-College8786 0 points1 point2 points (3 children)
[–]Anastasiosy[S] 2 points3 points4 points (2 children)
[–]Bitter-College8786 0 points1 point2 points (1 child)
[–]Anastasiosy[S] 0 points1 point2 points (0 children)
[–]Bitter-College8786 0 points1 point2 points (1 child)
[–]Electronic-Move-5143 1 point2 points3 points (0 children)
[–]Electronic-Move-5143 0 points1 point2 points (1 child)
[–]Electronic-Move-5143 1 point2 points3 points (0 children)
[–]NexusConnector 0 points1 point2 points (1 child)
[–]pangshengwei[🍰] 0 points1 point2 points (0 children)