How to run inference without vLLM? (e.g., with standard transformers)

#1
by Fezz04 - opened

Hi,

I'm really interested in using the LightOnOCR-1B-1025 model for a project. I've read the model card, but all the provided examples for inference rely on setting up a vLLM server.

Is it possible to run this model for inference directly using the standard Hugging Face transformers library (for example, by loading it with AutoModel or a specific model class)? The model card mentions "Transformers integration is coming soon for training," but I wasn't sure about inference.

If anyone has a working code snippet for running a single-image inference without vLLM, I would greatly appreciate it!

Thanks!

Hey there!

I see that the transformers integration is on its way: https://github.com/huggingface/transformers/pull/41621

When merged it would be possible to use it with standard transformers 🤗

Sign up or log in to comment