You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
We want to move from pickled objects saved by torch or torch.jit to safetensors format for the weights of docling-ibm-models. This has various advantages, such as better security, and also acts as a pre-requisite to achieve proper accelerator support across all models.
The text was updated successfully, but these errors were encountered:
Looking forward to this one! I'm working on a txtai integration for docling and the biggest downside is speed. For some PDFs that are a couple pages it takes 14s to extract vs 200ms with existing methods (Apache Tika). Obviously, the upside is all the formatting being preserved. But if 14s could go down to a couple seconds even it would be a big win.
We want to move from pickled objects saved by
torch
ortorch.jit
to safetensors format for the weights ofdocling-ibm-models
. This has various advantages, such as better security, and also acts as a pre-requisite to achieve proper accelerator support across all models.The text was updated successfully, but these errors were encountered: