You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
It is sufficient to only load small parts of a file into memory before upload.
It is possible to limit bandwidth usage by using a generator that provides chunks at a limited rate.
The Dropbox API of course already requires upload sessions (files/upload_session_start, files/upload_session_append and files/upload_session_finish) to upload files > 150 MB. However, this approach by itself does not replace chunked or streaming uploads because:
The request body should be ideally >= 4 MB to reduce the total number of API calls (both for efficiency and to not exhaust data transport API call limits).
Bandwidth control will be very coarse when performed on chunks of 4 MB compared for example 2 kB.
Memory usage will still be larger compared to 1 kB or 2 kB chunks, especially for parallel uploads.
Describe the solution you'd like
Requests supports streaming uploads by passing a file-like object as the request body and chunked uploads by passing a generator as the request body. However, the Python SDK explicitly prevents both by requiring the request body to be of type bytes:
# Disallow streams and file-like objects even though the underlying
# requests library supports them. This is to prevent incorrect
# behavior when a non-rewindable stream is read from, but the
# request fails and needs to be re-tried at a later time.
raiseTypeError('expected request_binary as binary type, got %s'%
type(request_binary))
It would be good to either completely drop this limitation, with appropriate warnings in the doc string, or at least allow chunked uploads (where requests handles retry / rewind logic) even when disallowing streaming uploads.
Describe alternatives you've considered
Not at present.
The text was updated successfully, but these errors were encountered:
If there is some consensus by the team this is worth considering, I'm happy to create a PR. But if this is not an approach that you want to pursue, do let me know and I'll look for other solutions.
We welcome PRs in general, but I can't say offhand if this in particular is something the team would or wouldn't want to support in the SDK. I'll ask them though to see if I can get some guidance on this.
Why is this feature valuable to you? Does it solve a problem you're having?
The requests library allows both streaming and chunked uploads (see https://requests.readthedocs.io/en/latest/user/advanced/#streaming-uploads and https://requests.readthedocs.io/en/latest/user/advanced/#chunk-encoded-requests). This has two benefits:
The Dropbox API of course already requires upload sessions (
files/upload_session_start
,files/upload_session_append
andfiles/upload_session_finish
) to upload files > 150 MB. However, this approach by itself does not replace chunked or streaming uploads because:Describe the solution you'd like
Requests supports streaming uploads by passing a file-like object as the request body and chunked uploads by passing a generator as the request body. However, the Python SDK explicitly prevents both by requiring the request body to be of type
bytes
:dropbox-sdk-python/dropbox/dropbox_client.py
Lines 533 to 539 in 9895d70
It would be good to either completely drop this limitation, with appropriate warnings in the doc string, or at least allow chunked uploads (where requests handles retry / rewind logic) even when disallowing streaming uploads.
Describe alternatives you've considered
Not at present.
The text was updated successfully, but these errors were encountered: