Details
-
Bug
-
Status: Closed
-
Major
-
Resolution: Not A Problem
-
3.0.0
-
None
-
This environment results in an error:
adlfs v0.6.0
fsspec 0.8.5
azure.storage.blob 12.6.0
adal 1.2.6
pandas 1.2.1
pyarrow 3.0.0
Description
https://github.com/dask/adlfs/issues/171
I am unable to save data to Azure Blob using ds.write_dataset() with pyarrow 3.0 and adlfs 0.6.0. Reverting to 0.5.9 fixes the issue, but I am not sure what the cause is - posting this here in case there were filesystem changes in pyarrow recently which are incompatible with changes made in adlfs.
File "pyarrow/_dataset.pyx", line 2343, in pyarrow._dataset._filesystemdataset_write File "pyarrow/_fs.pyx", line 1032, in pyarrow._fs._cb_create_dir File "/opt/conda/lib/python3.8/site-packages/pyarrow/fs.py", line 259, in create_dir self.fs.mkdir(path, create_parents=recursive) File "/opt/conda/lib/python3.8/site-packages/fsspec/asyn.py", line 121, in wrapper return maybe_sync(func, self, *args, **kwargs) File "/opt/conda/lib/python3.8/site-packages/fsspec/asyn.py", line 100, in maybe_sync return sync(loop, func, *args, **kwargs) File "/opt/conda/lib/python3.8/site-packages/fsspec/asyn.py", line 71, in sync raise exc.with_traceback(tb) File "/opt/conda/lib/python3.8/site-packages/fsspec/asyn.py", line 55, in f result[0] = await future File "/opt/conda/lib/python3.8/site-packages/adlfs/spec.py", line 1033, in _mkdir raise FileExistsError( FileExistsError: Cannot overwrite existing Azure container -- dev already exists.