Full basic Path documentation linked here: PathDocs.
A subclass of Path and PureS3Path, this class represents a concrete paths of AWS S3 Service. All actions are use boto3 as the SKD for AWS S3 Service:
>>> S3Path('/<bucket>/<key>')
S3Path('/<bucket>/<key>')
pathsegments are specified similarly to Path.
You can't use S3Path if you doesn't have boto3 installed in your environment:
>>> import boto3
Traceback (most recent call last):
File "<stdin>", line 1, in <module>
ModuleNotFoundError: No module named 'boto3'
>>> from s3path import S3Path
>>> S3Path('/<bucket>/<key>')
Traceback (most recent call last):
File "<stdin>", line 1, in <module>
File "pathlib.py", line 798, in __new__
% (cls.__name__,))
NotImplementedError: cannot instantiate 'S3Path' on your system
A subclass of S3Path and PureVersionedS3Path, this class represents a concrete path of the AWS S3 Service for buckets in which S3 versioning is enabled. All actions use boto3 as the SKD for AWS S3 Service:
>>> from s3path import VersionedS3Path
>>> VersionedS3Path('/<bucket>/<key>', version_id='<version_id>')
VersionedS3Path('/<bucket>/<key>', version_id='<version_id>')
S3Path and VersionedS3Path provide the following methods in addition to pure paths methods. All the methods below will raise a ValueError if the path isn't absolute. Many of these methods can raise a botocore.exceptions.ClientError if boto3 call fails (for example because the path doesn't exist).
NOTE: The following signatures are shown for S3Path but are equally valid for VersionedS3Path as well. Any behavioral differences between S3Path methods and their VersionedS3Path equivalents are explicitly detailed below (i.e. if a given VersionedS3Path method signature is not listed below, it is assumed that it behaves identically to its S3Path equivalent).
S3Path.stat(*, follow_symlinks=True)
Returns information about this path (similarly to boto3's ObjectSummary). For compatibility with pathlib, the returned object some similar attributes like os.stat_result. The result is looked up at each call to this method:
>>> path_stat = S3Path('/pypi-proxy/boto3/index.html').stat()
>>> path_stat
StatResult(size=188, last_modified=datetime.datetime(2018, 4, 4, 12, 26, 3, tzinfo=tzutc()), version_id=None)
>>> path_stat.st_size
188
>>> path_stat.st_mtime
1522833963.0
>>> print(path_stat.st_version_id)
None
>>> path_stat.st_atime
Traceback (most recent call last):
...
io.UnsupportedOperation: StatResult do not support st_atime attribute
NOTES:
follow_symlinks
option must be always set toTrue
.- The returned object will contain an additional
st_version_id
attribute that is not part of the
os.stat_result API. The value of st_version_id
will be None
.
VersionedS3Path.stat(*, follow_symlinks=True)
Behaves the same as S3Path.stat with the exception that the st_version_id
attribute of the
returned object will contain the version ID of the underlying S3 object.
Whether the path points to an existing Bucket, key or key prefix:
>>> S3Path('/pypi-proxy/boto3/index.html').exists()
True
>>> S3Path('/pypi-proxy/boto3/').exists()
True
>>> S3Path('/fake-bucket/').exists()
False
Behaves the same as S3Path.exists except that the version ID must match in addition to the bucket and key.
Glob the given relative pattern in the Bucket / key prefix represented by this path, yielding all matching files (of any kind):
>>> bucket_path = S3Path('/pypi-proxy/')
>>> [path for path in bucket_path.glob('boto*')]
[S3Path('/pypi-proxy/boto3/'), S3Path('/pypi-proxy/botocore/')]
>>> [path for path in bucket_path.glob('*/*.html')]
[S3Path('/pypi-proxy/requests/index.html'),
S3Path('/pypi-proxy/boto3/index.html'),
S3Path('/pypi-proxy/botocore/index.html')]]
The "**" pattern means "this Bucket / key prefix and all sub key prefixes, recursively". In other words, it enables recursive globbing:
>>> bucket_path = S3Path('/pypi-proxy/')
>>> list(bucket_path.glob('**/*.html'))
[S3Path('/pypi-proxy/requests/index.html'),
S3Path('/pypi-proxy/index.html'),
S3Path('/pypi-proxy/boto3/index.html'),
S3Path('/pypi-proxy/botocore/index.html')]
New in version 0.4.0: New Algorithm that better suited to s3 API. Especially for recursive searches.
To enable the old (pathlib common) Algorithm you can configure it like this:
register_configuration_parameter(path, glob_new_algorithm=False)
For more configuration details please see this `Advanced S3Path configuration`_
NOTE: Using the "**" pattern in large Buckets may consume an inordinate amount of time in the old algorithm.
Returns True
if the path points to a Bucket or a key prefix,
False
if it points to a full key path.
False
is also returned if the path doesn’t exist.
Other errors (such as permission errors) are propagated.
Returns True
if the path points to a Bucket key,
False
if it points to Bucket or a key prefix.
False
is also returned if the path doesn’t exist.
Other errors (such as permission errors) are propagated.
AWS S3 Service doesn't have mounting feature,
There for this method will always return False
AWS S3 Service doesn't have symlink feature,
There for this method will always return False
AWS S3 Service doesn't have sockets feature,
There for this method will always return False
AWS S3 Service doesn't have fifo feature,
There for this method will always return False
When the path points to a Bucket or a key prefix, yield path objects of the directory contents:
>>> bucket_path = S3Path('/pypi-proxy/')
>>> [path for path in bucket_path.iterdir() if path.is_dir()]
[S3Path('/pypi-proxy/requests/'),
S3Path('/pypi-proxy/boto3/'),
S3Path('/pypi-proxy/botocore/')]
>>> boto3_path = bucket_path.joinpath('boto3')
>>> [path for path in bucket_path.boto3_path()]
[S3Path('/pypi-proxy/boto3/boto3-1.4.1.tar.gz'), S3Path('/pypi-proxy/boto3/index.html')]
Opens the Bucket key pointed to by the path. This delegates to the smart_open library that handles the file streaming. returns a file like object that you can read or write with:
>>> with S3Path('/pypi-proxy/botocore/index.html').open() as f:
>>> print(f.read())
'<!DOCTYPE html>
<html>
<head>
<meta charset="UTF-8">
<title>Package Index</title>
</head>
<body>
<a href="botocore-1.4.93.tar.gz">botocore-1.4.93.tar.gz</a><br>
</body>
</html>'
Behaves the same as S3Path.open except that VersionedS3Path.version_id
will be used to open
the specified version of the object pointed to by the VersionedS3Path object.
Returns the name of the user owning the Bucket or key. Similarly to boto3's ObjectSummary owner attribute
Return the binary contents of the Bucket key as a bytes object:
>>> S3Path('/test_bucket/test.txt').write_bytes(b'Binary file contents')
>>> S3Path('/test_bucket/test.txt').read_bytes()
b'Binary file contents'
Returns the decoded contents of the Bucket key as a string:
>>> S3Path('/test_bucket/test.txt').write_text('Text file contents')
>>> S3Path('/test_bucket/test.txt').read_text()
'Text file contents'
Renames this file or Bucket / key prefix / key to the given target. If target exists and is a file, it will be replaced silently if the user has permission. If path is a key prefix, it will replace all the keys with the same prefix to the new target prefix. target can be either a string or another S3Path object:
>>> path = S3Path('/test_bucket/test.txt').write_text('Text file contents')
>>> target = S3Path('/test_bucket/new_test.txt')
>>> path.rename(target)
>>> target.read_text()
'Text file contents'
Renames this Bucket / key prefix / key to the given target. If target points to an existing Bucket / key prefix / key, it will be unconditionally replaced.
This is like calling S3Path.glob with "**/"
added in front of the given relative pattern:
>>> bucket_path = S3Path('/pypi-proxy/')
>>> list(bucket_path.rglob('*.html'))
[S3Path('/pypi-proxy/requests/index.html'),
S3Path('/pypi-proxy/index.html'),
S3Path('/pypi-proxy/botocore/index.html')]
New in version 0.4.0: New Algorithm that better suited to s3 API. Especially for recursive searches.
Removes this Bucket / key prefix. The Bucket / key prefix must be empty.
Removes this key from S3. Note that this will not remove directories or buckets, but will instead raise an IsADirectoryError. If the key does is not present in the given bucket, or if the bucket is not present, raises a `FileNotFoundError`_. If missing_ok is True then no exception will be raised.
Returns whether this path points to the same Bucket key as other_path, which can be either a Path object, or a string:
>>> path = S3Path('/test_bucket/test.txt')
>>> path.samefile(S3Path('/test_bucket/test.txt'))
True
>>> path.samefile('/test_bucket/fake')
False
Creates a key at this given path. If the key already exists, the function succeeds if exist_ok is true (and its modification time is updated to the current time), otherwise FileExistsError is raised.
Opens the key pointed to in bytes mode, write data to it, and close / save the key:
>>> S3Path('/test_bucket/test.txt').write_bytes(b'Binary file contents')
>>> S3Path('/test_bucket/test.txt').read_bytes()
b'Binary file contents'
Opens the key pointed to in text mode, writes data to it, and close / save the key:
>>> S3Path('/test_bucket/test.txt').write_text('Text file contents')
>>> S3Path('/test_bucket/test.txt').read_text()
'Text file contents'
NOTE: newline
option is only available on Python 3.10 and greater.
Create a path bucket.
AWS S3 Service doesn't support folders, therefore the mkdir method will only create the current bucket. If the bucket path already exists, FileExistsError is raised.
If exist_ok is false (the default), FileExistsError is raised if the target Bucket already exists.
If exist_ok is true, OSError exceptions will be ignored.
if parents is false (the default), mkdir will create the bucket only if this is a Bucket path.
if parents is true, mkdir will create the bucket even if the path have a Key path.
mode argument is ignored.
Returns a pre-signed url. Anyone with the url can make a GET request to get the file. You can set an expiration date with the expire_in argument (integer or timedelta object).
Note that generating a presigned url may require more information or setup than to use other S3Path functions. It's because it needs to know the exact aws region and use s3v4 as signature version. Meaning you may have to do this:
>>> import boto3
>>> from botocore.config import Config
>>> from s3path import S3Path, register_configuration_parameter
>>> resource = boto3.resource(
... "s3",
... config=Config(signature_version="s3v4"),
... region_name="the aws region name"
... )
>>> register_configuration_parameter(S3Path("/"), resource=resource)
Here is an example of using a presigned url:
>>> from s3path import S3Path
>>> import requests
>>> file = S3Path("/my-bucket/toto.txt")
>>> file.write_text("hello world")
>>> presigned_url = file.get_presigned_url()
>>> print(requests.get(presigned_url).content)
b"hello world"
Full basic PurePath documentation linked here: PurePathDocs.
A subclass of PurePath, this path flavour represents AWS S3 Service semantics.
>>> PureS3Path('/<bucket>/<key>')
PureS3Path('/<bucket>/<key>')
pathsegments are specified similarly to PurePath.
A subclass of PureS3Path, this path flavour represents AWS S3 Service semantics for buckets in which S3 versioning is enabled.
>>> from s3path import PureVersionedS3Path
>>> PureVersionedS3Path('/<bucket>/<key>', version_id='<version_id>')
PureVersionedS3Path('/<bucket>/<key>', version_id='<version_id>')
PureS3Path has a similar behavior to PurePosixPath, except for the below changes:
Double dots ('..'
) are treated as follows.
This is different then PurePath since AWS S3 Service doesn't support symbolic links:
>>> PureS3Path('foo/../bar')
PureS3Path('bar')
NOTE: All The methods below will raise ValueError if the path isn't absolute.
PureS3Path.joinpath(*other)
If the final element of other
is a PureVersionedS3Path instance, the resulting object will
also be a PureVersionedS3Path instance with version_id
set to other[-1].version_id
.
Otherwise, the resulting object will be a PureS3Path instance.
Represents the path as a AWS S3 URI:
>>> p = PureS3Path('/pypi-proxy/boto3/')
>>> p.as_uri()
's3://pypi-proxy/boto3/'
>>> p = PureS3Path('/pypi-proxy/boto3/index.html')
>>> p.as_uri()
's3://pypi-proxy/boto3/index.html'
Represents a AWS S3 URI as a PureS3Path:
>>> PureS3Path.from_uri('s3://pypi-proxy/boto3/index.html')
PureS3Path('/pypi-proxy/boto3/index.html')
This is a new class method.
Represents a AWS S3 Bucket and Key pairs as a PureS3Path:
>>> PureS3Path.from_bucket_key('pypi-proxy', 'boto3/index.html')
PureS3Path('/pypi-proxy/boto3/index.html')
This is a new class method.
New in version 0.3.0.
A string representing the AWS S3 Bucket name, if any:
>>> PureS3Path.from_uri('s3://pypi-proxy/boto3/').bucket
'pypi-proxy'
>>> PureS3Path('/').bucket
''
This is a new property.
A string representing the AWS S3 Key name, if any:
>>> PureS3Path('/pypi-proxy/boto3/').key
'boto3'
>>> PureS3Path('/pypi-proxy/boto3/index.html').key
'boto3/index.html'
>>> PureS3Path.from_uri('s3://pypi-proxy/').key
''
This is a new property.
PureVersionedS3Path has a similar behavior to PureS3Path, except for the below changes:
PureVersionedS3Path.from_uri(uri, *, version_id)
Represents a versioned AWS S3 URI as a PureVersionedS3Path:
>>> from s3path import PureVersionedS3Path
>>> PureVersionedS3Path.from_uri('s3://pypi-proxy/boto3/index.html', version_id='<version_id>')
PureVersionedS3Path('/pypi-proxy/boto3/index.html', version_id='<version_id>')
This is a new class method.
PureVersionedS3Path.from_bucket_key(bucket, key, *, version_id)
Represents a versioned AWS S3 Bucket and Key pairs as a PureVersionedS3Path:
>>> from s3path import PureVersionedS3Path
>>> PureVersionedS3Path.from_bucket_key('pypi-proxy', 'boto3/index.html', version_id='<version_id>')
PureVersionedS3Path('/pypi-proxy/boto3/index.html', version_id='<version_id>')
This is a new class method.
The division of PureVersionedS3Path instances with other objects will yield the following types:
PureVersionedS3Path / PureVersionedS3Path -> PureVersionedS3Path
PureS3Path / PureVersionedS3Path -> PureVersionedS3Path
str / PureVersionedS3Path -> PureVersionedS3Path
PureVersionedS3Path / PureS3Path -> PureS3Path
PureVersionedS3Path / str -> PureS3Path
>>> from s3path import S3Path, VersionedS3Path
>>> str_path = "example/path"
>>> s3_path = S3Path("example/path")
>>> versioned_s3_path = VersionedS3Path("example/path", version_id="<version_id>")
>>> type(versioned_s3_path / versioned_s3_path)
<<< s3path.VersionedS3Path
>>> type(s3_path / versioned_s3_path)
<<< s3path.VersionedS3Path
>>> type(str_path / versioned_s3_path)
<<< s3path.VersionedS3Path
>>> type(versioned_s3_path / s3_path)
<<< s3path.S3Path
>>> type(versioned_s3_path / str_path)
<<< s3path.S3Path
There are several methods that are not supported in S3Path. All of them will raise NotImplementedError.
For example AWS S3 Service doesn't have a current directory:
>>> S3Path('/test_bucket/test.txt').cwd()
Traceback (most recent call last):
File "<stdin>", line 1, in <module>
File "/home/lior/lior_env/s3path/s3path.py", line 235, in cwd
raise NotImplementedError(message)
NotImplementedError: PathNotSupportedMixin.cwd is unsupported on AWS S3 service
Here is a list of all unsupported methods:
- classmethod S3Path.cwd()
- classmethod S3Path.home()
- S3Path.chmod(mode, *, follow_symlinks=True)
- S3Path.expanduser()
- S3Path.lchmod(mode)
- S3Path.group()
- S3Path.is_block_device()
- S3Path.is_char_device()
- S3Path.lstat()
- S3Path.resolve()
- S3Path.symlink_to(target, target_is_directory=False)