Skip to content

Commit 4b9f270

Browse files
committed
Version 2.4.0 release preparation.
1 parent e95c49b commit 4b9f270

File tree

10 files changed

+22
-19
lines changed

10 files changed

+22
-19
lines changed

CONTRIBUTING.md

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -84,7 +84,7 @@ You can choose from three different environments to test your fixes/changes, bas
8484
### Mocked test environment
8585

8686
* Pick up a Linux or MacOS.
87-
* Install Python 3.6, 3.7 or 3.8
87+
* Install Python 3.7, 3.8 or 3.9
8888
* Fork the AWS Data Wrangler repository and clone that into your development environment
8989
* Go to the project's directory create a Python's virtual environment for the project
9090

@@ -115,7 +115,7 @@ or
115115
**DISCLAIMER**: Make sure you know what you are doing. These steps will charge some services on your AWS account and require a minimum security skill to keep your environment safe.
116116

117117
* Pick up a Linux or MacOS.
118-
* Install Python 3.6, 3.7 or 3.8
118+
* Install Python 3.7, 3.8 or 3.9
119119
* Fork the AWS Data Wrangler repository and clone that into your development environment
120120
* Go to the project's directory create a Python's virtual environment for the project
121121

@@ -168,7 +168,7 @@ or
168168
**DISCLAIMER**: This environment contains Aurora MySQL, Aurora PostgreSQL and Redshift (single-node) clusters which will incur cost while running.
169169

170170
* Pick up a Linux or MacOS.
171-
* Install Python 3.6, 3.7 and 3.8
171+
* Install Python 3.7, 3.8 or 3.9
172172
* Fork the AWS Data Wrangler repository and clone that into your development environment
173173
* Go to the project's directory create a Python's virtual environment for the project
174174

@@ -244,5 +244,5 @@ or
244244

245245
## Common Errors
246246

247-
Check the file below to check the common errors and solutions
247+
Check the file below to check the common errors and solutions
248248
[ERRORS](https://github.com/awslabs/aws-data-wrangler/blob/main/CONTRIBUTING_COMMON_ERRORS.md)

awswrangler/_utils.py

Lines changed: 4 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -63,9 +63,11 @@ def boto3_from_primitives(primitives: Optional[Boto3PrimitivesType] = None) -> b
6363
def default_botocore_config() -> botocore.config.Config:
6464
"""Botocore configuration."""
6565
retries_config: Dict[str, Union[str, int]] = {
66-
"max_attempts": int(os.getenv("AWS_MAX_ATTEMPTS", "3")),
67-
"mode": os.getenv("AWS_RETRY_MODE", "standard"),
66+
"max_attempts": int(os.getenv("AWS_MAX_ATTEMPTS", "5")),
6867
}
68+
mode: Optional[str] = os.getenv("AWS_RETRY_MODE")
69+
if mode:
70+
retries_config["mode"] = mode
6971
return botocore.config.Config(retries=retries_config, connect_timeout=10, max_pool_connections=10)
7072

7173

awswrangler/s3/__init__.py

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -5,6 +5,7 @@
55
from awswrangler.s3._describe import describe_objects, get_bucket_region, size_objects # noqa
66
from awswrangler.s3._download import download # noqa
77
from awswrangler.s3._list import does_object_exist, list_directories, list_objects # noqa
8+
from awswrangler.s3._merge_upsert_table import merge_upsert_table # noqa
89
from awswrangler.s3._read_excel import read_excel # noqa
910
from awswrangler.s3._read_parquet import read_parquet, read_parquet_metadata, read_parquet_table # noqa
1011
from awswrangler.s3._read_text import read_csv, read_fwf, read_json # noqa

building/build-wheel.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -3,5 +3,5 @@ set -ex
33

44
pushd ..
55
rm -rf *.egg-info build dist/*.whl
6-
python3.6 setup.py bdist_wheel
6+
python setup.py bdist_wheel
77
rm -rf *.egg-info build

building/lambda/build-lambda-layer.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -14,7 +14,7 @@ export ARROW_HOME=$(pwd)/dist
1414
export LD_LIBRARY_PATH=$(pwd)/dist/lib:$LD_LIBRARY_PATH
1515

1616
git clone \
17-
--branch apache-arrow-2.0.0 \
17+
--branch apache-arrow-3.0.0 \
1818
--single-branch \
1919
https://github.com/apache/arrow.git
2020

docs/source/install.rst

Lines changed: 3 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -57,12 +57,11 @@ AWS Glue PySpark Jobs
5757
Go to your Glue PySpark job and create a new *Job parameters* key/value:
5858

5959
* Key: ``--additional-python-modules``
60-
* Value: ``awswrangler``
61-
62-
To install a specific version, set the value for above Job parameter as follows:
63-
6460
* Value: ``awswrangler==2.3.0``
6561

62+
P.S. By now AWS Glue PySpark Jobs does not support PyArrow +3.0.0.
63+
Please use awswrangler==2.3.0 that uses PyArrow 2.0.0 to overcome this limitation.
64+
6665
`Official Glue PySpark Reference <https://docs.aws.amazon.com/glue/latest/dg/reduced-start-times-spark-etl-jobs.html#reduced-start-times-new-features>`_
6766

6867
Amazon SageMaker Notebook

requirements-dev.txt

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1,3 +1,4 @@
1+
wheel==0.36.2
12
isort==5.7.0
23
black==20.8b1
34
pylint==2.6.0

tests/test_config.py

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -129,16 +129,16 @@ def test_botocore_config(path):
129129
original = botocore.client.ClientCreator.create_client
130130

131131
# Default values for botocore.config.Config
132-
expected_max_retries_attempt = 3
132+
expected_max_retries_attempt = 5
133133
expected_connect_timeout = 10
134134
expected_max_pool_connections = 10
135-
expected_retry_mode = "standard"
135+
expected_retry_mode = None
136136

137137
def wrapper(self, **kwarg):
138138
assert kwarg["client_config"].retries["max_attempts"] == expected_max_retries_attempt
139139
assert kwarg["client_config"].connect_timeout == expected_connect_timeout
140140
assert kwarg["client_config"].max_pool_connections == expected_max_pool_connections
141-
assert kwarg["client_config"].retries["mode"] == expected_retry_mode
141+
assert kwarg["client_config"].retries.get("mode") == expected_retry_mode
142142
return original(self, **kwarg)
143143

144144
# Check for default values

tox.ini

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -11,12 +11,12 @@ deps =
1111
moto==1.3.16
1212
s3fs==0.4.2
1313
commands =
14-
pytest -n 16 --timeout=300 tests
14+
pytest -n 32 --timeout=300 tests
1515

1616
[testenv:py39]
1717
passenv = AWS_PROFILE AWS_DEFAULT_REGION AWS_ACCESS_KEY_ID AWS_SECRET_ACCESS_KEY
1818
deps =
1919
{[testenv]deps}
20-
pytest-cov==2.10.1
20+
pytest-cov==2.11.1
2121
commands =
22-
pytest -n 16 --timeout=300 --cov=awswrangler tests
22+
pytest -n 32 --timeout=300 --cov=awswrangler tests

tutorials/007 - Redshift, MySQL, PostgreSQL, SQL Server.ipynb

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -39,7 +39,7 @@
3939
"cell_type": "markdown",
4040
"metadata": {},
4141
"source": [
42-
"## Connect throught Glue Catalog Connections\n",
42+
"## Connect using the Glue Catalog Connections\n",
4343
"\n",
4444
"- [wr.redshift.connect()](https://aws-data-wrangler.readthedocs.io/en/stable/stubs/awswrangler.redshift.connect.html)\n",
4545
"- [wr.mysql.connect()](https://aws-data-wrangler.readthedocs.io/en/stable/stubs/awswrangler.mysql.connect.html)\n",

0 commit comments

Comments
 (0)