Skip to content

Commit da3f55b

Browse files
committed
formats code
1 parent 96aab27 commit da3f55b

File tree

2 files changed

+40
-42
lines changed

2 files changed

+40
-42
lines changed

tests/extract/test_incremental.py

Lines changed: 38 additions & 39 deletions
Original file line numberDiff line numberDiff line change
@@ -1,53 +1,52 @@
1-
import os
21
import asyncio
32
import inspect
3+
import os
44
import random
5-
from time import sleep
6-
from typing import Optional, Any
7-
from unittest import mock
85
from datetime import datetime # noqa: I251
96
from itertools import chain, count
7+
from time import sleep
8+
from typing import Any, Optional
9+
from unittest import mock
1010

1111
import duckdb
12+
import pyarrow as pa
1213
import pytest
1314

1415
import dlt
16+
from dlt.common import Decimal
17+
from dlt.common.configuration import ConfigurationValueError
1518
from dlt.common.configuration.container import Container
1619
from dlt.common.configuration.exceptions import InvalidNativeValue
17-
from dlt.common.configuration.specs.base_configuration import configspec, BaseConfiguration
18-
from dlt.common.configuration import ConfigurationValueError
20+
from dlt.common.configuration.specs.base_configuration import (
21+
BaseConfiguration,
22+
configspec,
23+
)
24+
from dlt.common.json import json
1925
from dlt.common.pendulum import pendulum, timedelta
20-
from dlt.common import Decimal
2126
from dlt.common.pipeline import NormalizeInfo, StateInjectableContext, resource_state
2227
from dlt.common.schema.schema import Schema
23-
from dlt.common.utils import uniq_id, digest128, chunks
24-
from dlt.common.json import json
25-
28+
from dlt.common.utils import chunks, digest128, uniq_id
2629
from dlt.extract import DltSource
27-
from dlt.extract.items import ValidateItem
28-
from dlt.extract.resource import DltResource
29-
from dlt.sources.helpers.transform import take_first
30-
from dlt.extract.incremental import IncrementalResourceWrapper, Incremental
30+
from dlt.extract.incremental import Incremental, IncrementalResourceWrapper
3131
from dlt.extract.incremental.exceptions import (
3232
IncrementalCursorInvalidCoercion,
33+
IncrementalCursorPathHasValueNone,
3334
IncrementalCursorPathMissing,
3435
IncrementalPrimaryKeyMissing,
35-
IncrementalCursorPathHasValueNone,
3636
)
37+
from dlt.extract.items import ValidateItem
38+
from dlt.extract.resource import DltResource
3739
from dlt.pipeline.exceptions import PipelineStepFailed
38-
40+
from dlt.sources.helpers.transform import take_first
3941
from tests.extract.utils import AssertItems, data_item_to_list
42+
from tests.pipeline.utils import assert_query_data
4043
from tests.utils import (
44+
ALL_TEST_DATA_ITEM_FORMATS,
45+
TestDataItemFormat,
4146
data_item_length,
4247
data_to_item_format,
43-
TestDataItemFormat,
44-
ALL_TEST_DATA_ITEM_FORMATS,
4548
)
4649

47-
from tests.pipeline.utils import assert_query_data
48-
49-
import pyarrow as pa
50-
5150

5251
@pytest.fixture(autouse=True)
5352
def switch_to_fifo():
@@ -919,39 +918,39 @@ def some_data(
919918
def test_cursor_path_not_nullable_arrow(
920919
item_type: TestDataItemFormat,
921920
) -> None:
922-
923921
@dlt.resource
924922
def some_data(
925923
invocation: int,
926924
created_at=dlt.sources.incremental("created_at", on_cursor_value_missing="include"),
927925
):
928926
if invocation == 1:
929927
data = [
930-
{"id": 1, "created_at": 1},
931-
{"id": 2, "created_at": 1},
932-
{"id": 3, "created_at": 2},
933-
]
928+
{"id": 1, "created_at": 1},
929+
{"id": 2, "created_at": 1},
930+
{"id": 3, "created_at": 2},
931+
]
934932
elif invocation == 2:
935933
data = [
936-
{"id": 4, "created_at": 1},
937-
{"id": 5, "created_at": 2},
938-
{"id": 6, "created_at": 3},
939-
]
934+
{"id": 4, "created_at": 1},
935+
{"id": 5, "created_at": 2},
936+
{"id": 6, "created_at": 3},
937+
]
940938

941-
schema = pa.schema([
942-
pa.field('id', pa.int32(), nullable=False),
943-
pa.field('created_at', pa.int32(), nullable=False)
944-
])
945-
id_array = pa.array([item['id'] for item in data], type=pa.int32())
946-
created_at_array = pa.array([item['created_at'] for item in data], type=pa.int32())
939+
schema = pa.schema(
940+
[
941+
pa.field("id", pa.int32(), nullable=False),
942+
pa.field("created_at", pa.int32(), nullable=False),
943+
]
944+
)
945+
id_array = pa.array([item["id"] for item in data], type=pa.int32())
946+
created_at_array = pa.array([item["created_at"] for item in data], type=pa.int32())
947947
if item_type == "arrow-table":
948948
source_items = [pa.Table.from_arrays([id_array, created_at_array], schema=schema)]
949949
elif item_type == "arrow-batch":
950950
source_items = [pa.RecordBatch.from_arrays([id_array, created_at_array], schema=schema)]
951951

952952
yield source_items
953953

954-
955954
p = dlt.pipeline(pipeline_name=uniq_id())
956955
p.run(some_data(1), destination="duckdb")
957956
p.run(some_data(2), destination="duckdb")
@@ -2523,7 +2522,7 @@ def test_type_3():
25232522

25242523
@pytest.mark.parametrize("yield_pydantic", (True, False))
25252524
def test_pydantic_columns_validator(yield_pydantic: bool) -> None:
2526-
from pydantic import BaseModel, Field, ConfigDict
2525+
from pydantic import BaseModel, ConfigDict, Field
25272526

25282527
# forbid extra fields so "id" in json is not a valid field BUT
25292528
# add alias for id_ that will serde "id" correctly

tests/utils.py

Lines changed: 2 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -3,7 +3,7 @@
33
import platform
44
import sys
55
from os import environ
6-
from typing import Any, Iterable, Iterator, List, Literal, Union, get_args
6+
from typing import Any, Iterable, Iterator, Literal, Union, get_args
77
from unittest.mock import patch
88

99
import pytest
@@ -18,15 +18,14 @@
1818
from dlt.common.configuration.specs.config_providers_context import (
1919
ConfigProvidersContext,
2020
)
21-
from dlt.common.pipeline import PipelineContext
21+
from dlt.common.pipeline import PipelineContext, SupportsPipeline
2222
from dlt.common.runtime.init import init_logging
2323
from dlt.common.runtime.telemetry import start_telemetry, stop_telemetry
2424
from dlt.common.schema import Schema
2525
from dlt.common.storages import FileStorage
2626
from dlt.common.storages.versioned_storage import VersionedStorage
2727
from dlt.common.typing import StrAny, TDataItem
2828
from dlt.common.utils import custom_environ, uniq_id
29-
from dlt.common.pipeline import PipelineContext, SupportsPipeline
3029

3130
TEST_STORAGE_ROOT = "_storage"
3231

0 commit comments

Comments
 (0)