Skip to content

Commit 43dd612

Browse files
chore: use gapic-generator-python 0.53.4 (#341)
- [ ] Regenerate this pull request now. docs: list oneofs in docstring fix(deps): require google-api-core >= 1.28.0 fix(deps): drop packaging dependency committer: busunkim96@ PiperOrigin-RevId: 406468269 Source-Link: googleapis/googleapis@83d81b0 Source-Link: googleapis/googleapis-gen@2ff001f Copy-Tag: eyJwIjoiLmdpdGh1Yi8uT3dsQm90LnlhbWwiLCJoIjoiMmZmMDAxZmJhY2I5ZTc3ZTcxZDczNGRlNWY5NTVjMDVmZGFlODUyNiJ9
1 parent af26939 commit 43dd612

26 files changed

Lines changed: 198 additions & 593 deletions

File tree

packages/google-cloud-bigquery-storage/google/cloud/bigquery_storage_v1/services/big_query_read/async_client.py

Lines changed: 12 additions & 10 deletions
Original file line numberDiff line numberDiff line change
@@ -19,13 +19,15 @@
1919
from typing import Dict, AsyncIterable, Awaitable, Sequence, Tuple, Type, Union
2020
import pkg_resources
2121

22-
import google.api_core.client_options as ClientOptions # type: ignore
22+
from google.api_core.client_options import ClientOptions # type: ignore
2323
from google.api_core import exceptions as core_exceptions # type: ignore
2424
from google.api_core import gapic_v1 # type: ignore
2525
from google.api_core import retry as retries # type: ignore
2626
from google.auth import credentials as ga_credentials # type: ignore
2727
from google.oauth2 import service_account # type: ignore
2828

29+
OptionalRetry = Union[retries.Retry, object]
30+
2931
from google.cloud.bigquery_storage_v1.types import arrow
3032
from google.cloud.bigquery_storage_v1.types import avro
3133
from google.cloud.bigquery_storage_v1.types import storage
@@ -168,12 +170,12 @@ def __init__(
168170

169171
async def create_read_session(
170172
self,
171-
request: storage.CreateReadSessionRequest = None,
173+
request: Union[storage.CreateReadSessionRequest, dict] = None,
172174
*,
173175
parent: str = None,
174176
read_session: stream.ReadSession = None,
175177
max_stream_count: int = None,
176-
retry: retries.Retry = gapic_v1.method.DEFAULT,
178+
retry: OptionalRetry = gapic_v1.method.DEFAULT,
177179
timeout: float = None,
178180
metadata: Sequence[Tuple[str, str]] = (),
179181
) -> stream.ReadSession:
@@ -202,7 +204,7 @@ async def create_read_session(
202204
caller.
203205
204206
Args:
205-
request (:class:`google.cloud.bigquery_storage_v1.types.CreateReadSessionRequest`):
207+
request (Union[google.cloud.bigquery_storage_v1.types.CreateReadSessionRequest, dict]):
206208
The request object. Request message for
207209
`CreateReadSession`.
208210
parent (:class:`str`):
@@ -301,11 +303,11 @@ async def create_read_session(
301303

302304
def read_rows(
303305
self,
304-
request: storage.ReadRowsRequest = None,
306+
request: Union[storage.ReadRowsRequest, dict] = None,
305307
*,
306308
read_stream: str = None,
307309
offset: int = None,
308-
retry: retries.Retry = gapic_v1.method.DEFAULT,
310+
retry: OptionalRetry = gapic_v1.method.DEFAULT,
309311
timeout: float = None,
310312
metadata: Sequence[Tuple[str, str]] = (),
311313
) -> Awaitable[AsyncIterable[storage.ReadRowsResponse]]:
@@ -319,7 +321,7 @@ def read_rows(
319321
reflecting the current state of the stream.
320322
321323
Args:
322-
request (:class:`google.cloud.bigquery_storage_v1.types.ReadRowsRequest`):
324+
request (Union[google.cloud.bigquery_storage_v1.types.ReadRowsRequest, dict]):
323325
The request object. Request message for `ReadRows`.
324326
read_stream (:class:`str`):
325327
Required. Stream to read rows from.
@@ -400,9 +402,9 @@ def read_rows(
400402

401403
async def split_read_stream(
402404
self,
403-
request: storage.SplitReadStreamRequest = None,
405+
request: Union[storage.SplitReadStreamRequest, dict] = None,
404406
*,
405-
retry: retries.Retry = gapic_v1.method.DEFAULT,
407+
retry: OptionalRetry = gapic_v1.method.DEFAULT,
406408
timeout: float = None,
407409
metadata: Sequence[Tuple[str, str]] = (),
408410
) -> storage.SplitReadStreamResponse:
@@ -421,7 +423,7 @@ async def split_read_stream(
421423
once the streams have been read to completion.
422424
423425
Args:
424-
request (:class:`google.cloud.bigquery_storage_v1.types.SplitReadStreamRequest`):
426+
request (Union[google.cloud.bigquery_storage_v1.types.SplitReadStreamRequest, dict]):
425427
The request object. Request message for
426428
`SplitReadStream`.
427429
retry (google.api_core.retry.Retry): Designation of what errors, if any,

packages/google-cloud-bigquery-storage/google/cloud/bigquery_storage_v1/services/big_query_read/client.py

Lines changed: 5 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -30,6 +30,8 @@
3030
from google.auth.exceptions import MutualTLSChannelError # type: ignore
3131
from google.oauth2 import service_account # type: ignore
3232

33+
OptionalRetry = Union[retries.Retry, object]
34+
3335
from google.cloud.bigquery_storage_v1.types import arrow
3436
from google.cloud.bigquery_storage_v1.types import avro
3537
from google.cloud.bigquery_storage_v1.types import storage
@@ -391,7 +393,7 @@ def create_read_session(
391393
parent: str = None,
392394
read_session: stream.ReadSession = None,
393395
max_stream_count: int = None,
394-
retry: retries.Retry = gapic_v1.method.DEFAULT,
396+
retry: OptionalRetry = gapic_v1.method.DEFAULT,
395397
timeout: float = None,
396398
metadata: Sequence[Tuple[str, str]] = (),
397399
) -> stream.ReadSession:
@@ -513,7 +515,7 @@ def read_rows(
513515
*,
514516
read_stream: str = None,
515517
offset: int = None,
516-
retry: retries.Retry = gapic_v1.method.DEFAULT,
518+
retry: OptionalRetry = gapic_v1.method.DEFAULT,
517519
timeout: float = None,
518520
metadata: Sequence[Tuple[str, str]] = (),
519521
) -> Iterable[storage.ReadRowsResponse]:
@@ -601,7 +603,7 @@ def split_read_stream(
601603
self,
602604
request: Union[storage.SplitReadStreamRequest, dict] = None,
603605
*,
604-
retry: retries.Retry = gapic_v1.method.DEFAULT,
606+
retry: OptionalRetry = gapic_v1.method.DEFAULT,
605607
timeout: float = None,
606608
metadata: Sequence[Tuple[str, str]] = (),
607609
) -> storage.SplitReadStreamResponse:

packages/google-cloud-bigquery-storage/google/cloud/bigquery_storage_v1/services/big_query_read/transports/base.py

Lines changed: 1 addition & 34 deletions
Original file line numberDiff line numberDiff line change
@@ -15,7 +15,6 @@
1515
#
1616
import abc
1717
from typing import Awaitable, Callable, Dict, Optional, Sequence, Union
18-
import packaging.version
1918
import pkg_resources
2019

2120
import google.auth # type: ignore
@@ -38,15 +37,6 @@
3837
except pkg_resources.DistributionNotFound:
3938
DEFAULT_CLIENT_INFO = gapic_v1.client_info.ClientInfo()
4039

41-
try:
42-
# google.auth.__version__ was added in 1.26.0
43-
_GOOGLE_AUTH_VERSION = google.auth.__version__
44-
except AttributeError:
45-
try: # try pkg_resources if it is available
46-
_GOOGLE_AUTH_VERSION = pkg_resources.get_distribution("google-auth").version
47-
except pkg_resources.DistributionNotFound: # pragma: NO COVER
48-
_GOOGLE_AUTH_VERSION = None
49-
5040

5141
class BigQueryReadTransport(abc.ABC):
5242
"""Abstract transport class for BigQueryRead."""
@@ -100,7 +90,7 @@ def __init__(
10090
host += ":443"
10191
self._host = host
10292

103-
scopes_kwargs = self._get_scopes_kwargs(self._host, scopes)
93+
scopes_kwargs = {"scopes": scopes, "default_scopes": self.AUTH_SCOPES}
10494

10595
# Save the scopes.
10696
self._scopes = scopes
@@ -133,29 +123,6 @@ def __init__(
133123
# Save the credentials.
134124
self._credentials = credentials
135125

136-
# TODO(busunkim): This method is in the base transport
137-
# to avoid duplicating code across the transport classes. These functions
138-
# should be deleted once the minimum required versions of google-auth is increased.
139-
140-
# TODO: Remove this function once google-auth >= 1.25.0 is required
141-
@classmethod
142-
def _get_scopes_kwargs(
143-
cls, host: str, scopes: Optional[Sequence[str]]
144-
) -> Dict[str, Optional[Sequence[str]]]:
145-
"""Returns scopes kwargs to pass to google-auth methods depending on the google-auth version"""
146-
147-
scopes_kwargs = {}
148-
149-
if _GOOGLE_AUTH_VERSION and (
150-
packaging.version.parse(_GOOGLE_AUTH_VERSION)
151-
>= packaging.version.parse("1.25.0")
152-
):
153-
scopes_kwargs = {"scopes": scopes, "default_scopes": cls.AUTH_SCOPES}
154-
else:
155-
scopes_kwargs = {"scopes": scopes or cls.AUTH_SCOPES}
156-
157-
return scopes_kwargs
158-
159126
def _prep_wrapped_messages(self, client_info):
160127
# Precompute the wrapped methods.
161128
self._wrapped_methods = {

packages/google-cloud-bigquery-storage/google/cloud/bigquery_storage_v1/services/big_query_read/transports/grpc_asyncio.py

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -20,7 +20,6 @@
2020
from google.api_core import grpc_helpers_async # type: ignore
2121
from google.auth import credentials as ga_credentials # type: ignore
2222
from google.auth.transport.grpc import SslCredentials # type: ignore
23-
import packaging.version
2423

2524
import grpc # type: ignore
2625
from grpc.experimental import aio # type: ignore

packages/google-cloud-bigquery-storage/google/cloud/bigquery_storage_v1/services/big_query_write/async_client.py

Lines changed: 19 additions & 17 deletions
Original file line numberDiff line numberDiff line change
@@ -28,13 +28,15 @@
2828
)
2929
import pkg_resources
3030

31-
import google.api_core.client_options as ClientOptions # type: ignore
31+
from google.api_core.client_options import ClientOptions # type: ignore
3232
from google.api_core import exceptions as core_exceptions # type: ignore
3333
from google.api_core import gapic_v1 # type: ignore
3434
from google.api_core import retry as retries # type: ignore
3535
from google.auth import credentials as ga_credentials # type: ignore
3636
from google.oauth2 import service_account # type: ignore
3737

38+
OptionalRetry = Union[retries.Retry, object]
39+
3840
from google.cloud.bigquery_storage_v1.types import storage
3941
from google.cloud.bigquery_storage_v1.types import stream
4042
from google.cloud.bigquery_storage_v1.types import table
@@ -181,11 +183,11 @@ def __init__(
181183

182184
async def create_write_stream(
183185
self,
184-
request: storage.CreateWriteStreamRequest = None,
186+
request: Union[storage.CreateWriteStreamRequest, dict] = None,
185187
*,
186188
parent: str = None,
187189
write_stream: stream.WriteStream = None,
188-
retry: retries.Retry = gapic_v1.method.DEFAULT,
190+
retry: OptionalRetry = gapic_v1.method.DEFAULT,
189191
timeout: float = None,
190192
metadata: Sequence[Tuple[str, str]] = (),
191193
) -> stream.WriteStream:
@@ -198,7 +200,7 @@ async def create_write_stream(
198200
received.
199201
200202
Args:
201-
request (:class:`google.cloud.bigquery_storage_v1.types.CreateWriteStreamRequest`):
203+
request (Union[google.cloud.bigquery_storage_v1.types.CreateWriteStreamRequest, dict]):
202204
The request object. Request message for
203205
`CreateWriteStream`.
204206
parent (:class:`str`):
@@ -280,7 +282,7 @@ def append_rows(
280282
self,
281283
requests: AsyncIterator[storage.AppendRowsRequest] = None,
282284
*,
283-
retry: retries.Retry = gapic_v1.method.DEFAULT,
285+
retry: OptionalRetry = gapic_v1.method.DEFAULT,
284286
timeout: float = None,
285287
metadata: Sequence[Tuple[str, str]] = (),
286288
) -> Awaitable[AsyncIterable[storage.AppendRowsResponse]]:
@@ -366,17 +368,17 @@ def append_rows(
366368

367369
async def get_write_stream(
368370
self,
369-
request: storage.GetWriteStreamRequest = None,
371+
request: Union[storage.GetWriteStreamRequest, dict] = None,
370372
*,
371373
name: str = None,
372-
retry: retries.Retry = gapic_v1.method.DEFAULT,
374+
retry: OptionalRetry = gapic_v1.method.DEFAULT,
373375
timeout: float = None,
374376
metadata: Sequence[Tuple[str, str]] = (),
375377
) -> stream.WriteStream:
376378
r"""Gets information about a write stream.
377379
378380
Args:
379-
request (:class:`google.cloud.bigquery_storage_v1.types.GetWriteStreamRequest`):
381+
request (Union[google.cloud.bigquery_storage_v1.types.GetWriteStreamRequest, dict]):
380382
The request object. Request message for
381383
`GetWriteStreamRequest`.
382384
name (:class:`str`):
@@ -448,18 +450,18 @@ async def get_write_stream(
448450

449451
async def finalize_write_stream(
450452
self,
451-
request: storage.FinalizeWriteStreamRequest = None,
453+
request: Union[storage.FinalizeWriteStreamRequest, dict] = None,
452454
*,
453455
name: str = None,
454-
retry: retries.Retry = gapic_v1.method.DEFAULT,
456+
retry: OptionalRetry = gapic_v1.method.DEFAULT,
455457
timeout: float = None,
456458
metadata: Sequence[Tuple[str, str]] = (),
457459
) -> storage.FinalizeWriteStreamResponse:
458460
r"""Finalize a write stream so that no new data can be appended to
459461
the stream. Finalize is not supported on the '_default' stream.
460462
461463
Args:
462-
request (:class:`google.cloud.bigquery_storage_v1.types.FinalizeWriteStreamRequest`):
464+
request (Union[google.cloud.bigquery_storage_v1.types.FinalizeWriteStreamRequest, dict]):
463465
The request object. Request message for invoking
464466
`FinalizeWriteStream`.
465467
name (:class:`str`):
@@ -528,10 +530,10 @@ async def finalize_write_stream(
528530

529531
async def batch_commit_write_streams(
530532
self,
531-
request: storage.BatchCommitWriteStreamsRequest = None,
533+
request: Union[storage.BatchCommitWriteStreamsRequest, dict] = None,
532534
*,
533535
parent: str = None,
534-
retry: retries.Retry = gapic_v1.method.DEFAULT,
536+
retry: OptionalRetry = gapic_v1.method.DEFAULT,
535537
timeout: float = None,
536538
metadata: Sequence[Tuple[str, str]] = (),
537539
) -> storage.BatchCommitWriteStreamsResponse:
@@ -543,7 +545,7 @@ async def batch_commit_write_streams(
543545
becomes available for read operations.
544546
545547
Args:
546-
request (:class:`google.cloud.bigquery_storage_v1.types.BatchCommitWriteStreamsRequest`):
548+
request (Union[google.cloud.bigquery_storage_v1.types.BatchCommitWriteStreamsRequest, dict]):
547549
The request object. Request message for
548550
`BatchCommitWriteStreams`.
549551
parent (:class:`str`):
@@ -613,10 +615,10 @@ async def batch_commit_write_streams(
613615

614616
async def flush_rows(
615617
self,
616-
request: storage.FlushRowsRequest = None,
618+
request: Union[storage.FlushRowsRequest, dict] = None,
617619
*,
618620
write_stream: str = None,
619-
retry: retries.Retry = gapic_v1.method.DEFAULT,
621+
retry: OptionalRetry = gapic_v1.method.DEFAULT,
620622
timeout: float = None,
621623
metadata: Sequence[Tuple[str, str]] = (),
622624
) -> storage.FlushRowsResponse:
@@ -632,7 +634,7 @@ async def flush_rows(
632634
BUFFERED.
633635
634636
Args:
635-
request (:class:`google.cloud.bigquery_storage_v1.types.FlushRowsRequest`):
637+
request (Union[google.cloud.bigquery_storage_v1.types.FlushRowsRequest, dict]):
636638
The request object. Request message for `FlushRows`.
637639
write_stream (:class:`str`):
638640
Required. The stream that is the

packages/google-cloud-bigquery-storage/google/cloud/bigquery_storage_v1/services/big_query_write/client.py

Lines changed: 8 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -30,6 +30,8 @@
3030
from google.auth.exceptions import MutualTLSChannelError # type: ignore
3131
from google.oauth2 import service_account # type: ignore
3232

33+
OptionalRetry = Union[retries.Retry, object]
34+
3335
from google.cloud.bigquery_storage_v1.types import storage
3436
from google.cloud.bigquery_storage_v1.types import stream
3537
from google.cloud.bigquery_storage_v1.types import table
@@ -374,7 +376,7 @@ def create_write_stream(
374376
*,
375377
parent: str = None,
376378
write_stream: stream.WriteStream = None,
377-
retry: retries.Retry = gapic_v1.method.DEFAULT,
379+
retry: OptionalRetry = gapic_v1.method.DEFAULT,
378380
timeout: float = None,
379381
metadata: Sequence[Tuple[str, str]] = (),
380382
) -> stream.WriteStream:
@@ -459,7 +461,7 @@ def append_rows(
459461
self,
460462
requests: Iterator[storage.AppendRowsRequest] = None,
461463
*,
462-
retry: retries.Retry = gapic_v1.method.DEFAULT,
464+
retry: OptionalRetry = gapic_v1.method.DEFAULT,
463465
timeout: float = None,
464466
metadata: Sequence[Tuple[str, str]] = (),
465467
) -> Iterable[storage.AppendRowsResponse]:
@@ -535,7 +537,7 @@ def get_write_stream(
535537
request: Union[storage.GetWriteStreamRequest, dict] = None,
536538
*,
537539
name: str = None,
538-
retry: retries.Retry = gapic_v1.method.DEFAULT,
540+
retry: OptionalRetry = gapic_v1.method.DEFAULT,
539541
timeout: float = None,
540542
metadata: Sequence[Tuple[str, str]] = (),
541543
) -> stream.WriteStream:
@@ -607,7 +609,7 @@ def finalize_write_stream(
607609
request: Union[storage.FinalizeWriteStreamRequest, dict] = None,
608610
*,
609611
name: str = None,
610-
retry: retries.Retry = gapic_v1.method.DEFAULT,
612+
retry: OptionalRetry = gapic_v1.method.DEFAULT,
611613
timeout: float = None,
612614
metadata: Sequence[Tuple[str, str]] = (),
613615
) -> storage.FinalizeWriteStreamResponse:
@@ -677,7 +679,7 @@ def batch_commit_write_streams(
677679
request: Union[storage.BatchCommitWriteStreamsRequest, dict] = None,
678680
*,
679681
parent: str = None,
680-
retry: retries.Retry = gapic_v1.method.DEFAULT,
682+
retry: OptionalRetry = gapic_v1.method.DEFAULT,
681683
timeout: float = None,
682684
metadata: Sequence[Tuple[str, str]] = (),
683685
) -> storage.BatchCommitWriteStreamsResponse:
@@ -754,7 +756,7 @@ def flush_rows(
754756
request: Union[storage.FlushRowsRequest, dict] = None,
755757
*,
756758
write_stream: str = None,
757-
retry: retries.Retry = gapic_v1.method.DEFAULT,
759+
retry: OptionalRetry = gapic_v1.method.DEFAULT,
758760
timeout: float = None,
759761
metadata: Sequence[Tuple[str, str]] = (),
760762
) -> storage.FlushRowsResponse:

0 commit comments

Comments
 (0)