· 6 years ago · Mar 05, 2019, 07:40 AM
1From d25f7ae3a562d1d747a575a0c30fcf7eb9ffadc9 Mon Sep 17 00:00:00 2001
2From: =?UTF-8?q?Kamil=20Bregu=C5=82a?= <kamil.bregula@polidea.com>
3Date: Sat, 1 Jan 2000 00:00:00 +0000
4Subject: Highlight format element
5
6---
7 airflow/contrib/hooks/bigquery_hook.py | 14 +++++++-------
8 airflow/contrib/hooks/gcp_sql_hook.py | 4 ++--
9 airflow/contrib/operators/bigquery_operator.py | 6 +++---
10 .../operators/bigquery_table_delete_operator.py | 2 +-
11 airflow/contrib/operators/bigquery_to_gcs.py | 2 +-
12 airflow/contrib/operators/gcs_to_bq.py | 4 ++--
13 6 files changed, 16 insertions(+), 16 deletions(-)
14
15diff --git a/airflow/contrib/hooks/bigquery_hook.py b/airflow/contrib/hooks/bigquery_hook.py
16index 4fb2d514..2b23769b 100644
17--- a/airflow/contrib/hooks/bigquery_hook.py
18+++ b/airflow/contrib/hooks/bigquery_hook.py
19@@ -347,9 +347,9 @@ class BigQueryBaseCursor(LoggingMixin):
20 for more details about these parameters.
21
22 :param external_project_dataset_table:
23- The dotted (<project>.|<project>:)<dataset>.<table>($<partition>) BigQuery
24+ The dotted ``(<project>.|<project>:)<dataset>.<table>($<partition>)`` BigQuery
25 table name to create external table.
26- If <project> is not included, project will be the
27+ If ``<project>`` is not included, project will be the
28 project defined in the connection json.
29 :type external_project_dataset_table: str
30 :param schema_fields: The schema field list as defined here:
31@@ -654,7 +654,7 @@ class BigQueryBaseCursor(LoggingMixin):
32
33 :param sql: The BigQuery SQL to execute.
34 :type sql: str
35- :param destination_dataset_table: The dotted <dataset>.<table>
36+ :param destination_dataset_table: The dotted ``<dataset>.<table>``
37 BigQuery table to save the query results.
38 :type destination_dataset_table: str
39 :param write_disposition: What to do if the table already exists in
40@@ -863,7 +863,7 @@ class BigQueryBaseCursor(LoggingMixin):
41
42 For more details about these parameters.
43
44- :param source_project_dataset_table: The dotted <dataset>.<table>
45+ :param source_project_dataset_table: The dotted ``<dataset>.<table>``
46 BigQuery table to use as the source data.
47 :type source_project_dataset_table: str
48 :param destination_cloud_storage_uris: The destination Google Cloud
49@@ -932,7 +932,7 @@ class BigQueryBaseCursor(LoggingMixin):
50 ``(project:|project.)<dataset>.<table>``
51 BigQuery tables to use as the source data. Use a list if there are
52 multiple source tables.
53- If <project> is not included, project will be the project defined
54+ If ``<project>`` is not included, project will be the project defined
55 in the connection json.
56 :type source_project_dataset_tables: list|string
57 :param destination_project_dataset_table: The destination BigQuery
58@@ -1015,8 +1015,8 @@ class BigQueryBaseCursor(LoggingMixin):
59 For more details about these parameters.
60
61 :param destination_project_dataset_table:
62- The dotted (<project>.|<project>:)<dataset>.<table>($<partition>) BigQuery
63- table to load data into. If <project> is not included, project will be the
64+ The dotted ``(<project>.|<project>:)<dataset>.<table>($<partition>)`` BigQuery
65+ table to load data into. If ``<project>`` is not included, project will be the
66 project defined in the connection json. If a partition is specified the
67 operator will automatically append the data, create a new partition or create
68 a new DAY partitioned table.
69diff --git a/airflow/contrib/hooks/gcp_sql_hook.py b/airflow/contrib/hooks/gcp_sql_hook.py
70index 7b386e66..c168ef7d 100644
71--- a/airflow/contrib/hooks/gcp_sql_hook.py
72+++ b/airflow/contrib/hooks/gcp_sql_hook.py
73@@ -407,9 +407,9 @@ class CloudSqlProxyRunner(LoggingMixin):
74 :param instance_specification: Specification of the instance to connect the
75 proxy to. It should be specified in the form that is described in
76 https://cloud.google.com/sql/docs/mysql/sql-proxy#multiple-instances in
77- -instances parameter (typically in the form of <project>:<region>:<instance>
78+ -instances parameter (typically in the form of ``<project>:<region>:<instance>``
79 for UNIX socket connections and in the form of
80- <project>:<region>:<instance>=tcp:<port> for TCP connections.
81+ ``<project>:<region>:<instance>=tcp:<port>`` for TCP connections.
82 :type instance_specification: str
83 :param gcp_conn_id: Id of Google Cloud Platform connection to use for
84 authentication
85diff --git a/airflow/contrib/operators/bigquery_operator.py b/airflow/contrib/operators/bigquery_operator.py
86index f3a43baa..b5d27f2e 100644
87--- a/airflow/contrib/operators/bigquery_operator.py
88+++ b/airflow/contrib/operators/bigquery_operator.py
89@@ -34,7 +34,7 @@ class BigQueryOperator(BaseOperator):
90 a list of str (sql statements), or reference to a template file.
91 Template reference are recognized by str ending in '.sql'.
92 :param destination_dataset_table: A dotted
93- (<project>.|<project>:)<dataset>.<table> that, if set, will store the results
94+ ``(<project>.|<project>:)<dataset>.<table>`` that, if set, will store the results
95 of the query. (templated)
96 :type destination_dataset_table: str
97 :param write_disposition: Specifies the action that occurs if the destination table
98@@ -357,8 +357,8 @@ class BigQueryCreateExternalTableOperator(BaseOperator):
99 table to. (templated)
100 If source_format is 'DATASTORE_BACKUP', the list must only contain a single URI.
101 :type source_objects: list
102- :param destination_project_dataset_table: The dotted (<project>.)<dataset>.<table>
103- BigQuery table to load data into (templated). If <project> is not included,
104+ :param destination_project_dataset_table: The dotted ``(<project>.)<dataset>.<table>``
105+ BigQuery table to load data into (templated). If ``<project>`` is not included,
106 project will be the project defined in the connection json.
107 :type destination_project_dataset_table: str
108 :param schema_fields: If set, the schema field list as defined here:
109diff --git a/airflow/contrib/operators/bigquery_table_delete_operator.py b/airflow/contrib/operators/bigquery_table_delete_operator.py
110index 45c48145..106afd1e 100644
111--- a/airflow/contrib/operators/bigquery_table_delete_operator.py
112+++ b/airflow/contrib/operators/bigquery_table_delete_operator.py
113@@ -27,7 +27,7 @@ class BigQueryTableDeleteOperator(BaseOperator):
114 Deletes BigQuery tables
115
116 :param deletion_dataset_table: A dotted
117- (<project>.|<project>:)<dataset>.<table> that indicates which table
118+ ``(<project>.|<project>:)<dataset>.<table>`` that indicates which table
119 will be deleted. (templated)
120 :type deletion_dataset_table: str
121 :param bigquery_conn_id: reference to a specific BigQuery hook.
122diff --git a/airflow/contrib/operators/bigquery_to_gcs.py b/airflow/contrib/operators/bigquery_to_gcs.py
123index ec6b937d..19b004fd 100644
124--- a/airflow/contrib/operators/bigquery_to_gcs.py
125+++ b/airflow/contrib/operators/bigquery_to_gcs.py
126@@ -32,7 +32,7 @@ class BigQueryToCloudStorageOperator(BaseOperator):
127
128 :param source_project_dataset_table: The dotted
129 ``(<project>.|<project>:)<dataset>.<table>`` BigQuery table to use as the
130- source data. If <project> is not included, project will be the project
131+ source data. If ``<project>`` is not included, project will be the project
132 defined in the connection json. (templated)
133 :type source_project_dataset_table: str
134 :param destination_cloud_storage_uris: The destination Google Cloud
135diff --git a/airflow/contrib/operators/gcs_to_bq.py b/airflow/contrib/operators/gcs_to_bq.py
136index c26b2fa1..f57b0d85 100644
137--- a/airflow/contrib/operators/gcs_to_bq.py
138+++ b/airflow/contrib/operators/gcs_to_bq.py
139@@ -43,8 +43,8 @@ class GoogleCloudStorageToBigQueryOperator(BaseOperator):
140 :param source_objects: List of Google cloud storage URIs to load from. (templated)
141 If source_format is 'DATASTORE_BACKUP', the list must only contain a single URI.
142 :type source_objects: list[str]
143- :param destination_project_dataset_table: The dotted (<project>.)<dataset>.<table>
144- BigQuery table to load data into. If <project> is not included,
145+ :param destination_project_dataset_table: The dotted ``(<project>.)<dataset>.<table>``
146+ BigQuery table to load data into. If ``<project>`` is not included,
147 project will be the project defined in the connection json. (templated)
148 :type destination_project_dataset_table: str
149 :param schema_fields: If set, the schema field list as defined here:
150--
1512.20.1