-
Notifications
You must be signed in to change notification settings - Fork 4.5k
Closed
Description
What happened?
When I tried to deploy python jdbc pipeline to dataflow from my local env, failed to deploy into dataflow and got connection error. seems to be python jdbc io trying to connect database from local env not only dataflow env.
I have checked connection and find trying to make connection from my pc.database can only accepting connection inside from dataflow net work so got connection error.
I have also checked java jdbc version and it worked fine. so python versions this behavior must be bug
class PostgresToBigQueryDataflow():
def __init__(self):
self._username = '<username>'
self._password = '<password>'
self._driver_class_name = 'org.postgresql.Driver'
self._query = "select id from beam_table;"
self._jdbc_url = 'jdbc:postgresql://<private_IP>:5432/beam'
self._project = '<project id>'
self._dataset = '<dataset>'
self._table = '<table>'
self._options = DebugOptions([
"--runner=DataflowRunner",
"--project=<project id>",
"--job_name=<job name>",
"--temp_location=gs://<project id>/tmp/",
"--region=us-central1",
"--experiments=use_runner_v2",
"--subnetwork=regions/us-central1/subnetworks/<subnet>",
])
def test(self):
JdbcToBigQuery(self._username, self._password, self._driver_class_name, self._query, self._jdbc_url, self._project, self._dataset,self._table, self._options).run()
Issue Priority
Priority: 2
Issue Component
Component: cross-language
Reactions are currently unavailable
