-
Notifications
You must be signed in to change notification settings - Fork 91
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Is it possible for the fhir-data-pipes
to sink directly into a Data Warehouse e.g. Google BigQuery?
#1191
Comments
Actually this feature is the long standing issue #455, i.e., adding BigQuery as a sink option. It should not be too hard to add this and I think it is a useful feature. The main reason we have not implemented it yet is that we have not heard much demand for it from our partners. If this is a useful feature for you and you can contribute for implementing it, I am willing to help. Side note 1: We have actually done some work in #454 to make the resulting schema similar to the BigQuery schema of Side note 2: You can import Parquet files into BigQuery; that's how the comparisons in #454 was done. |
@bashir2 I see. Initially I was thinking of using the JDBC driver for BigQuery and try and create a sample JDBC URL config for BigQuery in the DatabaseConfiguration Lines 58 to 62 in dc70755
and then make use of the fhir-data-pipes/pipelines/controller/config/application.yaml Lines 168 to 173 in dc70755
References |
@muhammad-levi your JDBC based idea can work but since we use Beam for our pipeline, I would first consider BigQueryIO; it is usually better to rely on Beam IOs when it is possible. That said, there are reasons not to use them; for example, in some places, we don't use ParquetIO for creating Parquet files (mostly because of Flink's memory overhead in the single-machine mode). |
Instead of:
fhir-data-pipes
-> Google Healthcare API FHIR Store -> Google BigQueryIt will be like:
fhir-data-pipes
-> Google BigQueryAs also suggested in this diagram
"Data Loaders" includes
fhir-data-pipes
.The text was updated successfully, but these errors were encountered: