-
Notifications
You must be signed in to change notification settings - Fork 307
Description
System information
TensorFlow version (you are using): 2.x
Are you willing to contribute it (Yes/No): Potentially as a tester / documenter
Describe the feature and the current behavior/state.
On one hand TensorFlow is currently recognized as one of the main machine learning libraries [1], in particular for neural network families of models (including auto-encoders) that integrates with additional services (e.g. TFX [2]) available on GCP.
On the other hand, BigTable is an existing GCP product that handles tables with interesting properties [3]:
- Wide table up to million of columns
- High sparsity due to schema-free definition and zero-cost saving
Training a machine learning model on such raw data would require to reduce dimension, which is one specific use case where TensorFlow is good at. Precisely, the training of deep and WIDE model would be in scope [4].
Combining both technologies seems to be a great fit.
Will this change the current api? How?
Can't say
Who will benefit with this feature?
Data engineering/data science community using GCP and sourcing BigTable data inside TensorFlow, typically for dimensionality reduction and sparse data use cases (e.g fraud detection that leverages dynamically aggregated values [5]).
[1] https://www.upgrad.com/blog/top-deep-learning-frameworks/
[2] https://www.tensorflow.org/tfx
[3] https://cloud.google.com/bigtable/docs/schema-design
[4] https://ai.googleblog.com/2016/06/wide-deep-learning-better-together-with.html
[5] https://cloud.google.com/solutions/machine-learning/minimizing-predictive-serving-latency-in-machine-learning#precomputing_and_caching_predictions