I am hoping someone can provide some architecture advice on a project I am working on. First I will outline what we are trying to achieve and then I will outline the direction that I am currently considering, and if anyone has any advice or constructive criticism that would be great.
High Level Objective
Using particle Boron’s to collect end-user data and send via the cellular network to an online database. This data will then need to cleaned to remove any obvious outliers. We will also be calculating averages and fluctuations in the data over time. This data will then be presented to the individual end users via a dashboard behind a login wall on our website.
Current Proposed Strategy
Collect Data via particle, publish as an event.
Transfer data from Particle cloud services to Google BigQuery using stitch and a webhook.
Use some Google cloud service to filter data and calculate metrics.
Potentially transfer data to clean database
Display relevant data to logged in user on our website.
I am a bit unsure if the google portion of the above is the best way to go about doing this. I am also looking into embedded BI softwares such as SISENSE, which, I believe, can connect to a data source, filter and display the data on our website.
I am hoping someone here can give some guidance as to whether I am going about this the right way, or if I am overly complicating things. Thanks in advance.