- Kevin Trisnadi Nugroho -
[email protected]
- Herlina Budiarti -
[email protected]
- Rastika Intan Nastiti -
[email protected]
-
The android app project is named
ENDF
, located inside Android folder in the repo.
So, if you have cloned our repo, in order to build the app, you have to import theENDF
folder into the Android Studio, not the parent repo folder:capstone
. -
The whole project is written in
Kotlin
. -
In order for the app to run, you have to provide the following before building the app:
-
This is required under the assumption that you have set up a collection on Firebase and populate it with data you want to visualize in the app. In our context, it is the earthquakes record from BMKG. This credential is needed to authenticate the request for fetching the data from the app. To setup the credential, place the google-services.json file inside the app folder of the project.
However, you can go with other alternatives on how and where you want to host your data. This way, the firebase credential is no longer needed.
-
Place the API key in
gradle.properties
andlocal.properties
file. Both are located inapp
folder. This API key is needed for two main servces:Markers
andPlaces Autocomplete
.For the detailed documentation, refer to:
-
- Install Python3 on your device.
- Run the following commands:
sudo apt install python3-pip
pip3 install pandas numpy tensorflow matplotlib sklearn
- Run train.py to generate new model. It will be saved as a tensorflow saved_model format.
- Create a GCP Project.
- Enable the necessary APIs and SDKs.
Firestore, Cloud Run, Cloud Storage, Compute Engine, Container Registry, Maps, Places
- Create an API key that will be used by your Android app to authenticate.
- Connect your Android app to Firebase, this is required in order to be able to use the Firebase SDK.
- Create a Compute Engine instance, this will be used to deploy the flask app and run the scraping scripts.
- Use the Dockerfile to build a custom image based on the TensorFlow Serving image >This Dockerfile will copy the ML folder which contains the model into the container and run the necessary commands
- Upload the image to the Container Registry.
- Deploy the image from Container Registry to Cloud Run.
- Make changes in the flask app accordingly.
- Copy the flask app into the compute engine instance.
- Set up the compute engine instance to run flask. >You may refer to this guide for AWS EC2: https://www.datasciencebytes.com/bytes/2015/02/24/running-a-flask-app-on-aws-ec2/
- Make changes in the scraping scripts accordingly.
- Copy the scripts into the compute engine instance.
- Run the scripts to start scraping.