What I learned from building a React Native iOS app - Part 3- 4 mins
Part 1 is mainly about setup and all the prep. Part 2 will be notes about problems I encountered while implementing the app.
Using the Camera
AWS has a similar image scene detector api, but I still prefer google’s image analysis api, it has many cool features like label detection, face detection, and OCR, for the app, my intention is to use the google api to identify items such as this is a can, that is a bottle etc. The usage of the api is pretty simple and straightforward, a sample request is like this:
I am gonna have to use the base64-encoded image string to request the api.
I remember I spent a big chunk of my time to make the camera flow work, but I encountered some native app specific issues. I used react-native-camera, not only you have to follow the doc to add the usage description key in xcode info.plist, I also had to do some change to link cameraroll in xcode(see this doc), and you can’t test if the camera is working with simulator, you have to connect to a real device 😅, I believe I spent more time on the configuration than actually coding for the camera flow, just another typical dev day.
The plan to use the api is: 1) take the picture 2) get the uri of the pic in photo library 3) resize the image to the minimum size to get both quality and speed 4) update the label data in redux.
This seems to be okay, but later when I got to work on the navigation, I realized that since the keyword search and image search are sharing the same result page, this page assumes the data is already there when user get to the result page. So I need the user to stay on the camera page with a loading spinner, and only navigate them to the result page when it gets the google api response, the detail of this will be discussed in another post.
Go back to the plan, I remember it felt easy to use the
yield in redux-saga, because all these file reading resizing processes are asynchronous, it just looks so much cleaner with the yield generator.
Another very native app issue is the permissions, I didn’t realize the necessity of handling logic when user denies the camera or photo library access until I was about to make the ipa file.
Find the result
copy the the src dir from node_modules/fuse.js/src and pasted it into my ./app/ dir but renamed it to fuse then use it like this
const Fuse = require('../libs/fuse');.
The search for one key word is easy but it’s kind of hard to decide the threshold parameter, because the google image api returns multiple labels, if the threshold is high the result will be very long, but if the threshold is too low, it won’t return good result, for instance, if the threshold is 0, the label from google api is
eyecare it doesn’t return
eyeware stuff like this.
The biggest lesson I learned from Navigation is you really need to have a design, it doesn’t have to be fancy, but it should at least have all the user flow so that you have a plan when you implement the routing.
I used react-native-navigation as the official RN docs said, BUT I saw a long thread discussing things to improve for this lib, and I did have some issues with it, I would say if you are familiar with other react routing libs, maybe compare the pros and cons before choosing which routing/navigation lib to use.
I mentioned earlier that after taking a picture I need to get the result first then navigate user to the result screen from camera screen, that means I need to call the navigation in saga instead of inside the component, unfortunately there is no api to use, I had to use the workaround in this thread.
The other thing I noticed is that the navigation transition sometime is slow, I am not sure if a customized transition can fix this or it’s just a performance issue of the lib.