Custom model support is not there in https://rnfirebase.io/
Does nay other project support ML Kit on custom model?
ML Kit may not have support for React Native in Custom Model, please consider to use TFLite directly.
There are a couple of projects doing it, for example:
https://www.npmjs.com/search?q=react%20native%20tensorflow%20lite
Related
I have an FYP project (a social media app like Instagram) that requires me to create a simple recommendation system. I've trained my dataset on cosine similarity using Python, but I'm at a loss on what to do next. How can I integrate my trained ml model to React native or if there is a better and easier way to make a recommendation system?
I tried reading documentation and watching videos. But I still don't seem to be able to grasp some difficult concepts. I would greatly appreciate it if you could give me instructions or steps on what to learn after training my model. Or if I have to use some library or packages etc. [not sure if this is the appropriate forum for this inquiry]
I wanted came across this guide https://developers.google.com/ml-kit/vision/pose-detection/classifying-poses and I wanted to develop a cross plateform app with pose classification using react native.
I haven't been able to find any wrapper from react-native ? Is this going to be developped one day ?
I thought about using the flutter one but it seems that it doesn't contain the pose detection library ?
ML Kit itself does not currently have a plan to provide React Native or Flutter wrappers. Some developers have come up their own wrappers, e.g. https://pub.dev/packages/google_ml_kit. They are not officially endorsed by ML Kit and your mileage may vary.
For React Native we have a wrapper for ML Kit https://github.com/a7medev/react-native-ml-kit but pose detection is not implemented yet.
I am trying to serve a deep learning model from mobiles. My react native app using tensorflow takes about a minute just to load. My model is about 175mb (about 30 million params). This is my first time trying to run a model on mobiles and I couldn't find any good performance data for tensorflow js on react native.
Is my model too large to expect a reasonably quick loading and inference time on react native? Is this because of hardware limitation or framework? I read that tfjs react native uses web-gl which would be slower than direct access the mobile's gpu so could I use core ml or something like that and expect a better time?
As an extra question (I'll also make a separate post), another route I'm considering is moving inference to a web browser for laptops/desktops. Could I expect a browser web-gl to perform as well as directly using the computer gpu?
Yes, you ML Kit and Core ML are much faster than TF.js.
Also you can create your model, convert it to TF lite then deploy it to firebase and use your model on ios and android in the native side, (currently #react-native-firebase/ml. does not support custom models).
I want to use Kaggle flower dataset (with 4242 images uploaded) for flower recognition app with react native tensorflow.
However, currently, I am sticking on a stupid mistake.
I am unable to change flowers.npz and label.npz files after data preparation followed by kaggle official website(https://www.kaggle.com/jrmst102/flow-ers-data-preparation/notebook) into tensorflow_graph.pb and tensorflow_label.txt formats.
So questions are as follows :
How to transform flowers.npz and labels.npz files into tensorflow_graph.pb and tensorflow_label.txt extensions respectively?
Or how to properly upload this pretrained flower dataset into react native application, for testing purposes with react native tensorflow?
Googling all over around did not help me.
Any advice is appreciated!
P.S. In case, I am developing an app using CRNA
Could somebody tell me whether it is possible to develop a chatbot using python ML frameworks such as tensorflow and deploy in Slack using Slack's apps?
As far as I have read we could develop some retrieval based model using node.js. But I'm looking for a generative model.
Anything to help me get started is much appreciated.
Thanks!
In order to use a generative model you have to use tensorflow sequence to sequence architecture . Here's the official TF tutorial from TF on sequence-to-sequence architecture .
So after training the model you can set this as an API which is easy. Then may be you can set it with the slack API for sure.