I've been trying to run simple scripts on my Intel Edison to try the different Grove Seeed Studio sensors and cannot manage to make them work.
My board is configured correctly, its firmware is up to date, and all cables have been connected correctly as shown on different tutorials. I can connect to the board through the IDE, via SSH, and with serial.
I've been using mainly the templates provided by the Intel-XDK IDE because I'd like to develop using javascript and nodejs. The "Onboard LED Blink" template has worked perfectly. My board successfully blinked the onboard LED light so the board is clearly listening to my machine, but all the other templates have failed to make my sensors work. For example, running the "Local Temperature" template which aims to just retrieve the local temperature with the temperature sensor, just prints Sample Reading Grove Kit Temperature Sensor and stays there without printing any temperatures in the console. This has been the case for other sensors and other templates like the "Touch Notifier" template which when running also just prints Sample Reading Touch Sensor and performs no other actions when touching the touch sensor.
I tried moving away from the Intel XDK templates and trying some of Johnny-Five templates like this one and still no success. After running that particular script the console just printed Looking for connected device. I found a related post here but nobody was able to solve the issue.
It almost seems as if I am not being able to properly connect to the sensors. I get no errors on the console, just not the desired outcome. What am I doing wrong? How can I even debug this?
The samples in the XDK are at first glance a bit deceiving. The "Local Temperature" and "Touch Notifier" are designed to work with HTML5 companion apps on a phone for example (you can see this by reading the template description when you select it). Hench when you see the xdk printing the message out, it is first listening for a connection to the HTML5 app which is where it will send the temperature for display. There is an article about the "Local Temperature" app that goes into further detail and how to set everything up here: https://software.intel.com/en-us/creating-a-temperature-monitoring-app-using-intel-xdk-iot-edition
And if you want code for individual sensors, I recommend using this link here to search for code by sensor (each one should have code for c++, javascript, and python code samples): https://software.intel.com/en-us/iot/hardware/sensors
Related
I would like to make a personal application to be installed on two iPhones. The first to be used as a webcam that transmits to the second via wifi.
Having no experience with xCode, I am looking for a code example to connect 2 devices via wifi and transmit a real-time video stream.
Unfortunately, the documentation and examples I found are deprecated or partial and inconsistent.
Where can I find some code examples to help me solve my problem, preferably in ObjectiveC (but also in Swift)?
Thank you
We have a LabVIEW application that is based on sound acquisition.
I connected a Custom USB Audio device (Microphone). It is set to be the default Recording device.
But while running the application, if there is another mic source connected, then sound is being recorded/acquired by the other mic and not the default one.
So in order to have the custom microphone do the acquisition, I tried to change the default Audio Recording device through LabVIEW. But all I can do is list the audio devices connected by running a PowerShell command through System Exec function.
Is there any possible way to change the default recording device using LabVIEW? Please do mention if any.
Back in the day, I had a similar requirement, I did the following!
Get All microphones connected on the PC.
Once I get all the Device IDs, I performed a search using "Device Name" to get the Device ID I needed!
Note: If the above VI returns only one microphone, then make sure you've installed DirectX on the PC.
I hope this helps :)
Hello am very late for this information, but I can provide you with a few things you need to do.
Tap setting
Choose my device
Difficult device
Select and activate
Or
Visit Cpanel select program and tap on device activety, now choose difficult setting.
To change difficult device follow each of the steps above:
1.Setting
2.Device
3. Select difficult program
May any changes you want then done.
Helloo all
I'm currently developing an app for surface pro 3 that should be capable of:
detecting whether touchscreen was tapped and where
getting device information (product ID, amount of RAM, CPu model etc.) done by launching console application first which is capable of getting this information and saves it to a specific folder, UWP app can read its results from there and log them.
accessing sensor data like: accelerometer, gyroscope and ambient light sensor
testing cameras as i can command an app to make picture using either front or rear camera
testing microphones (both front and back)
testing speakers (i made synthesizer that is able to make beeps at given frequency at given stereo mode (left or right if both)
testing wifi - so it can connect to desired wifi network
bluetooth (swill working on it...)
I have already figured out quite a lot and put out a lot of work into it already, all listed tests are already developed on that UWP application so switching to completely different platform means rewriting whole app which I don't have time anymore.
UWP was chosen because this can run on different windows 10 devices and after completing this app the same app (with minor modifications) will be used on other windows 10 devices (like other surfaces and many different windows 10 phones)
This app will be automatically installed on a factory-resetted surface pro 3 that has no special configurations enabled, so tinkering with its settings is resource-hungry and not recommended process at all.
Now I have some other serious issues regarding of device:
how can I test the functionality of all the buttons Surface pro 3 has ?
It has 3 buttons: Volume up, Volume down and power button
but pressing power button sets screen to clack and locks device.
Can I make app override basic functionality of a button so that if button is pressed it detects it and logs its result.
same question goes to volume up and down buttons.
Only similiar questions about this are here:
Another thread on StackOverflow
I also cannot use same solution as I did with getting device information because test must be repetitive while app is running. (and UWP app cannot launch console application by itself)
Any help regarding this topic is highly welcome.
First for the power button behavior try this :
https://www.windowscentral.com/how-customize-power-button-action-when-pressed-windows-10
I don't have a surface so I cannot test it
also, I do have some inputs and thoughts about your app:
it seems to me that you are doing some sort of sanity check software for pieces of hardware I'd suggest looking into this
https://support.microsoft.com/en-my/help/4037239/surface-fix-common-surface-problems-using-surface-diagnostic-toolkit
https://www.lovemysurface.net/surface-diagnostic-toolkit/
additional thoughts of mine :
Overriding hardware behavior programmatically ould be considered as a harmful action especially when it comes to prebuilt devices such as the surface and by extension, I don't foresee MS providing API's for such a capability also such button might be communicating to the hardware directly rather and doesn't go through the software, runtime or the OS at all, changing it manually using the link i provided might be reflecting some registry settings changes but since UWP apps run in containers and cannot directly edit registry there is a dirty workaround look into this
read/write registery key file in uwp
hope this would help
You can use SystemInformation helper class from windows community toolkit it gives you a lot of details about the device.
https://learn.microsoft.com/en-us/windows/communitytoolkit/helpers/systeminformation
I'm building a Kinect application in Unity (C#) that tracks multiple people. My stumbling block is how to test this without having several live 'actors' in front of the sensor.
I know Kinect Studio allows you to record live streams, but don't think this can be played back so that it feeds a Unity app. I am using the Kinect for Windows Unity package as a starting point.
https://dev.windows.com/en-us/kinect/tools
Any Kinect or Unity experts care to chime in?
Actually you can use recorded data to playback them without any "actors".
While you are running your app (and without the need of an actual Kinect device plugged in your PC), open Kinect Studio and chose Open (Read-Only) from File. Click then on the Connect icon under the Play tab:
After that, using the Play icon (the typical triangular-shaped one), your recordings will be send to the KinectService, so you can test your app without any actors.
I am trying to drive two Grasshopper cameras, from Point Grey Research, using a laptop. As my laptop (and generally majority laptop's) does not have any FireWire 800 port, I am using a FireWire 800 to Express Card device which contains two FireWire 800 ports.
I am trying to capture sample images using the example codes provided by FlyCapture (SDK of Point Grey). When I run these examples on my laptop, they show an error "Error starting isochronous stream". However, I am able to run the same examples on a desktop and could capture images successfully.
Please note that power supply to drive the cameras doesnot seem to be a problem as both the cameras show a glowing led.
I have also written a sample code using libdc1394 library to list the active cameras connected to the machine. And upon running this code on my laptop, it shows that two cameras are connected. (which may mean that the problem lie in flycapture)
Did anyone face a similar problem ? Or could any of you successfully drive one or two Grasshopper camera using a laptop ?
Thanks a lot for the incoming help.
This is Nina from Point Grey's Support team. I can assist you with your question.
The fact that the cameras work well on your desktop PC but generate Isochronous stream error on your laptop indicates the cameras are working well. I suspect that you are running into problems with the bandwidth when you run two camera synchronously through the express card on your laptop. As the bandwidth the express card can sustain is S800 if you try to run two grasshoppers at full frame rate and resolution you might be generating more data from the cameras than the express card can sustain. Please try reducing the frame rate or resolution on your two cameras or try running one camera only and let me know if it streams well.
What FlyCapture example code are you trying to run? Which grasshopper model are you using? When you tested the camera on your PC did you run the cameras on two separate firewire buses?
Please let me know. You can also contact us by sending us an email at support#ptgrey.com
Regards,
Nina