- Search Results
I’m working in a research-project trying to collect gaze data in a non-test scenario. (People are moving in front of the screen)
I’d like to get a more robust gaze detection, by restricting the track box, or manipulate the recorded images of the EyeTracker(X2-60).
The Analytics SDK 3.0 doesn’t support manipulating of the track box nor gives a possibility to get the images.
I look forward to hear from you.
Topic: OSC output
Has anyone seen or developed an OSC connection setup for the Tobii x2-60 ?
I am an interactive artist and such a connection is essential for a series of projects using different platforms like MaxMSP, Pure Data Touch Designer, Isadora etc.
Still if someone could guide me to a direction ?
My programming skills need a lot of work and a lot of theoretical clarification in general but i dont mind giving it a try and learing… if somone has any guidelines…
The thing is i have no idea where to start from
Which SDK for the Tobii X2-60 should i use?
I guess all the live streaming data (x & y positions ?) would be stored in some variables and then send as Osc messages using an OSC library for the platform (Python or C++) used.
Thanks in advance
From the Analytics SDK 3.0 Developers’ Guide, Section “Calibration Procedure” (p.20):
1. A small animated object should be shown on the screen to catch the user’s attention.
2. When it arrives at the calibration point, let the point rest for about 0.5 seconds to give the user a chance to focus. Shrink the object to focus the gaze.
3. When shrunk, tell the eye tracker to start collecting data for the specific calibration point.
4. Wait for the eye tracker to finish calibration data collection on the current position.
In the Dev Guide, the calibration procedure in the Analytics SDK 3.0 is described as above. However, all language bindings of the SDK seem to be missing functions to actually trigger recording of calibrationdata for a specified target. All samples just enter a calibration mode, in which the eyetracker presumably records all data it can get, regardless of which calibration target (if any at all) is actually displayed. After having displayed such a target the addCalibrationPoint(point2d) function is called which supposably retrieves a small well-fitting subset of the recorded data and assigns it to the passed calibration point.
Is there any way to get more control over what data is used in calibration (like setting start and endpoint for recording as in the above text)? What kind of logic is applied here to choose datapoints for the calibration procedure?
I am asking these things because I want to squeeze as much precision out of the X2-60 as possible and as we all know, it all starts with a good calibration. 🙂
Hi, I am using my X2-60 with the Analytics SDK because of its Matlab binding. Apparently the Matlab binding is the only one in that SDK that lacks implementations for getCalibration() and setCalibration(). It is really cumbersome having to recalibrate the eyetracker each and everytime I relaunch my program, so getting access to those functions would be highly appreciated.
The way I see it, we have three options here:
1) The aforementioned functions are added to the Matlab binding in a future version of the SDK, for which I will have to wait (hopefully not too long).
2) I get access to the source of the mex-file, which probably only consists of calls to the C-binding of the Analytics SDK anyway. (That means that nobody gets to see the confidential closed source portions, so everything is fine.) Of the top of my head, I could think of numerous other reasons, why having access to the mex-source would be useful. Missing Xconfiguration functions is only one further example.
3) I get frustrated because of not getting those functions. 😉
Topic: X2-60 and EyeX
I see in the requirements that the X2 should work with EyeX.
However, I can’t get it to work.
If I connect via the TCPIP (UTB-cable) I can work with tobii studio.
But EyeX does not recognise an EyeTracker (No EyeTracker connected)
When I connect the X2-60 Directly via USB to my laptop (any port) I see the message “Eye tracker Connected” as a balloon tooltip. But in the Eyex-settings it keeps “initialising”, also after several restarts, and other USB ports.
SO: is is possible in the first place to use the X2-60 and if yes, how should I install it? Thanks.
I have bought an EyeX controller and tried some demos on that. However the precision is not satisfying, so I want to know more about those for analytic use. Is there any specifications for the eye trackers like X1-Lite, X2-30, X2-60 and so on so that I can know their accuracy and precision?
Topic: WPF on EyeX SDK?
I have 2 questions please:
1. Does EyeX SDK work with X2-30 and X2-60 controllers?
2. EyeX Unity page says the SDK provides “Code samples for C#/Mono that demonstrate the use of the .NET/Mono EyeX API”; so, are there WPF samples included?
* Since Visual Studio 2010, it is very discouraged to use Windows-Forms to develop applications for Windows desktop, and surprisingly Gaze SDK only provides Windows-Forms samples!
Topic: Any WPF Sample?
I’m very new to Tobii, and I’m using Gaze SDK on X2-30 and X2-60 eye trackers to develop applications, using WPF technology.
My very first WPF test application has WinForm-converted codes. the application compiles but throws exception, as soon as reading Gaze data.
Is there any WPF sample application, anywhere on the planet? It’s strange I cannot find anything about it on Google! 😀
And if there is no, can anyone do this favor to me and take a look at my very short code to find the problem? (I’ll provide a link)
Viewing 15 results - 31 through 45 (of 45 total)
Viewing 15 results - 31 through 45 (of 45 total)