👓 neon


Year

user-7e49f7 01 July, 2025, 07:45:22

Hello, I had a question about the storage. In my cloud data it says I have a full database of 2 hours, but when I go to delete videos it still continues to tell me that I have full storage and that I need to manage it. I am not sure what is going on.

user-f43a29 01 July, 2025, 07:46:49

Hi @user-7e49f7 , have you also deleted these recordings from the Trash? See attached image.

Chat image

user-7e49f7 01 July, 2025, 08:03:01

Are questions about neon player okay here?

user-f43a29 01 July, 2025, 08:07:16

Of course!

user-7e49f7 01 July, 2025, 08:08:12

So I was downloading the video from neon player but I can see that my gaze data is completely empty and I was wondering what could be causing this?

user-f43a29 01 July, 2025, 08:11:36

Just to be sure I understand, do you mean "export the video from Neon Player" or "downloading the video for Neon Player"?

user-7e49f7 01 July, 2025, 08:12:45

I click to the left side of the neon player with the download symbol.

user-f43a29 01 July, 2025, 08:21:52

I see. And you are inspecting the gaze_positions.csv file in folders like recording_folder/neon_player/exports/000/?

user-7e49f7 01 July, 2025, 08:38:43

Did you need the video after I downloaded it from neon player?

user-f43a29 01 July, 2025, 08:51:30

I mean the original recording that you load into Neon Player, not the exports.

user-35255b 01 July, 2025, 10:43:34

What do I need to do when my reference image gives an error. I tried other images of the environment but none of them is working. Can anyone help me

user-f43a29 02 July, 2025, 22:50:06

Hi @user-35255b , apologies, your message feel off our radar.

If you could invite cloud-support@pupil-labs.com to your Workspace, then we can investigate. Please note that then we will be able to see all data in said Workspace until the investigation is complete.

user-f4b730 01 July, 2025, 19:55:43

Hello, we have an offline Windows 10 machine where we installed PsychoPy 2024, but we were unable to install (lots of network errors) the Pupil package using PyPI (https://pypi.org/project/psychopy-eyetracker-pupil-labs/#history). Is there a way to install the pupil lab package without internet access? Also, is there a list of all the necessary dependencies?

user-f43a29 02 July, 2025, 08:22:04

Hi @user-f4b730 , it is in principle possible to install Python packages without internet access. However, doing so for PsychoPy's plugin sub-system might require extra steps. The PsychoPy team is better positioned to answer this question and it might be best to check on their community forums .

user-b407ae 02 July, 2025, 13:19:18

Hi everyone! I’m working on synchronizing multiple data streams for a recording session, and among those streams I have three Pupil Neon devices. I’m currently using LabRecorder (LSL app) and successfully recording synchronized data from all three Neons into a single XDF file. However, I’d also like to synchronize the data stored on Pupil Cloud for the three devices. I’ve read about lsl-relay, but I’m wondering whether it’s effective when working with three Neon devices simultaneously. Is there a recommended way to handle this kind of multi-device synchronization between LSL recordings and Pupil Cloud data? Thanks a lot for any advice!

user-f43a29 02 July, 2025, 13:24:17

Hi @user-b407ae , the lsl-relay is deprecated in favor of the integarted LSL functionality in the Neon Companion app.

If you want to synchronize the Timeseries CSV data from Pupil Cloud, then you only need to send some Events to Neon during the recording, perhaps one every few seconds. These Events will be saved in Neon's recording timeline and LSL's. Then, after you have downloaded the Timeseries CSV data, you simply align the Events with their corresponding entries in the LSL stream and you have synchronized.

user-5be4bb 02 July, 2025, 13:24:38

Hello! I want to export the gaze data (fixations, saccades, blinks, etc.). I have two methods, either using pl-rec-export or from Pupil Cloud. But since Pupil Cloud creates additional grey frames at the beginning and end of the main data, and I don't prefer using it, because later on, I need to extract frames from the original raw video exported from the phone (which does not contain grey frames). When reading the caveats of the pl rec export function, they said: "Fixations and saccades detected locally might vary from those exported from Pupil Cloud. This is due to differences in real-time gaze inference when compared to the full 200 Hz gaze data available in the Cloud." After comparing, I found that the number of fixations and saccades slightly differ, along with their characteristics. My question is: Despite this difference in detection, can I use only data from pl rec export and consider it as reliable 100%? Because my main interest is to keep using the raw video file exported from the phone, not the one with grey frames from Pupil Cloud.

user-f43a29 02 July, 2025, 13:42:36

Hi @user-5be4bb , the pl-neon-recording library is now preferred for working with the raw data at the command-line or programmatically, rather than pl-rec-export. You can also use Neon Player to export the raw data to CSV.

The pl-rec-export library was reliable, but on the latest version of the Neon Companion app, fixations and blinks are computed in real-time and saved on the phone. Saccades are taken as the inter-fixation intervals. Aside from that, the pl-neon-recording library exposes a more ergonomic programming interface.

Please note though that gray frames do not interfere with analysis. You can simply skip over them. Otherwise, the Timeseries CSV data is in principle the same data as saved in the Native Recording Data, but in a more user-friendly format, and re-processed at 200Hz.

user-a4aa71 02 July, 2025, 16:33:24

Hello, I performed an acquisition with the Pupil Labs Neon. The recording on the cloud is fine, and I downloaded the “Native data recording,” but when I drag the folder into the Neon Player it opens briefly and then the program closes. I tried downloading the folder again, but the same problem persists. Could you tell me why this is happening? Thank you very much!

user-f43a29 02 July, 2025, 22:52:03

Hi @user-a4aa71 , could you provide the logs in the neon_player_settings folder? This is found in the user directory of your computer. You can send it via DM, if you prefer.

user-ccf2f6 03 July, 2025, 03:36:01

https://discord.com/channels/285728493612957698/733230031228370956/1385099425680064562

Hi Pupil Labs, We're trying out the new GPS app but when we click the white button to start recording, it turns to gray and displays the message "Initializing sensor ..." The message doesn't go away and if we click the recording button again, it displays the filepath where gps csv is stored. All the stored csvs are however blank. We've allowed location and other settings and the phone is connected to internet. What could be going wrong here?

user-f43a29 03 July, 2025, 05:58:36

Hi @user-ccf2f6 , great to hear you are using it! How long are you waiting? You may need to wait about 10 seconds for the GPS sensor to fully initialize and start providing data.

user-9a1aed 03 July, 2025, 05:38:58

For the event.csv, is there a way that I could define each trial with a unique name to send to the eye tracker? Currently, it does not indicate the unique trial name. I am using the plEvent component of the pupillab's plugin.

Chat image

user-f43a29 03 July, 2025, 09:10:43

Hi @user-9a1aed , similar to how my colleague @user-d407c1 explained here (https://discord.com/channels/285728493612957698/1047111711230009405/1385566090607984660), this is possible in PsychoPy. It requires the use of their variables and conditions files. Since this is more about creating a custom string in the context of PsychoPy's system and not specific to the Neon PsychoPy plugin, their Support team might be better positioned to help you with this request.

user-35255b 03 July, 2025, 11:28:44

Hi, I'm currently trying to enrich my recordings on the Pupil Cloud using a reference image of the environment, but I keep encountering an error that prevents me from proceeding. The error message I receive is: "The enrichment could not be computed based on the selected scanning video and reference image. Please refer to the setup instructions and try again with a different video or image."

All my recordings are successfully uploaded and visible on the Pupil Cloud. I've tried using multiple reference images — including screenshots taken directly from the video and separate photos of the environment — but none of them work. I also attempted the enrichment with different scanning videos and different timestamps within those videos, but the same error keeps appearing. Could you help me?

user-f43a29 03 July, 2025, 11:36:07

Hi @user-35255b , please see my message here: https://discord.com/channels/285728493612957698/1047111711230009405/1390102265414815866

user-5b6ea3 03 July, 2025, 11:50:33

For the neon eye-tracker, when the white LED is on, what does it indicate?

user-5b6ea3 03 July, 2025, 11:53:32

I am using the eye-tracker today, and the LED was on. After the recording session, I looked at the recording, and there was no fixation information (red circle) on the video. Does this mean the eye-tracking data is lost?

user-f43a29 03 July, 2025, 12:07:27

Hi @user-5b6ea3 , the data might be recoverable, but this a sign of a potential hardware issue with your Neon module. Could you open a Support Ticket in 🛟 troubleshooting ? Thanks.

user-5b6ea3 03 July, 2025, 11:55:34

Chat image

user-13d297 03 July, 2025, 15:34:17

Hello! I was hoping to get some advice on how to best use events. Each of my recordings contain 16 events with unique labels to preserve their order over the course of the task (e.g., Event 1, Event 2...). I then run the Reference Image Mapper on the full recording and draw AOIs. My goal is an output file with the fixations to each AOI during each event. However, I noticed the event labels are not included when I download the Reference Image Mapper report. I understand one solution would be to run a separate enrichment for each event, however this would be quite time intensive given the large number of events and recordings. What is the best way to obtain an output file with fixations to the AOIs in each event?

user-f43a29 03 July, 2025, 16:39:36

Hi @user-13d297 , sure! All Enriched data is on the same time axis as the rest of Neon's standard data. So, you can still run the Reference Image Mapper on the full recording, if you prefer that approach. Then, use the timestamps in the events.csv file in the Timeseries Data download to filter and choose subsections of the AOI fixation data to analyze or export to a separate file.

user-ffef53 04 July, 2025, 07:45:53

Hey you guys! Our lab just got our pair of Neon glasses and we’re testing them out today. We’re super excited about them! Just a quick question: do other people notice that they are a bit hot to the touch after using them for prolonged periods of time? After 20min or so they’re getting pretty warm. We plan to run experiments that are an hour or more. We’re curious about others’ experiences and how they might address this in their labs. Thanks so much!

user-f43a29 04 July, 2025, 08:28:50

Hi @user-ffef53 , are you referring to the Neon module or the phone? The module is expected to warm up during normal use and the phone can also warm up. Please see this message for more details, also if you'd like foam pads for your frame: https://discord.com/channels/285728493612957698/1047111711230009405/1383005671611433083

user-9a1aed 04 July, 2025, 08:50:05

The recordings on Pupil Cloud showing that some of the tags are detected while others are not --> is built using jsPsych.

user-f43a29 04 July, 2025, 08:52:55

Ok, thanks for clarifying. Then, would you be willing to share the psyexp file for that recording with us at [email removed] Because the discrepancy in your video between the gaze circle in the app and the real-time gaze cursor in the PsychoPy program is not expected, unless there was lag or transmission delay.

user-9a1aed 04 July, 2025, 08:53:45

Thx! I will send the file here

user-9a1aed 04 July, 2025, 08:54:44

I am connecting the program and the Neon eye tracker using Wifi. Not a hub. Will it cause the delay?

user-9a1aed 04 July, 2025, 08:54:08

object1_butterflyCar.psyexp

user-f43a29 04 July, 2025, 16:54:26

Hi @user-b407ae , you are welcome!

If the goal is simply to align the Neon and LSL streams for sync purposes, then you can use the basic Timestamp on arrival method.

And the second link directs to the Event Data Outlet section of our LSL Integration Documentation. That section is brief and shows at the top of the browser, just above Connection problems? as in the attached image.

Chat image

user-3c26e4 07 July, 2025, 08:27:33

Hi, I am using QR markers for years, but I was wondering whether it is possible to put the QR markers in the video in pupil cloud after the gaze measurements.

user-f43a29 07 July, 2025, 08:30:05

Hi @user-3c26e4 , do you mean edit the videos digitally to post-hoc insert "virtual" QR codes? If so, then this is not possible on Pupil Cloud. The QR codes must be physically present in the scene, at the time of recording.

If you are looking for a markerless approach, then we would recommend the Reference Image Mapper.

user-11dbde 07 July, 2025, 19:06:42

In one of our experiments using the pupillabs neon device, when checking the data collected in pupilcloud we observed that sometimes blinks where not detected for some participants while for others there was an ! error and no pupil diameter was calculated...Any ideas of what happened?

user-f43a29 07 July, 2025, 22:46:39

Hi @user-11dbde , can you open a Support Ticket in 🛟 troubleshooting ? One of us will respond in the morning.

user-a578d9 08 July, 2025, 15:17:31

Hi, We're conducting a study using the Neon eye tracking glasses and are collecting an hours worth of data per participant. However, we've reached the 2-hour Pupil Cloud storage cap and are looking for alternative ways to manage and store our recordings without losing the data we want to analyse. Any suggestions on the best way to mange this issue would be great.

user-f43a29 08 July, 2025, 15:34:10

Hi @user-a578d9 , first, note that once you have deleted a recording from your Trash on Pupil Cloud (or it is auto-deleted from Trash after 30 days), then it is permanently removed from Pupil Cloud (it will of course still be available on the phone, provided you did not delete it from there, too). So, if you are unable to obtain an Unlimited Storage Add-on and you will choose your current workflow, then be sure to make local backups, either by downloading the recordings from Pupil Cloud, or exporting the raw data from the phone via USB cable.

Then, with a free account, any time you delete a recording from Pupil Cloud, it frees up the queue for more recordings.

You can also do analysis on the raw data from the phone locally with Neon Player or even load up the raw data into Python with the pl-neon-recording library. Just note that features like Reference Image Mapper and Face Mapper are Pupil Cloud only.

user-03b75d 08 July, 2025, 15:28:39

Hello! Just received the neon glasses (very cool), have done a few recordings and it seems well. Unsure if I'm missing something so thought I'd come onto this and asked, but once exported the videos, the pupil-tracker (the lil circle on the screen itself), is no longer there (still on the raw recording on the neon app). Only have played around with the pair for 30mins so still quite new to it!

user-f43a29 08 July, 2025, 15:37:19

Hi @user-03b75d , great to hear you are enjoying it!

Your device is working fine and what you report is expected. To get a video with the gaze circle, try the Video Renderer Enrichment of Pupil Cloud or the Video Exporter plugin of Neon Player.

If you have not yet had it, I recommend signing up for your free 30-minute Onboarding call. Just send an email to info@pupil-labs.com with the original Order ID.

user-cc6fe4 08 July, 2025, 20:26:37

Hi! I'm trying to download data from the cloud, but I'm experiencing unstable download speeds — they fluctuate a lot (from 0 up to around 65), even though my internet connection is stable. Because of this, the download fails. Could you please help me troubleshoot this or suggest what I can do to fix it?

Thanks in advance!

user-f43a29 09 July, 2025, 07:48:14

Hi! I'm trying to download data from the

user-67b98a 10 July, 2025, 04:53:04

Hi everyone,

I'm working on an exploratory study focusing on shopper experience in a supermarket using Pupil neon eye tracking glasses. The goal is to understand attention and engagement patterns (e.g., what products shoppers look at, dwell times, etc.).

However, I don’t have reference images or a static scene to map gaze data onto. The supermarket is a dynamic, complex environment, so traditional surface detection and reference image-based AOIs aren't feasible.

I'd really appreciate help with the following:

How can I analyze gaze data without reference images? Are there best practices for working with raw gaze or world camera video in dynamic environments? What metrics can I extract in this context? For example: heatmaps Fixation durations Dwell time in broad regions (e.g., shelf vs. signage) Gaze distribution over time Any recommended tools or workflows for post-processing or annotating gaze data manually or semi-automatically? If anyone has worked on similar mobile eye-tracking studies in naturalistic settings, I’d love to hear how you approached the analysis.

Thanks in advance!

nmt 10 July, 2025, 07:24:20

Hi @user-67b98a! Thanks for your question. Could you elaborate more on which areas of the shopping environment you're investigating? You do mention shelf vs signage. For these, our reference image mapper enrichment will often work well. E.g. check out this supermarket shelving example: https://docs.pupil-labs.com/neon/pupil-cloud/enrichments/reference-image-mapper/#_3-supermarket-shelf

user-b57ada 10 July, 2025, 13:22:24

Hey @user-67b98a, I'm a fellow researcher. If you haven't come across them yet, I highly recommend the Fundamentals of Eye Tracking series of papers, written by experts especially for researchers who are just starting out with eye tracking. They write about choosing tools, for example, in this paper: https://link.springer.com/article/10.3758/s13428-024-02529-7 As a relative beginner with eye tracking, I have personally found this series extremely accessible and helpful, so perhaps you will too.

user-b3b1d3 10 July, 2025, 15:12:06

Hello Neon Team. Just to now if someone is having problems when dowloading timeseries CSV from neon cloud. I am experimenting really slow donwloads that don´´t come to an end

user-9a1aed 10 July, 2025, 16:09:02

Hi Team, I received this error when I was using the April tag frame. From the demo program, the experiment is using Height. But for my other program, I use pixels. So I converted the size [w,h] of the tag frame from [2,2] to the window size in pixels [1360,768]. May know if I did the conversion incorrectly to cause this error? thx!

Chat image

user-cdcab0 10 July, 2025, 16:36:26

Hi, @user-9a1aed - if your experiment is configured for pixels, I think you should leave the values for size and marker size as-is and instead change the spatial units (for both of those) to "Height"

user-b5b5ba 10 July, 2025, 16:32:26

Hello, I am master student at Ulm university. We are conducting a research. We have one problem. When we are trying to do a enrichment and after when we try to donwload this enriched video. All of the other videos datas are also coming within specific videos excel. We tried to change wearer but didnt work. You can see the screenshot. I would be very happy if you could help us download the data of just the video we selected.

Chat image

user-f43a29 10 July, 2025, 16:47:08

Hi @user-b5b5ba , do I understand correctly that you want to download the Enriched data for just one recording?

user-b5b5ba 10 July, 2025, 17:02:53

yes @user-f43a29

user-f43a29 10 July, 2025, 17:09:13

I see. To do this:

  • Go to the Downloads tab for that Project. You are currently looking at the Enrichments tab.
  • Click on the gray box that shows 32/34 or 28/34, etc in the "Recordings" column. In the attached image, I clicked on the 2/2 button.
  • Then, choose just the single Recording for which you want to download Enriched Data and click the Download button.

Note that changing the wearer on the recording will not have any effect in this regard. It would be recommend to revert the selected wearer for that recording to the profile of the person who wore Neon at the time of recording.

Chat image

user-b5b5ba 10 July, 2025, 17:23:04

@user-f43a29 it worked thank you so much

user-5a90f3 14 July, 2025, 03:06:25
user-5a90f3 14 July, 2025, 03:07:31

Hello, the following questions appeared during my recording, I would like to ask if my recording record can be found?

user-ffef53 14 July, 2025, 09:07:44

Hi there! Our lab just got the Neon glasses and we've been playing around with some mini experiments. I had a few questions for the community:

  1. Since we do not have Pupil Cloud, we’re writing scripts to convert all the raw files into CSVs on our own. Does anyone have experience with this? I tried looking for examples, but I don’t see much in the discord about this (outside of using the cloud)
  2. Related to this, in the Neon Player you can convert some files into CSV format, but I only see that it’s possible for raw gaze files. Can someone explain this further?
  3. I’m seeing some interesting differences between the folders when I export from the companion device. I see 2 folders, Neon and Neon Export, that contain basically the same info, but Neon Export has 3 more files (gaze_right ps1.raw and .time and info.json). Additionally, when I upload the folders to the Neon Player, I generate a new folder within the original raw data folder which contains even more files including world timestamps and a gaze.csv that would be super useful to have from the start. Does anyone have any best practices for organizing all this data? I think the workflow is a bit unclear. Its ok for now when we’re just running little tests, but in the future I’d like to know the best way of managing all of this raw data and converting it to CSVs without losing or misaligning anything.

Thanks in advance for your help!

user-d407c1 14 July, 2025, 09:57:20

Hi @user-ffef53 👋 ! Welcome! All of that is already possible, let me clarify some aspects and walk you through the workflows:

  • Transferring Recordings: The general workflow for exporting recordings is described here. Both the raw and the export folders would contain all the relevant information. Generally, we recommend exporting it and using the Neon Export folder to prevent affecting the normal app's operation, but if you are going to delete the data afterwards is also fine.

With the data already on your computer, you have different options to export the relevant CSV files.

  1. Using Neon Player:

You can load your recordings into Neon Player, then, go to the Plugins menu, enable the streams/plugins you want to export (e.g. gaze, pupil size, eyeball position) and click Export to generate the corresponding CSV files.

The general workflow is described here.

Just keep in mind: Only the enabled plugins will be included in the export.

2: Programmatic CSV export

If you're working in Python, and want to programmatically export the data to CSV files, you can leverage our pl-neon-recording library.
Here you can find a full example on how to export the data onto CSV files.

Let us know if you have any questions or get stuck somewhere.

user-f389a1 14 July, 2025, 13:21:07

Hello. I am a PhD student collecting data from the Neon glasses in children for a continuous recording of ~20 min. I have plotted the pupil size (collected from the 3d_eye_states.csv file) but am struggling to interpret where the blinks are - I am wondering if you have any advice? (I can send picture of the plot/data if that is helpful - there are many downward spikes, but they vary in their depth and occur at very irregular time periods - so am not sure whether they are blinks or not)

user-d407c1 14 July, 2025, 13:48:24

Hi @user-f389a1 👋 ! You can find the blink data in the blinks.csv file.

By default, NeonNet would try to infer pupil size at all times, even during blinks. You'll typically see a rapid drop in pupil size, though it may not reach zero. Gaze data is also still reported during blinks, you can see the signal turn noisy, but we don’t automatically remove it.

This is intentional, so that you have full control over how you define a blink. For example, you might prefer to base it on a different eyelid aperture rather than rely on a hard cutoff. Lemme know if you need help filtering them out or plotting them!

user-d2bc2a 14 July, 2025, 20:33:43

Hello! I am a Phd student and our lab currently owns a pupil core and currently looking into upgrading to a Pupil Neon. My current concerns are as follows: 1. Does Neon require eye camera adjustments like the core? 2. Can we save videos without the Pupil Cloud for Neon?

user-cdcab0 14 July, 2025, 21:26:40

Hi, @user-d2bc2a! Thanks for Considering Neon. To answer your questions 1. Neon does not require (or even allow) for eye camera adjustments. The eye cameras are in a fixed position on the module which is fixed to the frames. Our deep-learning approach and diverse training data allows Neon to work well on an incredible variety of facial geometries and wearing positions. Neon's eyetracking is even robust to slippage! 2. Uploads to Pupil Cloud are completely optional and can be easily disabled in the app. We provide open-source, offline analysis tools to help you accomplish your research goals without relying on our cloud service.

user-ffef53 15 July, 2025, 09:28:44

I should have the most up to date of both because we just received the device a few weeks ago. I just tried again on a new, much shorter recording and it worked, but for the longer ones its not working so maybe its a bug?

user-ffef53 15 July, 2025, 09:30:59

I think the processing just takes much longer, but maybe there's a way to optimize it that im missing?

user-d407c1 15 July, 2025, 09:36:46

Thanks for your quick reply! Fixations in the latest version of the app are computed directly on the device, as long as you have “Compute fixations” enabled in the Companion App settings.

That means Neon Player shouldn't take long to load them, since it doesn’t need to process them post-hoc.

Could you double-check that you’re using Neon Player 5.x?

user-037674 15 July, 2025, 09:51:54

Hi PupilCloud Team, could you let me know the dimensions of the red circle indicating gaze in the cloud - either in pixel or field of view (degrees)? Thanks!

user-d407c1 15 July, 2025, 14:10:01

Hi @user-037674 ! the Gaze overlay radius employed in Cloud with Neon is 45px but it scales with the video canvas.

user-33b314 15 July, 2025, 15:00:50

Hello! I work for a market research company, and in a project with Neon glasses we are having problems with two videos loading in the cloud, it says "processing", but they don't load, these videos can be played on the mobile. Do you know what I can do?

Chat image

user-f43a29 15 July, 2025, 20:17:26

Hi @user-33b314 , we also received your email and followed up there.

user-a5d1a8 15 July, 2025, 15:38:26

Hi, I must answer some questions to the "Datenschutzbeauftragte" (information security officer?) concerning the videos of the Iris/pupil. could they be identified to a special person? is the video "safe" - in aspects of the data-security and re-identification of a person? Is there any information about this issue? Maybe I could link to a document? Is it possible to remove the eyes-video without loosing information? Thanks in advance!

user-f43a29 15 July, 2025, 20:26:06

Hi @user-a5d1a8 , could you send an email to info@pupil-labs.com with this request? Then, we can provide you with the relevant Documentation.

Briefly:

  • It will be helpful in your email to clarify what is meant by "identified to a special person".
  • The videos that are uploaded to Pupl Cloud are securely transmitted & securely stored. Pupil Cloud is GDPR compliant and offers additional Anonymization features, such as face blurring and custom Data Processing Agreements.
  • It is not necessary to use Pupil Cloud. You can completely disable that function before making your first recording and completely disconnect the Companion Device from the Internet. Then, all data is saved only on the Companion Device. The data formats are open and we offer free & open-source tools for working with them.
  • After making a recording, you can remove the eye-videos. All necessary eyetracking info is computed and saved in real-time. The data is yours to make your own decisions.
user-878a5a 15 July, 2025, 18:51:16

ello everyone,

I recently used the Neon to record multiple videos. When I checked the cloud later, I noticed that six of the videos were not uploaded. I tried to upload them manually, but after a few seconds, the upload fails each time.

Is there any way to successfully upload these videos to the cloud?

Chat image

user-f43a29 15 July, 2025, 20:17:51

Hi @user-878a5a , we also received your email and followed up there.

user-1a3091 17 July, 2025, 08:48:01

I am using Psychopy 2025.1.1 for an eye tracking experiment with videos. I have downloaded the Pupil Labs Plugin. In settings, I select Pupil Labs (Neon) but it does not allow me to Modify the IP address and port. I don't see those options in the dropdown.

user-f43a29 17 July, 2025, 08:55:46

Hi @user-1a3091 , that version of PsychoPy is still in beta and with that version, the PsychoPy team has made changes to the plugin system that are not yet accommodated. It would be recommended to use the latest stable PsychoPy, v2024.2.4.

Note that when switching versions, you may want to first clear PsychoPy's cache of the plugin, as follows for Windows:

  • First, close all PsychoPy windows.
  • Then, delete the psychopy3 folder in %APPDATA% (just copy-paste that into Windows Explorer).
  • Then, uninstall PsychoPy and install the stable version.
  • Thereafter, download the plugin anew, following the steps here.

On Linux and MacOS, the .psychopy3 (note the dot, .) folder is found in the user's home directory and the process is otherwise the same.

user-21cddf 17 July, 2025, 11:01:22

Hi, has anyone been able to synch fNIRS aurora lsl triggers and the neon recording? My triggers arent showing up for the neon recording

user-f43a29 17 July, 2025, 14:10:41

Hi @user-21cddf , are you already using Neon's LSL integration and the standard LSL LabRecorder app?

user-45f4b0 17 July, 2025, 20:35:51

And also, does PupilLab have any devices that can output the gaze data in real time? So that we can combine the gaze data with other real time processing code?

nmt 18 July, 2025, 02:21:06

Yes! We have a fully-featured Realtime API for Neon that can do all of this and more. Check it out here: https://docs.pupil-labs.com/neon/real-time-api/#real-time-api

user-6eef09 17 July, 2025, 23:26:30

Hello. I am trying to connect monitor my data collection in real-time with the neon monitor app, but the link, http://neon.local:8080/, doesn't seem to be working. Both my companion device and my computer are connected to the same network. Could I get help troubleshooting?

nmt 18 July, 2025, 02:22:54

Hi @user-6eef09! Could you describe the network you're trying to use? Also, have you already seen the troubleshooting steps oulined here? https://docs.pupil-labs.com/neon/data-collection/monitor-app/#connection-problems

user-30f8d5 18 July, 2025, 05:11:55

Hello, I’m currently trying out the Pupil Labs Neon our research unit recently bought. I appear to have some problems with local streaming, that is, syncing my computer with the Neon.

I am currently on PsychoPy trying out your Builder demo, gaze_contingent_demo.psyexp. While I can get it to run when using MouseGaze as the eyetracker device, I am unable to do so using the Pupil Labs (Neon) plugin. I’ve done the following troubleshooting steps:

  1. Updated all packages included in the Pupil Labs plugin
  2. Checked companion address and port to ensure it is neon.local and 8080 respectively
  3. Also tried the Coder .py file version, which didn’t work
  4. Pinged the eyetracker via ping neon.local which did show a connection

Tldr; I would just like to try out the demo but unfortunately, it doesn’t work with my eyetracker. Is there any additional steps that I’m supposed to take?

user-f43a29 18 July, 2025, 05:26:07

Hi @user-30f8d5 , what kind of network are you using? Is it university WiFi?

user-30f8d5 18 July, 2025, 05:26:53

Hi @user-f43a29 I am at home, using a home wifi. I believe it is a wifi 6

user-f43a29 18 July, 2025, 05:29:58

Ok, could you also try with the hotspot of your personal cellphone? This will help clarify if it is a configuration issue with your router. See these Troubleshooting steps for more details. Let us know the result and we can help further.

user-30f8d5 18 July, 2025, 06:12:00

Hi @user-f43a29 , thanks for the quick reply. So my mobile hotspot did nothing to change it. I do realise that trying to access Neon Monitor via the URL, neon.local:8080 or the IP address provided on the Companion, results in some extremely slow lag, if not just screen freezing.

I’ve tried running dns-sd -B _services._dns-sd._udp via MacBook Terminal, and I couldn’t find anything related to the eyetracker. I do agree with you that it may be a router issue but my wifi and hotspot does allow for mDNS and UDP traffic. Any idea what I can do next?

user-f43a29 18 July, 2025, 06:22:49

Could you open a Support Ticket in 🛟 troubleshooting ? Thanks

user-878a5a 20 July, 2025, 12:15:18

Hello everyone, I’m having trouble uploading 7 videos to Pupil Cloud from the phone, as the uploads keep failing. I can extract the videos from the phone, so I was wondering if it’s possible to upload them to Pupil Cloud directly from a computer. This seems like my best option, since I’ve already contacted Pupil Labs support and fixing the issue on the phone appears to be difficult.

user-f43a29 20 July, 2025, 13:32:34

Hi @user-878a5a , the team is aware of and handling your issue. We should have an answer to your email by Monday.

Please note that uploading to Pupil Cloud from a computer is not a supported workflow, but we will resolve the issue for you.

user-21cddf 21 July, 2025, 11:56:59

@user-f43a29 Yes. We have now managed to make it work for one pair of neon glasses, but the experiment involves 2 pair of glasses recording at the same time. The api only seems to recognize one pair at a time

user-f43a29 21 July, 2025, 12:37:24

Hi @user-21cddf , can you describe your setup in more detail? It sounds like you rather receive Aurora‘s triggers programmatically and then convert those to Neon‘s Events in your own Python code?

user-a4aa71 21 July, 2025, 15:55:22

Hi everyone, I have a very simple Python script that remotely controls the start and stop of recordings. The code uses these two functions: recording_id = device.recording_start() and device.recording_stop_and_save(). I connect my PC and the smartphone connected to the NEON to my phone's hotspot (the code searches the device by IP). The problem is that sometimes (very often actually) I get this error --> Error: Python Error: DeviceError: (400, 'Cannot stop recording, not recording') etc ... I noticed that if I keep sending start and stop commands, the recordings either don't get saved or are corrupted (they contain no video of the scene), even though no error appears in the Python console. Moreover, after this error occurs, it's impossible to get valid recordings anymore—I had to reboot the device. How can I solve this issue? Next time I run experiments, I will make sure to use a router, but is this a connection problem or is it caused by something else? I have already used my phone's hotspot, and with the Pupil Labs Core I have never had this problem. Thank you very much.

user-f43a29 21 July, 2025, 16:28:10

Hi @user-a4aa71 , this is not a connection issue. May I ask if your code also uses device.close() at the end of each session, and whenever it is quit, including when it is quit by Ctrl+C?

user-430fc1 21 July, 2025, 16:01:37

Hi, does anyone know of any publications or code bases that focus on gaze contingent research with multiple neon devices?

user-f43a29 21 July, 2025, 16:55:02

my code is this one, very simple... (I'm

user-d086cf 21 July, 2025, 18:49:02

Hey guys. I may be doing things a bit unconventionally, but rather than having to load each of my recordings in neon player prior to analysis, I wrote a python script that will open up the .raw and .time files from all my exported recordings and spit out csvs for the data I need.

I have everything working for basic eye state and gaze data, but I am having trouble figuring out what the columns correspond to in the "fixations ps1.raw" file. It looks like there's 15 columns, but I am not sure whats what. Can someone with insider knowledge on this let me know the column names?

user-f43a29 21 July, 2025, 19:28:41

Hi @user-d086cf , you can save a lot of time by using our free & open-source Python library, pl-neon-recording. It does all this and more for you already and its code self-documents the structure of the binary files, such as the fixations files.

You may want to reference this example for exporting the data to CSV format.

user-a578d9 22 July, 2025, 11:12:40

Good afternoon, Quick question about wireless charging — we’re going to be doing several back-to-back hours of recordings using the Neon glasses, and I wanted to ask whether a wireless charger would work effectively with the companion phone while it's connected to the glasses. Has anyone tested this setup? Also, if you’ve got any recommendations for reliable Qi wireless chargers that have worked well with the phone and Neon during long sessions, I’d really appreciate it.

user-d407c1 22 July, 2025, 11:25:24

Hi @user-a578d9 👋 ! Whether wireless charging works depends on the companion device. Both the Samsung S25 and Moto Edge 40 Pro support Qi wireless charging.

Just a heads-up: they don’t attach magnetically, and wireless charging is generally less efficient, with the handicap that it tends to warm up the phone.

While this is not an official recommendation. I’ve been using a previous version of this, it helps keep things cool while charging 😉

user-e0a71c 22 July, 2025, 16:10:25

Good evening, I ran into an issue when trying to time-align my cloud exports with my xdf data when recording using the LabRecorder. As we were able to find a neon lsl stream that was coming from the mobile device connected to the glasses directly, we did not record using the lsl_relay on a separate laptop (how we did it before the app update). Now, after just recording using the stream coming directly from the neon phone, I am unable to time-sync gaze events using the lsl_relay_time_alignment. Is there another way to time-sync the streams without using the lsl_relay on a separate laptop? I could not find information on this topic in the Pupil Labs Companion LSL Relay documentation. This is the error message: WARNING Session id missing! Skipping incompatible Pupil Companion stream xdf_cloud_time_sync.py:107 [{'acquisition': [{'manufacturer': ['Pupil Labs'], 'model': ['Neon'], 'session_id': []}], 'channels': [{'channel': [{'label': ['Event'], 'format': ['string']}]}]}]
ERROR Could not extract any time alignments for [this is my path here]. No valid Pupil Companion event streams found!

user-cdcab0 22 July, 2025, 21:46:30

Hi, @user-e0a71c - yes, the LSL Relay app is a solution provided for our older eyetracker (Pupil Invisible), but it worked for Neon as well until we integrated the LSL outlet directly in the companion app, so you're correct to avoid the LSL Relay with Neon.

To sync data between LabRecorder XDF data and a Neon Recording, you just need one or more Events that have been recorded in both. Then you can simply examine the timestamps of the same event recorded in both systems to compute the offset. A quick search of this discord server should turn up a few different conversations on this exact topic

user-4b18ca 23 July, 2025, 07:27:05

Good morning! We are seeing some calibration issues in our setup. Person is sitting in front of a 27" (non-curved) screen with a distance of around 40 centimeters. Calibration routine is executed by displaying fixation points at the corners of the screen, asking person to look at it, and correcting mismatch on the phone (following https://docs.pupil-labs.com/neon/data-collection/offset-correction/).

Is this a problematic distance/setup for good calibration?

user-d407c1 23 July, 2025, 08:28:13

Hi @user-4b18ca 👋 Just to clarify, Gaze Offset is different from a traditional calibration.

On a traditional calibrated eye-tracker, you would look at different targets while the eyes are recorded. Then, a regression is built from detected features on the eye images (like dark pupil, Purkinje image, etc) and the location of the target in the scene camera or the screen. Such that you can estimate where the person looks, based on the location of these features on the eye images.

This is a bit frail, as if the person / headset moves the relationship can break. Neon, breaks with this paradigm, it uses a deep learning-based model (NeonNet) that does not require this step, it was trained on a large and diverse dataset, so it doesn’t require user-specific calibration. In other words, you can take the glasses off and put them back on, and you’ll continue receiving valid gaze data.

Although this deep learning approach has been trained with a vast amount of data, different faces, eyes, etc. For a small subset of participants, accuracy may be slightly off, in those cases, you can apply a gaze offset correction, which is a simple linear correction across the entire visual field.

As reported in our whitepaper, average gaze accuracy is 1.8° across a 60°×35° field of view. With an offset applied, this can improve to around 1.3°.

Because the correction is global, applying it multiple times (e.g., at different screen corners) just overwrites the previous offset, it’s not cumulative or spatially adaptive. So it does not make sense to apply it on multiple points.

user-d407c1 23 July, 2025, 08:28:24

You might now be wondering: - Does accuracy decrease in the periphery? Yes. That’s also reported in the whitepaper.

nmt 23 July, 2025, 11:35:18

[email removed] I work in Matlab and import

user-cc6fe4 24 July, 2025, 15:46:03

Hi 🙂 I have a couple of questions, some may be a bit silly but I want to make sure I am doing the correct thing. Thank you in advance. If I download the Timeseries CSV and Scene Video are the video and gaze data undistorted, I mean, are they already corrected for lens distortions? From what I see in the videos, they are not. So, I am following the Undistort Video and Gaze Data tutorial (https://docs.pupil-labs.com/alpha-lab/undistort/). What data should I be using for this (Timeseries CSV and Scene Video or Native Recording Data)? Also I am a bit confused about what timestamps of every world video frame means in the context of the world_timestamps.csv? Could you please elaborate.

user-f43a29 25 July, 2025, 08:24:01

Hi @user-cc6fe4 , not silly! This space is for such questions.

May I first ask why you need to undistort the data? Knowing that, I can better assist, as the distorted data are fine for a variety of use cases.

With respect to the scene camera timestamps, could you clarify a bit more what is confusing? Briefly, they are the time at which each scene camera frame was taken, specified in nanoseconds. You can directly compare them to the timestamps for Neon's gaze, IMU, etc data.

user-539d55 24 July, 2025, 18:25:53

What is the easiest way to calculate convergence from the eye state data?

user-f43a29 25 July, 2025, 10:14:09

Hi @user-539d55 , may I first ask, what is the end goal or what will be the application?

user-ffef53 25 July, 2025, 13:15:02

Hi another question here: in the Neon Player is there a way to adjust the offset at varying times and apply it to only a select portion of the video? In our experiments we will have participants looking at a table and later at a screen. The screen gazepoints seem right on the mark, but on the table when the participants are looking down it is a bit inaccurate and could use the offset. Thanks!

user-f43a29 25 July, 2025, 14:05:06

Hi @user-ffef53 , the Gaze Offset Correction applies for a whole recording. You could technically set the trim points in Neon Player to only export a subsection and then apply a separate Gaze Offset Correctlon for the export of each trim. However, may I ask if they are moving their heads to get a good view of the table or are they rather looking far down, such that gaze is towards the bottom edge of the scene camera video?

user-a4aa71 25 July, 2025, 15:37:56

Hello, one of my students tried this script: https://pupil-labs.github.io/pl-realtime-api/dev/methods/simple/streaming/eye-events/ from the simple API but the .receive_eye_events() method does not exist on the device object. In contrast, it works with the async API version. Does the previous code need to be modified in some way? Thank you very much!

user-f43a29 25 July, 2025, 15:39:09

Hi @user-a4aa71 , are they on the latest version of the Python Real-time API? They may need to upgrade it in their virtual environment:

pip install --upgrade pupil-labs-realtime-api
user-a84f25 25 July, 2025, 18:47:48

Hi - I'm trying to figure out the precise timing differences between the source neon phone and the client, in order to eventually estimate what the jitter and offsets we can expect when compared to EEG data. I've been looking at the packets over wireshark and am a bit confused. 1 - eye gaze data should be 77 bytes when gaze and eyestate are included, but the closest packets I can find have an 88 byte payload? 2 - Theres an RSTP packet that described data from /?camera=gaze?audioenable=on. Aren't the eye camera stream under /?camera=eyes ? What is that stream describing? 3 - Similarly, what are the /trackID=0 rtsp streams? 4 - There are a bunch of payloads of 200 bytes that wireshark can only identify as udp or unknown rtcp version. Are there any filters/decode as options that would help me analyze this data better? Thanks!

user-f43a29 25 July, 2025, 19:44:44

Hi @user-a84f25 , for your case, there's no need to go so deeply into the packets, although cool to see it!

For example, we already offer methods for precisely estimating the time difference between the Companion Device and the client. And, we have Lab Streaming Layer support directly integrated into the Neon Companion app.

Or, is it rather that you are implementing your own Real-time API integration in another language?

user-a84f25 25 July, 2025, 19:21:50

Also, is the timestamp returned by the API as the time the data was taken just the timestamp on the RTP packet?

user-a84f25 25 July, 2025, 20:28:27

Thanks for the clarification! The reason I was turning to looking at the packets is because when I wrote a python script to compare the timestamps on the sensor data to the client ingestion time, and compared that to the estimated offset and roundtrip times provided by the timeestimate function, I get close matches for the time offset seen when comparing ingestion time and sensor times for the gaze data (they are like 10 ms more), but the video data seems to be off by like 80 ms. This made me think there was some parsing delay in my python code (even using the async library), so I wanted to compare the times to when the packets were actually received by the computer. So is the timestamp of the sensor data also embedded in the payload of the RTP packet, or is the time attached to the RTP packet just the time the sensor took the data? I'm really sorry for my confusion

user-f43a29 25 July, 2025, 20:34:05

No problem.

Time offset is the offset between the time reported by the clock of Neon and the clock of your computer. It is independent of and not related to transmission + processing latency, which is what it sounds like you are measuring.

Are you looking to do something like real-time gaze contigent experiments or do you just need the data to be synced for post-hoc analysis, after an experimental session is completed?

Also, I'm not sure I understand the difference between:

  • The timestamp of the sensor data embedded in the payload
  • And, the time the sensor took the data

Could you clarify further?

user-a84f25 25 July, 2025, 21:02:59

Sure sorry for the confusion. This is mostly for post hoc analysis - aligning the neon data with an EEG or maybe other signals that are also synced via PTP. Ultimately I want to know the offset between the sensor clock and the computer/client clock. I know you provide that through the TimeEstimator class, but I'm essentially trying to self verify it as well. From what I understand, there are a couple of times to consider: the time the sensor took the data, the time the data is sent, and the time the data is received. From what you are saying, the time the sensor took the data is in the payload of the RTP packet, separate from the timestamp on the RTP packet (which is when the data was sent?). The time that the computer received the data should optimally be half the roundtrip time, which in my current setup should be < .5 ms according to your API. But there is probably some latency/processing which happens (ideally a smaller amount in the time wireshark received the packet compared to the python API). That latency (based on the time that wireshark received the packets, and the time on the RTP packet, and assuming a clock of 90000 which is in the RTSP description) is currently 2 - 5 ms. I'd like to be able to compare the time the sensor actually took the data to these other timestamps. But I don;'t know how to parse the RTP payloads. Through wireshark I have payloads that are 89 bytes, which matches up with the description of EyestateEyelidGazeData. But that doesn't include a timestamp in it? So is the timestamp on the RTP packet the time the sensor took the data? I.e. there is no difference between the time the sensor took the data, and the time on the RTP packet (not the payload of the packet)?

user-f43a29 25 July, 2025, 22:32:09

Sure sorry for the confusion. This is

user-9a1aed 28 July, 2025, 13:19:05

Hi Team, I am not sure what's going on here. This recording has been uploading for quite a while already. ID: bf342863-0439-4edb-88fb-5edcb73c00e3

Chat image

user-9a1aed 28 July, 2025, 13:19:46

Also, it says the cloud storage is full. If I deleted my cloud recordings, is it going to help? Or do I have to purchase additional storage?

Chat image

user-f43a29 28 July, 2025, 13:26:43

Hi @user-9a1aed , for the recording that is uploading, could you go to the Recordings panel in the Neon Companion app and use the menu to Pause uploading and then wait a few seconds before clicking the Upload button?

For the recording with the "quota surpassed" message, that is unrelated to the upload process. To work with that recording, you either need to activate an Unlimited Add-on or delete some recordings from Pupil Cloud to free up the quota. Please note that when you delete a recording, that will permanently delete it from Pupil Cloud, so be sure to have a local backup.

user-898b32 28 July, 2025, 15:16:27

Hi Team! Not sure if this topic has been touched on before here but I was wondering if there is a way to perform the same processing that is done in cloud on data that has been downloaded as native/raw data from a phone but has been deleted from the phone and is now only available on a computer? I am noticing that for 3 donwloaded recordings I am missing the data from the cloud but I have set up my analysis pipeline in a way that uses the fixations.csv and world_timestamps.csv files from the timeseries download. The neon player allows me to export a file that is also called fixations.csv, i assume this is the same as the one from the cloud agorithm. Can I somehow also create the world_timestamps.csv with the neon player? I only find a world_timestamps.npy and world_lookup.npy but these do not seem to be on the same scale as the world_timestamps.csv. Thanks!

user-f43a29 28 July, 2025, 15:23:55

Hi @user-898b32 , Neon Player's CSV formats match with Pupil Cloud's.

To get the world_timestamps.csv file, make sure to also activate the World Video Exporter plugin.

You can also use pl-neon-recording's CSV export script.

user-4369a0 29 July, 2025, 00:41:58

Hello 👋 New here. Not sure if this has been discussed, but can the Neon receive TTL pulses? Currently, trying to synchronise their data with intracranial EEG data, and couldn't find guides online. (P.S. Successfully followed this guide with my experiment: https://github.com/pupil-labs/pl-neon-matlab/blob/main/track-your-experiment-progress-using-events.md) Thank you in advance.

nmt 29 July, 2025, 02:14:26

Neon doesn't receive analog TTL signals. But it has a real-time API that can send/receive software events over a local network. That said, often the easiest way to sync Neon with EEG equipment is via our Lab Streaming Layer integration, if your EEG device supports that.

user-539d55 29 July, 2025, 10:59:13

[email removed] , there is the typical

user-ffc52e 29 July, 2025, 11:16:24

Hi, sorry I am having the following error on the screen: May I know why this error happens and if anyone knows a possible solution? Thank you

Chat image

user-f43a29 29 July, 2025, 16:09:28

Hi @user-ffc52e , I see you already opened a Support Ticket and are being helped. Thanks!

user-ffc52e 29 July, 2025, 11:17:04

I tried unpluging it and plug it in back again all morning, but keeps having this error

user-5a90f3 30 July, 2025, 07:16:41

Hello, I have a question, as shown in the picture is the frame of neon, the temples and frame are separated, on the model do they have a link structure, or do I need to link myself. Thank you so much.

Chat image

user-f43a29 31 July, 2025, 12:26:44

Hi @user-5a90f3, have you modified the CAD file after it was loaded into your software? It should look something like in the attached image when loaded. The CAD model is provided as reference geometry, to serve as a starting point for prototyping your own custom frames. When you say “link structure”, do you mean something in the CAD software that virtually assigns the temples to the front of the frames?

Chat image

user-f43a29 31 July, 2025, 14:10:46

Hi @user-5a90f3 , some more info from the team:

  • If you want to print that CAD model and use it as a frame, then you can fuse the arms to the front piece and print it as one solid frame.
  • If you were asking about the hinge mechanism that lets you fold the arms, then that is not available for purchase in small quantities.
  • If you rather need "dummy" frames, then you can reach out to us via email [email removed] and we can supply that.
user-37a2bd 31 July, 2025, 09:04:06

Hi Team, on the Pupil Cloud are we able to set the number of fixations we want to be plotted on the gaze plot? Initially when we ran it through the google collab workbook we were able to select the number of fixations to be plotted. is that option not available anymore?

user-f43a29 31 July, 2025, 12:27:03

Hi @user-37a2bd , may I ask which notebook you are referring to?

user-56cde0 31 July, 2025, 11:41:16

Hi, all--Has anyone recorded an individual with infantile nystagmus syndrome (a.k.a. congenital nystagmus) with the Neon? A head-mounted tracker with a wide gaze angle operating range would be really useful but it would be nice to see a sample of a recording before asking my department to get one. Thanks!

user-7413e1 31 July, 2025, 13:13:44

Hi, I am working with data collected with Pupil Neon and have computed enrichments using PupilCloud to record fixations on AOI. I noticed that my fixations include cases of duration <100ms. I understand that 100-200ms is the treshold for the actual definition of fixation (from your online documentation). So I was wondering how others have dealt with this and whether a simple filter would suffice to clean my data, or whether I am missing something else in the settings of how raw data is used to extract fixation (if so, should I work in PupilPlayer instead before downloading my data / computing enrichments)?

user-d407c1 31 July, 2025, 13:27:27

Hi @user-7413e1 👋 Just a quick note, I think there might be some confusion between Pupil Core and Neon.

For Neon, fixation and saccade data are available via the fixations stream which is now computed in realtime too. You can also find more detail in the fixation's detector whitepaper (PDF) where you can find a table noting that the minimum fixation duration for Neon is 70 ms.

user-d407c1 31 July, 2025, 13:31:31

Is this on the fixations.csv from the TimeSeries or somewhere else?

user-7413e1 31 July, 2025, 13:36:22

This is in the fixations.csv file and the aoi_fixations.csv file downloaded from the enrichment (marker mapper)

user-7b0de7 31 July, 2025, 15:26:44

Hi team, I'm running into some issues with the companion app and was hoping to get some insight: 1. When I start a recording, the duration shown is the time from button press, but it doesn't match the duration of scene capture. For example, a 9s recording will capture 3s of video. 2. The app becomes unresponsive randomly and will get rid of 1-3 of the previous recordings. I am using the Motorola edge 40 pro and I am trying to record 3sec clips back to back, but the cameras activation has been a various times, so I record 8sec and trim afterward. I would appreciate an help

user-f43a29 31 July, 2025, 15:33:20

Hi @user-7b0de7 , may I first ask why you make such short recordings? Is there a reason to not make one long recording and use Events to mark moments of interest?

user-241e99 31 July, 2025, 18:19:40

Hi everyone! I could use some help with a couple of issues I’m having with my Pupil Neon glasses.

  1. Connecting Pupil Neon to desktop software: I’ve used the glasses successfully with Pupil Cloud, but now I want to connect them to the desktop apps (like Pupil Service, Pupil Player, or Pupil Capture). I’m not sure how to do it, and I haven’t found clear instructions. Has anyone here managed to set it up? Any steps or tips would be appreciated!

  2. Blink detection not working on Pupil Cloud: Even though calibration works fine, blinks are not being detected or logged on Pupil Cloud. Is there something I’m missing or need to activate?

  3. Reference Image Mapper not detecting reference image: I’m also trying to use the Reference Image Mapper to track when a participant looks at a product image, but it doesn’t seem to detect the reference image during playback. Any idea why it might not be working or how to troubleshoot it?

user-f43a29 31 July, 2025, 20:19:46

Hi @user-241e99 , it is currently after hours for us, but has your group already had its free 30-minute Onboarding call? We can cover these and more questions a bit more efficiently that way. Then, you can also show us your Reference Image for direct feedback.

Also, to clarify the Pupil Capture/Player/Service software is for our original eyetracker, Pupil Core. With Neon, you connect to it from a computer via the Real-time API or Neon Monitor. Or, are you rather aiming to analyze data on your computer?

user-241e99 31 July, 2025, 19:47:34

Hello?

user-b442f7 31 July, 2025, 21:18:22

Hi Team, I am processing a few sessions of neon data locally (v4.8 player and running offline fixation detection since the data are collected with v2.8 companion app).

I noticed that fixation accuracy is heavily influenced by blinking in many cases. In the linked example, it seems like fixation is assigned based on the instantaneous gaze positions post-blink (note these gaze are marked as "blink" by the blink detector), while the more accurate position (with gaze stable for >500ms) is higher above.

Is there a way to customize the algorithm such that fixation estimation ignores the initial post-blink gazes? Or register these as two separate events?

Chat image Chat image

nmt 01 August, 2025, 02:40:08

Hi @user-b442f7! Would you be able to show a screen recording showing the progression of the fixation and gaze? That would be helpful for us to understand what you're describing.

End of July archive