-
Notifications
You must be signed in to change notification settings - Fork 163
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Revisiting space definition in BIDS (reference frames, coordsys) with respect to Motion extension #1488
Comments
This is a great effort. I think one important part that is not currently addressed is the synchronization with other modalities. Right now, I think the consensus is
I know this is slightly outside the scope of motion capture, but still important. |
But Arno, how are modality-shared 'sync' events to be recorded? For
example, in studies recording EEG and eye tracking, no such common events
are logged. We know that our own and the 100s of other studies now using
LabStreamingLayer avoid the need for such a hardware-enabled solution.
…On Tue, May 9, 2023 at 7:08 PM Arnaud Delorme ***@***.***> wrote:
This is a great effort. I think one important part that is not currently
addressed is the synchronization with other modalities. Right now, I think
the consensus is
1. Use the scan.tsv file "acq_time" variable to determine the relative
time in both data files (for example, motion and EEG) and align the first
sample.
2. Use common events in the event.tsv files of both modalities to
further align the recording and take into account small differences in
sampling rates between the different modalities. I believe these events
should have a specific type ("sync"?), or maybe there could be a separate
column with boolean values indicating if specific events can be used for
synching.
I know this is slightly outside the scope of motion capture, but still
important.
—
Reply to this email directly, view it on GitHub
<https://urldefense.com/v3/__https://github.com/bids-standard/bids-specification/issues/1488*issuecomment-1541000299__;Iw!!Mih3wA!D9vvHVfOhobujz0JRoGyHUOwtOMh2pBm5A0fYerGJYyI5_aZ1hR07Qrbhvq605DCZWqCzuO1Wg0BO37L5rXnJLZO$>,
or unsubscribe
<https://urldefense.com/v3/__https://github.com/notifications/unsubscribe-auth/AKN2SFTFEZKK3R4JGZL3LCLXFLE45ANCNFSM6AAAAAAX3LD77I__;!!Mih3wA!D9vvHVfOhobujz0JRoGyHUOwtOMh2pBm5A0fYerGJYyI5_aZ1hR07Qrbhvq605DCZWqCzuO1Wg0BO37L5lG9fR7Z$>
.
You are receiving this because you were mentioned.Message ID:
***@***.***>
--
Scott Makeig, Research Scientist and Director, Swartz Center for
Computational Neuroscience, Institute for Neural Computation, University of
California San Diego, La Jolla CA 92093-0559, http://sccn.ucsd.edu/~scott
|
Not everybody uses LSL. It is common to have multiple recording apparatus synched by TTL pulses. If the data is recorded with LSL, then such events can be created from the data streams. Computer clock frequencies are not exact. For example, when you want 300 Hz sampling frequency, the signal might be acquired at 300.01 Hz. Seems like nothing, but 0.01 Hz is 120 ms after one hour (0.01*3600 = 36 samples which represents 36/300=0.12 seconds), which is very large in terms of EEG synchronization standard and cannot be ignored. |
In a recent experiment @helenacockx and I used a combination of methods in an experiment where the task (LSL), NIRS (LSL), motion capture (TTL) and video+audio with camcorders (inaudible beep in one of the stereo channels) needed to be synchronized. We used the There are 4 experimental blocks or runs, each with one raw nirs and one synchronous motion recording. We are still discussing whether the motion needs to be split into two files, to accommodate the different reference frames in which the motion data is expressed as discussed in the 1st post. Let me know if anything is unclear w.r.t. synchronization, but preferably through another channel, as it is not specific to motion data. |
Arno - You write:
... synched by TTL pulses. If the data is
recorded with LSL, then such events can be created from the data streams.
Actually, LSL does create such 'virtual events' - e.g., one 'time_stamp
event' for every sample in each data stream. ... Naturally, including all
these 'virtual events' in the 'events.tsv' files for each modality is not
optimal; the time_samp latencies should be far more simply and effectively
maintained as a channel of type 'LSLtime' in the data matrix, or located
within the 'run/scan' folder for each recording modality. If, for some
modalities, there are occasional TTL pulse events as well, their latencies
might be conveniently noted in the events files.
…On Wed, May 10, 2023 at 12:15 AM Arnaud Delorme ***@***.***> wrote:
It is common to have multiple recording apparatus synched by TTL pulses.
If the data is recorded with LSL, then such events can be created from the
data streams. Computer clock frequencies are not exact. For example, when
you want 300 Hz sampling frequency, the signal might be acquired at 300.01
Hz. Seems like nothing, but 0.01 Hz is 120 ms after one hour (0.01*3600 =
36 samples which represents 36/300=0.12 seconds), which is very large in
terms of EEG synchronization standard and cannot be ignored.
—
Reply to this email directly, view it on GitHub
<https://urldefense.com/v3/__https://github.com/bids-standard/bids-specification/issues/1488*issuecomment-1541329714__;Iw!!Mih3wA!EAM_n44UguruHcBGxMcz4vyM6Jawfx_yeH_AKydZSEIzfHvCGX9Sr0JwvZbMOEa1MExTx4P0UYD6gzdpWwwgiFq1$>,
or unsubscribe
<https://urldefense.com/v3/__https://github.com/notifications/unsubscribe-auth/AKN2SFWZJFE4Q5KAZWTWGVTXFMI7ZANCNFSM6AAAAAAX3LD77I__;!!Mih3wA!EAM_n44UguruHcBGxMcz4vyM6Jawfx_yeH_AKydZSEIzfHvCGX9Sr0JwvZbMOEa1MExTx4P0UYD6gzdpW9aa9_m3$>
.
You are receiving this because you were mentioned.Message ID:
***@***.***>
--
Scott Makeig, Research Scientist and Director, Swartz Center for
Computational Neuroscience, Institute for Neural Computation, University of
California San Diego, La Jolla CA 92093-0559, http://sccn.ucsd.edu/~scott
|
Thanks Robert. For the scan.tsv files, it seems that it is also possible to specify milliseconds. It was not done in your example, you used seconds, but milliseconds may be used right? Scott, yes an additional channel for synchronization is would work as well. |
@smakeig @arnodelorme Yeah in response to related discussions in the past we included Latency channel here, which can be especially useful for LSL recorded data sets. |
So earlier today @helenacockx @robertoostenveld @JuliusWelzel and I had another meeting to further discuss the space related standards, how the current spec can be already limiting future developments. The example data set (XSens) in Robert's comment above actually illustrates a scenario where having one "SpatialAxes" (as well as RotationOrder and RotationRule) defined per .json can be inappropriate. A suggested solution is
The motivation to do this is to have some structure in place for the potential future developments to be plugged into, instead of having a half-informative standardized field (i.e. the current Specification itself, example data set, and bids-validator will then have to be updated accordingly. |
I support taking a conservative approach and not prematurely committing to a format and specification that locks future developments out. The HED Working Group struggled with specifying coordinate systems and frames of reference in HED tags and eventually concluded that this is a very hard problem that needed to be tackled more globally. Thanks for your work on this. |
Hey @arnodelorme, yes, a channel of type LATENCY does allow sub-second precision. |
And in the |
I am happy to see that practical dataset curation has revealed some issues and that we get a chance to solve them before a release of the spec. +1 to go ahead with the changes in a PR, which can then be reviewed and discussed.
I am not so happy about any solutions that involve .bidsignore. Why can't the definitions go into the respective |
There is currently no such thing as channels.json, or do you mean to create a new file channels.json? |
I do like the solution suggested by Stefan, completely forgot about channels.json. But as Helena points out I somehow can't find the official documentation about channels.json specifically, although I remember it being brought up in some discussions. I think it is inferred from this bit on general principles for tsv files. |
Agree about |
yes, that is what I was referring to:
|
Helena just proposed in her draft dataset
I suggest the following
which makes the two more equal/balanced. Rather than using an "Object of objects" (aka nested objects) for I also added the initial calibration position (I hope I have that correct) and removed an "s" from "channels" in |
@robertoostenveld This solution would be more informal when data is shared and relevant information is not standardized. The previously introduced fields in
Overall I like the solutions using |
For me both solutions are fine. I could still add this to the CoordinateSystemDescription of the global reference frame: "Global coordinate system relative to the initial calibration position of the participant at the start of the session." |
@sappelhoff do you know whether Julius' suggestion above would pass the validator? It uses a nested objects under |
I know @sjeung had a solution for nested .json files in the validator at some point of Motion-BIDS |
we would have to make adjustments to the JSON schema for this particular case, but it'd be doable. And for the new schema-based validator it shouldn't be a problem either (after the structure is implemented). |
That would be wonderful @sappelhoff! Would we prefer to have the channels.json on the top level, or in each motion folder? I can imagine that a channels.json on the top level can lead to confusion as it does not apply to the channels.tsv for the other folders (e.g., nirs, eeg). |
As I am personally in favor of standardizing to some degree, this seems like a good solution, with a clearer definition of applicable situations. And yes, nested json for validator had been implemented in the past and it is doable. |
Regarding requirement level, I would only leave in the three fields as "recommended" in the standard (SpatialAxes, RotationOrder, RotationRule) because only these can have entries that are machine readable, and allow people to add optional additional fields such as CoordinateSystemDescription, and leave its exact naming for future updataes. |
Sein - Your proposal seems quite workable to me. Now, re nested reference
frames (head < torso < room; screen<room; finger<hand<torso<room; etc.) can
there at least be a standard for specifying how connecting frames are
linked?
For example, the (2D) screen is centered in the room at (x,y,z) and angled
outward wrt room coordinates (mu,theta), producing a json description as
"screen_frame" { ...}
Scott
…On Fri, May 12, 2023 at 8:19 AM Sein Jeung ***@***.***> wrote:
@robertoostenveld
<https://urldefense.com/v3/__https://github.com/robertoostenveld__;!!Mih3wA!Cz-Y_WV2ZpUCAoaJC7XBK9L7Cho5ujvnsckTxFKj3cyGEvYV4oPavtGSk3HFWr0Zu0AZCbDGdevfvVzJ6y9GaSCu$>
This solution would be more informal when data is shared and relevant
information is not standardized. The previously introduced fields
<https://urldefense.com/v3/__https://bids-specification.readthedocs.io/en/latest/modality-specific-files/motion.html*motion-specific-fields__;Iw!!Mih3wA!Cz-Y_WV2ZpUCAoaJC7XBK9L7Cho5ujvnsckTxFKj3cyGEvYV4oPavtGSk3HFWr0Zu0AZCbDGdevfvVzJ65Vp_uGc$>
in *_motion.json regarding a reference_frame would be dropped. This would
lead to a lower level of standardization, even for the cases where the
previously introduced fields work. I would suggest moving fields regarding
reference frames (SpatialAxes, RotationOrder, RotationRule) from the
*_motion.json to the *_channels.json and adding the option n/a if
something is not specified like elsewhere in BIDS
<https://urldefense.com/v3/__https://bids-specification.readthedocs.io/en/stable/appendices/coordinate-systems.html*ieeg-specific-coordinate-systems__;Iw!!Mih3wA!Cz-Y_WV2ZpUCAoaJC7XBK9L7Cho5ujvnsckTxFKj3cyGEvYV4oPavtGSk3HFWr0Zu0AZCbDGdevfvVzJ67VRlPDs$>.
This would be my suggestion for Helenas dataset:
{
"reference_frame": {
"LongName": "reference frame",
"Description": "reference frame in which the channels is represented.",
"Levels": {
"global": {
"CoordinateSystemDescription": "n/a",
"SpatialAxes": "ALS",
"RotationOrder": "ZXY",
"RotationRule": "right-hand"
},
"local": {
"CoordinateSystemDescription": "joint angles are described following the ISB-based coordinate system, with a local reference frame attached to the body segment. See Wu and Cavanagh (1995), Wu et al. (2002), Wu et al. (2005), and the Xsens MVN Awinda user manual for more information.",
"SpatialAxes": "n/a",
"RotationOrder": "n/a",
"RotationRule": "n/a"
}
}
}
}
Overall I like the solutions using *_channels.json here.
As I am personally in favor of standardizing to some degree, this seems
like a good solution, with a clearer definition of applicable situations.
But if we do some standardization here (to make the validator work and to
be useful for applicable cases) again I have to revive the discussion
around ALS versus FLU. From what I understood this data set has axes
definition bound to external space not body coordinate system. So why not
FLU? We can make the FLU thing a "recommended" standard now and add in ALS
system for situations when axes have anatomical meaning (but users can
still use it now if we allow freeform specification).
And yes, nested json for validator had been implemented in the past and it
is doable.
—
Reply to this email directly, view it on GitHub
<https://urldefense.com/v3/__https://github.com/bids-standard/bids-specification/issues/1488*issuecomment-1545655512__;Iw!!Mih3wA!Cz-Y_WV2ZpUCAoaJC7XBK9L7Cho5ujvnsckTxFKj3cyGEvYV4oPavtGSk3HFWr0Zu0AZCbDGdevfvVzJ68vuQVkN$>,
or unsubscribe
<https://urldefense.com/v3/__https://github.com/notifications/unsubscribe-auth/AKN2SFX3VOONT2WWGJHS633XFYTD7ANCNFSM6AAAAAAX3LD77I__;!!Mih3wA!Cz-Y_WV2ZpUCAoaJC7XBK9L7Cho5ujvnsckTxFKj3cyGEvYV4oPavtGSk3HFWr0Zu0AZCbDGdevfvVzJ6-lrI2Cn$>
.
You are receiving this because you were mentioned.Message ID:
***@***.***>
--
Scott Makeig, Research Scientist and Director, Swartz Center for
Computational Neuroscience, Institute for Neural Computation, University of
California San Diego, La Jolla CA 92093-0559, http://sccn.ucsd.edu/~scott
|
Hi Scott, for sure this is on the list of aspects we want to consider in the next version. But as Kay also mentioned we should probably take more time thinking about this and hopefully collect more community input |
I do not really see how using FLU instead of ALS would solve the problem of not having a meaningful description of up/down, forward/backward. Anterior is just the Latin translation for 'to the front', and superior is just the Latin translation for 'up' or 'above'. I neither see a big problem for misinterpretation of the axis definition. For instance, if the global reference frame of my dataset is described as |
Sure, when it comes to the exact meaning it makes no difference. But I wanted to consider the conventional usage of labels for those axes (as detailed in the first post of this thread, and also the point pro ALS being about anatomical conventions in the first place - which are mostly about static anatomical locations) and make the association less restrictive. I can imagine how ALS planes will be more appropriate if there is a data set where everything is referenced to the individual's torso but other than that description of motion doesn't have to be body-bound. This is of course even more so in spatial cognition studies or pointing tasks that I am used to which are not full-on biomechanics. In any case, I think the standard related decisions don't have to be made in a hurry and thanks for testing the ideas around with your data set so far. |
@helenacockx In any case, even if you finalize your data set now with ALS string it will mostly likely be still compatible with the spec because we don't intend to be exclusive about how it should be describe as of now, so the FRU style description will only be recommendation, provided it gets implemented that way. |
Hello everyone, I am currently reworking the specifications to match the discussion above. I noticed, that MEG, EEG, and iEEG have their "own" definition of coordinate systems within BIDS. So far there is also a section for Image-based Coordinate Systems. To move forward, I would like to use on one of two solutions:
Let me know what you think. Best, |
I am in favor of solution 1: making |
Hi everyone, after more discussion and thinking we are converging to the approach of using Concretely
Will get to opening all related PRs unless there is any foreseeable problem with this approach. |
which will be standardized later, see: bids-standard/bids-specification#1488 (comment)
@sjeung @JuliusWelzel could you please summarize this thread and tell us what's still needed from your side, now that this one has been merged? |
Hello @sappelhoff, this thread started out because the information about a coordinate system was previously stored in the
To summarize, this approach allows the user to define the coordinate system per channel for a tracking system. Additionally, we allow a more free-form description if needed. What we did not cover: |
Great, thank you very much for the summary @JuliusWelzel -- I will close this issue now and leave it up to you and Sein to decide to reopen a new, concise and targeted issue that could be about the coordinate systems. |
Dear everyone in BIDS community,
In the past couple of days, I have been in corresponding with @robertoostenveld to revisit some points regarding space definition, within the broader context of entire BIDS, not just motion tracking data. He had very helpful insights especially in terms of some pitfalls to beware of when contributing an extension to BIDS. Some of these points were brought up in the past, and in the very first drafts of Motion-BIDS we were trying to be more proactive in defining things such as nested reference frames, but later on decided to refrain from being specific. Still we are hoping for future developments to gradually resolve these problems and here I would like to put together what gaps we have identified in the process, reasonings behind why some ideas were taken or rejected, and make sure we can generally agree on how we deal with this as of now.
Please chime in to let us know what you think and especially whether you can agree on “suggestions” at the end of this post.
Current specification
_coordsys.json
file. See this page on coordinate systems._motion.json
which describes spatial axes used, using generic keywords for space such as forward-backward.Conflicts & issues
BIDS extension for Motion data is not making reference to already defined concepts within BIDS. Specifically, following definitions can be relevant :
a. Axis labels such as
ALS
(anterior-left-superior) used in anatomical locations, notFLU
(forward-left-up) in _motion.jsonb.
Anatomical
space versus non-anatomicaldevice
(or sensor) space, the distinction has already been madeWithin the same tracking system, motion channels can have distinct reference frames defined in different spaces, and currently there is no way to communicate this in neither
_motion.json
nor_channels.tsv
a. Anatomical space (origin of the space itself can move): Anatomical space can be hierarchical. For instance, joint angles can be expressed around the axis tied to the forearm. This value would not change, no matter which way the participant is facing in the global space. Or when the head moves, and the point between ears also move with the head, then the electrode locations defined with respect to the head reference frame will remain static throughout.
b. Sensor (Device) space : Here the origin can be a single fixed point within the room or multiple fixed points (IMUs with respective positions at the onset of recording)
Until the release of next BIDS version, the goal is not to immediately solve these issues, but to make sure the current specification will not be hindrance to future developments. Considering backwards compatibility, we should check what may be required for “re-attaching the head back to the body” within BIDS. Now, it is difficult to know ahead of time, what can cause compatibility issues with future solutions (for example anatomical library of sensor positions), without thinking about how these solutions will look like and how they will be plugged into BIDS. Ideally, when the space definitions for head (electrodes, MR images…) and body parts are connected seamlessly, we should be able to map the static sensor/electrode locations to the global space, following the movement of the participant (regardless of whether this is a useful thing to do at the moment).
Suggestions from my side
Motion-BIDS does not map the axes to the anatomical reference frame yet.
a. Anatomical reference frame is what is referred to in EEG, MEG .etc. At the moment, description of MOTION coordinate system in anatomical terms (ALS, origin somewhere…) is not meaningful, because the definition of those “origins” or “fiducials” are missing. If this were to be used, a significant amount of data wrangling (inverse kinematics, etc.) should be accompanied.
b. It is true that when we visualize or try to interpret motion data, there are at least three scenarios, as Robert pointed out. Sometimes the origins of each channel will all be the same axes (floor center). Other times, like in IMU, the origins will be shifted from each other. Or the origin itself will be moving throughout recording, when the finger is treated as the “child” object of hand within the hierarchy and that “local” transform is recorded with respect to the center of the hand. Unless we know how those differing origins or moving object ID can be communicated, it is not meaningful to make this distinction in the current spec.
c.
_channels.tsv
for motion data has a column “tracked points”, which now contains user-defined labels of whatever is being tracked. This can in the future refer to an external library and using this information, one can infer where the anatomical origin used in neuroimaging is (e.g., we find out where the center point between ears is, with respect to moving body parts, through inverse kinematics), then really start talking about anatomical coordinate systems.We will update the description for field “SpatialAxes” in
_motion.json
, so that it is only given, when the axes (in non-anatomical device space) have a physical meaning (say, when one can really answer to “What’s up?” in the shared data - down is where the floor is in a normal physical room and the direction of the gravitational pull). That way, it will not misguide users to think that IMU axes should have some physical meaning.a. Admittedly, even for room-scale recordings within global space with single origin, definition of left-right and forward-backward is arbitrary. But sharing a string like FRU it will enable users to reasonably visualize the data so that we can make out the shape of a person standing or walking, for instance, without mixing up vertical and horizontal plane or flipping sides relative to each other. When this coherence conveniently exists in the data, we should enable people to share that.
b. In case within the same recording from the same device, some channel subgroup has a differing spatial axes (like joint angle channels using anatomical reference frame), this is to be separated as an independent
_motion.tsv
file with a separate_motion.json
.c. In the future, more fine-grained information about arbitrary spatial axes (where up-down does not mean anything because the axis is the forearm or the axis was tilted with the IMU) can be stored in the
_motion.json
file (as a set of vectors, or some set of points defined on a 3D template of human body model…or even as a separately stored image file - establishment of this convention is what I think goes beyond the scope of BIDS).We will make these points more explicit in the documentation of the motion spec, coordsys page, as well as the manuscript we are working on for Motion-BIDS.
It became a very long post so thank you very much for reading this through. I think the difficulty comes from the fact that motion tracking is mostly not done in anatomical space. It might be a very trivial statement but when anatomy is about the static, unchanging property of what is inherently attached to what, motion tracking (at least in use cases I am familiar with) is simply way more flexible and relational, making it difficult to interpret in terms of anterior-posterior (what is anterior to an animal or human when their looking direction is orthogonal to walking direction?). All of this is why I imagine motion data in literature is not described along anatomical axes. So when asking “which body part was tracked? (= how was the sensor placed?)” it will be useful to use anatomical planes but when asking “how did this body part move?” it’s often enough to have a generic description of spatial axes.
I will wait for your feedback, tagging @JuliusWelzel @helenacockx @sappelhoff @Remi-Gau @guiomar @smakeig @neuromechanist among many others.
The text was updated successfully, but these errors were encountered: