Skip to content

BUG - MAG L1D jobs failing and duplicate runs #2518

@alastairtree

Description

@alastairtree

Description of the issue

The MAG L1D processing seems to have issues, for example Dec 11th there are no L1D files despite us having data. The job failure logs show a couple of issues:

  • L1D Jobs seems to be running more than once, starting at the same time for same duration. Some jobs do the upload while the duplicate jobs shows skipped uploads
  • errors relating to the spin data not being available that then run later - is L1D running too soon?
  • Multiple versions for L1D files where there does not need to be so many versions

Sample of the errors:

/usr/local/lib/python3.12/site-packages/xarray/namedarray/core.py:261: UserWarning: Duplicate dimension names present: dimensions {'dim0'} appear more than once in dims=('dim0', 'dim0'). We do not yet support duplicate dimension names, but we do allow initial construction of the object. We recommend you rename the dims immediately to become distinct, as most xarray functionality is likely to fail silently if you do not. To rename the dimensions you will need to set the ``.dims`` attribute of each variable, ``e.g. var.dims=('x0', 'x1')``.
  self._dims = self._parse_dimensions(dims)
/usr/local/lib/python3.12/site-packages/imap_processing/spice/spin.py:210: SettingWithCopyWarning: 
A value is trying to be set on a copy of a slice from a DataFrame.
Try using .loc[row_indexer,col_indexer] = value instead
See the caveats in the documentation: https://pandas.pydata.org/pandas-docs/stable/user_guide/indexing.html#returning-a-view-versus-a-copy
  out_df["sc_spin_phase"] = spin_phases
2025-12-12 20:49:49 - WARNING:cdflib.logging:ISTP Compliance Warning: range is listed as the DEPEND_0 for variable URFTOORFO, but the dimensions do not match.
2025-12-12 20:49:49 - WARNING:cdflib.logging:ISTP Compliance Warning: range is listed as the DEPEND_0 for variable URFTOORFI, but the dimensions do not match.
2025-12-12 20:49:49 - WARNING:cdflib.logging:Warning: Variable valid_start_datetime listed DEPEND_0 as , but no variable by that name was found.
2025-12-12 20:49:49 - WARNING:cdflib.logging:Warning: Variable gradiometer_factor listed DEPEND_0 as , but no variable by that name was found.
/usr/local/lib/python3.12/site-packages/xarray/namedarray/core.py:261: UserWarning: Duplicate dimension names present: dimensions {'dim0'} appear more than once in dims=('dim0', 'dim0'). We do not yet support duplicate dimension names, but we do allow initial construction of the object. We recommend you rename the dims immediately to become distinct, as most xarray functionality is likely to fail silently if you do not. To rename the dimensions you will need to set the ``.dims`` attribute of each variable, ``e.g. var.dims=('x0', 'x1')``.
  self._dims = self._parse_dimensions(dims)
2025-12-12 20:49:49 - WARNING:cdflib.logging:Warning: Variable spin_average_application_factor listed DEPEND_0 as , but no variable by that name was found.
2025-12-12 20:49:49 - WARNING:cdflib.logging:Warning: Variable number_of_spins listed DEPEND_0 as , but no variable by that name was found.
2025-12-12 20:49:49 - WARNING:cdflib.logging:Warning: Variable quality_flag_threshold listed DEPEND_0 as , but no variable by that name was found.
2025-12-12 20:49:49 - WARNING:cdflib.logging:Warning: Variable URFTOORFO listed DEPEND_0 as range, but they have different dimension lengths.
2025-12-12 20:49:49 - WARNING:cdflib.logging:Warning: Variable URFTOORFI listed DEPEND_0 as range, but they have different dimension lengths.
File "/usr/local/lib/python3.12/site-packages/imap_processing/spice/spin.py", line 170, in interpolate_spin_data
    raise ValueError(
ValueError: Query times, [5.03107205e+08 5.03107206e+08 5.03107206e+08 ... 5.03193604e+08
 5.03193604e+08 5.03193605e+08] are outside of the spin data range, (np.float64(503086046.34021), np.float64(503156536.025827)).

Duplicate example

Image

Days where we are missing L1D data:
2025-12-10
2025-12-11
2025-12-14 and onwards

Please can we have data for the missing days reprocessed?

Going forwards, while we wait for a proper long term fix, should we just request reprocessing for missing days? Could this be automated somehow?

Steps to reproduce the issue

No response

Expected behavior (What should happen)

No response

Actual behavior (What does happen)

No response

Code Snippet:

Code

Additional notes, affected areas, and suggested fixes

No response

Metadata

Metadata

Assignees

Labels

Ins: MAGRelated to the MAG instrumentbugSomething isn't working

Type

No type

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions