Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 2 additions & 0 deletions docs/source/whats_new.rst
Original file line number Diff line number Diff line change
Expand Up @@ -20,6 +20,7 @@ Version 1.6 (Source - GitHub)
Enhancements
~~~~~~~~~~~~
- Add unified interactive macro table for dataset summary page with 58 metadata columns, SearchPanes filtering, paradigm distribution bar, and CSV export (:gh:`1043`).
- Add rich HTML repr for preprocessing pipeline transformers: pipelines now render as interactive sklearn diagrams in Jupyter notebooks and sphinx-gallery docs, showing the three-stage flow (Raw, Epochs, Array) with readable step names, event lists, and key parameters instead of raw dict dumps (by `Zach Munro`_ and `Bruno Aristimunha`_).
- Expose ``motor_imagery`` and ``mental_arithmetic`` keyword-only parameters on :class:`moabb.datasets.Shin2017A` (default: MI=True, MA=False) and :class:`moabb.datasets.Shin2017B` (default: MI=False, MA=True), allowing users to load both conditions simultaneously while preserving backward compatibility (by `Bruno Aristimunha`_)
- Add resting state annotations and EMG channel support to :class:`moabb.datasets.Lee2019` resting state runs for BIDS export compatibility (by `Bruno Aristimunha`_)
- Skip zip extraction in :class:`moabb.datasets.GuttmannFlury2025` when files are already extracted, with ``/scratch`` fallback for NFS filesystems on compute nodes (by `Bruno Aristimunha`_)
Expand Down Expand Up @@ -868,3 +869,4 @@ API changes
.. _Katelyn Begany: https://github.com/kbegany
.. _Sarthak Tayal: https://github.com/tayal-sarthak
.. _Benedetto Leto: https://github.com/ben9809
.. _Zach Munro: https://github.com/zmunro
4 changes: 2 additions & 2 deletions examples/advanced_examples/plot_pre_processing_steps.py
Original file line number Diff line number Diff line change
Expand Up @@ -90,8 +90,8 @@
pre_procesing_filter_bank_steps = paradigm_filterbank.make_process_pipelines(dataset)

# By default, we have six filter banks, and each filter bank has the same steps.
for i, step in enumerate(pre_procesing_filter_bank_steps):
print(f"Filter bank {i}: {step}")
# Let's display the first one:
pre_procesing_filter_bank_steps[0]

##############################################################################
# How to include extra steps?
Expand Down
Loading
Loading