Skip to content

Commit 44b8792

Browse files
committed
more Heberto feedback
1 parent 405deca commit 44b8792

File tree

1 file changed

+34
-33
lines changed

1 file changed

+34
-33
lines changed

neo/rawio/intanrawio.py

Lines changed: 34 additions & 33 deletions
Original file line numberDiff line numberDiff line change
@@ -75,59 +75,60 @@ def _parse_header(self):
7575
raise FileNotFoundError(f"{filename} does not exist")
7676

7777
if self.filename.endswith(".rhs"):
78-
self.file_type = 'header-attached'
78+
self.file_format = 'header-attached'
7979
self._global_info, self._ordered_channels, data_dtype, header_size, self._block_size = read_rhs(
8080
self.filename
8181
)
82-
# 3 possibilities for rhd files, one combines the header and the data in the same file with suffix `rhd` while the other two separates the data from the header which is always called `info.rhd`
82+
# 3 possibilities for rhd files, one combines the header and the data in the same file with suffix `rhd` while
83+
# the other two separates the data from the header which is always called `info.rhd`
8384
# attached to the actual binary file with data
8485
elif self.filename.endswith(".rhd"):
8586
if filename.name == "info.rhd":
8687
# first we have one-file-per-signal which is where one neo stream/file is saved as .dat files
8788
if any((filename.parent / file).exists() for file in one_file_per_signal_filenames):
88-
self.file_type = 'one-file-per-signal'
89+
self.file_format = 'one-file-per-signal'
8990
raw_file_paths_dict = create_one_file_per_signal_dict(filename.parent)
9091
# then there is one-file-per-channel where each channel in a neo stream is in its own .dat file
9192
else:
92-
self.file_type = 'one-file-per-channel'
93+
self.file_format = 'one-file-per-channel'
9394
raw_file_paths_dict = create_one_file_per_channel_dict(filename.parent)
9495
# finally the format with the header-attached to the binary file as one giant file
9596
else:
96-
self.file_type = 'header-attached'
97+
self.file_format = 'header-attached'
9798

9899
self._global_info, self._ordered_channels, data_dtype, header_size, self._block_size = read_rhd(
99-
self.filename, self.file_type
100+
self.filename, self.file_format
100101
)
101102

102103
# memmap raw data with the complicated structured dtype
103104
# if header-attached there is one giant memory-map
104-
if self.file_type == "header-attached":
105+
if self.file_format == "header-attached":
105106
self._raw_data = np.memmap(self.filename, dtype=data_dtype, mode="r", offset=header_size)
106107
else:
107108
self._raw_data = {}
108-
for stream_index, (stream_index_key, sub_datatype) in enumerate(data_dtype.items()):
109+
for stream_index, (stream_index_key, stream_datatype) in enumerate(data_dtype.items()):
109110
# for 'one-file-per-signal' we have one memory map / neo stream
110-
if self.file_type == "one-file-per-signal":
111+
if self.file_format == "one-file-per-signal":
111112
self._raw_data[stream_index] = np.memmap(
112-
raw_file_paths_dict[stream_index_key], dtype=sub_datatype, mode="r"
113+
raw_file_paths_dict[stream_index_key], dtype=stream_datatype, mode="r"
113114
)
114115
# for one-file-per-channel we have one memory map / channel stored as a list / neo stream
115116
else:
116117
self._raw_data[stream_index] = []
117-
for channel_index, datatype in enumerate(sub_datatype):
118+
for channel_index, channel_datatype in enumerate(stream_datatype):
118119
self._raw_data[stream_index].append(
119120
np.memmap(
120121
raw_file_paths_dict[stream_index_key][channel_index],
121-
dtype=[datatype],
122+
dtype=[channel_datatype],
122123
mode="r",
123124
)
124125
)
125126

126127
# check timestamp continuity
127-
if self.file_type == "header-attached":
128+
if self.file_format == "header-attached":
128129
timestamp = self._raw_data["timestamp"].flatten()
129130
# timestamps are always the last stream
130-
elif self.file_type == "one-file-per-signal":
131+
elif self.file_format == "one-file-per-signal":
131132
timestamp = self._raw_data[max(self._raw_data.keys())]["timestamp"].flatten()
132133
else:
133134
timestamp = self._raw_data[max(self._raw_data.keys())][0]["timestamp"].flatten()
@@ -168,9 +169,9 @@ def _parse_header(self):
168169

169170
self._max_sampling_rate = np.max(signal_channels["sampling_rate"])
170171

171-
if self.file_type == 'header-attached':
172+
if self.file_format == 'header-attached':
172173
self._max_sigs_length = self._raw_data.size * self._block_size
173-
elif self.file_type == 'one-file-per-signal':
174+
elif self.file_format == 'one-file-per-signal':
174175
self._max_sigs_length = max(
175176
[
176177
raw_data.size * self._block_size
@@ -216,9 +217,9 @@ def _get_signal_size(self, block_index, seg_index, stream_index):
216217
signal_channels = self.header["signal_channels"][mask]
217218
channel_names = signal_channels["name"]
218219
chan_name0 = channel_names[0]
219-
if self.file_type == "header-attached":
220+
if self.file_format == "header-attached":
220221
size = self._raw_data[chan_name0].size
221-
elif self.file_type == 'one-file-per-signal':
222+
elif self.file_format == 'one-file-per-signal':
222223
size = self._raw_data[stream_index][chan_name0].size
223224
else:
224225
size = self._raw_data[stream_index][0][chan_name0].size
@@ -245,9 +246,9 @@ def _get_analogsignal_chunk(self, block_index, seg_index, i_start, i_stop, strea
245246
channel_indexes_are_none = True
246247
channel_names = signal_channels["name"][channel_indexes]
247248

248-
if self.file_type == 'header-attached':
249+
if self.file_format == 'header-attached':
249250
shape = self._raw_data[channel_names[0]].shape
250-
elif self.file_type == 'one-file-per-signal':
251+
elif self.file_format == 'one-file-per-signal':
251252
shape = self._raw_data[stream_index][channel_names[0]].shape
252253
else:
253254
if channel_indexes_are_none:
@@ -268,9 +269,9 @@ def _get_analogsignal_chunk(self, block_index, seg_index, i_start, i_stop, strea
268269

269270
sigs_chunk = np.zeros((i_stop - i_start, len(channel_names)), dtype="uint16")
270271
for i, chan_name in enumerate(channel_names):
271-
if self.file_type == 'header-attached':
272+
if self.file_format == 'header-attached':
272273
data_chan = self._raw_data[chan_name]
273-
elif self.file_type == 'one-file-per-signal':
274+
elif self.file_format == 'one-file-per-signal':
274275
data_chan = self._raw_data[stream_index][chan_name]
275276
else:
276277
if channel_indexes_are_none:
@@ -537,14 +538,14 @@ def read_rhs(filename):
537538
}
538539

539540

540-
def read_rhd(filename, file_type: str):
541+
def read_rhd(filename, file_format: str):
541542
"""Function for reading the rhd file header
542543
543544
Parameters
544545
----------
545546
filename: str | Path
546547
The filename of the *.rhd file
547-
file_type: 'header-attached' | 'one-file-per-signal' | 'one-file-per-channel'
548+
file_format: 'header-attached' | 'one-file-per-signal' | 'one-file-per-channel'
548549
Whether the header is included with the rest of the data ('header-attached')
549550
Or as a standalone file ('one-file-per-signal' or 'one-file-per-channel')
550551
"""
@@ -578,7 +579,7 @@ def read_rhd(filename, file_type: str):
578579

579580
# read channel group and channel header
580581
channels_by_type = {k: [] for k in [0, 1, 2, 3, 4, 5]}
581-
if not file_type == "header-attached":
582+
if not file_format == "header-attached":
582583
data_dtype = {k: [] for k in range(7)} # 5 streams + 6 for timestamps for not header attached
583584
for g in range(global_info["nb_signal_group"]):
584585
group_info = read_variable_header(f, rhd_signal_group_header)
@@ -602,12 +603,12 @@ def read_rhd(filename, file_type: str):
602603
ordered_channels = []
603604

604605
if version >= V("1.2"):
605-
if file_type == "header-attached":
606+
if file_format == "header-attached":
606607
data_dtype = [("timestamp", "int32", BLOCK_SIZE)]
607608
else:
608609
data_dtype[6] = [("timestamp", "int32", BLOCK_SIZE)]
609610
else:
610-
if file_type == "header-attached":
611+
if file_format == "header-attached":
611612
data_dtype = [("timestamp", "uint32", BLOCK_SIZE)]
612613
else:
613614
data_dtype[6] = [("timestamp", "uint32", BLOCK_SIZE)]
@@ -620,7 +621,7 @@ def read_rhd(filename, file_type: str):
620621
chan_info["gain"] = 0.195
621622
chan_info["offset"] = -32768 * 0.195
622623
ordered_channels.append(chan_info)
623-
if file_type == "header-attached":
624+
if file_format == "header-attached":
624625
data_dtype += [(name, "uint16", BLOCK_SIZE)]
625626
else:
626627
data_dtype[0] += [(name, "uint16", BLOCK_SIZE)]
@@ -633,7 +634,7 @@ def read_rhd(filename, file_type: str):
633634
chan_info["gain"] = 0.0000374
634635
chan_info["offset"] = 0.0
635636
ordered_channels.append(chan_info)
636-
if file_type == "header-attached":
637+
if file_format == "header-attached":
637638
data_dtype += [(name, "uint16", BLOCK_SIZE // 4)]
638639
else:
639640
data_dtype[1] += [(name, "uint16", BLOCK_SIZE // 4)]
@@ -646,7 +647,7 @@ def read_rhd(filename, file_type: str):
646647
chan_info["gain"] = 0.0000748
647648
chan_info["offset"] = 0.0
648649
ordered_channels.append(chan_info)
649-
if file_type == "header-attached":
650+
if file_format == "header-attached":
650651
data_dtype += [(name, "uint16")]
651652
else:
652653
data_dtype[1] += [(name, "uint16", BLOCK_SIZE // 4)]
@@ -677,7 +678,7 @@ def read_rhd(filename, file_type: str):
677678
chan_info["gain"] = 0.0003125
678679
chan_info["offset"] = -32768 * 0.0003125
679680
ordered_channels.append(chan_info)
680-
if file_type == "header-attached":
681+
if file_format == "header-attached":
681682
data_dtype += [(name, "uint16", BLOCK_SIZE)]
682683
else:
683684
data_dtype[3] += [(name, "uint16", BLOCK_SIZE)]
@@ -696,7 +697,7 @@ def read_rhd(filename, file_type: str):
696697
chan_info["gain"] = 1.0
697698
chan_info["offset"] = 0.0
698699
ordered_channels.append(chan_info)
699-
if file_type == "header-attached":
700+
if file_format == "header-attached":
700701
data_dtype += [(name, "uint16", BLOCK_SIZE)]
701702
else:
702703
data_dtype[sig_type] += [(name, "uint16", BLOCK_SIZE)]
@@ -706,7 +707,7 @@ def read_rhd(filename, file_type: str):
706707
else:
707708
global_info["notch_filter_applied"] = False
708709

709-
if not file_type == "header-attached":
710+
if not file_format == "header-attached":
710711
# filter out dtypes without any values
711712
data_dtype = {k:v for (k,v) in data_dtype.items() if len(v) > 0}
712713

0 commit comments

Comments
 (0)