---------------------------------------------------------------------------
MaybeEncodingError Traceback (most recent call last)
/tmp/ipykernel_5614/40083620.py in <module>
----> 1 cktest.fit(feats)
~/miniconda3/envs/standard/lib/python3.9/site-packages/deeptime/base.py in __call__(self, *args, **kwargs)
329 # here we invoke the immutable setting context manager.
330 with self:
--> 331 return self.fit_method(*args, **kwargs)
332
333
~/miniconda3/envs/standard/lib/python3.9/site-packages/deeptime/decomposition/_vamp.py in fit(self, data, n_jobs, progress, **kw)
536
537 def fit(self, data, n_jobs=None, progress=None, **kw):
--> 538 return super().fit(data, n_jobs, progress, _vamp_estimate_model_for_lag, **kw)
~/miniconda3/envs/standard/lib/python3.9/site-packages/deeptime/util/_validation.py in fit(self, data, n_jobs, progress, estimate_model_for_lag, **kw)
215 estimated_models = [None for _ in range(len(args))]
216 with joining(get_context("spawn").Pool(processes=n_jobs)) as pool:
--> 217 for result in progress(pool.imap_unordered(_imap_wrapper, args),
218 total=len(lags_for_estimation), leave=False):
219 estimated_models[result[0]] = result[1]
~/miniconda3/envs/standard/lib/python3.9/site-packages/deeptime/util/platform.py in __iter__(self)
45
46 def __iter__(self):
---> 47 for x in self._x:
48 yield x
49
~/miniconda3/envs/standard/lib/python3.9/multiprocessing/pool.py in next(self, timeout)
868 if success:
869 return value
--> 870 raise value
871
872 __next__ = next # XXX
MaybeEncodingError: Error sending result: '(0, CovarianceKoopmanModel-140689055304384:cov=CovarianceModel-140689055772624:bessels_correction=False,
cov_00=array([[ 0.14842, -0.04204, 0.07373, -0.043 , 0.07192],
[-0.04204, 0.04164, -0.05627, -0.00437, -0.03761],
[ 0.07373, -0.05627, 0.24251, -0.0284 , 0.07516],
[-0.043 , -0.00437, -0.0284 ,...ean_t=array([ 0.11889, 0.06795, 0.19451, -0.07669, -0.01361]),
symmetrized=False],
dim=None, epsilon=1e-06,
instantaneous_coefficients=array([[ 1.22836, 0.10245, 0.86968, -0.85672, 2.75446],
[ 0.78748, -0.74543, 6.44766, 1.68509, 1.40851],
[ 1.80051, -0.18341, 1.1264 , 0.28173, -1.27747],
[-0.13282, -1.704 , 0.79938, -1.14148, 0.05633],
[-0.51658, -2.02416, 0.04631, 1.5041 , 0.30593]]),
instantaneous_obs=<deeptime.basis._base.Concatenation object at 0x7ff4b930a9d0>,
rank_0=None, rank_t=None, scaling=None,
singular_values=array([0.99284, 0.96986, 0.96327, 0.93159, 0.92537]),
timelagged_coefficients=array([[ 1.21888, 0.10375, 0.86981, -0.86834, 2.75489],
[ 0.80913, -0.73404, 6.47044, 1.63579, 1.41194],
[ 1.79924, -0.16619, 1.13919, 0.26392, -1.26038],
[-0.12339, -1.69904, 0.80932, -1.14899, 0.06429],
[-0.50191, -2.02019, 0.0393 , 1.48308, 0.30815]]),
timelagged_obs=<deeptime.basis._base.Concatenation object at 0x7ff4b930aa30>,
var_cutoff=None])'. Reason: 'RuntimeError('Attempted to send CUDA tensor received from another process; this is not currently supported. Consider cloning before sending.')'