The dataset viewer is not available for this split.
Error code: StreamingRowsError
Exception: TypeError
Message: Couldn't cast array of type
struct<final_loss: double, final_grad_norm: double, iterations: int64, margin: double, final_hinge_value: double, final_W_frobenius_max: double, val_loss_history_len: int64, best_val_iteration: int64>
to
{'loss_kind': Value('string'), 'iterations': Value('int64'), 'final_W_frobenius_max': Value('float64')}
Traceback: Traceback (most recent call last):
File "/src/services/worker/src/worker/utils.py", line 99, in get_rows_or_raise
return get_rows(
^^^^^^^^^
File "/src/libs/libcommon/src/libcommon/utils.py", line 272, in decorator
return func(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^
File "/src/services/worker/src/worker/utils.py", line 77, in get_rows
rows_plus_one = list(itertools.islice(ds, rows_max_number + 1))
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/datasets/iterable_dataset.py", line 2690, in __iter__
for key, example in ex_iterable:
^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/datasets/iterable_dataset.py", line 2227, in __iter__
for key, pa_table in self._iter_arrow():
^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/datasets/iterable_dataset.py", line 2251, in _iter_arrow
for key, pa_table in self.ex_iterable._iter_arrow():
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/datasets/iterable_dataset.py", line 494, in _iter_arrow
for key, pa_table in iterator:
^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/datasets/iterable_dataset.py", line 384, in _iter_arrow
for key, pa_table in self.generate_tables_fn(**gen_kwags):
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/datasets/packaged_modules/json/json.py", line 295, in _generate_tables
self._cast_table(pa_table, json_field_paths=json_field_paths),
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/datasets/packaged_modules/json/json.py", line 128, in _cast_table
pa_table = table_cast(pa_table, self.info.features.arrow_schema)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/datasets/table.py", line 2321, in table_cast
return cast_table_to_schema(table, schema)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/datasets/table.py", line 2255, in cast_table_to_schema
cast_array_to_feature(
File "/usr/local/lib/python3.12/site-packages/datasets/table.py", line 1804, in wrapper
return pa.chunked_array([func(chunk, *args, **kwargs) for chunk in array.chunks])
^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/datasets/table.py", line 2011, in cast_array_to_feature
_c(array.field(name) if name in array_fields else null_array, subfeature)
File "/usr/local/lib/python3.12/site-packages/datasets/table.py", line 1806, in wrapper
return func(array, *args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/datasets/table.py", line 2101, in cast_array_to_feature
raise TypeError(f"Couldn't cast array of type\n{_short_str(array.type)}\nto\n{_short_str(feature)}")
TypeError: Couldn't cast array of type
struct<final_loss: double, final_grad_norm: double, iterations: int64, margin: double, final_hinge_value: double, final_W_frobenius_max: double, val_loss_history_len: int64, best_val_iteration: int64>
to
{'loss_kind': Value('string'), 'iterations': Value('int64'), 'final_W_frobenius_max': Value('float64')}Need help to make the dataset viewer work? Make sure to review how to configure the dataset viewer, and open a discussion for direct support.
YAML Metadata Warning:empty or missing yaml metadata in repo card
Check out the documentation for more information.
AtlasZ — Signed Milestones + Framing Artifacts
Private archive of signed milestone JSONs and framing artifacts from the AtlasZ Universal Forecasting Pipeline — a Calibrator-pluggable substrate-agnostic discrimination architecture. Inherits the PaperRoute Phase 1-3 lineage; AtlasZ activates the universal-manifold thesis from Phase 5+.
Contents
verdict/milestones/ signed empirical-result JSONs (signal_name, dataset_hash,
gate-evaluation report, outcome_label, content_hash, timestamp)
HONEST_CLAIMS.md rolling honesty ledger across phases (versioned;
walks back over-claims with empirical receipts)
docs/discipline.md OPSEC + annotation framework (Gates 1-8)
docs/specs/ AtlasZ Universal Forecasting Pipeline design spec
docs/plans/ Phase 5.0 implementation plan (T61-T82)
Solution-space atlas
Every artifact is a vector in a calibrated atlas of "what we tried." Outcomes are labeled A/B/C/D per spec §6.2:
- A: all gates pass (Outcome A on calibrated truth = architecture validated)
- B: most gates pass (architecture works, threshold or tuning needs refinement)
- C: mixed (split-stochastic; flag for multi-comparison check)
- D: most gates fail (falsified — pre-registered escalation triggers)
Lineage
| Phase | Outcome | Anchor |
|---|---|---|
| 2.6 | falsified macro-augmentation hypothesis | 27-stalk macro sheaf |
| 2.7 | bimodal ceiling robust across detector architectures | val→test cascade |
| 2.8 | bimodal ceiling BROKEN on test (first partial confirm) | matched-filter against canonical templates |
| 3.0 | falsified deformation-first multi-horizon | 8-layer L0-L7 trade-trajectory |
| 5.0 | SPY readout Outcome D (Calibrator tuning issue) | first Calibrator-pluggable milestone |
| 5.0a | quark_gluon Outcome B (architecture validated) | cross-domain validation |
| 5.0b | quark_gluon Outcome A 3/3 (Cohen's d substrate-agnostic) | universality empirically anchored |
OPSEC
Private-only. No public surface. Future-agent atlas-walk only. CC-BY-NC-style spirit (private archive; no redistribution). No malfeasance-operationalizable codification.
- Downloads last month
- 49