File size: 32,023 Bytes
6beeb32
 
 
7afa652
a739565
 
 
 
 
bfe37a4
 
3bacc03
 
 
 
 
 
 
d5e074f
7a692b2
7afa652
8e50edc
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
7a692b2
 
 
 
 
 
 
a89084a
 
 
 
 
ee0772a
 
 
 
a89084a
 
99e3cfe
 
0e033f3
 
 
 
 
99e3cfe
 
 
 
f73316e
5e11e95
c252009
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
620c133
fdc079a
 
 
 
 
 
 
 
 
 
 
620c133
 
 
9dfb74d
fdc079a
517e4c8
 
 
 
a94b239
517e4c8
6beeb32
43f961a
 
 
 
 
 
 
 
 
648c108
 
 
43f961a
af4d03c
 
92861de
af4d03c
 
 
 
831220c
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
cadb7b8
 
 
831220c
 
 
 
 
 
 
ea3b079
a72c82f
 
 
 
99e4135
 
 
 
4cd963b
 
 
 
 
cca415d
 
 
 
 
 
 
6beeb32
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
---
license: apache-2.0
---
# Prototype v12 - Patch 16 - 128x128 images - 1.2m imagenet images

Well the 64x64 image set worked just fine, so it's time to upgrade and test the limits of the architecture.

Can it simply... scale? ooooor do we need more solvers along the way to compensate?

benjamin-paine/imagenet-1k-128x128

Can we actually solve it

YES WE DID!

![image](https://cdn-uploads.huggingface.co/production/uploads/630cf55b15433862cfc9556f/ALPrNBXk6oGTZ7HX3SkxX.png)

It seems geometric manifolds learn... differently than standard manifolds, don't they.


# Prototype V11 - Patch16 - MSE 0.0005 - 64x64 tiny imagenet

I'd say it works.

The images show it works. It works.


![image](https://cdn-uploads.huggingface.co/production/uploads/630cf55b15433862cfc9556f/XwBzrSvL1HL3wwUkIDQlz.png)


```
Using geolip-core SVD (fp64 Gram+eigh (FL=available, N<=12))
PatchSVAE - 16 patches of 16×16
  Dataset: tiny_imagenet (64×64, 200 classes)
  Per-patch: (256, 16) = 4096 elements, rows on S^15
  Encoder/Decoder: hidden=768, depth=4 (residual blocks)
  Cross-attention: 2 layers on S vectors (2,272 params)
  Soft hand: boost=1.5x near CV=0.125, penalty=0.3 far
  Total params: 16,942,419
===============================================================================================
  ep |    loss   recon  t/ep |   t_rec |     S0     SD ratio erank |  row_cv  prox    rw | S_delta
-----------------------------------------------------------------------------------------------
   1 |  0.2595  0.1806  12.2 |  0.1024 |  5.036  3.254  1.55 15.87 |  0.2007 0.905  1.45 | 0.09694 a:0.0242/0.0247
   2 |  0.1216  0.0845  12.3 |  0.0675 |  5.071  3.298  1.54 15.88 |  0.2018 0.885  1.44 | 0.17411 a:0.0251/0.0257
   3 |  0.0847  0.0587  12.3 |  0.0470 |  5.093  3.312  1.54 15.88 |  0.2046 0.869  1.43 | 0.19894 a:0.0258/0.0265
   4 |  0.0623  0.0432  12.3 |  0.0430 |  5.115  3.323  1.54 15.88 |  0.2006 0.864  1.43 | 0.20848 a:0.0264/0.0272
   6 |  0.0359  0.0248  12.3 |  0.0198 |  5.129  3.332  1.54 15.88 |  0.2006 0.907  1.45 | 0.21832 a:0.0273/0.0281
   8 |  0.0225  0.0155  12.2 |  0.0196 |  5.149  3.341  1.54 15.87 |  0.2017 0.876  1.44 | 0.22351 a:0.0279/0.0287
  10 |  0.0170  0.0116  12.3 |  0.0100 |  5.151  3.352  1.54 15.88 |  0.2035 0.924  1.46 | 0.22671 a:0.0283/0.0290
  12 |  0.0141  0.0096  12.3 |  0.0114 |  5.159  3.354  1.54 15.88 |  0.2009 0.909  1.45 | 0.22924 a:0.0285/0.0293
  14 |  0.0121  0.0082  12.3 |  0.0073 |  5.156  3.362  1.53 15.88 |  0.2018 0.855  1.43 | 0.23137 a:0.0288/0.0296
  16 |  0.0105  0.0072  12.3 |  0.0108 |  5.161  3.363  1.53 15.88 |  0.2003 0.860  1.43 | 0.23316 a:0.0290/0.0298
  18 |  0.0094  0.0064  12.3 |  0.0055 |  5.158  3.365  1.53 15.88 |  0.2017 0.879  1.44 | 0.23467 a:0.0292/0.0300
  20 |  0.0086  0.0058  12.3 |  0.0050 |  5.157  3.367  1.53 15.88 |  0.2023 0.805  1.40 | 0.23601 a:0.0293/0.0301
  22 |  0.0079  0.0054  12.4 |  0.0045 |  5.157  3.369  1.53 15.88 |  0.1996 0.872  1.44 | 0.23726 a:0.0295/0.0303
  24 |  0.0074  0.0050  12.2 |  0.0064 |  5.146  3.380  1.52 15.88 |  0.2044 0.879  1.44 | 0.23848 a:0.0296/0.0305
  26 |  0.0068  0.0046  12.4 |  0.0039 |  5.155  3.372  1.53 15.88 |  0.2036 0.884  1.44 | 0.23955 a:0.0297/0.0306
  28 |  0.0063  0.0042  12.3 |  0.0036 |  5.155  3.378  1.53 15.88 |  0.2077 0.841  1.42 | 0.24057 a:0.0299/0.0307
  30 |  0.0058  0.0038  12.3 |  0.0038 |  5.155  3.380  1.53 15.88 |  0.2027 0.911  1.46 | 0.24149 a:0.0300/0.0309
  32 |  0.0055  0.0036  12.2 |  0.0032 |  5.150  3.383  1.52 15.88 |  0.2045 0.807  1.40 | 0.24239 a:0.0301/0.0310
  34 |  0.0054  0.0036  12.3 |  0.0037 |  5.145  3.388  1.52 15.88 |  0.1996 0.875  1.44 | 0.24329 a:0.0302/0.0311
  36 |  0.0049  0.0032  12.3 |  0.0031 |  5.154  3.385  1.52 15.88 |  0.2054 0.828  1.41 | 0.24409 a:0.0303/0.0312
  38 |  0.0046  0.0030  12.3 |  0.0027 |  5.152  3.390  1.52 15.88 |  0.2038 0.847  1.42 | 0.24490 a:0.0304/0.0313
  40 |  0.0044  0.0029  12.3 |  0.0032 |  5.155  3.392  1.52 15.89 |  0.2046 0.855  1.43 | 0.24566 a:0.0305/0.0314
  42 |  0.0043  0.0028  12.3 |  0.0024 |  5.152  3.395  1.52 15.89 |  0.2064 0.905  1.45 | 0.24637 a:0.0305/0.0315
  44 |  0.0042  0.0027  12.3 |  0.0023 |  5.150  3.395  1.52 15.89 |  0.2084 0.844  1.42 | 0.24705 a:0.0306/0.0316
  46 |  0.0039  0.0025  12.3 |  0.0022 |  5.149  3.400  1.51 15.89 |  0.2057 0.868  1.43 | 0.24776 a:0.0307/0.0317
  48 |  0.0037  0.0024  12.3 |  0.0024 |  5.152  3.403  1.51 15.89 |  0.2138 0.831  1.42 | 0.24843 a:0.0308/0.0318
  50 |  0.0038  0.0024  12.3 |  0.0025 |  5.149  3.406  1.51 15.89 |  0.2078 0.810  1.40 | 0.24906 a:0.0309/0.0319
  52 |  0.0034  0.0021  12.3 |  0.0019 |  5.154  3.405  1.51 15.89 |  0.2082 0.872  1.44 | 0.24965 a:0.0309/0.0320
  54 |  0.0033  0.0020  12.2 |  0.0019 |  5.156  3.406  1.51 15.89 |  0.2085 0.894  1.45 | 0.25022 a:0.0310/0.0320
  56 |  0.0033  0.0020  12.4 |  0.0019 |  5.150  3.412  1.51 15.89 |  0.2058 0.866  1.43 | 0.25079 a:0.0311/0.0321
  58 |  0.0031  0.0019  12.3 |  0.0033 |  5.147  3.416  1.51 15.89 |  0.2071 0.774  1.39 | 0.25135 a:0.0311/0.0322
  60 |  0.0030  0.0018  12.4 |  0.0017 |  5.153  3.415  1.51 15.89 |  0.2134 0.840  1.42 | 0.25187 a:0.0312/0.0323
  62 |  0.0030  0.0018  12.3 |  0.0016 |  5.155  3.416  1.51 15.89 |  0.2080 0.764  1.38 | 0.25235 a:0.0313/0.0323
  64 |  0.0028  0.0017  12.2 |  0.0014 |  5.156  3.416  1.51 15.89 |  0.2100 0.666  1.33 | 0.25285 a:0.0313/0.0324
  66 |  0.0028  0.0017  12.3 |  0.0017 |  5.151  3.419  1.51 15.89 |  0.2101 0.865  1.43 | 0.25333 a:0.0314/0.0324
  68 |  0.0026  0.0015  12.3 |  0.0014 |  5.158  3.419  1.51 15.89 |  0.2078 0.838  1.42 | 0.25381 a:0.0314/0.0325
  70 |  0.0025  0.0015  12.4 |  0.0021 |  5.160  3.422  1.51 15.89 |  0.2112 0.806  1.40 | 0.25428 a:0.0315/0.0326
  72 |  0.0026  0.0015  12.2 |  0.0013 |  5.158  3.422  1.51 15.89 |  0.2126 0.835  1.42 | 0.25471 a:0.0316/0.0326
  74 |  0.0024  0.0014  12.2 |  0.0015 |  5.154  3.427  1.50 15.89 |  0.2143 0.838  1.42 | 0.25514 a:0.0316/0.0327
  76 |  0.0024  0.0014  12.2 |  0.0012 |  5.161  3.424  1.51 15.89 |  0.2151 0.847  1.42 | 0.25553 a:0.0317/0.0327
  78 |  0.0023  0.0013  12.3 |  0.0014 |  5.157  3.428  1.50 15.89 |  0.2121 0.686  1.34 | 0.25592 a:0.0317/0.0328
  80 |  0.0024  0.0013  12.3 |  0.0012 |  5.160  3.428  1.51 15.89 |  0.2068 0.824  1.41 | 0.25630 a:0.0317/0.0328
  82 |  0.0027  0.0016  12.2 |  0.0015 |  5.146  3.394  1.52 15.89 |  0.2065 0.899  1.45 | 0.25687 a:0.0318/0.0329
  84 |  0.0022  0.0013  12.2 |  0.0013 |  5.156  3.405  1.51 15.89 |  0.2092 0.875  1.44 | 0.25709 a:0.0319/0.0329
  86 |  0.0022  0.0012  12.3 |  0.0022 |  5.154  3.413  1.51 15.89 |  0.2091 0.835  1.42 | 0.25726 a:0.0319/0.0329
  88 |  0.0021  0.0012  12.3 |  0.0014 |  5.154  3.417  1.51 15.89 |  0.2074 0.840  1.42 | 0.25740 a:0.0319/0.0329
  90 |  0.0022  0.0013  12.3 |  0.0030 |  5.147  3.416  1.51 15.89 |  0.2191 0.848  1.42 | 0.25753 a:0.0319/0.0329
  92 |  0.0021  0.0011  12.3 |  0.0012 |  5.157  3.418  1.51 15.89 |  0.2123 0.775  1.39 | 0.25766 a:0.0319/0.0329
  94 |  0.0021  0.0011  12.3 |  0.0010 |  5.156  3.419  1.51 15.89 |  0.2117 0.710  1.35 | 0.25779 a:0.0319/0.0329
  96 |  0.0020  0.0011  12.3 |  0.0013 |  5.154  3.420  1.51 15.89 |  0.2166 0.940  1.47 | 0.25793 a:0.0319/0.0330
  98 |  0.0019  0.0011  12.2 |  0.0010 |  5.156  3.421  1.51 15.89 |  0.2143 0.762  1.38 | 0.25807 a:0.0320/0.0330
 100 |  0.0020  0.0010  12.3 |  0.0009 |  5.155  3.422  1.51 15.89 |  0.2173 0.642  1.32 | 0.25821 a:0.0320/0.0330
 102 |  0.0020  0.0010  12.2 |  0.0009 |  5.156  3.423  1.51 15.89 |  0.2165 0.868  1.43 | 0.25835 a:0.0320/0.0330
 104 |  0.0019  0.0010  12.4 |  0.0009 |  5.157  3.423  1.51 15.89 |  0.2125 0.788  1.39 | 0.25850 a:0.0320/0.0330
 106 |  0.0019  0.0009  12.3 |  0.0009 |  5.156  3.424  1.51 15.89 |  0.2219 0.666  1.33 | 0.25866 a:0.0320/0.0331
 108 |  0.0019  0.0009  12.3 |  0.0009 |  5.153  3.425  1.50 15.89 |  0.2202 0.671  1.34 | 0.25881 a:0.0321/0.0331
 110 |  0.0020  0.0009  12.3 |  0.0011 |  5.153  3.427  1.50 15.89 |  0.2163 0.726  1.36 | 0.25896 a:0.0321/0.0331
 112 |  0.0019  0.0009  12.4 |  0.0009 |  5.155  3.427  1.50 15.89 |  0.2205 0.837  1.42 | 0.25911 a:0.0321/0.0331
 114 |  0.0019  0.0008  12.3 |  0.0008 |  5.155  3.427  1.50 15.89 |  0.2220 0.803  1.40 | 0.25926 a:0.0321/0.0332
 116 |  0.0018  0.0008  12.3 |  0.0009 |  5.155  3.427  1.50 15.89 |  0.2211 0.852  1.43 | 0.25942 a:0.0321/0.0332
 118 |  0.0019  0.0008  12.3 |  0.0008 |  5.153  3.429  1.50 15.89 |  0.2207 0.694  1.35 | 0.25957 a:0.0321/0.0332
 120 |  0.0018  0.0008  12.2 |  0.0008 |  5.156  3.429  1.50 15.89 |  0.2271 0.664  1.33 | 0.25972 a:0.0322/0.0332
 122 |  0.0018  0.0008  12.3 |  0.0008 |  5.154  3.429  1.50 15.89 |  0.2266 0.658  1.33 | 0.25986 a:0.0322/0.0333
 124 |  0.0017  0.0007  12.3 |  0.0008 |  5.154  3.431  1.50 15.89 |  0.2201 0.771  1.39 | 0.26000 a:0.0322/0.0333
 126 |  0.0017  0.0007  12.4 |  0.0008 |  5.157  3.430  1.50 15.89 |  0.2253 0.862  1.43 | 0.26014 a:0.0322/0.0333
 128 |  0.0018  0.0007  12.3 |  0.0008 |  5.153  3.431  1.50 15.89 |  0.2222 0.638  1.32 | 0.26027 a:0.0322/0.0333
 130 |  0.0018  0.0007  12.3 |  0.0007 |  5.155  3.431  1.50 15.89 |  0.2255 0.786  1.39 | 0.26040 a:0.0322/0.0333
 132 |  0.0018  0.0007  12.3 |  0.0007 |  5.153  3.432  1.50 15.89 |  0.2325 0.778  1.39 | 0.26053 a:0.0323/0.0333
 134 |  0.0017  0.0007  12.3 |  0.0007 |  5.154  3.433  1.50 15.89 |  0.2250 0.876  1.44 | 0.26065 a:0.0323/0.0334
 136 |  0.0017  0.0006  12.3 |  0.0007 |  5.157  3.432  1.50 15.89 |  0.2269 0.866  1.43 | 0.26077 a:0.0323/0.0334
 138 |  0.0017  0.0006  12.3 |  0.0007 |  5.156  3.433  1.50 15.89 |  0.2247 0.760  1.38 | 0.26088 a:0.0323/0.0334
 140 |  0.0016  0.0006  12.3 |  0.0006 |  5.157  3.433  1.50 15.89 |  0.2242 0.725  1.36 | 0.26099 a:0.0323/0.0334
 142 |  0.0016  0.0006  12.3 |  0.0006 |  5.157  3.433  1.50 15.89 |  0.2241 0.909  1.45 | 0.26109 a:0.0323/0.0334
 144 |  0.0017  0.0006  12.3 |  0.0006 |  5.157  3.433  1.50 15.89 |  0.2287 0.815  1.41 | 0.26119 a:0.0323/0.0334
 146 |  0.0016  0.0006  12.3 |  0.0007 |  5.158  3.434  1.50 15.90 |  0.2205 0.722  1.36 | 0.26128 a:0.0324/0.0334
 148 |  0.0016  0.0006  12.3 |  0.0008 |  5.157  3.434  1.50 15.90 |  0.2286 0.691  1.35 | 0.26137 a:0.0324/0.0335
 150 |  0.0016  0.0006  12.3 |  0.0006 |  5.158  3.434  1.50 15.90 |  0.2259 0.845  1.42 | 0.26146 a:0.0324/0.0335
 152 |  0.0017  0.0006  12.3 |  0.0006 |  5.158  3.434  1.50 15.90 |  0.2295 0.757  1.38 | 0.26154 a:0.0324/0.0335
 154 |  0.0016  0.0005  12.3 |  0.0006 |  5.159  3.435  1.50 15.90 |  0.2304 0.751  1.38 | 0.26162 a:0.0324/0.0335
 156 |  0.0018  0.0005  12.3 |  0.0006 |  5.159  3.435  1.50 15.90 |  0.2264 0.796  1.40 | 0.26169 a:0.0324/0.0335
 158 |  0.0017  0.0005  12.3 |  0.0006 |  5.160  3.434  1.50 15.90 |  0.2282 0.788  1.39 | 0.26176 a:0.0324/0.0335
 160 |  0.0017  0.0005  12.3 |  0.0005 |  5.161  3.434  1.50 15.90 |  0.2291 0.766  1.38 | 0.26183 a:0.0324/0.0335
 162 |  0.0016  0.0005  12.3 |  0.0005 |  5.161  3.434  1.50 15.90 |  0.2282 0.716  1.36 | 0.26189 a:0.0324/0.0335
 164 |  0.0016  0.0005  12.3 |  0.0005 |  5.161  3.435  1.50 15.90 |  0.2344 0.792  1.40 | 0.26196 a:0.0324/0.0335
 166 |  0.0016  0.0005  12.3 |  0.0006 |  5.162  3.434  1.50 15.90 |  0.2305 0.707  1.35 | 0.26202 a:0.0324/0.0335
 168 |  0.0016  0.0005  12.3 |  0.0005 |  5.162  3.434  1.50 15.90 |  0.2353 0.816  1.41 | 0.26207 a:0.0324/0.0335
 170 |  0.0016  0.0005  12.3 |  0.0005 |  5.163  3.434  1.50 15.90 |  0.2296 0.756  1.38 | 0.26213 a:0.0325/0.0335
 172 |  0.0018  0.0005  12.4 |  0.0005 |  5.163  3.434  1.50 15.90 |  0.2391 0.742  1.37 | 0.26218 a:0.0325/0.0336
 174 |  0.0016  0.0005  12.4 |  0.0005 |  5.163  3.434  1.50 15.90 |  0.2307 0.863  1.43 | 0.26224 a:0.0325/0.0336
 176 |  0.0016  0.0005  12.3 |  0.0005 |  5.163  3.434  1.50 15.90 |  0.2329 0.854  1.43 | 0.26228 a:0.0325/0.0336
 178 |  0.0017  0.0005  12.3 |  0.0005 |  5.164  3.434  1.50 15.90 |  0.2287 0.803  1.40 | 0.26233 a:0.0325/0.0336
 180 |  0.0017  0.0005  12.3 |  0.0005 |  5.164  3.434  1.50 15.90 |  0.2361 0.819  1.41 | 0.26237 a:0.0325/0.0336
 182 |  0.0018  0.0005  12.3 |  0.0005 |  5.164  3.434  1.50 15.90 |  0.2360 0.729  1.36 | 0.26241 a:0.0325/0.0336
 184 |  0.0016  0.0005  12.3 |  0.0005 |  5.164  3.434  1.50 15.90 |  0.2395 0.774  1.39 | 0.26245 a:0.0325/0.0336
 186 |  0.0015  0.0005  12.3 |  0.0005 |  5.165  3.434  1.50 15.90 |  0.2325 0.777  1.39 | 0.26248 a:0.0325/0.0336
 188 |  0.0018  0.0005  12.3 |  0.0005 |  5.165  3.434  1.50 15.90 |  0.2430 0.669  1.33 | 0.26250 a:0.0325/0.0336
 190 |  0.0016  0.0005  12.3 |  0.0005 |  5.165  3.434  1.50 15.90 |  0.2293 0.781  1.39 | 0.26252 a:0.0325/0.0336
 192 |  0.0017  0.0005  12.3 |  0.0005 |  5.165  3.434  1.50 15.90 |  0.2246 0.763  1.38 | 0.26254 a:0.0325/0.0336
 194 |  0.0019  0.0005  12.3 |  0.0005 |  5.165  3.434  1.50 15.90 |  0.2324 0.764  1.38 | 0.26255 a:0.0325/0.0336
 196 |  0.0016  0.0005  12.2 |  0.0005 |  5.165  3.434  1.50 15.90 |  0.2301 0.868  1.43 | 0.26256 a:0.0325/0.0336
 198 |  0.0016  0.0005  12.3 |  0.0005 |  5.165  3.434  1.50 15.90 |  0.2332 0.696  1.35 | 0.26256 a:0.0325/0.0336
 200 |  0.0017  0.0005  12.3 |  0.0005 |  5.165  3.434  1.50 15.90 |  0.2397 0.808  1.40 | 0.26256 a:0.0325/0.0336

==========================================================================================
FINAL ANALYSIS
==========================================================================================

  PatchSVAE: 16 patches × (256, 16)
  Target CV: 0.125
  Recon MSE: 0.000489 +/- 0.000743
  Row CV: 0.2397
  Cross-attention S delta: 0.26256

  Learned alpha per mode (coordination strength):
    Layer 0: mean=0.0327  max=0.0336  min=0.0323
      α[ 0]: 0.0324  ######################################
      α[ 1]: 0.0323  ######################################
      α[ 2]: 0.0327  ######################################
      α[ 3]: 0.0326  ######################################
      α[ 4]: 0.0325  ######################################
      α[ 5]: 0.0326  ######################################
      α[ 6]: 0.0332  #######################################
      α[ 7]: 0.0336  #######################################
      α[ 8]: 0.0326  ######################################
      α[ 9]: 0.0324  ######################################
      α[10]: 0.0326  ######################################
      α[11]: 0.0325  ######################################
      α[12]: 0.0328  #######################################
      α[13]: 0.0324  ######################################
      α[14]: 0.0331  #######################################
      α[15]: 0.0326  ######################################
    Layer 1: mean=0.0323  max=0.0327  min=0.0315
      α[ 0]: 0.0324  #######################################
      α[ 1]: 0.0326  #######################################
      α[ 2]: 0.0323  #######################################
      α[ 3]: 0.0326  #######################################
      α[ 4]: 0.0327  #######################################
      α[ 5]: 0.0326  #######################################
      α[ 6]: 0.0320  #######################################
      α[ 7]: 0.0315  ######################################
      α[ 8]: 0.0324  #######################################
      α[ 9]: 0.0327  #######################################
      α[10]: 0.0325  #######################################
      α[11]: 0.0324  #######################################
      α[12]: 0.0321  #######################################
      α[13]: 0.0324  #######################################
      α[14]: 0.0317  ######################################
      α[15]: 0.0322  #######################################

  Coordinated singular value profile:
    S[ 0]:   5.1650  cum=  9.2%  #############################
    S[ 1]:   4.9525  cum= 17.6%  ############################
    S[ 2]:   4.8142  cum= 25.6%  ###########################
    S[ 3]:   4.6335  cum= 32.9%  ##########################
    S[ 4]:   4.5199  cum= 40.0%  ##########################
    S[ 5]:   4.4203  cum= 46.7%  #########################
    S[ 6]:   4.3376  cum= 53.2%  #########################
    S[ 7]:   4.2448  cum= 59.3%  ########################
    S[ 8]:   4.1641  cum= 65.3%  ########################
    S[ 9]:   4.0915  cum= 71.1%  #######################
    S[10]:   4.0086  cum= 76.6%  #######################
    S[11]:   3.9144  cum= 81.9%  ######################
    S[12]:   3.7995  cum= 86.8%  ######################
    S[13]:   3.6926  cum= 91.5%  #####################
    S[14]:   3.5949  cum= 95.9%  ####################
    S[15]:   3.4336  cum=100.0%  ###################

  Saving reconstruction grid...
  Saved to /content/svae_patch_recon.png
```


# Prototype V10.3 - Patch16 - The VIT size.

Might need some tweaks but I don't think so. We're approaching actual vit prototype accuracy now.

Lets see how the SVAE performs. 

# Prototype V10.2 - Patch32 - Patchwork Cross-Attention with Edge Smoothing

This eliminates the edge cutting of the last version, and in the process the recon accuracy has gone up.

Model still escapes the discharge within 2 epochs and has robust recon.

Defeated the last version.
![image](https://cdn-uploads.huggingface.co/production/uploads/630cf55b15433862cfc9556f/d-HNotWpeM2vwEKkmxObP.png)


![image](https://cdn-uploads.huggingface.co/production/uploads/630cf55b15433862cfc9556f/KO3wWJyfcSGxk2er70Pm5.png)

# Prototype V10.1 Patchwork Cross-Attention - Stabilized

The patchwork has stabilized, and the output is more accurate than the original now that it supports SVD 32 with more accuracy and higher speed

Epoch 28 hit the unstable point, but the gradient clipped attention was the ticket that ensured solidity.

The discharge recovered immediately.

![image](https://cdn-uploads.huggingface.co/production/uploads/630cf55b15433862cfc9556f/9t2jk03_E4CEZK3O4g51U.png)


Give or take 97% accurate recall, lets get those numbers up before we move onto more powerful image sets. Roughly 28m params.

![image](https://cdn-uploads.huggingface.co/production/uploads/630cf55b15433862cfc9556f/TLkaOkOxv_4OK_SHinbQC.png)

```
 174 |  0.0471  0.0315   8.0 |  0.0318 |  4.038  2.092  1.93 31.52 |  0.1271 0.995  1.50 | 0.26864 a:0.0471/0.0476
 176 |  0.0471  0.0314   7.9 |  0.0318 |  4.038  2.092  1.93 31.52 |  0.1343 1.000  1.50 | 0.26874 a:0.0471/0.0476
 178 |  0.0471  0.0314   7.9 |  0.0318 |  4.038  2.093  1.93 31.52 |  0.1313 0.995  1.50 | 0.26883 a:0.0471/0.0477
 180 |  0.0471  0.0314   8.0 |  0.0317 |  4.038  2.093  1.93 31.52 |  0.1312 1.000  1.50 | 0.26892 a:0.0471/0.0477
 182 |  0.0471  0.0314   7.9 |  0.0317 |  4.038  2.093  1.93 31.52 |  0.1310 0.995  1.50 | 0.26899 a:0.0471/0.0477
 184 |  0.0471  0.0314   7.9 |  0.0317 |  4.038  2.092  1.93 31.52 |  0.1350 0.993  1.50 | 0.26906 a:0.0472/0.0477
 186 |  0.0470  0.0314   7.9 |  0.0317 |  4.038  2.092  1.93 31.52 |  0.1338 1.000  1.50 | 0.26911 a:0.0472/0.0477
 188 |  0.0470  0.0314   8.0 |  0.0317 |  4.038  2.093  1.93 31.52 |  0.1305 0.999  1.50 | 0.26916 a:0.0472/0.0477
 190 |  0.0470  0.0314   8.0 |  0.0317 |  4.038  2.093  1.93 31.52 |  0.1358 0.999  1.50 | 0.26919 a:0.0472/0.0477
 192 |  0.0470  0.0314   8.0 |  0.0317 |  4.038  2.092  1.93 31.52 |  0.1354 0.999  1.50 | 0.26922 a:0.0472/0.0477
 194 |  0.0470  0.0314   7.9 |  0.0317 |  4.038  2.093  1.93 31.52 |  0.1301 0.992  1.50 | 0.26923 a:0.0472/0.0477
 196 |  0.0470  0.0314   7.9 |  0.0317 |  4.038  2.093  1.93 31.52 |  0.1330 0.998  1.50 | 0.26924 a:0.0472/0.0477
 198 |  0.0470  0.0314   7.9 |  0.0317 |  4.038  2.093  1.93 31.52 |  0.1312 1.000  1.50 | 0.26925 a:0.0472/0.0477
 200 |  0.0470  0.0314   7.9 |  0.0317 |  4.038  2.093  1.93 31.52 |  0.1300 1.000  1.50 | 0.26925 a:0.0472/0.0477

==========================================================================================
FINAL ANALYSIS
==========================================================================================

  PatchSVAE: 4 patches × (256, 32)
  Target CV: 0.125
  Recon MSE: 0.031701 +/- 0.024789
  Row CV: 0.1300
  Cross-attention S delta: 0.26925

  Learned alpha per mode (coordination strength):
    Layer 0: mean=0.0471  max=0.0477  min=0.0466
      α[ 0]: 0.0470  #######################################
      α[ 1]: 0.0473  #######################################
      α[ 2]: 0.0474  #######################################
      α[ 3]: 0.0473  #######################################
      α[ 4]: 0.0471  #######################################
      α[ 5]: 0.0474  #######################################
      α[ 6]: 0.0469  #######################################
      α[ 7]: 0.0472  #######################################
      α[ 8]: 0.0470  #######################################
      α[ 9]: 0.0475  #######################################
      α[10]: 0.0467  #######################################
      α[11]: 0.0471  #######################################
      α[12]: 0.0477  #######################################
      α[13]: 0.0466  #######################################
      α[14]: 0.0471  #######################################
      α[15]: 0.0472  #######################################
      α[16]: 0.0472  #######################################
      α[17]: 0.0471  #######################################
      α[18]: 0.0470  #######################################
      α[19]: 0.0475  #######################################
      α[20]: 0.0466  #######################################
      α[21]: 0.0477  #######################################
      α[22]: 0.0470  #######################################
      α[23]: 0.0469  #######################################
      α[24]: 0.0472  #######################################
      α[25]: 0.0472  #######################################
      α[26]: 0.0471  #######################################
      α[27]: 0.0471  #######################################
      α[28]: 0.0474  #######################################
      α[29]: 0.0472  #######################################
      α[30]: 0.0466  #######################################
      α[31]: 0.0475  #######################################
    Layer 1: mean=0.0472  max=0.0477  min=0.0466
      α[ 0]: 0.0474  #######################################
      α[ 1]: 0.0472  #######################################
      α[ 2]: 0.0470  #######################################
      α[ 3]: 0.0471  #######################################
      α[ 4]: 0.0474  #######################################
      α[ 5]: 0.0470  #######################################
      α[ 6]: 0.0474  #######################################
      α[ 7]: 0.0473  #######################################
      α[ 8]: 0.0473  #######################################
      α[ 9]: 0.0470  #######################################
      α[10]: 0.0477  #######################################
      α[11]: 0.0472  #######################################
      α[12]: 0.0466  #######################################
      α[13]: 0.0477  #######################################
      α[14]: 0.0473  #######################################
      α[15]: 0.0471  #######################################
      α[16]: 0.0472  #######################################
      α[17]: 0.0471  #######################################
      α[18]: 0.0476  #######################################
      α[19]: 0.0470  #######################################
      α[20]: 0.0475  #######################################
      α[21]: 0.0470  #######################################
      α[22]: 0.0472  #######################################
      α[23]: 0.0475  #######################################
      α[24]: 0.0472  #######################################
      α[25]: 0.0471  #######################################
      α[26]: 0.0475  #######################################
      α[27]: 0.0474  #######################################
      α[28]: 0.0472  #######################################
      α[29]: 0.0469  #######################################
      α[30]: 0.0476  #######################################
      α[31]: 0.0466  #######################################

  Coordinated singular value profile:
    S[ 0]:   4.0376  cum=  5.3%  #############################
    S[ 1]:   3.9321  cum= 10.3%  #############################
    S[ 2]:   3.8501  cum= 15.1%  ############################
    S[ 3]:   3.7785  cum= 19.8%  ############################
    S[ 4]:   3.7092  cum= 24.2%  ###########################
    S[ 5]:   3.6414  cum= 28.5%  ###########################
    S[ 6]:   3.5771  cum= 32.7%  ##########################
    S[ 7]:   3.5158  cum= 36.7%  ##########################
    S[ 8]:   3.4554  cum= 40.6%  #########################
    S[ 9]:   3.3961  cum= 44.3%  #########################
    S[10]:   3.3371  cum= 48.0%  ########################
    S[11]:   3.2788  cum= 51.5%  ########################
    S[12]:   3.2230  cum= 54.8%  #######################
    S[13]:   3.1681  cum= 58.1%  #######################
    S[14]:   3.1141  cum= 61.2%  #######################
    S[15]:   3.0607  cum= 64.3%  ######################
    S[16]:   3.0088  cum= 67.2%  ######################
    S[17]:   2.9568  cum= 70.1%  #####################
    S[18]:   2.9075  cum= 72.8%  #####################
    S[19]:   2.8572  cum= 75.5%  #####################
    S[20]:   2.8067  cum= 78.0%  ####################
    S[21]:   2.7584  cum= 80.5%  ####################
    S[22]:   2.7075  cum= 82.9%  ####################
    S[23]:   2.6574  cum= 85.2%  ###################
    S[24]:   2.6060  cum= 87.4%  ###################
    S[25]:   2.5535  cum= 89.5%  ##################
    S[26]:   2.4991  cum= 91.5%  ##################
    S[27]:   2.4413  cum= 93.5%  ##################
    S[28]:   2.3770  cum= 95.3%  #################
    S[29]:   2.2906  cum= 97.0%  #################
    S[30]:   2.2012  cum= 98.6%  ################
    S[31]:   2.0926  cum=100.0%  ###############

  Saving reconstruction grid...
  Saved to /content/svae_patch_recon.png
```


# Prototype V10 Patchwork Cross-Attention - Unstable

Tiny Imagenet can't draw enough information from a single monotonic MLP projection, so I'm breaking the structure into quadrant-based mlp patches
with cross-attention for a prototype.

Each patch is 32x32 and they have svd 24 independently represented each with patchwork cross-attention. Similar to a vit,
so I'm building it to a full vit structure over time to ensure solidity and solidarity.

![image](https://cdn-uploads.huggingface.co/production/uploads/630cf55b15433862cfc9556f/fRbpbpDJN2ASwI-pLJZSn.png)

Current proto is more stable but requires a bit more oomph.

The CV is enjoying it's drift a BIT too much

I'll try attention alpha rather than rigid alpha. 4 patches is a bit unstable, so lets get some stability.


# Prototype V9 prod

Should run on colab. Install the necessary repos.

https://huggingface.co/AbstractPhil/geolip-SVAE/blob/main/prototype_v9_prod.py


# Prototype V8 Soft Hand Loss

Stable prototype found. Scaling with the CV ratio within this band is a stable attractor to the structural response.

The soft hand loss is acting like a stable attractant. Correct utilization of this behavior can directly attenuate
a model's structural internals to align to certain trajectory-based routes.

The alignment can be directly tuned at runtime, shifted to learn implicit rules, altered to teach specific behaviors, and more.

0.034 mse, which is a different gauge of loss entirely.
![image](https://cdn-uploads.huggingface.co/production/uploads/630cf55b15433862cfc9556f/oQfFDMiuo8FAJMuuYRqVk.png)


# Prototype V7

Normalized spherical without magnitude, expected considerably faster with less accuracy at first stages.

![image](https://cdn-uploads.huggingface.co/production/uploads/630cf55b15433862cfc9556f/Hrhmlf1uvSUv51Mw9PThF.png)


# What happens if you train with the wrong CV value?

```
Using geolip-core SVD (Gram + eigh)
SVAE - V=96, D=24 (Validated: CV=0.3668)
  Matrix: (96, 24) = 2304 elements
  SVD: geolip-core Gram+eigh
  Losses: recon + CV(w=0.1, target=0.3668)
  Params: 6,036,736
=====================================================================================
 ep |    loss   recon    cv_l  t/ep |   t_rec |     S0     SD ratio erank |  row_cv
-------------------------------------------------------------------------------------
  1 |  0.4174  0.4169  0.0037   7.3 |  0.2843 |   5.39  1.977  2.73 23.15 |  0.3039
  2 |  0.2492  0.2489  0.0031   7.3 |  0.2286 |   5.43  1.978  2.75 23.14 |  0.3148
  3 |  0.2096  0.2093  0.0030   7.3 |  0.1946 |   5.50  1.982  2.77 23.13 |  0.3352
  4 |  0.1858  0.1855  0.0021   7.2 |  0.1812 |   5.48  1.980  2.77 23.13 |  0.3460
  6 |  0.1586  0.1581  0.0046   7.3 |  0.1541 |   5.31  1.873  2.83 23.09 |  0.3938
  8 |  0.1419  0.1407  0.0096   7.3 |  0.1377 |   5.33  1.815  2.93 23.03 |  0.4565
 10 |  0.1314  0.1283  0.0385   7.3 |  0.1279 |   5.42  1.778  3.05 22.97 |  0.5373
 12 |  0.1226  0.1160  0.0599   7.2 |  0.1162 |   5.67  1.738  3.26 22.86 |  0.6060
 14 |  0.1189  0.1087  0.0847   7.1 |  0.1109 |   5.78  1.705  3.39 22.79 |  0.6643
 16 |  0.1175  0.1014  0.1935   7.1 |  0.0996 |   6.17  1.701  3.63 22.67 |  0.7598
 18 |  0.1170  0.0952  0.2238   7.2 |  0.0974 |   6.50  1.671  3.89 22.52 |  0.8211
 20 |  0.1173  0.0905  0.1539   7.2 |  0.0907 |   6.69  1.649  4.06 22.43 |  0.8383
 22 |  0.1200  0.0852  0.3335   7.1 |  0.0903 |   7.11  1.655  4.30 22.30 |  0.9128
 24 |  0.1233  0.0817  0.2770   7.2 |  0.0831 |   7.51  1.646  4.56 22.15 |  0.9654
 26 |  0.1286  0.0785  0.3243   7.1 |  0.0778 |   7.71  1.646  4.68 22.09 |  1.0196
 28 |  0.1328  0.0752  0.4244   7.2 |  0.0780 |   7.84  1.636  4.80 22.02 |  1.1002
 30 |  0.1373  0.0726  0.8786   7.1 |  0.0752 |   8.24  1.631  5.05 21.87 |  1.1243
 32 |  0.1437  0.0703  0.6946   7.2 |  0.0704 |   8.52  1.631  5.23 21.76 |  1.2061
 34 |  0.6025  0.6020  0.0062   7.1 |  0.5194 |  28.25 10.261  2.75 23.14 |  0.2935
 36 |  0.4995  0.4990  0.0062   7.2 |  0.4949 |  29.82 10.939  2.73 23.15 |  0.2982
 38 |  0.4947  0.4942  0.0058   7.2 |  0.4915 |  28.37 10.433  2.72 23.15 |  0.2988
 40 |  0.4579  0.4574  0.0053   7.2 |  0.4557 |  26.31  9.585  2.74 23.14 |  0.3041
 42 |  0.4333  0.4328  0.0051   7.1 |  0.4259 |  22.03  7.996  2.75 23.14 |  0.2984
 44 |  0.4057  0.4054  0.0038   7.1 |  0.3880 |  21.15  7.656  2.76 23.15 |  0.3177
 46 |  0.3670  0.3667  0.0024   7.2 |  0.3634 |  19.33  6.943  2.78 23.13 |  0.3280
 48 |  0.3495  0.3493  0.0005   7.1 |  0.3457 |  18.34  6.569  2.79 23.13 |  0.3336
 50 |  0.3341  0.3340  0.0010   7.3 |  0.3326 |  17.55  6.298  2.79 23.13 |  0.3424
 52 |  0.3205  0.3204  0.0003   7.2 |  0.3182 |  16.94  6.069  2.79 23.12 |  0.3549
```

SNAP. right there at epoch 34. The tension was too strong, the model simply snapped.
I had it set to around 0.366, and it requires that value there where it snapped to. 0.2935

The actual value as of the bulk embedding tests show;  CV=0.2992 is the stable attractor, almost precisely where the model snapped to.

The effect was so strong, that the entire model had a forced reset when it realized the fundamental invalidity.

Why? I don't know yet.

# Models

V4 5m SVD+EIGH 100 epochs 48x24

![image](https://cdn-uploads.huggingface.co/production/uploads/630cf55b15433862cfc9556f/4V_kfCqNIexGlOEjbie8A.png)

V4 111m 200x24 SVD+EIGH KL_DIV - Undercooked, needs more epochs -> sequel faulty, collapse

![image](https://cdn-uploads.huggingface.co/production/uploads/630cf55b15433862cfc9556f/VQkIDgdcrxTdL2xbFVpnA.png)


V3 v1024 - SVD 24


![image](https://cdn-uploads.huggingface.co/production/uploads/630cf55b15433862cfc9556f/sGKEH4LWTimmR6xHcaRQU.png)

V2 16 modes

![image](https://cdn-uploads.huggingface.co/production/uploads/630cf55b15433862cfc9556f/tt_mwpCqlQxuMfQUPfkun.png)



V1 8 modes
![image](https://cdn-uploads.huggingface.co/production/uploads/630cf55b15433862cfc9556f/sUbWtH4lKDkHS8plnkkxa.png)