Comment by mnky9800n

Comment by mnky9800n 4 days ago

6 replies

Why not, as each new task comes up, and then weights are revalued, save those weights and keep them for reference as priors for similar future tasks? As the model is exposed to new data the average of the set of priors of things the model thinks is similar might move closer to the posterior making the model quicker and more able to arrive at good outcomes. I suppose storage might be an issue.

magospietato 3 days ago

I'm wondering if you could fine tune the model on an aggregate of a temporal slice of revalued weights? Something analogous to REM sleep's involvement in embedding the days events into long term memory.

  • Jerrrry 3 days ago

    Sieve the temporary backprop interim weights as a function of its loss of varentrophy relative to its place in the revalued weights.

    Remove the bottom weights dynamically based on the local gradient in varentrophy so that internal dissonance ("doubt") can be selected against.

    "Preference Optimization" but with more opportunities for meta-optimization.

QuadmasterXLII 3 days ago

thats just mixture of experts

  • mnky9800n 3 days ago

    i thought mixture of experts didn't update itself with new sets of weights and was just a collection of already trained networks/weights? I could be wrong.

    • QuadmasterXLII 3 days ago

      Well, that depends in whether you keep training it

      • mnky9800n 3 days ago

        perhaps they should always be training and never static. haha. i allegedly grow wiser in my age, why not neural networks?