[ENHANCEMENT] Dynamic binning + distance-aware label smoothing #251
Killer3048
started this conversation in
Ideas
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Dynamic binning + distance-aware label smoothing
Why this matters
[-15..+15]
for quantization after mean scaling. This can cause severe overflow (if the real data extends beyond +15) or underflow (if data is confined way below ±15) in zero-shot scenarios.chronos.py
, the classMeanScaleUniformBins
specifically relies onlow_limit
andhigh_limit
to buildself.centers
andself.boundaries
. While that works fine for data roughly within[-15..+15]
, new domains can push well outside that range.torch.bucketize
without distinguishing “close” vs. “far” bins.Proposed enhancement
Dynamic range determination
MeanScaleUniformBins
to compute (at inference or per training batch) a local min/max (or percentiles, like p10/p90) for the scaled data.self.centers
andself.boundaries
so that we cover[actual_min..actual_max]
, clamping if needed (e.g.,[-50..50]
).Distance-aware label smoothing
b
is the correct bin, allocate ~80% probability tob
, ~10% tob-1
, and ~10% tob+1
(splitting among neighbors).ChronosModel.forward
, if we have direct control of the loss.Estimated impact on accuracy & inference
Zero-shot improvements:
Few-shot improvements:
Inference overhead:
Why it improves zero-/few-shot
Beta Was this translation helpful? Give feedback.
All reactions