Skip to content

A novel methodology for improving the needle in a haystack capabilities of chat SLMs.

Notifications You must be signed in to change notification settings

georgepullen/multi-contextual-token-sampling

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

6 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

This methodology has been superseded by Batched Multi-Contextual Token Sampling.


Multi-Contextual Token Sampling

A novel methodology for improving the needle in a haystack capabilities of chat SLMs.

mcts

The Method


  • Each context window of previous messages responds as if it is the sole context
  • The highest probability next token across all context windows is sampled
  • Each context window is concatenated with the same next token
  • Only tokens that have appeared in previous chat logs can be used in the response
  • This gives the agent the ability to adapt to the users vocabulary and style of conversation overtime

Justification


Small Language Models struggle to generate accurate responses in long-context settings, such as chat modelling. With this novel methodology, the logits for the next token are computed for each chunk of context separately. Furthermore, the masking of unseen tokens further increases accuracy, particularly with respect to information such as dates and times.

About

A novel methodology for improving the needle in a haystack capabilities of chat SLMs.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages