Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Better alignment than "per-token" for SMC #1

Open
alex-lew opened this issue May 20, 2023 · 0 comments
Open

Better alignment than "per-token" for SMC #1

alex-lew opened this issue May 20, 2023 · 0 comments

Comments

@alex-lew
Copy link
Contributor

alex-lew commented May 20, 2023

When to resample in SMC? Currently particles are aligned by number-of-tokens, so when we resample, all particles have the same number of tokens (unless some have already hit EOS). But this isn't really fair. For example:

  • When intersecting "My favorite physicist is" and "My favorite writer is", we end up comparing particles that say, e.g., " Richard Feynman. He was" and " Neil deGrasse Tyson" -- when we really want to compare " Richard Feynman" to " Neil deGrasse Tyson".
  • When intersecting "A great personal finance tip is" and "A great tip for healthy living is", we end up comparing particles that say, e.g., " to avoid eating out" and " to make sure you're". The former loses out, intuitively because its weight already factors in the semantic constraints whereas they largely 'withhold judgment' on the vaguer latter particle.

It would be great to find a clear theoretical framework for thinking about these intermediate distributions, and other heuristics (or principled strategies) for alignment.

One heuristic worth trying might be to resample at syntax-directed points -- at the end of each sentence, clause, or some other grammatical element.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant