Replies: 2 comments
-
That's an excellent question, and very much a research question at this point :) |
Beta Was this translation helpful? Give feedback.
0 replies
-
Implementing #479 may make it easier to answer this question. Note that you can already get the logits at each step by building the sequence generator with |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Hi everyone,
I am interested in the calibration of LLMs to ensure their trustworthiness. We saw with GPT4 paper that RLHF for instance destroyed calibrated answers.
I am interested in measuring calibration of models that were trained on general text generation and not classification specifically, e.g. GPT4.
For instance, imagine I want the model to do classification by using outlines to only output "Cat" or "Dog".
Can I use outline to provide an accurate measure of output probability without having to fine-tune the model for classification? Aka, does zero-ing all tokens that don't produce the class I am interested in, can be a good way to provide output probability similar to what I would get if I were to fine-tune a classification head?
Thanks for the help :)
Beta Was this translation helpful? Give feedback.
All reactions