Finish off completions direct modules
Some checks failed
ci/woodpecker/push/author-tests Pipeline failed

This commit is contained in:
Ryan Voots 2023-11-23 14:51:54 -05:00
parent 1304eb119f
commit 950e83017a
2 changed files with 59 additions and 0 deletions

View file

@ -30,6 +30,12 @@ A result from a completion request, L<OpenAIAsync::Types::Request::Completion>
id of the completion response, used for tracking duplicate responses or reporting issues to the service id of the completion response, used for tracking duplicate responses or reporting issues to the service
=head1 choices
An array of L<OpenAIAsync::Types::Results::CompletionChoices> objects. If you asked for more than 1 response with the request parameter C<n> then they will be present here.
You likely just want to get ->text from the first result, as demonstrated in the synopsis but see the ::CompletionChoices docs for more detailed information.
=head2 model =head2 model
The model that was used to generate the response. Usually will be what you requested, The model that was used to generate the response. Usually will be what you requested,

View file

@ -0,0 +1,53 @@
=pod
=head1 NAME
OpenAIAsync::Types::Results::CompletionChoices
=head1 DESCRIPTION
A choice from a completion request, L<OpenAIAsync::Types::Request::Completion> as part of L<OpenAIAsync::Types::Result::Completion>
=head1 SYNOPSIS
use OpenAIAsync::Client;
use IO::Async::Loop;
my $loop = IO::Async::Loop->new();
my $client = OpenAIAsync::Client->new();
$loop->add($client)
my $output_future = $client->completion({max_tokens => 1024, prompt => "Tell a story about a princess named Judy and her princess sister Emmy"});
my $result = $output_future->get();
print $result->choices->[0]->text;
=head1 Fields
=head2 text
The contents of the response, very likely all you want or need
=head2 index
Index of the choice? I believe this will just always be the same as it's position in the array.
=head2 logprobs
Logit probabilities, see L<OpenAIAsync::Types::Results::LogProbs> for details
=head2 finish_reason
What made the model stop generating. Could be from hitting a stop token, or running into max tokens.
=head1 SEE ALSO
L<OpenAIAsync::Types::Request::Completion>, L<OpenAIAsync::Types::Result::Completion>, L<OpenAIAsync::Client>
=head1 AUTHOR
Ryan Voots ...
=cut