use v5.38.0; use Object::Pad; use IO::Async::SSL; # We're not directly using it but I want to enforce that we pull it in when detecting dependencies, since openai itself is always https use Future::AsyncAwait; class OpenAIAsync::Client :repr(HASH) :isa(IO::Async::Notifier) { use JSON::MaybeXS qw/encode_json decode_json/; use Net::Async::HTTP; use Feature::Compat::Try; use URI; field $_http_max_in_flight :param(http_max_in_flight) = 2; field $_http_max_redirects :param(http_max_redirects) = 3; field $_http_max_connections_per_host :param(http_max_connections_per_host) = 2; field $_http_timeout :param(http_timeout) = 120; # My personal server is kinda slow, use a generous default field $_http_stall_timeout :param(http_stall_timeout) = 600; # generous for my slow personal server field $_http_proxy_host :param(http_proxy_host) = undef; field $_http_proxy_port :param(http_proxy_port) = undef; field $_http_proxy_path :param(http_proxy_path) = undef; field $http; field $api_base :param(api_base) = $ENV{OPENAI_API_BASE} // "https://api.openai.com/v1"; field $api_key :param(api_key) = $ENV{OPENAI_API_KEY}; method configure(%params) { # We require them to go this way my %io_async_params = ($params{io_async_notifier_params} // {})->%*; IO::Async::Notifier::configure($self, ); } method __make_http() { die "Missing API Key for OpenAI" unless $api_key; return Net::Async::HTTP->new( user_agent => "SNN OpenAI Client 1.0", +headers => { "Authorization" => "Bearer $api_key", "Content-Type" => "application/json", }, max_redirects => $_http_max_redirects, max_connections_per_host => $_http_max_connections_per_host, max_in_flight => $_http_max_in_flight, timeout => $_http_timeout, stall_timeout => $_http_stall_timeout, # TODO proxy stuff ) } ADJUST { $http = $self->__make_http; $api_base =~ s|/$||; # trim an accidental final / since we will be putting it on the endpoints } async method _make_request($endpoint, $data) { my $json = encode_json($data); my $url = URI->new($api_base . $endpoint ); my $result = await $http->do_request( uri => $url, method => "POST", content => $json, content_type => 'application/json', ); if ($result->is_success) { my $json = $result->decoded_content; my $out_data = decode_json($json); return $out_data; } else { die "Failure in talking to OpenAI service: ".$result->status_line.": ".$result->decoded_content; } } method _add_to_loop($loop) { $loop->add($http); } method _remove_from_loop($loop) { $loop->remove($http); $http = $self->__make_http; # overkill? want to make sure we have a clean one } # This is the legacy completion api async method completion($prompt, $model = "gpt-3.5-turbo", $max_tokens = 500, $temperature = 1, $top_p = 0.9, $seed = -1) { my $request_body = { model => $model, # ignored by my local service prompt => $prompt, max_tokens => $max_tokens, temperature => $temperature, top_p => $top_p, seed => $seed }; my $res = await $self->_make_request("/completions", $request_body); use Data::Dumper; print Dumper($res); die "FUCK"; } async method chat($prompt) { ... } async method embedding($input) { } async method image_generate($input) { ... } async method text_to_speech($text) { ... } async method speech_to_text($sound_data) { ... } async method vision($image, $prompt) { ... } }