mirror of
https://github.com/nomic-ai/gpt4all
synced 2024-11-18 03:25:46 +00:00
a153cc5b25
Signed-off-by: Tare Ebelo <75279482+TareHimself@users.noreply.github.com> Signed-off-by: jacob <jacoobes@sern.dev> Signed-off-by: Jared Van Bortel <jared@nomic.ai> Co-authored-by: jacob <jacoobes@sern.dev> Co-authored-by: Jared Van Bortel <jared@nomic.ai>
60 lines
1.4 KiB
C++
60 lines
1.4 KiB
C++
#ifndef PREDICT_WORKER_H
|
|
#define PREDICT_WORKER_H
|
|
|
|
#include "napi.h"
|
|
#include "llmodel_c.h"
|
|
#include "llmodel.h"
|
|
#include <thread>
|
|
#include <mutex>
|
|
#include <iostream>
|
|
#include <atomic>
|
|
#include <memory>
|
|
|
|
struct TokenCallbackInfo
|
|
{
|
|
int32_t tokenId;
|
|
std::string total;
|
|
std::string token;
|
|
};
|
|
|
|
struct LLModelWrapper
|
|
{
|
|
LLModel *llModel = nullptr;
|
|
LLModel::PromptContext promptContext;
|
|
~LLModelWrapper() { delete llModel; }
|
|
};
|
|
|
|
struct PromptWorkerConfig
|
|
{
|
|
Napi::Function tokenCallback;
|
|
bool bHasTokenCallback = false;
|
|
llmodel_model model;
|
|
std::mutex * mutex;
|
|
std::string prompt;
|
|
llmodel_prompt_context context;
|
|
std::string result;
|
|
};
|
|
|
|
class PromptWorker : public Napi::AsyncWorker
|
|
{
|
|
public:
|
|
PromptWorker(Napi::Env env, PromptWorkerConfig config);
|
|
~PromptWorker();
|
|
void Execute() override;
|
|
void OnOK() override;
|
|
void OnError(const Napi::Error &e) override;
|
|
Napi::Promise GetPromise();
|
|
|
|
bool ResponseCallback(int32_t token_id, const std::string token);
|
|
bool RecalculateCallback(bool isrecalculating);
|
|
bool PromptCallback(int32_t tid);
|
|
|
|
private:
|
|
Napi::Promise::Deferred promise;
|
|
std::string result;
|
|
PromptWorkerConfig _config;
|
|
Napi::ThreadSafeFunction _tsfn;
|
|
};
|
|
|
|
#endif // PREDICT_WORKER_H
|