@themaximalist/llm.js
Preparing search index...
LLM
Class LLM
LLM Base Class
Hierarchy (
View Summary
)
LLM
Anthropic
Ollama
OpenAI
Google
APIv1
Index
Constructors
constructor
Properties
base
Url
extended
json?
max_
thinking_
tokens?
max_
tokens
messages
model
model
Usage
options
parser?
quality
Filter
service
stream
temperature?
think
tools?
DEFAULT_
BASE_
URL
DEFAULT_
MODEL
is
Bearer
Auth
is
Local
parsers
service
Accessors
api
Key
chat
Url
is
Local
llm
Headers
llm
Options
models
Url
parsers
Methods
abort
add
Message
assistant
chat
fetch
Models
filter
Quality
Model
get
Chat
Url
get
Default
Model
Usage
get
Models
get
Models
Url
get
Quality
Models
parse
Content
parse
Content
Chunk
parse
Messages
parse
Model
parse
Options
parse
Thinking
parse
Thinking
Chunk
parse
Token
Usage
parse
Tools
parse
Tools
Chunk
parse
Usage
refresh
Model
Usage
response
send
system
thinking
tool
Call
user
verify
Connection
Constructors
constructor
new
LLM
(
input
?:
Options
|
Input
,
options
?:
Options
)
:
LLM
Parameters
Optional
input
:
Options
|
Input
options
:
Options
= {}
Returns
LLM
Properties
base
Url
baseUrl
:
string
extended
extended
:
boolean
Optional
json
json
?:
boolean
Optional
max_
thinking_
tokens
max_thinking_tokens
?:
number
max_
tokens
max_tokens
:
number
messages
messages
:
Message
[]
model
model
:
string
model
Usage
modelUsage
:
ModelUsage
options
options
:
Options
Optional
parser
parser
?:
ParserResponse
quality
Filter
qualityFilter
:
QualityFilter
service
service
:
string
stream
stream
:
boolean
Optional
temperature
temperature
?:
number
think
think
:
boolean
Optional
tools
tools
?:
Tool
[]
Static
DEFAULT_
BASE_
URL
DEFAULT_BASE_URL
:
string
Static
DEFAULT_
MODEL
DEFAULT_MODEL
:
string
Static
is
Bearer
Auth
isBearerAuth
:
boolean
= false
Static
is
Local
isLocal
:
boolean
= false
Static
parsers
parsers
:
parsers
= parsers
Static
Readonly
service
service
:
string
Accessors
api
Key
get
apiKey
()
:
undefined
|
null
|
string
Returns
undefined
|
null
|
string
chat
Url
get
chatUrl
()
:
string
Returns
string
is
Local
get
isLocal
()
:
boolean
Returns
boolean
llm
Headers
get
llmHeaders
()
:
Record
<
string
,
string
>
Returns
Record
<
string
,
string
>
llm
Options
get
llmOptions
()
:
Options
Returns
Options
models
Url
get
modelsUrl
()
:
string
Returns
string
parsers
get
parsers
()
:
Parsers
Returns
Parsers
Methods
abort
abort
()
:
void
Returns
void
add
Message
addMessage
(
role
:
MessageRole
,
content
:
any
)
:
void
Parameters
role
:
MessageRole
content
:
any
Returns
void
assistant
assistant
(
content
:
string
)
:
void
Parameters
content
:
string
Returns
void
chat
chat
(
input
:
string
,
options
?:
Options
,
)
:
Promise
<
|
string
|
Response
|
PartialStreamResponse
|
AsyncGenerator
<
string
,
any
,
any
>
,
>
Parameters
input
:
string
Optional
options
:
Options
Returns
Promise
<
|
string
|
Response
|
PartialStreamResponse
|
AsyncGenerator
<
string
,
any
,
any
>
,
>
fetch
Models
fetchModels
()
:
Promise
<
Model
[]
>
Returns
Promise
<
Model
[]
>
filter
Quality
Model
filterQualityModel
(
model
:
Model
)
:
boolean
Parameters
model
:
Model
Returns
boolean
get
Chat
Url
getChatUrl
(
opts
:
Options
)
:
string
Parameters
opts
:
Options
Returns
string
get
Default
Model
Usage
getDefaultModelUsage
(
model
:
Model
)
:
ModelUsageType
Parameters
model
:
Model
Returns
ModelUsageType
get
Models
getModels
(
quality_filter
?:
QualityFilter
)
:
Promise
<
Model
[]
>
Parameters
quality_filter
:
QualityFilter
= {}
Returns
Promise
<
Model
[]
>
get
Models
Url
getModelsUrl
()
:
string
Returns
string
get
Quality
Models
getQualityModels
()
:
Promise
<
Model
[]
>
Returns
Promise
<
Model
[]
>
parse
Content
parseContent
(
data
:
any
)
:
string
Parameters
data
:
any
Returns
string
parse
Content
Chunk
parseContentChunk
(
chunk
:
any
)
:
string
Parameters
chunk
:
any
Returns
string
parse
Messages
parseMessages
(
messages
:
Message
[]
)
:
Message
[]
Parameters
messages
:
Message
[]
Returns
Message
[]
parse
Model
parseModel
(
model
:
any
)
:
Model
Parameters
model
:
any
Returns
Model
parse
Options
parseOptions
(
options
:
Options
)
:
Options
Parameters
options
:
Options
Returns
Options
parse
Thinking
parseThinking
(
data
:
any
)
:
string
Parameters
data
:
any
Returns
string
parse
Thinking
Chunk
parseThinkingChunk
(
chunk
:
any
)
:
string
Parameters
chunk
:
any
Returns
string
parse
Token
Usage
parseTokenUsage
(
usage
:
any
)
:
null
|
InputOutputTokens
Parameters
usage
:
any
Returns
null
|
InputOutputTokens
parse
Tools
parseTools
(
data
:
any
)
:
ToolCall
[]
Parameters
data
:
any
Returns
ToolCall
[]
parse
Tools
Chunk
parseToolsChunk
(
chunk
:
any
)
:
ToolCall
[]
Parameters
chunk
:
any
Returns
ToolCall
[]
parse
Usage
parseUsage
(
tokenUsage
:
InputOutputTokens
)
:
Usage
Parameters
tokenUsage
:
InputOutputTokens
Returns
Usage
refresh
Model
Usage
refreshModelUsage
()
:
Promise
<
void
>
Returns
Promise
<
void
>
response
response
(
data
:
any
)
:
string
Parameters
data
:
any
Returns
string
send
send
(
options
?:
Options
,
)
:
Promise
<
|
string
|
Response
|
PartialStreamResponse
|
AsyncGenerator
<
string
,
any
,
any
>
,
>
Parameters
Optional
options
:
Options
Returns
Promise
<
|
string
|
Response
|
PartialStreamResponse
|
AsyncGenerator
<
string
,
any
,
any
>
,
>
system
system
(
content
:
string
)
:
void
Parameters
content
:
string
Returns
void
thinking
thinking
(
content
:
string
)
:
void
Parameters
content
:
string
Returns
void
tool
Call
toolCall
(
tool
:
ToolCall
)
:
void
Parameters
tool
:
ToolCall
Returns
void
user
user
(
content
:
string
)
:
void
Parameters
content
:
string
Returns
void
verify
Connection
verifyConnection
()
:
Promise
<
boolean
>
Returns
Promise
<
boolean
>
Settings
Member Visibility
Inherited
External
Theme
OS
Light
Dark
On This Page
Constructors
constructor
Properties
base
Url
extended
json
max_
thinking_
tokens
max_
tokens
messages
model
model
Usage
options
parser
quality
Filter
service
stream
temperature
think
tools
DEFAULT_
BASE_
URL
DEFAULT_
MODEL
is
Bearer
Auth
is
Local
parsers
service
Accessors
api
Key
chat
Url
is
Local
llm
Headers
llm
Options
models
Url
parsers
Methods
abort
add
Message
assistant
chat
fetch
Models
filter
Quality
Model
get
Chat
Url
get
Default
Model
Usage
get
Models
get
Models
Url
get
Quality
Models
parse
Content
parse
Content
Chunk
parse
Messages
parse
Model
parse
Options
parse
Thinking
parse
Thinking
Chunk
parse
Token
Usage
parse
Tools
parse
Tools
Chunk
parse
Usage
refresh
Model
Usage
response
send
system
thinking
tool
Call
user
verify
Connection
@themaximalist/llm.js
Loading...
LLM Base Class