Compare commits

...

10 Commits

Author SHA1 Message Date
e696343a73 a crumb of changes 2025-08-12 19:07:22 +00:00
88a0710c55 update system prompt to hopefully make it easier for Lexi to understand 2025-08-04 21:08:47 +00:00
75fa4cea8b jorkin my preanits 2025-08-04 11:56:06 +00:00
733a41a35c revert to more "chat" based api calls 2025-08-04 00:21:10 +00:00
ed3467b213 bump version 2025-08-03 23:32:48 +00:00
0f178fcfa9 beta release conversation context 2025-08-03 23:31:56 +00:00
0bfff52fd0 Merge branch 'main' into implement-conversation-context 2025-08-03 21:25:04 +00:00
8e90e8b71e add context response type 2025-08-03 21:24:40 +00:00
566d6ae518 update types 2025-08-03 20:59:53 +00:00
2ec367f203 fix typo 2025-08-03 20:05:21 +00:00
9 changed files with 167 additions and 43 deletions

View File

@ -2,6 +2,7 @@ DATABASE_URL="file:../dev.db" # SQLite database relative to the ./prisma path
PLEROMA_INSTANCE_URL="https://instance.tld" # Pleroma instance full URL including scheme PLEROMA_INSTANCE_URL="https://instance.tld" # Pleroma instance full URL including scheme
PLEROMA_INSTANCE_DOMAIN="instance.tld" # used if you want to only want to respond to people from a particular instance PLEROMA_INSTANCE_DOMAIN="instance.tld" # used if you want to only want to respond to people from a particular instance
PLEROMA_ACCOUNT_ID="" # obtained from /api/v1/accounts/{nickname} - used so we don't spam mentions when not directly addressed PLEROMA_ACCOUNT_ID="" # obtained from /api/v1/accounts/{nickname} - used so we don't spam mentions when not directly addressed
REPLY_WITH_CONTEXT="" # set to true or false whether you want the bot to fetch context or not
ONLY_WHITELIST="true" # change to "false" if you want to accept prompts from any and all domains - *** USE WITH CAUTION *** ONLY_WHITELIST="true" # change to "false" if you want to accept prompts from any and all domains - *** USE WITH CAUTION ***
WHITELISTED_DOMAINS="" # comma separated list of domains you want to allow the bot to accept prompts from (i.e. poa.st,nicecrew.digital,detroitriotcity.com,decayable.ink) WHITELISTED_DOMAINS="" # comma separated list of domains you want to allow the bot to accept prompts from (i.e. poa.st,nicecrew.digital,detroitriotcity.com,decayable.ink)
OLLAMA_URL="http://localhost:11434" # OLLAMA connection URL OLLAMA_URL="http://localhost:11434" # OLLAMA connection URL

2
.gitignore vendored
View File

@ -1,6 +1,6 @@
node_modules node_modules
# Keep environment variables out of version control # Keep environment variables out of version control
.env .env*
*.log *.log
*.db *.db
/dist /dist

View File

@ -1,6 +1,6 @@
{ {
"name": "pleroma-ollama-bot", "name": "pleroma-ollama-bot",
"version": "1.0.7", "version": "1.1.0",
"main": "index.js", "main": "index.js",
"scripts": { "scripts": {
"start": "tsc && node -r dotenv/config dist/main.js", "start": "tsc && node -r dotenv/config dist/main.js",

View File

@ -1,5 +1,5 @@
import { envConfig, prisma } from "./main.js"; import { envConfig, prisma } from "./main.js";
import { PleromaEmoji, Notification } from "../types.js"; import { PleromaEmoji, Notification, ContextResponse } from "../types.js";
const getNotifications = async () => { const getNotifications = async () => {
const { bearerToken, pleromaInstanceUrl } = envConfig; const { bearerToken, pleromaInstanceUrl } = envConfig;
@ -22,6 +22,32 @@ const getNotifications = async () => {
} }
}; };
const getStatusContext = async (statusId: string) => {
const { bearerToken, pleromaInstanceUrl } = envConfig;
try {
const response = await fetch(
`${pleromaInstanceUrl}/api/v1/statuses/${statusId}/context`,
{
method: "GET",
headers: {
Authorization: `Bearer ${bearerToken}`,
},
}
);
if (!response.ok) {
throw new Error(
`Could not get conversation context: ${response.status} - ${response.statusText}`
);
}
const data: ContextResponse = await response.json();
return data;
} catch (error: unknown) {
if (error instanceof Error) {
throw new Error(error.message);
}
}
};
const getInstanceEmojis = async () => { const getInstanceEmojis = async () => {
const { bearerToken, pleromaInstanceUrl } = envConfig; const { bearerToken, pleromaInstanceUrl } = envConfig;
try { try {
@ -72,4 +98,9 @@ const deleteNotification = async (notification: Notification) => {
} }
}; };
export { deleteNotification, getInstanceEmojis, getNotifications }; export {
deleteNotification,
getInstanceEmojis,
getNotifications,
getStatusContext,
};

View File

@ -2,25 +2,25 @@ import {
NewStatusBody, NewStatusBody,
Notification, Notification,
OllamaConfigOptions, OllamaConfigOptions,
// OllamaChatRequest, OllamaChatRequest,
// OllamaChatResponse, OllamaChatResponse,
OllamaRequest, PostAncestorsForModel,
OllamaResponse,
} from "../types.js"; } from "../types.js";
// import striptags from "striptags"; // import striptags from "striptags";
import { PrismaClient } from "../generated/prisma/client.js"; import { PrismaClient } from "../generated/prisma/client.js";
import { import {
getInstanceEmojis, // getInstanceEmojis,
deleteNotification, deleteNotification,
getNotifications, getNotifications,
getStatusContext,
} from "./api.js"; } from "./api.js";
import { storeUserData, storePromptData } from "./prisma.js"; import { storeUserData, storePromptData } from "./prisma.js";
import { import {
isFromWhitelistedDomain, isFromWhitelistedDomain,
alreadyRespondedTo, alreadyRespondedTo,
recordPendingResponse, recordPendingResponse,
trimInputData, // trimInputData,
selectRandomEmoji, // selectRandomEmoji,
shouldContinue, shouldContinue,
} from "./util.js"; } from "./util.js";
@ -44,13 +44,14 @@ export const envConfig = {
? parseInt(process.env.RANDOM_POST_INTERVAL) ? parseInt(process.env.RANDOM_POST_INTERVAL)
: 3600000, : 3600000,
botAccountId: process.env.PLEROMA_ACCOUNT_ID, botAccountId: process.env.PLEROMA_ACCOUNT_ID,
replyWithContext: process.env.REPLY_WITH_CONTEXT === "true" ? true : false,
}; };
const ollamaConfig: OllamaConfigOptions = { const ollamaConfig: OllamaConfigOptions = {
temperature: 0.6, temperature: 0.9,
top_p: 0.85, top_p: 0.85,
top_k: 40, top_k: 60,
num_ctx: 2048, num_ctx: 16384, // maximum context window for Llama 3.1
repeat_penalty: 1.1, repeat_penalty: 1.1,
}; };
@ -59,9 +60,14 @@ const ollamaConfig: OllamaConfigOptions = {
const generateOllamaRequest = async ( const generateOllamaRequest = async (
notification: Notification notification: Notification
): Promise<OllamaResponse | undefined> => { ): Promise<OllamaChatResponse | undefined> => {
const { whitelistOnly, ollamaModel, ollamaSystemPrompt, ollamaUrl } = const {
envConfig; whitelistOnly,
ollamaModel,
ollamaSystemPrompt,
ollamaUrl,
replyWithContext,
} = envConfig;
try { try {
if (shouldContinue(notification)) { if (shouldContinue(notification)) {
if (whitelistOnly && !isFromWhitelistedDomain(notification)) { if (whitelistOnly && !isFromWhitelistedDomain(notification)) {
@ -73,20 +79,59 @@ const generateOllamaRequest = async (
} }
await recordPendingResponse(notification); await recordPendingResponse(notification);
await storeUserData(notification); await storeUserData(notification);
const ollamaRequestBody: OllamaRequest = { let conversationHistory: PostAncestorsForModel[] = [];
if (replyWithContext) {
const contextPosts = await getStatusContext(notification.status.id);
if (!contextPosts?.ancestors || !contextPosts) {
throw new Error(`Unable to obtain post context ancestors.`);
}
conversationHistory = contextPosts.ancestors.map((ancestor) => {
const mentions = ancestor.mentions.map((mention) => mention.acct);
return {
account_fqn: ancestor.account.fqn,
mentions,
plaintext_content: ancestor.pleroma.content["text/plain"],
};
});
// console.log(conversationHistory);
}
// Simplified user message (remove [/INST] as it's not needed for Llama 3)
const userMessage = `${notification.status.account.fqn} says to you: \"${notification.status.pleroma.content["text/plain"]}\".`;
let systemContent = ollamaSystemPrompt;
if (replyWithContext) {
// Simplified context instructions (avoid heavy JSON; summarize for clarity)
systemContent = `${ollamaSystemPrompt}\n\nPrevious conversation context:\n${conversationHistory
.map(
(post) =>
`${post.account_fqn} (said to ${post.mentions.join(", ")}): ${
post.plaintext_content
}`
)
.join(
"\n"
)}\nReply to the user who addressed you (you are Lexi, also known as nice-ai or nice-ai@nicecrew.digital). Examine the context of the entire conversation and make references to topics or information where appropriate. Prefix usernames with '@' when addressing them. Assume if there is no domain in the username, the domain is @nicecrew.digital (for example @matty would be @matty@nicecrew.digital)`;
}
// Switch to chat request format (messages array auto-handles Llama 3 template)
const ollamaRequestBody: OllamaChatRequest = {
model: ollamaModel, model: ollamaModel,
prompt: `${notification.status.account.fqn} says: ${trimInputData( messages: [
notification.status.content { role: "system", content: systemContent as string },
)}`, { role: "user", content: userMessage },
system: ollamaSystemPrompt, ],
stream: false, stream: false,
options: ollamaConfig, options: ollamaConfig,
}; };
const response = await fetch(`${ollamaUrl}/api/generate`, {
// Change endpoint to /api/chat
const response = await fetch(`${ollamaUrl}/api/chat`, {
method: "POST", method: "POST",
body: JSON.stringify(ollamaRequestBody), body: JSON.stringify(ollamaRequestBody),
}); });
const ollamaResponse: OllamaResponse = await response.json(); const ollamaResponse: OllamaChatResponse = await response.json();
await storePromptData(notification, ollamaResponse); await storePromptData(notification, ollamaResponse);
return ollamaResponse; return ollamaResponse;
} }
@ -97,19 +142,19 @@ const generateOllamaRequest = async (
const postReplyToStatus = async ( const postReplyToStatus = async (
notification: Notification, notification: Notification,
ollamaResponseBody: OllamaResponse ollamaResponseBody: OllamaChatResponse
) => { ) => {
const { pleromaInstanceUrl, bearerToken } = envConfig; const { pleromaInstanceUrl, bearerToken } = envConfig;
const emojiList = await getInstanceEmojis(); // const emojiList = await getInstanceEmojis();
let randomEmoji; // let randomEmoji;
if (emojiList) { // if (emojiList) {
randomEmoji = selectRandomEmoji(emojiList); // randomEmoji = selectRandomEmoji(emojiList);
} // }
try { try {
let mentions: string[]; let mentions: string[];
const statusBody: NewStatusBody = { const statusBody: NewStatusBody = {
content_type: "text/markdown", content_type: "text/markdown",
status: `${ollamaResponseBody.response} :${randomEmoji}:`, status: `${ollamaResponseBody.message.content}`,
in_reply_to_id: notification.status.id, in_reply_to_id: notification.status.id,
}; };
if ( if (
@ -149,26 +194,28 @@ const createTimelinePost = async () => {
ollamaUrl, ollamaUrl,
pleromaInstanceUrl, pleromaInstanceUrl,
} = envConfig; } = envConfig;
const ollamaRequestBody: OllamaRequest = { const ollamaRequestBody: OllamaChatRequest = {
model: ollamaModel, model: ollamaModel,
prompt: "Say something random.", messages: [
system: ollamaSystemPrompt, { role: "system", content: ollamaSystemPrompt as string },
{ role: "user", content: "Say something random." },
],
stream: false, stream: false,
// options: ollamaConfig, options: ollamaConfig,
}; };
try { try {
const response = await fetch(`${ollamaUrl}/api/generate`, { const response = await fetch(`${ollamaUrl}/api/chat`, {
method: "POST", method: "POST",
body: JSON.stringify(ollamaRequestBody), body: JSON.stringify(ollamaRequestBody),
}); });
if (!response.ok) if (!response.ok)
throw new Error("Error generating ad-hoc Ollama response"); throw new Error("Error generating ad-hoc Ollama response");
const ollamaResponse: OllamaResponse = await response.json(); const ollamaResponse: OllamaChatResponse = await response.json();
const newStatusBody: NewStatusBody = { const newStatusBody: NewStatusBody = {
content_type: "text/markdown", content_type: "text/markdown",
status: ollamaResponse.response, status: ollamaResponse.message.content,
}; };
const pleromaResponse = await fetch( const pleromaResponse = await fetch(

View File

@ -1,16 +1,16 @@
import { Notification, OllamaResponse } from "../types.js"; import { Notification, OllamaChatResponse } from "../types.js";
import { trimInputData } from "./util.js"; import { trimInputData } from "./util.js";
import { prisma } from "./main.js"; import { prisma } from "./main.js";
const storePromptData = async ( const storePromptData = async (
notification: Notification, notification: Notification,
ollamaResponseBody: OllamaResponse ollamaResponseBody: OllamaChatResponse
) => { ) => {
try { try {
await prisma.response.updateMany({ await prisma.response.updateMany({
where: { pleromaNotificationId: notification.id }, where: { pleromaNotificationId: notification.id },
data: { data: {
response: ollamaResponseBody.response, response: ollamaResponseBody.message.content,
request: trimInputData(notification.status.content), request: trimInputData(notification.status.content),
to: notification.account.fqn, to: notification.account.fqn,
isProcessing: false, isProcessing: false,

View File

@ -40,7 +40,7 @@ const shouldContinue = (notification: Notification) => {
const { botAccountId } = envConfig; const { botAccountId } = envConfig;
const statusContent = trimInputData(notification.status.content); const statusContent = trimInputData(notification.status.content);
if ( if (
notification.status.visibility !== "private" && // notification.status.visibility !== "private" &&
!notification.account.bot && !notification.account.bot &&
notification.type === "mention" notification.type === "mention"
) { ) {

View File

@ -5,7 +5,7 @@ After=network-online.target
[Service] [Service]
Type=simple Type=simple
User=USERNAME_HERE User=bot
Restart=always Restart=always
RestartSec=3 RestartSec=3
ExecStart=/usr/bin/screen -L -DmS pleroma-ollama-bot /home/bot/.nvm/versions/node/v22.11.0/bin/npm run start ExecStart=/usr/bin/screen -L -DmS pleroma-ollama-bot /home/bot/.nvm/versions/node/v22.11.0/bin/npm run start

45
types.d.ts vendored
View File

@ -6,6 +6,41 @@ export interface Notification {
created_at: string; created_at: string;
} }
export interface ContextResponse {
ancestors: ContextObject[];
descendents: ContextObject[];
}
export interface PostAncestorsForModel {
account_fqn: string;
mentions: string[];
plaintext_content: string;
}
interface ContextAccountObject {
acct: string;
avatar: string;
bot: boolean;
display_name: string;
followers_count: number;
following_count: number;
fqn: string;
id: string;
}
export interface ContextObject {
content: string;
id: string;
in_reply_to_account_id: string | null;
in_reply_to_id: string | null;
media_attachments: string[];
mentions: Mention[];
pleroma: PleromaObjectInResponse;
visibility: "public" | "private" | "unlisted";
uri: string;
account: ContextAccountObject;
}
export interface NewStatusBody { export interface NewStatusBody {
content_type: "application/json" | "text/markdown"; content_type: "application/json" | "text/markdown";
in_reply_to_id?: string; in_reply_to_id?: string;
@ -94,9 +129,19 @@ export interface Status {
in_reply_to_account_id: string; // account ID of the reply in_reply_to_account_id: string; // account ID of the reply
in_reply_to_id: string; // status that the user has replied to in_reply_to_id: string; // status that the user has replied to
mentions: Mention[]; // array of mentions mentions: Mention[]; // array of mentions
pleroma: PleromaObjectInResponse;
visibility: "private" | "public" | "unlisted"; visibility: "private" | "public" | "unlisted";
} }
interface PleromaObjectInResponse {
content: { "text/plain": string };
context: string;
conversation_id: number;
direct_conversation_id: number | null;
local: boolean;
in_reply_to_account_acct: string;
}
export interface Mention { export interface Mention {
acct: string; acct: string;
id: string; id: string;