PXAI
v8.42
Feed
Viral
World
Politics
Technology
Daily Briefing
Sources
|
ToS
Version History
Version
Description
Date
8.42
[ADDED SOURCE REGISTRY LIST VIEW]
04/03/2026 16:19
8.41
[LOCALIZED INDEX.PHP MENU AND MODALS TO US ENGLISH]
04/03/2026 16:11
8.40
[TRANSLATED FRONTEND TO US ENGLISH]
04/03/2026 16:10
8.41
[RESTORED DETAILED SOURCES VIEW (ARTICLES + EVENTS) IN US ENGLISH]
04/03/2026 15:50
8.41
[RESTORED DETAILED SOURCES VIEW (ARTICLES + EVENTS) IN US ENGLISH]
04/03/2026 15:46
8.40
[REVERTED TO STANDARD TABLE LAYOUT (US ENGLISH)]
04/03/2026 15:40
8.39
[FIXED DB CONNECTION SCOPE IN SOURCES LOGIC]
04/03/2026 15:38
8.38
[TRANSLATED SOURCES VIEW TO US ENGLISH]
04/03/2026 15:01
8.37
[FULL FRONTEND TRANSLATION (MENU, FEED, COMMENTS, TTS) TO US ENGLISH.]
04/03/2026 14:58
8.67
[MANUAL OVERRIDE OF GREEK MENU ITEMS]
04/03/2026 13:36
8.66
[NGINX OPTIMIZED MENU FIX]
04/03/2026 13:33
8.65
[FIXED GREEK MENU ITEMS]
04/03/2026 13:32
8.60
[MENU & UI LOCALIZATION TO US ENGLISH]
04/03/2026 13:30
8.50
[FULL TRANSLATION TO US ENGLISH]
04/03/2026 13:27
9.75
[INJECTED EVENT FUSION_SUMMARY INTO FEED LOOPS TO DISPLAY AI TAGS CORRECTLY]
26/02/2026 14:44
PXAI Audio Feed
+5
ΟΛΑ
25/03 15:24
ente.com
Local LLM by Ente
Local LLM
Ente C
privacy
offline AI
lightweight model
NLP
24/03 23:35
dev.to
Urgent Security Alerts & Self-Hosted Swarm: Building Local LLM Infra Safely
security
LiteLLM
LM Studio
Docker Swarm
Komodo v2
local LLM infra
24/03 18:51
dev.to
How to Run NemoClaw with a Local LLM & Connect to Telegram (Without Losing Your Mind)
NemoClaw
local LLM
Telegram integration
WSL2
RTX 4080
AI agent
24/03 10:52
dev.to
# Reading YC-Backed Code #1: claude-mem — Great Idea, Poor Implementation
claude‑mem
persistent memory
Claude Code
Y Combinator
implementation critique
local LLM
24/03 05:38
dev.to
Self-Hosting AI in 2026: Privacy, Control, and the Case for Running Your Own
self‑hosting AI
privacy
control
local LLM
cloud models
user autonomy
22/03 14:31
dev.to
AI Agents in 2026: Building Autonomous Workflows with Local LLMs and Open-Source Frameworks
AI agents
local LLMs
open‑source frameworks
market growth
data privacy
compliance
22/03 14:21
dev.to
Running Qwen2.5-32B on RTX 4060 8GB — Beating M4 at 10.8 t/s with llama.cpp
Qwen2.5-32B
RTX 4060
llama.cpp
hybrid inference
local LLM
token throughput
22/03 14:19
dev.to
llama.swap Model Switcher Quickstart for OpenAI-Compatible Local LLMs
llama‑swap
local LLM
OpenAI API
proxy
model switching
vLLM
20/03 07:42
dev.to
Building a Perplexity Clone for Local LLMs in 50 Lines of Python
local LLM
web search
citation engine
Python pipeline
Perplexity clone
RAG
20/03 07:42
dev.to
Building a Perplexity Clone for Local LLMs in 50 Lines of Python
local LLM
web search
citation engine
Python pipeline
Perplexity clone
RAG
Comments
Loading...
Send
Dev Changelog
v8.42
No logs found in database.
0
Display Settings
Size
Aa
Brightness
Theme
Dark
Comments