TokenSlim

Compress LLM tokens by 30-50%

Paste any JSON. See instant SLIM compression. Ship it with one import change.

Try it now
30-50%
Token savings
93.8%
LLM accuracy
1 line
Code change
JSON351 chars
[
  {
    "employeeId": "EMP-001",
    "firstName": "Alice",
    "department": "Engineering",
    "salary": 125000
  },
  {
    "employeeId": "EMP-002",
    "firstName": "Bob",
    "department": "Marketing",
    "salary": 95000
  },
  {
    "employeeId": "EMP-003",
    "firstName": "Carol",
    "department": "Engineering",
    "salary": 130000
  }
]
SLIM193 chars
@keys{employeeId:_0,firstName:_1,department:_2}

[3]{department,employeeId,firstName,salary}:
  Engineering,EMP-001,Alice,125000
  Marketing,EMP-002,Bob,95000
  Engineering,EMP-003,Carol,130000

Live Playground

Paste JSON or pick a sample to see SLIM compression in real time.

Samples:
Delimiter:
JSON Input
SLIM Output
SLIM output appears here...

Cost Calculator

Estimate how much you'll save per month with TokenSlim.

10%70%
Monthly savings
$50.00
Yearly savings
$600.00
Monthly cost (before)
$125.00
Monthly cost (after)
$75.00

Integrate in 1 minute

Change one import. Keep your existing code.

npm install @tokenslim/openai openai
Before
import OpenAI from 'openai';

const client = new OpenAI({
  apiKey: process.env.OPENAI_API_KEY,
});

const response = await client.chat.completions.create({
  model: 'gpt-4o',
  messages: [
    { role: 'user', content: JSON.stringify(data) },
  ],
});
After (TokenSlim)
import { TokenSlimOpenAI as OpenAI } from '@tokenslim/openai';

const client = new OpenAI({
  apiKey: process.env.OPENAI_API_KEY,
});

const response = await client.chat.completions.create({
  model: 'gpt-4o',
  messages: [
    { role: 'user', content: JSON.stringify(data) },
  ],
});

That's it. All SDK features, types, and error handling work identically.