Run LLMs locally without API costs using Ollama:
# Install Ollama
curl -fsSL https://ollama.ai/install.sh | sh
# Pull a model
ollama pull llama2
ollama pull codellama
Node.js Integration
npm install ollama
import { Ollama } from 'ollama';
const ollama = new Ollama();
// Simple completion
const response = await ollama.chat({
model: 'llama2',
messages: [{ role: 'user', content: 'Explain closures in JS' }]
});
console.log(response.message.content);
// Streaming
const stream = await ollama.chat({
model: 'codellama',
messages: [{ role: 'user', content: 'Write a fibonacci function' }],
stream: true
});
for await (const chunk of stream) {
process.stdout.write(chunk.message.content);
}
Memory Requirements
- 7B models: 8GB RAM
- 13B models: 16GB RAM
- 70B models: 64GB+ RAM
