---
title: DeepSeek LLM | Developer Documentation
---

[DeepSeek Platform](https://platform.deepseek.com/)

## Usage

```
import { Settings } from "llamaindex";
import { DeepSeekLLM } from "@llamaindex/deepseek";


Settings.llm = new DeepSeekLLM({
  apiKey: "<YOUR_API_KEY>",
  model: "deepseek-coder", // or "deepseek-chat"
});
```

## Example

```
import { Document, VectorStoreIndex, Settings } from "llamaindex";
import { DeepSeekLLM } from "@llamaindex/deepseek";


const deepseekLlm = new DeepSeekLLM({
  apiKey: "<YOUR_API_KEY>",
  model: "deepseek-coder", // or "deepseek-chat"
});


async function main() {
  const response = await llm.deepseekLlm.chat({
    messages: [
      {
        role: "system",
        content: "You are an AI assistant",
      },
      {
        role: "user",
        content: "Tell me about San Francisco",
      },
    ],
    stream: false,
  });
  console.log(response);
}
```

# Limitations

Currently does not support function calling.

[Currently does not support json-output param while still is very good at json generating.](https://platform.deepseek.com/api-docs/faq#does-your-api-support-json-output)

## API Reference

- [DeepSeekLLM](/typescript/framework-api-reference/classes/deepseekllm/index.md)
