Skip to content

Hugging Face

Jan supports Hugging Face models through two methods: the new HF Router (recommended) and Inference Endpoints. Both methods require a Hugging Face token and billing to be set up.

HuggingFace Inference Providers

The HF Router provides access to models from multiple providers (Replicate, Together AI, SambaNova, Fireworks, Cohere, and more) through a single endpoint.

Visit Hugging Face Settings > Access Tokens and create a token. Make sure you have billing set up on your account.

  1. Go to Settings > Model Providers > HuggingFace
  2. Enter your HF token
  3. Use this URL: https://router.huggingface.co/v1

Jan HF Setup

You can find out more about the HF Router here.

Jan comes with three HF Router models pre-configured. Select one and start chatting immediately.

For more control over specific models and deployment configurations, you can use Hugging Face Inference Endpoints.

Visit the Hugging Face Model Hub (make sure you are logged in) and pick the model you want to use.

HuggingFace Model Hub

After you have selected the model you want to use, click on the Deploy button and select a deployment method. We will select HF Inference Endpoints for this one.

HuggingFace Deployment

This will take you to the deployment set up page. For this example, we will leave the default settings as they are under the GPU tab and click on Create Endpoint.

HuggingFace Deployment

Once your endpoint is ready, test that it works on the Test your endpoint tab.

HuggingFace Deployment

If you get a response, you can click on Copy to copy the endpoint URL and API key.

If you do not have an API key you can create one under Settings > Access Tokens here. Once you finish, copy the token and add it to Jan alongside your endpoint URL at Settings > Model Providers > HuggingFace.

3.1 HF Token Get Token

3.2 HF Endpoint URL Endpoint URL

3.3 Jan Settings Jan Settings

3.4 Add Model Details Add Model Details

Now you can start using the model in any chat.

Start Using the Model

If you want to learn how to use Jan Nano with MCP, check out the guide here.

Option 1 (HF Router): Access to models from multiple providers as shown in the providers image above.

Option 2 (Inference Endpoints): You can follow the steps above with a large amount of models on Hugging Face and bring them to Jan. Check out other models in the Hugging Face Model Hub.

Common issues and solutions:

1. Started a chat but the model is not responding

  • Verify your API_KEY/HF_TOKEN is correct and not expired
  • Ensure you have billing set up on your HF account
  • For Inference Endpoints: Ensure the model you’re trying to use is running again since, after a while, they go idle so that you don’t get charged when you are not using it

Model Running

2. Connection Problems

  • Check your internet connection
  • Verify Hugging Face’s system status
  • Look for error messages in Jan’s logs

3. Model Unavailable

  • Confirm your API key has access to the model
  • Check if you’re using the correct model ID
  • Verify your Hugging Face account has the necessary permissions

Need more help? Join our Discord community or check the Hugging Face’s documentation.