# QuantizedBiEncoderEmbeddings

> **Class** in `langchain_community`

📖 [View in docs](https://reference.langchain.com/python/langchain-community/embeddings/optimum_intel/QuantizedBiEncoderEmbeddings)

Quantized bi-encoders embedding models.

Please ensure that you have installed optimum-intel and ipex.

## Signature

```python
QuantizedBiEncoderEmbeddings(
    self,
    model_name: str,
    max_seq_len: int = 512,
    pooling_strategy: str = 'mean',
    query_instruction: Optional[str] = None,
    document_instruction: Optional[str] = None,
    padding: bool = True,
    model_kwargs: Optional[Dict] = None,
    encode_kwargs: Optional[Dict] = None,
    **kwargs: Any = {},
)
```

## Description

**Input:**

model_name: str = Model name.
max_seq_len: int = The maximum sequence length for tokenization. (default 512)
pooling_strategy: str =
    "mean" or "cls", pooling strategy for the final layer. (default "mean")
query_instruction: Optional[str] =
    An instruction to add to the query before embedding. (default None)
document_instruction: Optional[str] =
    An instruction to add to each document before embedding. (default None)
padding: Optional[bool] =
    Whether to add padding during tokenization or not. (default True)
model_kwargs: Optional[Dict] =
    Parameters to add to the model during initialization. (default {})
encode_kwargs: Optional[Dict] =
    Parameters to add during the embedding forward pass. (default {})

Example:

from langchain_community.embeddings import QuantizedBiEncoderEmbeddings

model_name = "Intel/bge-small-en-v1.5-rag-int8-static"
encode_kwargs = {'normalize_embeddings': True}
hf = QuantizedBiEncoderEmbeddings(
    model_name,
    encode_kwargs=encode_kwargs,
    query_instruction="Represent this sentence for searching relevant passages: "
)

## Extends

- `BaseModel`
- `Embeddings`

## Constructors

```python
__init__(
    self,
    model_name: str,
    max_seq_len: int = 512,
    pooling_strategy: str = 'mean',
    query_instruction: Optional[str] = None,
    document_instruction: Optional[str] = None,
    padding: bool = True,
    model_kwargs: Optional[Dict] = None,
    encode_kwargs: Optional[Dict] = None,
    **kwargs: Any = {},
) -> None
```

| Name | Type |
|------|------|
| `model_name` | `str` |
| `max_seq_len` | `int` |
| `pooling_strategy` | `str` |
| `query_instruction` | `Optional[str]` |
| `document_instruction` | `Optional[str]` |
| `padding` | `bool` |
| `model_kwargs` | `Optional[Dict]` |
| `encode_kwargs` | `Optional[Dict]` |


## Properties

- `model_name_or_path`
- `max_seq_len`
- `pooling`
- `padding`
- `encode_kwargs`
- `model_kwargs`
- `normalize`
- `batch_size`
- `query_instruction`
- `document_instruction`
- `model_config`

## Methods

- [`load_model()`](https://reference.langchain.com/python/langchain-community/embeddings/optimum_intel/QuantizedBiEncoderEmbeddings/load_model)
- [`embed_documents()`](https://reference.langchain.com/python/langchain-community/embeddings/optimum_intel/QuantizedBiEncoderEmbeddings/embed_documents)
- [`embed_query()`](https://reference.langchain.com/python/langchain-community/embeddings/optimum_intel/QuantizedBiEncoderEmbeddings/embed_query)

---

[View source on GitHub](https://github.com/langchain-ai/langchain-community/blob/a6a6079511ac8a5c1293337f88096b8641562e77/libs/community/langchain_community/embeddings/optimum_intel.py#L7)