Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

partners: Fixed the procedure of initializing pad_token_id #29500

Open
wants to merge 1 commit into
base: master
Choose a base branch
from

Conversation

tishizaki
Copy link
Contributor

Example code is followings:

from langchain_huggingface.llms import HuggingFacePipeline

hf = HuggingFacePipeline.from_model_id(
    model_id="meta-llama/Llama-3.2-3B-Instruct",
    task="text-generation",
    pipeline_kwargs={"max_new_tokens": 10},
)

from langchain_core.prompts import PromptTemplate

template = """Question: {question}

Answer: Let's think step by step."""
prompt = PromptTemplate.from_template(template)

chain = prompt | hf

question = "What is electroencephalography?"

print(chain.invoke({"question": question}))

Copy link

vercel bot commented Jan 30, 2025

The latest updates on your projects. Learn more about Vercel for Git ↗︎

1 Skipped Deployment
Name Status Preview Comments Updated (UTC)
langchain ⬜️ Ignored (Inspect) Visit Preview Jan 30, 2025 1:34pm

@dosubot dosubot bot added size:S This PR changes 10-29 lines, ignoring generated files. 🤖:bug Related to a bug, vulnerability, unexpected error with an existing feature labels Jan 30, 2025
Add to check pad_token_id and eos_token_id of model config.
It seems that this is the same bug as the HuggingFace TGI bug.

In additon, fix lint errors of test.
@dosubot dosubot bot added size:M This PR changes 30-99 lines, ignoring generated files. and removed size:S This PR changes 10-29 lines, ignoring generated files. labels Jan 30, 2025
@tishizaki
Copy link
Contributor Author

@ccurme I tried to add test program, but I didn't know how to refer the
tokenizer's member variables after doing HuggingFacePipeline.from_model_id() .
Sorry, the test program is incomplete.

Following is the program I'm currently considering to add to libs/partners/huggingface/tests/unit_tests/test_huggingface_pipeline.py.
But now, it's hard-coded program with the parameter of Llama-3.2-3B-Instruct.

@patch("transformers.AutoTokenizer.from_pretrained")
@patch("transformers.AutoModelForCausalLM.from_pretrained")
@patch("transformers.pipeline")
def test_initialization_with_from_model_id_pad_token_id(
    mock_pipeline: MagicMock, mock_model: MagicMock, mock_tokenizer: MagicMock
) -> None:
    """Test initialization with the from_model_id method"""

    mock_tokenizer.return_value = MagicMock(pad_token=None, eos_token_id=128009)
    mock_model.return_value = MagicMock()
    mock_model.return_value.config.pad_token_id = None
    mock_model.return_value.config.eos_token_id = [128001, 128008, 128009]

    mock_pipe = MagicMock()
    mock_pipe.task = "text-generation"
    mock_pipe.model = mock_model.return_value
    mock_pipeline.return_value = mock_pipe

    llm = HuggingFacePipeline.from_model_id(
        model_id="mock-model-id",
        task="text-generation",
    )

    assert [XXX.tokenizer.pad_token_id == 128009]

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
🤖:bug Related to a bug, vulnerability, unexpected error with an existing feature size:M This PR changes 30-99 lines, ignoring generated files.
Projects
Status: Triage
Development

Successfully merging this pull request may close these issues.

1 participant