Skip to content

unexpected response when using llama2-7b-chat #3

@kaishxu

Description

@kaishxu

Hello!

I'm trying to use your pre-trained model with this command:
CUDA_VISIBLE_DEVICES=4,5,6,7 python inference.py -i -m llama-2-7b-chat --eval_name concat_recur

However, there is an unexpected generation stop when inputting the query:
help me list popular songs written by Taylor Swift.

The result is shown as follows:
Screenshot 2024-04-17 at 21 26 19

It stops generating more content and outputs </s> instead.

Are there any other settings I missed?

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions