Movatterモバイル変換


[0]ホーム

URL:


Skip to content

Navigation Menu

Sign in
Appearance settings

Search code, repositories, users, issues, pull requests...

Provide feedback

We read every piece of feedback, and take your input very seriously.

Saved searches

Use saved searches to filter your results more quickly

Sign up
Appearance settings

Fixing block size for Mistral-7B.#141

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to ourterms of service andprivacy statement. We’ll occasionally send you account related emails.

Already on GitHub?Sign in to your account

Open
Artyom17 wants to merge3 commits intometa-pytorch:main
base:main
Choose a base branch
Loading
fromSesameAILabs:art/fix-mistral

Conversation

@Artyom17
Copy link
Contributor

According to Mistral's paper the block size for Mistral-7B should be 8192 (ref:https://arxiv.org/pdf/2310.06825.pdf,https://huggingface.co/docs/transformers/en/model_doc/mistral). But currently it is set to the default value (2048).

learning-chip reacted with thumbs up emoji
@facebook-github-botfacebook-github-bot added the CLA SignedThis label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed. labelMar 19, 2024
@Artyom17
Copy link
ContributorAuthor

It also saves some memory on 'freq_cis' tensor when the large block_size is used with relatively small max_seq_length.

Sign up for freeto join this conversation on GitHub. Already have an account?Sign in to comment

Reviewers

No reviews

Assignees

No one assigned

Labels

CLA SignedThis label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed.

Projects

None yet

Milestone

No milestone

Development

Successfully merging this pull request may close these issues.

2 participants

@Artyom17@facebook-github-bot

[8]ページ先頭

©2009-2025 Movatter.jp