Skip to content

Conversation

changqi1
Copy link
Contributor

@changqi1 changqi1 commented Apr 9, 2024

No description provided.

@changqi1 changqi1 requested a review from pujiang2018 April 9, 2024 05:33
@pujiang2018
Copy link
Contributor

@abenmao Could you please review this PR?


ctx->resize(1, numTokens, 0);
llama_mlp->forward(ctx, (float *)const_cast<void *>(input), (float *)output, inputStride, outputStride, false);
} else if (dt == DataType::fp16) {
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Can these duplicated or similar code lines for fp16 be reused for the bf16 section?

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thinked, and the answer is no.

@changqi1 changqi1 merged commit 280a915 into intel:main Apr 15, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

4 participants