Skip to content

Conversation

mohitmundhragithub
Copy link
Contributor

No description provided.

@mohitmundhragithub mohitmundhragithub requested review from anhappdev and a team as code owners August 26, 2025 05:41
Copy link

github-actions bot commented Aug 26, 2025

MLCommons CLA bot All contributors have signed the MLCommons CLA ✍️ ✅

…lemented performance benchmark for LLM pipeline
…y input and issue_query only handles output tokens
@farook-edev farook-edev changed the title Feat llm LLM pipeline implementation Sep 2, 2025
@farook-edev farook-edev linked an issue Sep 2, 2025 that may be closed by this pull request
Copy link

sonarqubecloud bot commented Sep 9, 2025

Copy link
Contributor Author

@mohitmundhragithub mohitmundhragithub left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

.

namespace mobile {

// A method to be called by the backend as soon as the first token is generated (only for token based benchmarks)
static void FirstTokenCallback(void* context) {
Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

what's the use of context?

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

the context is the arguments that get passed to loadGen, these are created by the driver and sent to the backend. Backend only needs to pass those to the callback without reading/modifying them.

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

@freedomtan to check it.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

Check what to do to bring in LLM models
3 participants