Skip to content

[examples/SplitLlama] Add the appendToken ,generatedToken and sendToClient interfaces#10

Merged
FloatingcloudKnight merged 3 commits intoFloatingcloudKnight:mainfrom
asdf1113:dev
Jun 3, 2025
Merged

[examples/SplitLlama] Add the appendToken ,generatedToken and sendToClient interfaces#10
FloatingcloudKnight merged 3 commits intoFloatingcloudKnight:mainfrom
asdf1113:dev

Conversation

@asdf1113
Copy link
Collaborator

@asdf1113 asdf1113 commented May 28, 2025

The main modifications include:

In llamaInput.h, extracting the functionality of appending the inferred token to the input sequence into the appendToken function.

In llamaOutput.h, extracting the functionality of generating a token based on the inference results and appending it to the output sequence into the generatedToken function.

Extract the functionality to send data to other modules from all header files into the sendToClient function

Additionally, the tokenizeInput and send_data function from llamaInput.h has been moved to BaseDisModel.h.

@asdf1113 asdf1113 changed the title [examples/SplitLlama] Add the appendToken and generatedToken interfaces [examples/SplitLlama] Add the appendToken ,generatedToken and sendToClient interfaces May 29, 2025
@FloatingcloudKnight FloatingcloudKnight merged commit b303115 into FloatingcloudKnight:main Jun 3, 2025
1 check failed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants