Skip to content

Conversation

abrown
Copy link
Collaborator

@abrown abrown commented Aug 13, 2024

Recent conversations (especially in the machine learning working group) suggest a need for a more specialized version of wasi-nn targeting LLMs. This change proposes such an interface, allowing users to access an ML graph directly using prompt strings instead of through the original tensor-based interface (i.e., inference).

Recent conversations (especially in the [machine learning working
group]) suggest a need for a more specialized version of wasi-nn
targeting LLMs. This change proposes such an interface, allowing users
to access an ML `graph` directly using prompt strings instead of through
the original tensor-based interface (i.e., `inference`).

[machine learning working group]: https://github.com/bytecodealliance/meetings/tree/main/machine-learning/
@abrown abrown marked this pull request as draft August 13, 2024 22:43
@shschaefer
Copy link
Contributor

Are you assuming that all LLMs have intrinsic tokenization? Not all foundation models are transforms from string to string.

Is the implication is that kv-cache and other stateful items will be kept opaquely by this context rather than maintained by the caller?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants