Monday, March 11, 2024

Show HN: Prompts as (WASM) Programs https://ift.tt/d2NMF4S

Show HN: Prompts as (WASM) Programs AICI is a proposed common interface between LLM inference engines (llama.cpp, vLLM, HF Transformers, etc.) and "controllers" - programs that can constrain the LLM output according to regexp, grammar, or custom logic, as well as control the generation process (forking, backtracking, etc.). AICI is based on Wasm, and is designed to be fast (runs on CPU while GPU is busy), secure (can run in multi-tenant cloud deployments), and flexible (allow libraries like Guidance, LMQL, Outlines, etc. to work on top of it). We (Microsoft Research) have released it recently, and would love feedback on the design of the interface, as well as our Rust AICI runtime. I'm the lead developer on this project and happy to answer any questions! https://ift.tt/ovVEht7 March 11, 2024 at 10:30PM

No comments:

Post a Comment

Show HN: MCP Server for Tally – Create and Manage Forms with Claude https://ift.tt/JOfDkWh

Show HN: MCP Server for Tally – Create and Manage Forms with Claude I've built an MCP server for Tally that bridges the gap between thei...