You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I am a beginner in LLM and am new to learn structure generation with Xgrammer.
I find that you have provided the benchmark results for Llama.cpp in the blog post and paper.
However, I do not find the benchmark experiment in the open source Xgrammer repo: examples/benchmark/bench_grammar_compile_mask_gen.py (I think it should be written here?)
If so, would you please add the test code snapshot for benchmarking Llama.cpp and show how to integrate it with Xgrammer?
Thanks.
Another question is when I run the python bench_grammar_compile_mask_gen.py --backend lmformatenforcer , I got the following error
with the same dataset in the file downloaded from huggingface.
What might be the problem?
The text was updated successfully, but these errors were encountered:
Hi @zlwu92, thanks for asking questions about beginning to use XGrammar and testing about llama.cpp.
For beginners, I would suggest following our tutorial that describes how to use xgrammar and huggingface transformer to guide the generation process. It's easy to learn and a very useful application scenario.
Regrading benchmark, the benchmark of llama.cpp and its internal grammar engine was done on our own fork because we needed to measure the speed of grammar initialization and mask generation.
show how to integrate it with Xgrammer
We do have a plan to integrate XGrammar into llama.cpp because we have a C++ API with complete features. That would come later.
Other baselines have changed a bit since we did our benchmark. We will update the script accordingly to make it work.
Currently, does the open-sourced xgrammer include scripts for the two benchmarking experiments (1. speed of masking logits and 2. end-to-end evaluation for the LLM inference engine efficiency in serving scenarios) in the paper or not?
Hi,
I am a beginner in LLM and am new to learn structure generation with Xgrammer.
I find that you have provided the benchmark results for Llama.cpp in the blog post and paper.
However, I do not find the benchmark experiment in the open source Xgrammer repo:
examples/benchmark/bench_grammar_compile_mask_gen.py
(I think it should be written here?)If so, would you please add the test code snapshot for benchmarking Llama.cpp and show how to integrate it with Xgrammer?
Thanks.
Another question is when I run the
python bench_grammar_compile_mask_gen.py --backend lmformatenforcer
, I got the following errorWhat might be the problem?
The text was updated successfully, but these errors were encountered: