Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add generic inference engine to allow dynamic selection by the user #1226

Merged
merged 4 commits into from
Sep 19, 2024

Conversation

eladven
Copy link
Member

@eladven eladven commented Sep 19, 2024

The generic inference engine enables users to set the actual inference engine dynamically via the inference_engine environment variable. Additionally, the class includes a default parameter, which specifies the inference engine to use if the environment variable is not set.

For example, when the inference_engine variable is not set, the engine will
default to ibm_gen_ai.llama_3_8b_instruct. However, if
inference_engine=ollama.llama2 is set, it will switch to the Ollama model.

This update allows lm-eval users to run the UniTXT dataset with LLMaaJ using any supported inference engine.

The generic inference engine enables users to set the actual inference
engine dynamically via the inference_engine environment variable.
Additionally, the class includes a default parameter, which specifies the
inference engine to use if the environment variable is not set.

For example, when the inference_engine variable is not set, the engine will
 default to ibm_gen_ai.llama_3_8b_instruct. However, if
inference_engine=ollama.llama2 is set, it will switch to the Ollama model.

This update allows lm-eval users to run the UniTXT dataset with LLMaaJ
using any supported inference engine.
@eladven eladven merged commit 433ef2e into main Sep 19, 2024
8 checks passed
@eladven eladven deleted the genric_inference_engine branch September 19, 2024 13:05
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants