filipealmeida
commited on
Commit
•
9c6b8fd
1
Parent(s):
18d0ebf
Command line argument for context size
Browse files
app.py
CHANGED
@@ -47,6 +47,7 @@ def generate_text(prompt, example):
|
|
47 |
parser = argparse.ArgumentParser()
|
48 |
parser.add_argument("--model", help="Path to the model file")
|
49 |
parser.add_argument("--gpu", help="How many GPU layers to use", default=0, type=int)
|
|
|
50 |
args = parser.parse_args()
|
51 |
|
52 |
if args.model:
|
@@ -55,7 +56,7 @@ else:
|
|
55 |
model_path = download_model()
|
56 |
|
57 |
# TODO n_gpu_layers doesn't seem to be working
|
58 |
-
llm = Llama(model_path=model_path, n_gpu_layers=args.gpu)
|
59 |
|
60 |
# Create a Gradio interface
|
61 |
interface = gr.Interface(
|
|
|
47 |
parser = argparse.ArgumentParser()
|
48 |
parser.add_argument("--model", help="Path to the model file")
|
49 |
parser.add_argument("--gpu", help="How many GPU layers to use", default=0, type=int)
|
50 |
+
parser.add_argument("--ctx", help="How many context tokens to use", default=512, type=int)
|
51 |
args = parser.parse_args()
|
52 |
|
53 |
if args.model:
|
|
|
56 |
model_path = download_model()
|
57 |
|
58 |
# TODO n_gpu_layers doesn't seem to be working
|
59 |
+
llm = Llama(model_path=model_path, n_gpu_layers=args.gpu, n_ctx=args.ctx)
|
60 |
|
61 |
# Create a Gradio interface
|
62 |
interface = gr.Interface(
|