Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
@@ -12,7 +12,12 @@ from transformers import AutoProcessor, Gemma3ForConditionalGeneration, TextIter
|
|
12 |
For more information on `huggingface_hub` Inference API support, please check the docs: https://huggingface.co/docs/huggingface_hub/v0.22.2/en/guides/inference
|
13 |
"""
|
14 |
# client = InferenceClient("HuggingFaceH4/zephyr-7b-beta")
|
15 |
-
|
|
|
|
|
|
|
|
|
|
|
16 |
duration=None
|
17 |
|
18 |
login(token = os.getenv('gemma'))
|
|
|
12 |
For more information on `huggingface_hub` Inference API support, please check the docs: https://huggingface.co/docs/huggingface_hub/v0.22.2/en/guides/inference
|
13 |
"""
|
14 |
# client = InferenceClient("HuggingFaceH4/zephyr-7b-beta")
|
15 |
+
print(f"Is CUDA available: {torch.cuda.is_available()}")
|
16 |
+
print(f"CUDA device: {torch.cuda.get_device_name(torch.cuda.current_device())}")
|
17 |
+
print(f"CUDA version: {torch.version.cuda}")
|
18 |
+
print(f"Python version: {platform.python_version()}")
|
19 |
+
print(f"Pytorch version: {torch.__version__}")
|
20 |
+
print(f"Gradio version: {gr. __version__}")
|
21 |
duration=None
|
22 |
|
23 |
login(token = os.getenv('gemma'))
|