Llama-3.1-8B-DALv0.1
/
venv
/lib
/python3.12
/site-packages
/torch
/nn
/quantizable
/modules
/activation.py
# flake8: noqa: F401 | |
r"""Quantizable Modules. | |
This file is in the process of migration to `torch/ao/nn/quantizable`, and | |
is kept here for compatibility while the migration process is ongoing. | |
If you are adding a new entry/functionality, please, add it to the | |
appropriate file under the `torch/ao/nn/quantizable/modules`, | |
while adding an import statement here. | |
""" | |
from torch.ao.nn.quantizable.modules.activation import MultiheadAttention | |