code
stringlengths 501
5.19M
| package
stringlengths 2
81
| path
stringlengths 9
304
| filename
stringlengths 4
145
|
---|---|---|---|
{{ fullname | escape | underline }}
.. automodule:: {{ fullname }}
{% block functions %}
{% if functions %}
.. rubric:: Functions
.. autosummary::
:toctree: {{ objname }}
{% for item in functions %}
{{ item }}
{%- endfor %}
{% endif %}
{% endblock %}
{% block classes %}
{% if classes %}
.. rubric:: Classes
.. autosummary::
:toctree: {{ objname }}
:template: class.rst
{% for item in classes %}
{{ item }}
{%- endfor %}
{% endif %}
{% endblock %}
{% block exceptions %}
{% if exceptions %}
.. rubric:: Exceptions
.. autosummary::
{% for item in exceptions %}
{{ item }}
{%- endfor %}
{% endif %}
{% endblock %}
| 21cmFAST | /21cmFAST-3.3.1.tar.gz/21cmFAST-3.3.1/docs/templates/module.rst | module.rst |
```
%matplotlib inline
import matplotlib.pyplot as plt
from py21cmmc import run_lightcone
import numpy as np
import matplotlib.pyplot as plt
%matplotlib inline
import py21cmmc as p21
from powerbox import get_power
init, perturb, xHI, Tb = p21.run_coeval(
redshift= 8.0,
user_params = dict(
HII_DIM = 100,
BOX_LEN = 300.0
)
)
plt.imshow(init.lowres_density[:,:,0], extent=(0,300,0,300));
plt.imshow(Tb.brightness_temp[:,:,0], extent=(0,300,0,300))
p, k = get_power(Tb.brightness_temp, boxlength=Tb.user_params.BOX_LEN)
plt.plot(k,p*k**3)
plt.xscale('log')
plt.yscale('log')
lc = run_lightcone(6.5, max_redshift=12.0, user_params={"HII_DIM":100, "DIM":300})
plt.figure(figsize=(10,4))
plt.imshow(lc.brightness_temp[0], origin='lower')
plt.colorbar()
```
| 21cmFAST | /21cmFAST-3.3.1.tar.gz/21cmFAST-3.3.1/devel/visualise_boxes_and_power.ipynb | visualise_boxes_and_power.ipynb |
import matplotlib.pyplot as plt
import numpy as np
from py21cmfast import AstroParams, CosmoParams, FlagOptions, UserParams
from py21cmfast._memory import estimate_memory_lightcone
user_params = UserParams(
{
"HII_DIM": 250,
"BOX_LEN": 2000.0,
"DIM": 1000,
"N_THREADS": 8,
"USE_RELATIVE_VELOCITIES": True,
"POWER_SPECTRUM": 5,
"USE_FFTW_WISDOM": True,
"PERTURB_ON_HIGH_RES": True,
"USE_INTERPOLATION_TABLES": True,
"FAST_FCOLL_TABLES": True,
}
)
flag_options = FlagOptions(
{
"INHOMO_RECO": True,
"USE_MASS_DEPENDENT_ZETA": True,
"USE_TS_FLUCT": True,
"USE_MINI_HALOS": True,
},
USE_VELS_AUX=True,
)
h2dim = np.array(list(range(200, 1500, 100)))
mems4 = []
for h2 in range(200, 1500, 100):
user_params.update(HII_DIM=h2, DIM=4 * h2)
mem = estimate_memory_lightcone(
max_redshift=1420 / 50 - 1,
redshift=1420 / 200 - 1,
user_params=user_params,
flag_options=flag_options,
cosmo_params=CosmoParams(),
astro_params=AstroParams(),
)
mems4.append(mem)
peaks4 = np.array([m["peak_memory"] / (1024**3) for m in mems4])
mems3 = []
for h2 in range(200, 1500, 100):
user_params.update(HII_DIM=h2, DIM=3 * h2)
mem = estimate_memory_lightcone(
max_redshift=1420 / 50 - 1,
redshift=1420 / 200 - 1,
user_params=user_params,
flag_options=flag_options,
cosmo_params=CosmoParams(),
astro_params=AstroParams(),
)
mems3.append(mem)
peaks3 = np.array([m["peak_memory"] / (1024**3) for m in mems3])
mems2 = []
for h2 in range(200, 1500, 100):
user_params.update(HII_DIM=h2, DIM=2 * h2)
mem = estimate_memory_lightcone(
max_redshift=1420 / 50 - 1,
redshift=1420 / 200 - 1,
user_params=user_params,
flag_options=flag_options,
cosmo_params=CosmoParams(),
astro_params=AstroParams(),
)
mems2.append(mem)
peaks2 = np.array([m["peak_memory"] / (1024**3) for m in mems2])
user_params.update(HII_DIM=1300, DIM=2600) # what we actually did.
mem = estimate_memory_lightcone(
max_redshift=1420 / 50 - 1,
redshift=1420 / 200 - 1,
user_params=user_params,
flag_options=flag_options,
cosmo_params=CosmoParams(),
astro_params=AstroParams(),
)
plt.plot(h2dim, peaks2, label="DIM=2*HII_DIM")
plt.plot(h2dim, peaks3, label="DIM=3*HII_DIM")
plt.plot(h2dim, peaks4, label="DIM=4*HII_DIM")
plt.scatter(
1300,
mem["peak_memory"] / 1024**3,
marker="*",
color="k",
label="Proposed Sims",
s=100,
zorder=3,
)
plt.scatter(
1300, 2924.847, marker="+", color="r", label="Measured Peak RAM", s=75, zorder=3
)
theory = h2dim**3
plt.plot(
h2dim,
theory * (peaks3.max() / theory.max()) * 1.2,
label="HII_DIM^3 scaling",
color="grey",
ls="--",
)
plt.axhline(4000, linestyle="dashed", color="k")
plt.text(h2dim[0], 4300, "Avail RAM on Bridges2-EM")
plt.legend()
plt.yscale("log")
plt.xscale("log")
plt.ylabel("Peak memory (GB)")
plt.xlabel("HII_DIM")
plt.savefig("peak_memory_usage.png") | 21cmFAST | /21cmFAST-3.3.1.tar.gz/21cmFAST-3.3.1/devel/estimated_mem_usage_plot.py | estimated_mem_usage_plot.py |
---
title: '21cmFAST v3: A Python-integrated C code for generating 3D realizations of the cosmic 21cm signal.'
tags:
- Python
- astronomy
- cosmology
- simulation
authors:
- name: Steven G. Murray
orcid: 0000-0003-3059-3823
affiliation: 1
- name: Bradley Greig
orcid: 0000-0002-4085-2094
affiliation: 2, 3
- name: Andrei Mesinger
orcid: 0000-0003-3374-1772
affiliation: 4
- name: Julian B. Muñoz
orcid: 0000-0002-8984-0465
affiliation: 5
- name: Yuxiang Qin
orcid: 0000-0002-4314-1810
affiliation: 4
- name: Jaehong Park
orcid: 0000-0003-3095-6137
affiliation: 4, 7
- name: Catherine A. Watkinson
orcid: 0000-0003-1443-3483
affiliation: 6
affiliations:
- name: School of Earth and Space Exploration, Arizona State University, Phoenix, USA
index: 1
- name: ARC Centre of Excellence for All-Sky Astrophysics in 3 Dimensions (ASTRO 3D)
index: 2
- name: School of Physics, University of Melbourne, Parkville, VIC 3010, Australia
index: 3
- name: Scuola Normale Superiore, Piazza dei Cavalieri 7, 56126 Pisa, Italy
index: 4
- name: Department of Physics, Harvard University, 17 Oxford St., Cambridge, MA, 02138, USA
index: 5
- name: School of Physics and Astronomy, Queen Mary University of London, G O Jones Building, 327 Mile End Road, London, E1 4NS, UK
index: 6
- name: School of Physics, Korea Institute for Advanced Study, 85 Hoegiro, Dongdaemun-gu, Seoul, 02455, Republic of Korea
index: 7
date: 24 Sep 2020
bibliography: paper.bib
---
# Summary
The field of 21-cm cosmology -- in which the hyperfine spectral line of neutral hydrogen
(appearing at the rest-frame wavelength of 21 cm) is mapped over large swathes of the
Universe's history -- has developed radically over the last decade.
The promise of the field is to revolutionize our
knowledge of the first stars, galaxies, and black holes through the timing and patterns
they imprint on the cosmic 21-cm signal.
In order to interpret the eventual observational data, a range of physical models have
been developed -- from simple analytic models of the global history of hydrogen reionization,
through to fully hydrodynamical simulations of the 3D evolution of the brightness
temperature of the spectral line.
Between these extremes lies an especially versatile middle-ground: fast semi-numerical
models that approximate the full 3D evolution of the relevant fields: density, velocity,
temperature, ionization, and radiation (Lyman-alpha, neutral hydrogen 21-cm, etc.).
These have the advantage of being comparable to the full first-principles
hydrodynamic simulations, but significantly quicker to run; so much so that they can
be used to produce thousands of realizations on scales comparable to those observable
by upcoming low-frequency radio telescopes, in order to explore the very wide
parameter space that still remains consistent with the data.
Amongst practitioners in the field of 21-cm cosmology, the `21cmFAST` program has become
the *de facto* standard for such semi-numerical simulators.
`21cmFAST` [@mesinger2007; @mesinger2010] is a high-performance C code that uses the
excursion set formalism [@furlanetto2004] to
identify regions of ionized hydrogen atop a cosmological density field evolved using
first- or second-order Lagrangian perturbation theory [@zeldovich1970; @scoccimarro2002],
tracking the thermal and ionization state of the intergalactic medium, and computing
X-ray, soft UV and ionizing UV cosmic radiation fields based on parametrized galaxy models.
For example, the following figure contains slices of lightcones (3D fields in which one
axis corresponds to both spatial *and* temporal evolution) for the various
component fields produced by `21cmFAST`.
{height=450px}
However, `21cmFAST` is a highly specialized code, and its implementation has been
quite specific and relatively inflexible.
This inflexibility makes it difficult to modify the behaviour of the code without detailed
knowledge of the full system, or disrupting its workings.
This lack of modularity within the code has led to
widespread code "branching" as researchers hack new physical features of interest
into the C code; the lack of a streamlined API has led derivative codes which run
multiple realizations of `21cmFAST` simulations [such as the Monte Carlo simulator,
`21CMMC`, @greig2015] to re-write large portions of the code in order to serve their purpose.
It is thus of critical importance, as the field moves forward in its understanding -- and
the range and scale of physical models of interest continues to increase -- to
reformulate the `21cmFAST` code in order to provide a fast, modular, well-documented,
well-tested, stable simulator for the community.
# Features of 21cmFAST v3
This paper presents `21cmFAST` v3+, which is formulated to follow these essential
guiding principles.
While keeping the same core functionality of previous versions of `21cmFAST`, it has
been fully integrated into a Python package, with a simple and intuitive interface, and
a great deal more flexibility.
At a higher level, in order to maintain best practices, a community of users and
developers has coalesced into a formal collaboration which maintains the project via a
Github organization.
This allows the code to be consistently monitored for quality, maintaining high test
coverage, stylistic integrity, dependable release strategies and versioning,
and peer code review.
It also provides a single point-of-reference for the community to obtain the code,
report bugs and request new features (or get involved in development).
A significant part of the work of moving to a Python interface has been the
development of a robust series of underlying Python structures which handle the passing
of data between Python and C via the `CFFI` library.
This foundational work provides a platform for future versions to extend the scientific
capabilities of the underlying simulation code.
The primary *new* usability features of `21cmFAST` v3+ are:
* Convenient (Python) data objects which simplify access to and processing of the various
fields that form the brightness temperature.
* Enhancement of modularity: the underlying C functions for each step of the simulation
have been de-coupled, so that arbitrary functionality can be injected into the process.
* Conversion of most global parameters to local structs to enable this modularity, and
also to obviate the requirement to re-compile in order to change parameters.
* Simple `pip`-based installation.
* Robust on-disk caching/writing of data, both for efficiency and simplified reading of
previously processed data (using HDF5).
* Simple high-level API to generate either coeval cubes (purely spatial 3D fields defined
at a particular time) or full lightcone data (i.e. those coeval cubes interpolated over
cosmic time, mimicking actual observations).
* Improved exception handling and debugging.
* Convenient plotting routines.
* Simple configuration management, and also more intuitive management for the
remaining C global variables.
* Comprehensive API documentation and tutorials.
* Comprehensive test suite (and continuous integration).
* Strict semantic versioning^[https://semver.org].
While in v3 we have focused on the establishment of a stable and extendable infrastructure,
we have also incorporated several new scientific features, appearing in separate papers:
* Generate transfer functions using the `CLASS` Boltzmann code [@Lesgourgues2011].
* Simulate the effects of relative velocities between dark matter and Baryons [@munoz2019a; @munoz2019b].
* Correction for non-conservation of ionizing photons (Park, Greig et al., *in prep*).
* Include molecularly cooled galaxies with distinct properties [@qin2020]
* Calculate rest-frame UV luminosity functions based on parametrized galaxy models.
`21cmFAST` is still in very active development.
Amongst further usability and performance improvements,
future versions will see several new physical models implemented,
including milli-charged dark matter models [@Munoz2018] and forward-modelled CMB
auxiliary data [@qin2020a].
In addition, `21cmFAST` will be incorporated into large-scale inference codes, such as
`21CMMC`, and is being used to create large data-sets for inference via machine learning.
We hope that with this new framework, `21cmFAST` will remain an important component
of 21-cm cosmology for years to come.
# Examples
`21cmFAST` supports installation using `conda`, which means installation is as simple
as typing `conda install -c conda-forge 21cmFAST`. The following example can then
be run in a Python interpreter.
In-depth examples can be found in the official documentation.
As an example of the simplicity with which a full lightcone may be produced with
the new `21cmFAST` v3, the following may be run in a Python interpreter (or Jupyter
notebook):
```python
import py21cmfast as p21c
lightcone = p21c.run_lightcone(
redshift=6.0, # Minimum redshift of lightcone
max_redshift=30.0,
user_params={
"HII_DIM": 150, # N cells along side in output cube
"DIM": 400, # Original high-res cell number
"BOX_LEN": 300, # Size of the simulation in Mpc
},
flag_options={
"USE_TS_FLUCT": True, # Don't assume saturated spin temp
"INHOMO_RECO": True, # Use inhomogeneous recombinations
},
lightcone_quantities=( # Components to store as lightcones
"brightness_temp",
"xH_box",
"density"
),
global_quantities=( # Components to store as mean
"xH_box", # values per redshift
"brightness_temp"
),
)
# Save to a unique filename hashing all input parameters
lightcone.save()
# Make a lightcone sliceplot
p21c.plotting.lightcone_sliceplot(lightcone, "brightness_temp")
```
{height=300px}
```python
# Plot a global quantity
p21c.plotting.plot_global_history(lightcone, "xH")
```
{height=300px}
# Performance
Despite being a Python code, `21cmFAST` v3 does not diminish the performance of previous
pure-C versions. It utilises `CFFI` to provide the interface to the C-code through
Python, which is managed by some custom Python classes that oversee the construction and
memory allocation of each C `struct`.
OpenMP parallelization is enabled within the C-code, providing excellent speed-up for
large simulations when performed on high-performance machines.
A simple performance comparison between v3 and v2.1 (the last pure-C version), running
a light-cone simulation over a redshift range between 35 and 5 (92 snapshots) with spin
temperature fluctuations (`USE_TS_FLUCT`), inhomogeneous recombinations
(`INHOMO_RECO`), FFTW Wisdoms (`USE_FFTW_WISDOM`) and interpolation tables
(`USE_INTERPOLATION_TABLES`),
with a resolution of `HII_DIM=250` cells, and `DIM=1000` cells for the initial conditions,
on an Intel(R) Xeon(R) CPU (E5-4657L v2 @ 2.40GHz) with 16 shared-memory cores, reveals
that a clock time of 7.63(12.63) hours and a maximum RAM of 224(105) gigabytes are needed
for v3(v2.1).
Note that while a full light-cone simulation can be expensive to perform,
it only takes 2-3min to calculate a Coeval box (excluding the initial conditions).
For instance, the aforementioned timing for v3 includes 80 minutes to generate the
initial condition, which also dominates the maximum RAM required, with an additional
~4 minutes per snapshot to calculate all required fields of perturbation, ionization,
spin temperature and brightness temperature.
To guide the user, we list some performance benchmarks for variations on this simulation,
run with `21cmFAST` v3.0.2. Note that these benchmarks are subject to change as new
minor versions are delivered; in particular, operational modes that reduce maximum
memory consumption are planned for the near future.
| Variation | Time (hr) | Memory (GB) |
| ----------------------------------------------- | --------- | ----------- |
| Reference | 7.63 | 224 |
| Single Core | 14.77 | 224 |
| 4 Shared-memory Cores | 7.42 | 224 |
| 64 Shared-memory Cores | 9.60 | 224 |
| Higher Resolution <br />(HII_DIM=500, DIM=2000) | 68.37 | 1790 |
| Lower Resolution <br />(HII_DIM=125, DIM=500) | 0.68 | 28 |
| No Spin Temperature | 4.50 | 224 |
| Use Mini-Halos | 11.57 | 233 |
| No FFTW Wisdoms | 7.33 | 224 |
At this time, the `21cmFAST` team suggests using 4 or fewer shared-memory cores.
However, it is worth noting that as performance does vary on different machines,
users are recommended to calculate their own scalability.
# Acknowledgements
This work was supported in part by the European Research Council
(ERC) under the European Union’s Horizon 2020 research
and innovation programme (AIDA – #638809). The results
presented here reflect the authors’ views; the ERC is not
responsible for their use. JBM was partially supported by NSF grant AST-1813694.
Parts of this research were supported by the European Research Council under ERC grant
number 638743-FIRSTDAWN. Parts of this research were supported by the Australian Research
Council Centre of Excellence for
All Sky Astrophysics in 3 Dimensions (ASTRO 3D), through project number CE170100013.
JP was supported in part by a KIAS individual Grant (PG078701) at Korea Institute for
Advanced Study.
# References
| 21cmFAST | /21cmFAST-3.3.1.tar.gz/21cmFAST-3.3.1/joss-paper/paper.md | paper.md |
MIT License
Copyright (c) 2022 [The_Robin_Hood]
Permission is hereby granted, free of charge, to any person obtaining a copy
of this software and associated documentation files (the "Software"), to deal
in the Software without restriction, including without limitation the rights
to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
copies of the Software, and to permit persons to whom the Software is
furnished to do so, subject to the following conditions:
The above copyright notice and this permission notice shall be included in all
copies or substantial portions of the Software.
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
SOFTWARE.
| 233-misc | /233_misc-0.0.3.tar.gz/233_misc-0.0.3/LICENSE.md | LICENSE.md |
import argparse
parser = argparse.ArgumentParser(description='Process some integers.')
parser.add_argument('--input', help='A 23andme data file', required=True)
parser.add_argument('--output', help='Output VCF file', required=True)
parser.add_argument('--fasta', help='An uncompressed reference genome GRCh37 fasta file', required=True)
parser.add_argument('--fai', help='The fasta index for for the reference', required=True)
def load_fai(args):
index = {}
with open(args.fai) as f:
for line in f:
toks = line.split('\t')
chrom = 'chr' + toks[0]
if chrom == 'chrMT':
chrom = 'chrM'
length = int(toks[1])
start = int(toks[2])
linebases = int(toks[3])
linewidth = int(toks[4])
index[chrom] = (start, length, linebases, linewidth)
return index
def get_vcf_records(pos_list, fai, args):
with open(args.fasta) as f:
def get_alts(ref, genotype):
for x in genotype:
assert x in 'ACGT'
if len(genotype) == 1:
if ref in genotype:
return []
return [genotype]
if ref == genotype[0] and ref == genotype[1]:
return []
if ref == genotype[0]:
return [genotype[1]]
if ref == genotype[1]:
return [genotype[0]]
return [genotype[0], genotype[1]]
for (rsid, chrom, pos, genotype) in pos_list:
start, _, linebases, linewidth = fai[chrom]
n_lines = int(pos / linebases)
n_bases = pos % linebases
n_bytes = start + n_lines * linewidth + n_bases
f.seek(n_bytes)
ref = f.read(1)
alts = get_alts(ref, genotype)
pos = str(pos + 1)
diploid = len(genotype) == 2
assert ref not in alts
assert len(alts) <= 2
if diploid:
if len(alts) == 2:
if alts[0] == alts[1]:
yield (chrom, pos, rsid, ref, alts[0], '.', '.', '.', 'GT', '1/1')
else:
yield (chrom, pos, rsid, ref, alts[0], '.', '.', '.', 'GT', '1/2')
yield (chrom, pos, rsid, ref, alts[1], '.', '.', '.', 'GT', '2/1')
elif len(alts) == 1:
yield (chrom, pos, rsid, ref, alts[0], '.', '.', '.', 'GT', '0/1')
elif len(alts) == 1:
yield (chrom, pos, rsid, ref, alts[0], '.', '.', '.', 'GT', '1')
def load_23andme_data(input):
with open(input) as f:
for line in f:
if line.startswith('#'): continue
if line.strip():
rsid, chrom, pos, genotype = line.strip().split('\t')
if chrom == 'MT':
chrom = 'M'
chrom = 'chr' + chrom
if genotype != '--':
skip = False
for x in genotype:
if x not in 'ACTG':
skip = True
if not skip:
yield rsid, chrom, int(pos) - 1, genotype # subtract one because positions are 1-based indices
def write_vcf_header(f):
f.write(
"""##fileformat=VCFv4.2
##source=23andme_to_vcf
##reference=GRCh37
##FORMAT=<ID=GT,Number=1,Type=String,Description="Genotype">
#CHROM POS ID REF ALT QUAL FILTER INFO FORMAT SAMPLE
""")
def write_vcf(outfile, records):
with open(outfile, 'w') as f:
write_vcf_header(f)
for record in records:
f.write('\t'.join(record) + '\n')
def main():
args = parser.parse_args()
fai = load_fai(args)
snps = load_23andme_data(args.input)
records = get_vcf_records(snps, fai, args)
write_vcf(args.output, records) | 23andme-to-vcf | /23andme_to_vcf-0.0.3-py3-none-any.whl/_23andme_to_vcf/driver.py | driver.py |
import argparse
parser = argparse.ArgumentParser(description='Process some integers.')
parser.add_argument('--input', help='A 23andme data file', required=True)
parser.add_argument('--output', help='Output VCF file', required=True)
def load_fai():
index = {}
with open('./data/GRCh37.fa.fai') as f:
for line in f:
toks = line.split('\t')
chrom = 'chr' + toks[0]
if chrom == 'chrMT':
chrom = 'chrM'
length = int(toks[1])
start = int(toks[2])
linebases = int(toks[3])
linewidth = int(toks[4])
index[chrom] = (start, length, linebases, linewidth)
return index
def get_vcf_records(pos_list, fai):
with open('./data/GRCh37.fa') as f:
def get_alts(ref, genotype):
for x in genotype:
assert x in 'ACTG'
if len(genotype) == 1:
if ref in genotype:
return []
return [genotype]
if ref == genotype[0] and ref == genotype[1]:
return []
if ref == genotype[0]:
return [genotype[1]]
if ref == genotype[1]:
return [genotype[0]]
return [genotype[0], genotype[1]]
for (rsid, chrom, pos, genotype) in pos_list:
start, _, linebases, linewidth = fai[chrom]
n_lines = int(pos / linebases)
n_bases = pos % linebases
n_bytes = start + n_lines * linewidth + n_bases
f.seek(n_bytes)
ref = f.read(1)
alts = get_alts(ref, genotype)
pos = str(pos + 1)
diploid = len(genotype) == 2
assert ref not in alts
assert len(alts) <= 2
if diploid:
if len(alts) == 2:
if alts[0] == alts[1]:
yield (chrom, pos, rsid, ref, alts[0], '.', '.', '.', 'GT', '1/1')
else:
yield (chrom, pos, rsid, ref, alts[0], '.', '.', '.', 'GT', '1/2')
yield (chrom, pos, rsid, ref, alts[1], '.', '.', '.', 'GT', '2/1')
elif len(alts) == 1:
yield (chrom, pos, rsid, ref, alts[0], '.', '.', '.', 'GT', '0/1')
elif len(alts) == 1:
yield (chrom, pos, rsid, ref, alts[0], '.', '.', '.', 'GT', '1')
def load_23andme_data(input):
with open(input) as f:
for line in f:
if line.startswith('#'): continue
if line.strip():
rsid, chrom, pos, genotype = line.strip().split('\t')
if chrom == 'MT':
chrom = 'M'
chrom = 'chr' + chrom
if genotype != '--':
skip = False
for x in genotype:
if x not in 'ACTG':
skip = True
if not skip:
yield rsid, chrom, int(pos) - 1, genotype # subtract one because positions are 1-based indices
def write_vcf_header(f):
f.write(
"""
##fileformat=VCFv4.2
##source=23andme_to_vcf
##reference=GRCh37
##FORMAT=<ID=GT,Number=1,Type=String,Description="Genotype">
#CHROM POS ID REF ALT QUAL FILTER INFO FORMAT SAMPLE
""")
def write_vcf(outfile, records):
with open(outfile, 'w') as f:
write_vcf_header(f)
for record in records:
f.write('\t'.join(record) + '\n')
def main():
args = parser.parse_args()
fai = load_fai()
snps = load_23andme_data(args.input)
records = get_vcf_records(snps, fai)
write_vcf(args.output, records) | 23andme-to-vcf | /23andme_to_vcf-0.0.3-py3-none-any.whl/23andme_to_vcf/driver.py | driver.py |
import os
import sys
import time
import signal
import math
import random
import itertools
import csv
def calc(a, o, b):
P = {'+': 1, '-': 1, '*': 2, '/': 2}
ea, eb = a['e'], b['e']
va, vb = eval(a['e']), eval(b['e'])
c = a['c'] + b['c']
if 'o' in a and P[a['o']] < P[o]:
ea = '(' + a['e'] + ')'
if 'o' in b and P[o] > P[b['o']]:
eb = '(' + b['e'] + ')'
if 'o' in b and P[o] == P[b['o']]:
if (o == '+' and b['o'] == '+') or (o == '*' and b['o'] == '*'):
pass
else:
eb = '(' + b['e'] + ')'
e = ea + o + eb
if o == '+':
c += 1
if va + vb == 24:
pass
else: #和越大越复杂
t = va + vb
if t <= 10:
pass
else:
c += t*1.2
elif o == '-':
c += 1
t = va - vb
if t < 0: #减出负数
c += 100
elif o == '*':
c += 1
if va == 1 and vb == 1: #与1相乘不增加复杂度
pass
elif va * vb == 24:
pass
else: #积越大越复杂
t = va * vb
if t <= 10:
pass
else:
c += t
elif o == '/':
c += 2
if va == vb: #相同数相除不增加复杂度
pass
else:
c += 3
y = math.modf(va / vb)[0]
if y > 0: # 商为小数
c += 800
if math.modf(y * 100000)[0] > 0: #无理数
c += 800
return {'r': eval(e), 'e': e, 'o': o, 'c': int(c)}
def get1(a):
return [{'r': a, 'e': str(a), 'c': 0}]
def get2(a, b):
A = a if isinstance(a, list) else get1(a)
B = b if isinstance(b, list) else get1(b)
z = [ ]
for a in A:
for b in B:
if a['c'] < b['c']:
z.append(calc(b, '+', a))
z.append(calc(b, '*', a))
else:
z.append(calc(a, '+', b))
z.append(calc(a, '*', b))
z.append(calc(a, '-', b))
z.append(calc(b, '-', a))
if b['r'] != 0: z.append(calc(a, '/', b))
if a['r'] != 0: z.append(calc(b, '/', a))
return z
def get3(a, b, c):
return get2(a, get2(b, c)) + \
get2(b, get2(a, c)) + \
get2(c, get2(a, b))
def get4(a, b, c, d):
return get2(get2(a, b), get2(c, d)) + \
get2(get2(a, c), get2(b, d)) + \
get2(get2(a, d), get2(b, c)) + \
get2(a, get3(b, c, d)) + \
get2(b, get3(a, c, d)) + \
get2(c, get3(a, b, d)) + \
get2(d, get3(a, b, c))
def fx(target, values):
n = len(values)
if n == 4:
z = get4(values[0], values[1], values[2], values[3])
elif n == 3:
z = get3(values[0], values[1], values[2])
return list(filter(lambda x: math.isclose(x['r'], target, rel_tol=1e-10), z))
def fy(z, complex_fold=True):
z.sort(key=lambda x: x['c'])
e = []
c = []
d = []
for a in z:
if a['e'] not in e:
if complex_fold and a['c'] in c: continue
e.append(a['e'])
c.append(a['c'])
d.append({'e': str(round(a['r'])) + '=' + a['e'], 'c': a['c']})
return d
################################
C = [1, 2, 3, 4, 5, 6, 7, 8, 9, 10] * 4
def All():
D = list(itertools.combinations(C,4))
S = []
for v in D:
u = [v[0], v[1], v[2], v[3]]
u.sort()
if u not in S:
S.append(u)
return S
def sigint_handler(signum, frame):
print ('catched interrupt signal!')
exit()
status = ''
start = 0
result = []
def new_question(space):
global status, start, result
v, b = [], []
while True:
v = list(random.choice(space))
b = fx(24, v)
if len(b) > 0: break
os.system('clear')
print('')
print('\u2661', '\u2664', ' {} {} {} {} '.format(v[0], v[1], v[2], v[3]), '\u2667', '\u2662')
status = "new question"
start = int(time.time())
return b
def show_answer(d, n, f, delta):
global status, start, result
r = fy(d, f)
if n == 1:
print(r[0]["e"])
print("")
print("complexity", r[0]["c"])
print("total", len(r), "solutions")
print("cost", delta, "seconds")
else:
for i, a in enumerate(r):
if i < n:
print('{0:<16}{1}'.format(a['e'], a['c']))
status = "show answer"
def main():
global status, start, result
if len(sys.argv) >= 2:
if sys.argv[1] in ['-h', '-H', '--help']:
print('24p 1 2 3 4 # for query')
print('24p # for play game')
return
signal.signal(signal.SIGINT, sigint_handler)
signal.signal(signal.SIGHUP, sigint_handler)
signal.signal(signal.SIGTERM, sigint_handler)
v = list(map(lambda x: int(x), sys.argv[1:]))
n = len(v)
if n == 0:
space = list(itertools.combinations(C,4))
result = new_question(space)
while True:
Q = input('')
if Q == 'q': exit()
if status == "new question":
delta = int(time.time()) - start
if Q == "":
show_answer(result, 1, True, delta)
else:
if Q == "a": show_answer(result, 99999, False, delta)
elif Q == "b": show_answer(result, 99999, True, delta)
elif Q == "n": show_answer(result, 99999, True, delta)
print("cost ", delta, " seconds")
elif status == "show answer":
if Q == "a": show_answer(result, 99999, False, 0)
elif Q == "b": show_answer(result, 99999, True, 0)
else: result = new_question(space)
if n == 1:
Q = int(sys.argv[1])
S = All()
for i, v in enumerate(S):
b = fx(24, v)
b.sort(key=lambda x: x['c'])
if len(b) > 0:
a = b[0]
if Q < a['c']:
print('{}%'.format(int(i/len(S) * 100)), v, '{0:>4}={1:<16}{2}'.format(int(a['r']), a['e'], a['c']))
if n == 2:
Q1 = int(sys.argv[1])
Q2 = int(sys.argv[2])
S = All()
for i, v in enumerate(S):
b = fx(24, v)
b.sort(key=lambda x: x['c'])
if len(b) > 0:
a = b[0]
if Q1 < a['c'] and a['c'] < Q2:
print('{}%'.format(int(i/len(S) * 100)), v, '{0:>4}={1:<16}{2}'.format(int(a['r']), a['e'], a['c']))
if n == 3:
for a in ['3', '4', '6', '8', '12', '1/3', '1/4', '1/6', '1/8', '24', '48', '72', '96']:
r = eval('1.0*' + a)
b = fx(r, v)
b.sort(key=lambda x: x['c'])
if len(b) > 0:
print(a +'='+ b[0]['e'])
if n == 4:
show_answer(fx(24, v), 99999, False, 0) | 24p | /24p-0.2.0-py3-none-any.whl/p24/p24.py | p24.py |
import math
import matplotlib.pyplot as plt
from .Generaldistribution import Distribution
class Gaussian(Distribution):
""" Gaussian distribution class for calculating and
visualizing a Gaussian distribution.
Attributes:
mean (float) representing the mean value of the distribution
stdev (float) representing the standard deviation of the distribution
data_list (list of floats) a list of floats extracted from the data file
"""
def __init__(self, mu=0, sigma=1):
Distribution.__init__(self, mu, sigma)
def calculate_mean(self):
"""Function to calculate the mean of the data set.
Args:
None
Returns:
float: mean of the data set
"""
avg = 1.0 * sum(self.data) / len(self.data)
self.mean = avg
return self.mean
def calculate_stdev(self, sample=True):
"""Function to calculate the standard deviation of the data set.
Args:
sample (bool): whether the data represents a sample or population
Returns:
float: standard deviation of the data set
"""
if sample:
n = len(self.data) - 1
else:
n = len(self.data)
mean = self.calculate_mean()
sigma = 0
for d in self.data:
sigma += (d - mean) ** 2
sigma = math.sqrt(sigma / n)
self.stdev = sigma
return self.stdev
def plot_histogram(self):
"""Function to output a histogram of the instance variable data using
matplotlib pyplot library.
Args:
None
Returns:
None
"""
plt.hist(self.data)
plt.title('Histogram of Data')
plt.xlabel('data')
plt.ylabel('count')
def pdf(self, x):
"""Probability density function calculator for the gaussian distribution.
Args:
x (float): point for calculating the probability density function
Returns:
float: probability density function output
"""
return (1.0 / (self.stdev * math.sqrt(2*math.pi))) * math.exp(-0.5*((x - self.mean) / self.stdev) ** 2)
def plot_histogram_pdf(self, n_spaces = 50):
"""Function to plot the normalized histogram of the data and a plot of the
probability density function along the same range
Args:
n_spaces (int): number of data points
Returns:
list: x values for the pdf plot
list: y values for the pdf plot
"""
mu = self.mean
sigma = self.stdev
min_range = min(self.data)
max_range = max(self.data)
# calculates the interval between x values
interval = 1.0 * (max_range - min_range) / n_spaces
x = []
y = []
# calculate the x values to visualize
for i in range(n_spaces):
tmp = min_range + interval*i
x.append(tmp)
y.append(self.pdf(tmp))
# make the plots
fig, axes = plt.subplots(2,sharex=True)
fig.subplots_adjust(hspace=.5)
axes[0].hist(self.data, density=True)
axes[0].set_title('Normed Histogram of Data')
axes[0].set_ylabel('Density')
axes[1].plot(x, y)
axes[1].set_title('Normal Distribution for \n Sample Mean and Sample Standard Deviation')
axes[0].set_ylabel('Density')
plt.show()
return x, y
def __add__(self, other):
"""Function to add together two Gaussian distributions
Args:
other (Gaussian): Gaussian instance
Returns:
Gaussian: Gaussian distribution
"""
result = Gaussian()
result.mean = self.mean + other.mean
result.stdev = math.sqrt(self.stdev ** 2 + other.stdev ** 2)
return result
def __repr__(self):
"""Function to output the characteristics of the Gaussian instance
Args:
None
Returns:
string: characteristics of the Gaussian
"""
return "mean {}, standard deviation {}".format(self.mean, self.stdev) | 29082022-distributions | /29082022_distributions-0.1.tar.gz/29082022_distributions-0.1/29082022_distributions/Gaussiandistribution.py | Gaussiandistribution.py |
import math
import matplotlib.pyplot as plt
from .Generaldistribution import Distribution
class Binomial(Distribution):
""" Binomial distribution class for calculating and
visualizing a Binomial distribution.
Attributes:
mean (float) representing the mean value of the distribution
stdev (float) representing the standard deviation of the distribution
data_list (list of floats) a list of floats to be extracted from the data file
p (float) representing the probability of an event occurring
n (int) number of trials
TODO: Fill out all functions below
"""
def __init__(self, prob=.5, size=20):
self.n = size
self.p = prob
Distribution.__init__(self, self.calculate_mean(), self.calculate_stdev())
def calculate_mean(self):
"""Function to calculate the mean from p and n
Args:
None
Returns:
float: mean of the data set
"""
self.mean = self.p * self.n
return self.mean
def calculate_stdev(self):
"""Function to calculate the standard deviation from p and n.
Args:
None
Returns:
float: standard deviation of the data set
"""
self.stdev = math.sqrt(self.n * self.p * (1 - self.p))
return self.stdev
def replace_stats_with_data(self):
"""Function to calculate p and n from the data set
Args:
None
Returns:
float: the p value
float: the n value
"""
self.n = len(self.data)
self.p = 1.0 * sum(self.data) / len(self.data)
self.mean = self.calculate_mean()
self.stdev = self.calculate_stdev()
def plot_bar(self):
"""Function to output a histogram of the instance variable data using
matplotlib pyplot library.
Args:
None
Returns:
None
"""
plt.bar(x = ['0', '1'], height = [(1 - self.p) * self.n, self.p * self.n])
plt.title('Bar Chart of Data')
plt.xlabel('outcome')
plt.ylabel('count')
def pdf(self, k):
"""Probability density function calculator for the gaussian distribution.
Args:
x (float): point for calculating the probability density function
Returns:
float: probability density function output
"""
a = math.factorial(self.n) / (math.factorial(k) * (math.factorial(self.n - k)))
b = (self.p ** k) * (1 - self.p) ** (self.n - k)
return a * b
def plot_bar_pdf(self):
"""Function to plot the pdf of the binomial distribution
Args:
None
Returns:
list: x values for the pdf plot
list: y values for the pdf plot
"""
x = []
y = []
# calculate the x values to visualize
for i in range(self.n + 1):
x.append(i)
y.append(self.pdf(i))
# make the plots
plt.bar(x, y)
plt.title('Distribution of Outcomes')
plt.ylabel('Probability')
plt.xlabel('Outcome')
plt.show()
return x, y
def __add__(self, other):
"""Function to add together two Binomial distributions with equal p
Args:
other (Binomial): Binomial instance
Returns:
Binomial: Binomial distribution
"""
try:
assert self.p == other.p, 'p values are not equal'
except AssertionError as error:
raise
result = Binomial()
result.n = self.n + other.n
result.p = self.p
result.calculate_mean()
result.calculate_stdev()
return result
def __repr__(self):
"""Function to output the characteristics of the Binomial instance
Args:
None
Returns:
string: characteristics of the Gaussian
"""
return "mean {}, standard deviation {}, p {}, n {}".\
format(self.mean, self.stdev, self.p, self.n) | 29082022-distributions | /29082022_distributions-0.1.tar.gz/29082022_distributions-0.1/29082022_distributions/Binomialdistribution.py | Binomialdistribution.py |
# Background and Validation
This wiki is built in Notion. Here are all the tips you need to contribute.
# General Background
<img src="https://github.com/Fluidentity/2D_Panel-CFD-/blob/main/img/Map-1_Step-.jpg" alt="Flow over a cylinder" width="700"/>
Flow over a cylinder
---
> **The project has been started as a Open Source repository for CFD solvers. The motive is to provide handy easy to understand code with multitude of CFD schemes for cfd developers. Also, needs to remain functional as an easy to setup open source solver for users. This release only comprises of a terminal sequential prompt, simple and effective. We have immediate plans of implementing a PyQT GUI to it.**
>
> Head to the notion page for more information on how to add to this project:
>
[https://florentine-hero-1e6.notion.site/2D_Panel-CFD-ad63baa924ee4a32af8a52b8134c0360](https://www.notion.so/2D_Panel-CFD-ad63baa924ee4a32af8a52b8134c0360)
> **This version comprises of a 2D Staggered Grid with Inlet, Outlet & Wall Boundary conditions. Obstacles can be imported & transformed with a list of points or with the inbuilt elliptical geometries.**
>
> **First order Upwind Scheme is used for Velocity with very good results for the benchmark Lid Driven Cavity problem when compared to results in Ghia etal.**
>
> **The SIM runs stable with terminal-python for <10000 Cells after which Residual plotting becomes laggy. spyder (Anaconda IDE) provides great speed-ups with multi-core utilisation & also improves the post-processing experience. The Sequential prompts based model is based on a GUI approach and will be ported to it in the next update.**
>
> **The lack of multi-threading support in python trumps the ease of accessibility of matplotlib library. We will be looking to port into C++ immediately utilizing vtk libraries with paraview & blender for visualization.**
>
> **The framework is designed to test new FVM schemes, & Coupling solvers. All popular convection schemes will be added soon. Multiple solvers will be available in the next updates, the likes of SIMPLER, PISO, Pimple etc. Future plans also include Unsteady & VOF solvers.**
>
> **The program works as a sequential prompt, for SIM Parameters. The prompts are designed keeping in mind a GUI approach, which will be available in the next update. There are frequent Check Cycles to render the result & modify any inputs. We'll go through an exemplary First Run in the next Section.**
>
# Installation
### Method: 1
To install using pip Run:
```python
python3 -m pip install 2D_Panel-CFD
```
Or:
### Method: 2
[https://github.com/Fluidentity/2D_Panel-CFD](https://github.com/Fluidentity/2D_Panel-CFD)
- Clone github `[RUN_package](https://github.com/Fluidentity/2D_Panel-CFD.git)` to anywhere in your machine from:
```tsx
cd /insert/folder/address/cfd
git clone https://github.com/Fluidentity/2D_Panel-CFD.git
```
- Set it to PYTHONPATH with:
```python
export PYTHONPATH="${PYTHONPATH}:/insert/folder/address/cfd/RUN_package"
```
It's advisable to run this package from [RUN-spyder.py](http://RUN-spyder.py) through an IDE like spyder for ease of use, and prolonged variable storage. Also, spyder has some great plotting interface.
## Executable
<aside>
💡 The source directory should be set up as PYTHONPATH if not installed using pip
</aside>
### Method: 1
Open python environment with: (in terminal)
```python
python3
```
or (if python —version is >3)
```python
python
```
then insert:
```python
from RUN_package import RUN
```
- **RUN.py** is meant to be run from terminal.
### Method: 2
Run on IDE by cloning [RUN_package](https://github.com/Fluidentity/2D_Panel-CFD/tree/main/RUN_package) from Github.
> Open python IDE like spyder from RUN_package directory:
>
> Run RUN-spyder.py
>
The cells for pre-processor, solver & post processors are different. Need to run all.
- **RUN_spyder.py** can be run with an IDE, such as spyder to improve multi-Core Utilisation & post-processing experience. ****
# Validation of Solver
<img src="https://github.com/Fluidentity/2D_Panel-CFD-/blob/main/img/ezgif.com-gif-maker(3).gif" alt="Vortex Shedding flow over a cylinder" width="700"/>
Vortex Shedding flow over a cylinder
---
> For validation of the solver laid out, following strategies are used:
>
1. Comparison with Benchmark Problem Lid Driven Cavity
1. Reference study Ghia etal. Re = 100, 1000, 5000
## Lid Driven Cavity Benchmark Ghia etal.
### **Residuals**
<img src="https://github.com/Fluidentity/2D_Panel-CFD-/blob/main/img/Untitled.png" alt="Untitled" width="400"/>
### **Benchmark Test at Re=100**
- First Order Upwind scheme
<img src="https://github.com/Fluidentity/2D_Panel-CFD-/blob/main/img/Untitled%201.png" alt="Untitled" width="400"/>
<img src="https://github.com/Fluidentity/2D_Panel-CFD-/blob/main/img/Untitled%202.png" alt="Untitled" width="400"/>
### **Benchmark Test at Re=1000**
- First Order Upwind scheme
<img src="https://github.com/Fluidentity/2D_Panel-CFD-/blob/main/img/Untitled%203.png" alt="Untitled" width="400"/>
<img src="https://github.com/Fluidentity/2D_Panel-CFD-/blob/main/img/Untitled%204.png" alt="Untitled" width="400"/>
### **Benchmark Test at Re=5000**
- First Order Upwind scheme
<img src="https://github.com/Fluidentity/2D_Panel-CFD-/blob/main/img/Untitled%205.png" alt="Untitled" width="400"/>
<img src="https://github.com/Fluidentity/2D_Panel-CFD-/blob/main/img/Untitled%206.png" alt="Untitled" width="400"/>
### Conclusion
First order UPWIND Scheme is good for low Reynolds no. but is only first order accurate to capture higher gradient.
## Fully developed flow between Parallel Plates
### Velocity Profile [at X=0.8*Lx and Y=0.5*Ly]
<img src="https://github.com/Fluidentity/2D_Panel-CFD-/blob/main/img/Untitled%207.png" alt="Untitled" width="400"/>
<img src="https://github.com/Fluidentity/2D_Panel-CFD-/blob/main/img/Untitled%208.png" alt="Untitled" width="400"/>
<img src="https://github.com/Fluidentity/2D_Panel-CFD-/blob/main/img/Map-1_Step-200.jpg" alt="Map-1 Step-[200].jpg" width="1500"/>
### Conclusion
The Umax Velocity comes close to 1.5 feactor for steady flow between parallel plates. First order UPWIND Scheme with high y-gradient.
| 2D-Panel-CFD | /2D_Panel-CFD-0.0.1.tar.gz/2D_Panel-CFD-0.0.1/README.md | README.md |
import numpy as np
import matplotlib
from matplotlib import pyplot as plt
import matplotlib.animation as animation
matplotlib.rcParams.update({'font.size': 15})
def postProcessor(P_values, uxq_values, uyq_values, xMin, xMax, yMin, yMax, nx, ny, dx, dy, Points_x, Points_y, caseFolder, Res_U_values, Res_V_values, Res_Mass_values, max_iter, min_Res, max_Res):
('plasma')
flag_generateResults = int(input("\n1. Generate Results\n2. Abort\n"))
flag_newMap = 'y'
flag_newMapOK = 2
if(flag_generateResults==1):
flag_generateResidualsPlot = input("Generate Residuals plot ?? [y/n]")
if(flag_generateResidualsPlot=='y'):
fig, ax = fig, ax = plt.subplots(figsize =(10, 5))
iter_ = np.linspace(0, max_iter, max_iter)
ax.plot(iter_, Res_U_values, color = 'red', label = 'Ux-Mom Residual')
ax.plot(iter_, Res_V_values, color = 'blue', label = 'Uy-Mom Residual')
ax.plot(iter_, Res_Mass_values, color = 'green', label = 'Mass Residual')
ax.set_ylim(min_Res, max_Res)
ax.set_xlim(0, max_iter)
ax.set_xlabel('Iterations')
ax.legend()
plt.savefig(caseFolder+"/Residuals.jpg")
flag_generateCSV = int(input("\n[1]- Generate CSV files of Residuals \n[2]- Generate CSV files of U, V and P\n[3]- No CSV Results\n"))
if(flag_generateCSV==1):
np.savetxt(caseFolder+"/U_Residual.csv", Res_U_values, delimiter=",")
np.savetxt(caseFolder+"/V-Residual.csv", Res_V_values, delimiter=",")
np.savetxt(caseFolder+"/Mass_Residual.csv", Res_Mass_values, delimiter=",")
elif(flag_generateCSV==2):
np.savetxt(caseFolder+"/U.csv", uxq_values[max_iter-1], delimiter=",")
np.savetxt(caseFolder+"/V.csv", uyq_values[max_iter-1], delimiter=",")
np.savetxt(caseFolder+"/P.csv", P_values[max_iter], delimiter=",")
mapCounter = 1
matplotlib.rcParams.update({'font.size': 35})
while(flag_newMap=='y'):
print("Generating Map-", mapCounter)
flag_resultColorMap = input("Add ColorMap ?? [y/n]")
flag_resultGeometry = input("Add Black Obstacle Geometry ?? [y/n]")
if(flag_resultColorMap=='y'):
flag_resultColorMapMarker = int(input("\nSelect Variable\n[1]- Pressure\n[2]- Velocity\n[3]- X-Velocity\n[4]- Y-Velocity\n"))
flag_resultColorMapContourlines = input("\nWith Labelled Contourlines ?? [y/n]\n")
if(flag_resultColorMapContourlines=='y'):
ContourlinesDensity = int(20/int(input("\nDensity of Contourlines ??\n")))
min_resultColorMap = float(input("Min value of variable"))
max_resultColorMap = float(input("Max value of variable"))
flag_resultStreamlines = input("Add Streamlines ?? [y/n]")
if(flag_resultStreamlines=='y'):
flag_resultStreamlinesDensity = int(input("Streamlines Density ??"))
flag_resultVectors = input("Add Vectors ?? [y/n]")
xq = np.linspace(xMin, xMax, num=2*nx+1)
yq = np.linspace(yMax, yMin, num=2*ny+1)
xqs = np.linspace(xMin, xMax, num=2*nx+1)
yqs = np.linspace(yMin, yMax, num=2*ny+1)
xp = np.linspace(xMin+dx/2, xMax-dx/2, num=nx)
yp = np.linspace(yMax-dy/2, yMin+dy/2, num=ny)
Xq, Yq = np.meshgrid(xqs, yqs)
# creating plot
Lx = xMax-xMin
Ly = yMax-yMin
fig, ax = plt.subplots(figsize =(Lx*10, Ly*10))
if(flag_resultGeometry=='y'):
ax.fill(Points_x, Points_y, color='black')
if(flag_resultColorMap=='y'):
levels = np.linspace(min_resultColorMap, max_resultColorMap, num=20)
if(flag_resultColorMapMarker==1):
contf = ax.contourf(xp, yp, P_values[max_iter-1], levels=levels, cmap=plt.cm.plasma)
if(flag_resultColorMapContourlines=='y'):
cont = ax.contour(contf, levels=contf.levels[::ContourlinesDensity], colors='r')
# ax.clabel(cont, colors='w', fontsize=14)
cbar = fig.colorbar(contf)
cbar.set_label('P')
elif(flag_resultColorMapMarker==2):
contf = ax.contourf(xq, yq, np.sqrt(uxq_values[max_iter-1]**2+uyq_values[max_iter-1]**2), levels=levels, cmap=plt.cm.plasma)
if(flag_resultColorMapContourlines=='y'):
cont = ax.contour(contf, levels=contf.levels[::ContourlinesDensity], colors='r')
# ax.clabel(cont, colors='w', fontsize=14)
cbar = fig.colorbar(contf)
cbar.set_label('Vel')
elif(flag_resultColorMapMarker==3):
contf = ax.contourf(xq, yq, uxq_values[max_iter-1], levels=levels, cmap=plt.cm.plasma)
if(flag_resultColorMapContourlines=='y'):
cont = ax.contour(contf, levels=contf.levels[::ContourlinesDensity], colors='r')
# ax.clabel(cont, colors='w', fontsize=14)
cbar = fig.colorbar(contf)
cbar.set_label('Vel_x')
elif(flag_resultColorMapMarker==4):
contf = ax.contourf(xq, yq, uyq_values[max_iter-1], levels=levels, cmap=plt.cm.plasma)
if(flag_resultColorMapContourlines=='y'):
cont = ax.contour(contf, levels=contf.levels[::ContourlinesDensity], colors='r')
# ax.clabel(cont, colors='w', fontsize=14)
cbar = fig.colorbar(contf)
cbar.set_label('Vel_y')
if(flag_resultStreamlines=='y'):
ax.streamplot(Xq, Yq, np.flip(uxq_values[max_iter-1], 0) ,np.flip(uyq_values[max_iter-1], 0), density=flag_resultStreamlinesDensity, color='black', linewidth=1)
ax.xaxis.set_ticks([])
ax.yaxis.set_ticks([])
ax.axis([xMin, xMax, yMin, yMax])
ax.set_aspect('equal')
elif(flag_resultVectors=='y'):
ax.quiver(Xq, Yq, np.flip(uxq_values[max_iter-1], 0) ,np.flip(uyq_values[max_iter-1], 0))
ax.xaxis.set_ticks([])
ax.yaxis.set_ticks([])
ax.axis([xMin, xMax, yMin, yMax])
ax.set_aspect('equal')
# show plot
plt.show()
flag_newMapOK = int(input("\n1. Continue with Map \n2. Generate again \n"))
if(flag_newMapOK==2):
continue
flag_plotall = int(input("\n0. Plot Last frame only [n] - [jpg]\n1. Plot every n iterations - [GIF]\n"))
if(flag_plotall!=0):
xq = np.linspace(xMin, xMax, num=2*nx+1)
yq = np.linspace(yMin, yMax, num=2*ny+1)
xp = np.linspace(xMin+dx/2, xMax-dx/2, num=nx)
yp = np.linspace(yMax-dy/2, yMin+dy/2, num=ny)
Xq, Yq = np.meshgrid(xq, yq)
Lx = xMax-xMin
Ly = yMax-yMin
fig, ax = plt.subplots(figsize =(Lx*10, Ly*10))
if(flag_resultColorMap=='y'):
levels = np.linspace(min_resultColorMap, max_resultColorMap, num=20)
if(flag_resultColorMapMarker==1):
contf = ax.contourf(xp, yp, P_values[0], levels=levels, cmap=plt.cm.plasma)
cbar = fig.colorbar(contf)
cbar.set_label('P')
elif(flag_resultColorMapMarker==2):
contf = ax.contourf(xq, yq, np.sqrt(uxq_values[0]**2+uyq_values[0]**2), levels=levels, cmap=plt.cm.plasma)
cbar = fig.colorbar(contf)
cbar.set_label('Vel')
elif(flag_resultColorMapMarker==3):
contf = ax.contourf(xq, yq, uxq_values[0], levels=levels, cmap=plt.cm.plasma)
cbar = fig.colorbar(contf)
cbar.set_label('Vel_x')
elif(flag_resultColorMapMarker==4):
contf = ax.contourf(xq, yq, uyq_values[0], levels=levels, cmap=plt.cm.plasma)
cbar = fig.colorbar(contf)
cbar.set_label('Vel_y')
def animate(i):
ax.clear()
if(flag_resultGeometry=='y'):
ax.fill(Points_x, Points_y, color='black')
if(flag_resultColorMap=='y'):
levels = np.linspace(min_resultColorMap, max_resultColorMap, num=20)
if(flag_resultColorMapMarker==1):
contf = ax.contourf(xp, yp, P_values[i], levels=levels, cmap=plt.cm.plasma)
if(flag_resultColorMapContourlines=='y'):
cont = ax.contour(contf, levels=contf.levels[::ContourlinesDensity], colors='r')
# ax.clabel(cont, colors='w', fontsize=14)
# cbar = fig.colorbar(contf)
# cbar.set_label('P')
elif(flag_resultColorMapMarker==2):
contf = ax.contourf(xq, yq, np.sqrt(uxq_values[i]**2+uyq_values[i]**2), levels=levels, cmap=plt.cm.plasma)
if(flag_resultColorMapContourlines=='y'):
cont = ax.contour(contf, levels=contf.levels[::ContourlinesDensity], colors='r')
# ax.clabel(cont, colors='w', fontsize=14)
# cbar = fig.colorbar(contf)
# cbar.set_label('Vel')
elif(flag_resultColorMapMarker==3):
contf = ax.contourf(xq, yq, uxq_values[i], levels=levels, cmap=plt.cm.plasma)
if(flag_resultColorMapContourlines=='y'):
cont = ax.contour(contf, levels=contf.levels[::ContourlinesDensity], colors='r')
# ax.clabel(cont, colors='w', fontsize=14)
# cbar = fig.colorbar(contf)
# cbar.set_label('Vel_x')
elif(flag_resultColorMapMarker==4):
contf = ax.contourf(xq, yq, uyq_values[i], levels=levels, cmap=plt.cm.plasma)
if(flag_resultColorMapContourlines=='y'):
cont = ax.contour(contf, levels=contf.levels[::ContourlinesDensity], colors='r')
# ax.clabel(cont, colors='w', fontsize=14)
# cbar = fig.colorbar(contf)
# cbar.set_label('Vel_y')
if(flag_resultStreamlines=='y'):
ax.streamplot(Xq, Yq, np.flip(uxq_values[i], 0) ,np.flip(uyq_values[i], 0), density=flag_resultStreamlinesDensity, color='black', linewidth=1)
elif(flag_resultVectors=='y'):
ax.quiver(Xq, Yq, np.flip(uxq_values[i], 0) ,np.flip(uyq_values[i], 0))
ax.set_title('Map-{} Step-[{}]'.format(mapCounter, i))
interval = 0.1
ani = animation.FuncAnimation(fig ,animate, max_iter, interval=interval*1e+3,blit=False)
writergif = animation.PillowWriter(fps=10)
ani.save(caseFolder+"/Map-{}.gif".format(mapCounter), writer = writergif)
if(flag_plotall==0):
fig.savefig(caseFolder+"/Map-{} Step-[{}].jpg".format(mapCounter, max_iter))
flag_newMap = input("Generate another Map ?? [y/n]")
mapCounter+=1 | 2D-Panel-CFD | /2D_Panel-CFD-0.0.1.tar.gz/2D_Panel-CFD-0.0.1/src/RUN_package/postProcessing.py | postProcessing.py |
import numpy as np
from matplotlib import pyplot as plt
import os
def domain(caseFolder):
# Domain input
xMin = float(input("\nEnter X_min value\n"))
xMax = float(input("\nEnter X_max value\n"))
dx_temp = float(input("\nEnter Cell size in X-axis [dx]\n"))
nx = int((xMax-xMin)/dx_temp)
dx = (xMax-xMin)/nx
print("\n--- No. of elements in X-axis [{}] ---\n".format(nx))
yMin = float(input("\nEnter Y_min value\n"))
yMax = float(input("\nEnter Y_max value\n"))
dy_temp = float(input("\nEnter Cell size in Y-axis [dy]\n"))
ny = int((yMax-yMin)/dy_temp)
dy = (yMax-yMin)/ny
print("\n--- No. of elements in Y-axis [{}] ---\n".format(ny))
temp = "/boundary.txt"
boundary_info = open(caseFolder+temp, 'w')
# Boundary revealing Matrix
B1 = 2*np.ones((ny+1, nx+1))
# Scalar Matrix [U, V, P]
B2 = -1000.5*np.ones((ny+1, nx+1))
# Domain Matrix
D = []
D.append(B1)
D.append(B2)
D.append(B2.copy())
D.append(B2.copy())
# Boundary Definition
flag_Boundary = 0
flag_SouthBoundary = 0
flag_NorthBoundary = 0
flag_WestBoundary = 0
flag_EastBoundary = 0
flag_Outlet = 0
# Boundary Defining Loop
while (flag_Boundary < 1):
# Southern Boundary Defining Loop
while (flag_SouthBoundary < 1):
flag_Inlet_SouthBoundary = 0
flag_Outlet_SouthBoundary = 0
flag_SlidingWall_SouthBoundary = 0
flag = input("\nInlet in South Boundary ?? [y/n]\n")
if(flag=='y'):
flag_Inlet_SouthBoundary = 1
flag = input("\nSliding Wall in South Boundary ?? [y/n]\n")
if(flag=='y'):
flag_SlidingWall_SouthBoundary = 1
flag = input("\nOutlet in South Boundary ?? [y/n]\n")
if(flag=='y'):
flag_Outlet_SouthBoundary = 1
# Initialising Boundary Markers
south_Boundary = np.zeros((3, 6))
south_Boundary[1, 0] = 1
south_Boundary[2, 0] = -1
# Temporary Wall
south_Boundary[0, 1] = xMin
south_Boundary[0, 2] = xMax
D[0][-1, :] = 0
D[1][-1, :] = 0
D[2][-1, :] = 0
# South Boundary Sliding U-velocity
if(flag_SlidingWall_SouthBoundary==1):
south_Boundary[0, 3] = float(input("\nEnter U-Velocity for Sliding Wall in South Boundary\n"))
D[1][-1, 1:nx] = south_Boundary[0, 3]
# Inlet definition in South Boundary
if(flag_Inlet_SouthBoundary==1):
south_Boundary[1, 1] = float(input("\nEnter xmin for Inlet in South Boundary\n"))
southInlet_min_index = int((south_Boundary[1, 1]-xMin)/dx)
if(southInlet_min_index==0):
southInlet_min_index+=1
south_Boundary[1, 2] = float(input("\nEnter xmax for Inlet in South Boundary\n"))
southInlet_max_index = int((south_Boundary[1, 2]-xMin)/dx)
if(southInlet_max_index==nx):
southInlet_max_index-=1
south_Boundary[1, 4] = float(input("\nEnter V-Velocity for Inlet in South Boundary\n"))
# Boundary Marker
D[0][-1, southInlet_min_index:southInlet_max_index+1] = 1.0
# U-vel - Initialised to 0
D[1][-1, southInlet_min_index:southInlet_max_index+1] = 0
# V-vel - User Initialised
D[2][-1, southInlet_min_index:southInlet_max_index+1] = south_Boundary[1, 4]
# Outlet definition in South Boundary
if(flag_Outlet_SouthBoundary==1):
flag_Outlet = 1
south_Boundary[2, 1] = float(input("\nEnter xmin for Outlet in South Boundary\n"))
southOutlet_min_index = int((south_Boundary[2, 1]-xMin)/dx)
if(southOutlet_min_index==0):
southOutlet_min_index+=1
south_Boundary[2, 2] = float(input("\nEnter xmax for Outlet in South Boundary\n"))
southOutlet_max_index = int((south_Boundary[2, 2]-xMin)/dx)
if(southOutlet_max_index==nx):
southOutlet_max_index-=1
south_Boundary[2, 5] = float(input("\nEnter Pressure for Outlet in South Boundary\n"))
# Boundary Marker
D[0][-1, southOutlet_min_index:southOutlet_max_index+1] = -1.0
# U-vel - Uninitialised
D[1][-1, southOutlet_min_index:southOutlet_max_index+1] = -1000.5
# V-vel - Uninitialised
D[2][-1, southOutlet_min_index:southOutlet_max_index+1] = -1000.5
# Pressure - User Initialised
D[3][-1, southOutlet_min_index:southOutlet_max_index+1] = south_Boundary[2, 5]
print("\n\n------------------SOUTH BOUNDARY------------------\n")
boundary_info.write("\n\n------------------SOUTH BOUNDARY------------------\n")
ite=0
while ite<nx+1:
if(D[0][-1, ite]==0):
ite1=ite+1
if(ite1==nx+1 and D[0][-1, ite-1]!=0):
print("\n[Wall]\n")
print("\nxMin = [{}] Starting Index = [{}]\n ".format(xMin+ite*dx, ite))
print("\nxMax = [{}] Ending Index = [{}]\n ".format(xMin+(ite1-1)*dx, ite1-1))
print("\nNo. of elements = [{}] \n".format(ite1-1-ite+1))
boundary_info.write("\n[Wall]\n")
boundary_info.write("\nxMin = [{}] Starting Index = [{}]\n ".format(xMin+ite*dx, ite))
boundary_info.write("\nxMax = [{}] Ending Index = [{}]\n ".format(xMin+(ite1-1)*dx, ite1-1))
boundary_info.write("\nNo. of elements = [{}] \n".format(ite1-1-ite+1))
while ite1<nx+1:
if(D[0][-1, ite1]!=0):
print("\n[Wall]\n")
print("\nxMin = [{}] Starting Index = [{}]\n ".format(xMin+ite*dx, ite))
print("\nxMax = [{}] Ending Index = [{}]\n ".format(xMin+(ite1-1)*dx, ite1-1))
print("\nNo. of elements = [{}] \n".format(ite1-1-ite+1))
boundary_info.write("\n[Wall]\n")
boundary_info.write("\nxMin = [{}] Starting Index = [{}]\n ".format(xMin+ite*dx, ite))
boundary_info.write("\nxMax = [{}] Ending Index = [{}]\n ".format(xMin+(ite1-1)*dx, ite1-1))
boundary_info.write("\nNo. of elements = [{}] \n".format(ite1-1-ite+1))
if(flag_SlidingWall_SouthBoundary==1):
print("\nSliding Wall Velocity {}\n".format(D[1][-1, ite1-1]))
boundary_info.write("\nSliding Wall Velocity {}\n".format(D[1][-1, ite1-1]))
ite=ite1-1
break
if(ite1==nx):
print("\n[Wall]\n")
print("\nxMin = [{}] Starting Index = [{}]\n ".format(xMin+ite*dx, ite))
print("\nxMax = [{}] Ending Index = [{}]\n ".format(xMin+(ite1)*dx, ite1))
print("\nNo. of elements = [{}] \n".format(ite1-ite+1))
boundary_info.write("\n[Wall]\n")
boundary_info.write("\nxMin = [{}] Starting Index = [{}]\n ".format(xMin+ite*dx, ite))
boundary_info.write("\nxMax = [{}] Ending Index = [{}]\n ".format(xMin+(ite1)*dx, ite1))
boundary_info.write("\nNo. of elements = [{}] \n".format(ite1-ite+1))
if(flag_SlidingWall_SouthBoundary==1):
print("\nSliding Wall Velocity {}\n".format(D[1][-1, ite1-1]))
boundary_info.write("\nSliding Wall Velocity {}\n".format(D[1][-1, ite1-1]))
ite=ite1-1
break
ite1=ite1+1
if(D[0][-1, ite]==1):
ite1 = ite+1
while ite1<nx+1:
if(D[0][-1, ite1]!=1):
print("\n[Inlet]\n")
print("\nxMin = [{}] Starting Index = [{}]\n ".format(xMin+ite*dx, ite))
print("\nxMax = [{}] Ending Index = [{}]\n ".format(xMin+(ite1-1)*dx, ite1-1))
print("\nNo. of elements = [{}] \n".format(ite1-1-ite+1))
print("\nInlet V-Velocity {}\n".format(D[2][-1, ite1-1]))
boundary_info.write("\n[Inlet]\n")
boundary_info.write("\nxMin = [{}] Starting Index = [{}]\n ".format(xMin+ite*dx, ite))
boundary_info.write("\nxMax = [{}] Ending Index = [{}]\n ".format(xMin+(ite1-1)*dx, ite1-1))
boundary_info.write("\nNo. of elements = [{}] \n".format(ite1-1-ite+1))
boundary_info.write("\nInlet V-Velocity {}\n".format(D[2][-1, ite1-1]))
ite=ite1-1
break
if(ite1==nx):
print("\n[Inlet]\n")
print("\nxMin = [{}] Starting Index = [{}]\n ".format(xMin+ite*dx, ite))
print("\nxMax = [{}] Ending Index = [{}]\n ".format(xMin+(ite1)*dx, ite1))
print("\nNo. of elements = [{}] \n".format(ite1-ite+1))
print("\nInlet V-Velocity {}\n".format(D[2][-1, ite1-1]))
boundary_info.write("\n[Inlet]\n")
boundary_info.write("\nxMin = [{}] Starting Index = [{}]\n ".format(xMin+ite*dx, ite))
boundary_info.write("\nxMax = [{}] Ending Index = [{}]\n ".format(xMin+(ite1)*dx, ite1))
boundary_info.write("\nNo. of elements = [{}] \n".format(ite1-ite+1))
boundary_info.write("\nInlet V-Velocity {}\n".format(D[2][-1, ite1-1]))
ite=ite1-1
break
ite1=ite1+1
if(D[0][-1, ite]==-1):
ite1=ite+1
while ite1<nx+1:
if(D[0][-1, ite1]!=-1):
print("\n[Outlet]\n")
print("\nxMin = [{}] Starting Index = [{}]\n ".format(xMin+ite*dx, ite))
print("\nxMax = [{}] Ending Index = [{}]\n ".format(xMin+(ite1-1)*dx, ite1-1))
print("\nNo. of elements = [{}] \n".format(ite1-1-ite+1))
print("\nOutlet Pressure {}\n".format(D[3][-1, ite1-1]))
boundary_info.write("\n[Outlet]\n")
boundary_info.write("\nxMin = [{}] Starting Index = [{}]\n ".format(xMin+ite*dx, ite))
boundary_info.write("\nxMax = [{}] Ending Index = [{}]\n ".format(xMin+(ite1-1)*dx, ite1-1))
boundary_info.write("\nNo. of elements = [{}] \n".format(ite1-1-ite+1))
boundary_info.write("\nOutlet Pressure {}\n".format(D[3][-1, ite1-1]))
ite=ite1-1
break
if(ite1==nx):
print("\n[Outlet]\n")
print("\nxMin = [{}] Starting Index = [{}]\n ".format(xMin+ite*dx, ite))
print("\nxMax = [{}] Ending Index = [{}]\n ".format(xMin+(ite1)*dx, ite1))
print("\nNo. of elements = [{}] \n".format(ite1-ite+1))
print("\nOutlet Pressure {}\n".format(D[3][-1, ite1-1]))
boundary_info.write("\n[Outlet]\n")
boundary_info.write("\nxMin = [{}] Starting Index = [{}]\n ".format(xMin+ite*dx, ite))
boundary_info.write("\nxMax = [{}] Ending Index = [{}]\n ".format(xMin+(ite1)*dx, ite1))
boundary_info.write("\nNo. of elements = [{}] \n".format(ite1-ite+1))
boundary_info.write("\nOutlet Pressure {}\n".format(D[3][-1, ite1-1]))
ite=ite1-1
break
ite1=ite1+1
ite=ite+1
flag = input("\nFinished with South Boundary Condition ?? [y/n]\n")
if(flag=='y'):
flag_SouthBoundary = 1
print("\n--------------------------------------------------\n")
boundary_info.write("\n--------------------------------------------------\n")
# Northern Boundary Defining Loop
while (flag_NorthBoundary < 1):
flag_Inlet_NorthBoundary = 0
flag_Outlet_NorthBoundary = 0
flag_SlidingWall_NorthBoundary = 0
flag = input("\nInlet in North Boundary ?? [y/n]\n")
if(flag=='y'):
flag_Inlet_NorthBoundary = 1
flag = input("\nSliding Wall in North Boundary ?? [y/n]\n")
if(flag=='y'):
flag_SlidingWall_NorthBoundary = 1
flag = input("\nOutlet in North Boundary ?? [y/n]\n")
if(flag=='y'):
flag_Outlet_NorthBoundary = 1
# Initialising Boundary Markers
north_Boundary = np.zeros((3, 6))
north_Boundary[1, 0] = 1
north_Boundary[2, 0] = -1
# Temporary Wall
north_Boundary[0, 1] = xMin
north_Boundary[0, 2] = xMax
D[0][0, :] = 0
D[1][0, :] = 0
D[2][0, :] = 0
# North Boundary Sliding U-velocity
if(flag_SlidingWall_NorthBoundary==1):
north_Boundary[0, 3] = float(input("\nEnter U-Velocity for Sliding Wall in North Boundary\n"))
D[1][0, 1:nx] = north_Boundary[0, 3]
# Inlet definition in North Boundary
if(flag_Inlet_NorthBoundary==1):
north_Boundary[1, 1] = float(input("\nEnter xmin for Inlet in North Boundary\n"))
northInlet_min_index = int((north_Boundary[1, 1]-xMin)/dx)
if(northInlet_min_index==0):
northInlet_min_index+=1
north_Boundary[1, 2] = float(input("\nEnter xmax for Inlet in North Boundary\n"))
northInlet_max_index = int((north_Boundary[1, 2]-xMin)/dx)
if(northInlet_max_index==nx):
northInlet_max_index-=1
north_Boundary[1, 4] = float(input("\nEnter V-Velocity for Inlet in North Boundary\n"))
# Boundary Marker
D[0][0, northInlet_min_index:northInlet_max_index+1] = 1.0
# U-vel Initialised to 0
D[1][0, northInlet_min_index:northInlet_max_index+1] = 0.0
# V-vel User Initialised
D[2][0, northInlet_min_index:northInlet_max_index+1] = north_Boundary[1, 4]
# Outlet definition in North Boundary
if(flag_Outlet_NorthBoundary==1):
flag_Outlet = 1
north_Boundary[2, 1] = float(input("\nEnter xmin for Outlet in North Boundary\n"))
northOutlet_min_index = int((north_Boundary[2, 1]-xMin)/dx)
if(northOutlet_min_index==0):
northOutlet_min_index+=1
north_Boundary[2, 2] = float(input("\nEnter xmax for Outlet in North Boundary\n"))
northOutlet_max_index = int((north_Boundary[2, 2]-xMin)/dx)
if(northOutlet_max_index==nx):
northOutlet_max_index-=1
north_Boundary[2, 5] = float(input("\nEnter Pressure for Outlet in North Boundary\n"))
# Boundary Marker
D[0][0, northOutlet_min_index:northOutlet_max_index+1] = -1.0
# U-vel Uninitialised
D[1][0, northOutlet_min_index:northOutlet_max_index+1] = -1000.5
# V-vel Uninitialised
D[2][0, northOutlet_min_index:northOutlet_max_index+1] = -1000.5
# Pressure User Initialised
D[3][0, northOutlet_min_index:northOutlet_max_index+1] = north_Boundary[2, 5]
print("\n\n------------------NORTH BOUNDARY------------------\n")
boundary_info.write("\n\n------------------NORTH BOUNDARY------------------\n")
ite=0
while ite<nx+1:
if(D[0][0, ite]==0):
ite1=ite+1
if(ite1==nx+1 and D[0][0, ite-1]!=0):
print("\n[Wall]\n")
print("\nxMin = [{}] Starting Index = [{}]\n ".format(xMin+ite*dx, ite))
print("\nxMax = [{}] Ending Index = [{}]\n ".format(xMin+(ite1-1)*dx, ite1-1))
print("\nNo. of elements = [{}] \n".format(ite1-1-ite+1))
boundary_info.write("\n[Wall]\n")
boundary_info.write("\nxMin = [{}] Starting Index = [{}]\n ".format(xMin+ite*dx, ite))
boundary_info.write("\nxMax = [{}] Ending Index = [{}]\n ".format(xMin+(ite1-1)*dx, ite1-1))
boundary_info.write("\nNo. of elements = [{}] \n".format(ite1-1-ite+1))
while ite1<nx+1:
if(D[0][0, ite1]!=0):
print("\n[Wall]\n")
print("\nxMin = [{}] Starting Index = [{}]\n ".format(xMin+ite*dx, ite))
print("\nxMax = [{}] Ending Index = [{}]\n ".format(xMin+(ite1-1)*dx, ite1-1))
print("\nNo. of elements = [{}] \n".format(ite1-1-ite+1))
boundary_info.write("\n[Wall]\n")
boundary_info.write("\nxMin = [{}] Starting Index = [{}]\n ".format(xMin+ite*dx, ite))
boundary_info.write("\nxMax = [{}] Ending Index = [{}]\n ".format(xMin+(ite1-1)*dx, ite1-1))
boundary_info.write("\nNo. of elements = [{}] \n".format(ite1-1-ite+1))
if(flag_SlidingWall_NorthBoundary==1):
print("\nSliding Wall Velocity {}\n".format(D[1][0, ite1-1]))
boundary_info.write("\nSliding Wall Velocity {}\n".format(D[1][0, ite1-1]))
ite=ite1-1
break
if(ite1==nx):
print("\n[Wall]\n")
print("\nxMin = [{}] Starting Index = [{}]\n ".format(xMin+ite*dx, ite))
print("\nxMax = [{}] Ending Index = [{}]\n ".format(xMin+(ite1)*dx, ite1))
print("\nNo. of elements = [{}] \n".format(ite1-ite+1))
boundary_info.write("\n[Wall]\n")
boundary_info.write("\nxMin = [{}] Starting Index = [{}]\n ".format(xMin+ite*dx, ite))
boundary_info.write("\nxMax = [{}] Ending Index = [{}]\n ".format(xMin+(ite1)*dx, ite1))
boundary_info.write("\nNo. of elements = [{}] \n".format(ite1-ite+1))
if(flag_SlidingWall_NorthBoundary==1):
print("\nSliding Wall Velocity {}\n".format(D[1][0, ite1-1]))
boundary_info.write("\nSliding Wall Velocity {}\n".format(D[1][0, ite1-1]))
ite=ite1-1
break
ite1=ite1+1
if(D[0][0, ite]==1):
ite1 = ite+1
while ite1<nx+1:
if(D[0][0, ite1]!=1):
print("\n[Inlet]\n")
print("\nxMin = [{}] Starting Index = [{}]\n ".format(xMin+ite*dx, ite))
print("\nxMax = [{}] Ending Index = [{}]\n ".format(xMin+(ite1-1)*dx, ite1-1))
print("\nNo. of elements = [{}] \n".format(ite1-1-ite+1))
print("\nInlet V-Velocity {}\n".format(D[2][0, ite1-1]))
boundary_info.write("\n[Inlet]\n")
boundary_info.write("\nxMin = [{}] Starting Index = [{}]\n ".format(xMin+ite*dx, ite))
boundary_info.write("\nxMax = [{}] Ending Index = [{}]\n ".format(xMin+(ite1-1)*dx, ite1-1))
boundary_info.write("\nNo. of elements = [{}] \n".format(ite1-1-ite+1))
boundary_info.write("\nInlet V-Velocity {}\n".format(D[2][0, ite1-1]))
ite=ite1-1
break
if(ite1==nx):
print("\n[Inlet]\n")
print("\nxMin = [{}] Starting Index = [{}]\n ".format(xMin+ite*dx, ite))
print("\nxMax = [{}] Ending Index = [{}]\n ".format(xMin+(ite1)*dx, ite1))
print("\nNo. of elements = [{}] \n".format(ite1-ite+1))
print("\nInlet V-Velocity {}\n".format(D[2][0, ite1-1]))
boundary_info.write("\n[Inlet]\n")
boundary_info.write("\nxMin = [{}] Starting Index = [{}]\n ".format(xMin+ite*dx, ite))
boundary_info.write("\nxMax = [{}] Ending Index = [{}]\n ".format(xMin+(ite1)*dx, ite1))
boundary_info.write("\nNo. of elements = [{}] \n".format(ite1-ite+1))
boundary_info.write("\nInlet V-Velocity {}\n".format(D[2][0, ite1-1]))
ite=ite1-1
break
ite1=ite1+1
if(D[0][0, ite]==-1):
ite1=ite+1
while ite1<nx+1:
if(D[0][0, ite1]!=-1):
print("\n[Outlet]\n")
print("\nxMin = [{}] Starting Index = [{}]\n ".format(xMin+ite*dx, ite))
print("\nxMax = [{}] Ending Index = [{}]\n ".format(xMin+(ite1-1)*dx, ite1-1))
print("\nNo. of elements = [{}] \n".format(ite1-1-ite+1))
print("\nOutlet Pressure {}\n".format(D[3][0, ite1-1]))
boundary_info.write("\n[Outlet]\n")
boundary_info.write("\nxMin = [{}] Starting Index = [{}]\n ".format(xMin+ite*dx, ite))
boundary_info.write("\nxMax = [{}] Ending Index = [{}]\n ".format(xMin+(ite1-1)*dx, ite1-1))
boundary_info.write("\nNo. of elements = [{}] \n".format(ite1-1-ite+1))
boundary_info.write("\nOutlet Pressure {}\n".format(D[3][0, ite1-1]))
ite=ite1-1
break
if(ite1==nx):
print("\n[Outlet]\n")
print("\nxMin = [{}] Starting Index = [{}]\n ".format(xMin+ite*dx, ite))
print("\nxMax = [{}] Ending Index = [{}]\n ".format(xMin+(ite1)*dx, ite1))
print("\nNo. of elements = [{}] \n".format(ite1-ite+1))
print("\nOutlet Pressure {}\n".format(D[3][0, ite1-1]))
boundary_info.write("\n[Outlet]\n")
boundary_info.write("\nxMin = [{}] Starting Index = [{}]\n ".format(xMin+ite*dx, ite))
boundary_info.write("\nxMax = [{}] Ending Index = [{}]\n ".format(xMin+(ite1)*dx, ite1))
boundary_info.write("\nNo. of elements = [{}] \n".format(ite1-ite+1))
boundary_info.write("\nOutlet Pressure {}\n".format(D[3][0, ite1-1]))
ite=ite1-1
break
ite1=ite1+1
ite=ite+1
flag = input("\nFinished with North Boundary Condition ?? [y/n]\n")
if(flag=='y'):
flag_NorthBoundary = 1
print("\n--------------------------------------------------\n")
boundary_info.write("\n--------------------------------------------------\n")
# Western Boundary Defining Loop
while (flag_WestBoundary < 1):
flag_Inlet_WestBoundary = 0
flag_Outlet_WestBoundary = 0
flag_SlidingWall_WestBoundary = 0
flag = input("\nInlet in West Boundary ?? [y/n]\n")
if(flag=='y'):
flag_Inlet_WestBoundary = 1
flag = input("\nSliding Wall in West Boundary ?? [y/n]\n")
if(flag=='y'):
flag_SlidingWall_WestBoundary = 1
flag = input("\nOutlet in West Boundary ?? [y/n]\n")
if(flag=='y'):
flag_Outlet_WestBoundary = 1
# Initialising Boundary Markers
west_Boundary = np.zeros((3, 6))
west_Boundary[1, 0] = 1
west_Boundary[2, 0] = -1
# Temporary Wall
west_Boundary[0, 1] = yMin
west_Boundary[0, 2] = yMax
D[0][:, 0] = 0
D[1][:, 0] = 0
D[2][:, 0] = 0
# West Boundary Sliding U-velocity
if(flag_SlidingWall_WestBoundary==1):
west_Boundary[0, 4] = float(input("\nEnter V-Velocity for Sliding Wall in West Boundary\n"))
D[2][1:ny, 0] = west_Boundary[0, 4]
# Inlet definition in West Boundary
if(flag_Inlet_WestBoundary==1):
west_Boundary[1, 1] = float(input("\nEnter ymin for Inlet in West Boundary\n"))
westInlet_bottom_index = int((yMax-west_Boundary[1, 1])/dy)
if(westInlet_bottom_index==ny):
westInlet_bottom_index-=1
west_Boundary[1, 2] = float(input("\nEnter ymax for Inlet in West Boundary\n"))
westInlet_top_index = int((yMax-west_Boundary[1, 2])/dy)
if(westInlet_top_index==0):
westInlet_top_index+=1
west_Boundary[1, 3] = float(input("\nEnter U-Velocity for Inlet in West Boundary\n"))
# Boundary Marker
D[0][westInlet_top_index:westInlet_bottom_index+1, 0] = 1.0
# U-vel User Initialised
D[1][westInlet_top_index:westInlet_bottom_index+1, 0] = west_Boundary[1, 3]
# V-vel Initialised to 0
D[2][westInlet_top_index:westInlet_bottom_index+1, 0] = 0.0
# Outlet definition in West Boundary
if(flag_Outlet_WestBoundary==1):
flag_Outlet = 1
west_Boundary[2, 1] = float(input("\nEnter ymin for Outlet in West Boundary\n"))
westOutlet_bottom_index = int((yMax-west_Boundary[2, 1])/dy)
if(westOutlet_bottom_index==ny):
westOutlet_bottom_index-=1
west_Boundary[2, 2] = float(input("\nEnter ymax for Outlet in West Boundary\n"))
westOutlet_top_index = int((yMax-west_Boundary[2, 2])/dy)
if(westOutlet_top_index==0):
westOutlet_top_index+=1
west_Boundary[2, 5] = float(input("\nEnter Pressure for Outlet in West Boundary\n"))
# Boundary Marker
D[0][westOutlet_top_index:westOutlet_bottom_index+1, 0] = -1.0
# U-vel Uninitialised
D[1][westOutlet_top_index:westOutlet_bottom_index+1, 0] = -1000.5
# V-vel Uninitialised
D[2][westOutlet_top_index:westOutlet_bottom_index+1, 0] = -1000.5
# Pressure User Initialised
D[3][westOutlet_top_index:westOutlet_bottom_index+1, 0] = west_Boundary[2, 5]
print("\n\n------------------WEST BOUNDARY-----------------\n")
boundary_info.write("\n\n------------------WEST BOUNDARY-----------------\n")
ite=ny
while ite>-1:
if(D[0][ite, 0]==0):
ite1=ite-1
if(ite1==-1 and D[0][ite+1, 0]!=0):
print("\n[Wall]\n")
print("\nyMin = [{:.2f}] Bottom Index = [{}]\n ".format(yMax-(ite)*dy, ite))
print("\nyMax = [{:.2f}] Top Index = [{}]\n ".format(yMax-(ite1+1)*dy, ite1+1))
print("\nNo. of elements = [{}] \n".format(ite-ite1))
boundary_info.write("\n[Wall]\n")
boundary_info.write("\nyMin = [{:.2f}] Bottom Index = [{}]\n ".format(yMax-(ite)*dy, ite))
boundary_info.write("\nyMax = [{:.2f}] Top Index = [{}]\n ".format(yMax-(ite1+1)*dy, ite1+1))
boundary_info.write("\nNo. of elements = [{}] \n".format(ite-ite1))
while ite1>-1:
if(D[0][ite1, 0]!=0):
print("\n[Wall]\n")
print("\nyMin = [{:.2f}] Bottom Index = [{}]\n ".format(yMax-(ite)*dy, ite))
print("\nyMax = [{:.2f}] Top Index = [{}]\n ".format(yMax-(ite1+1)*dy, ite1+1))
print("\nNo. of elements = [{}] \n".format(ite-ite1))
boundary_info.write("\n[Wall]\n")
boundary_info.write("\nyMin = [{:.2f}] Bottom Index = [{}]\n ".format(yMax-(ite)*dy, ite))
boundary_info.write("\nyMax = [{:.2f}] Top Index = [{}]\n ".format(yMax-(ite1+1)*dy, ite1+1))
boundary_info.write("\nNo. of elements = [{}] \n".format(ite-ite1))
if(flag_SlidingWall_WestBoundary==1):
print("\nSliding Wall Velocity {}\n".format(D[2][ite1+1, 0]))
boundary_info.write("\nSliding Wall Velocity {}\n".format(D[2][ite1+1, 0]))
ite=ite1+1
break
if(ite1==0):
print("\n[Wall]\n")
print("\nyMin = [{:.2f}] Bottom Index = [{}]\n ".format(yMax-(ite)*dy, ite))
print("\nyMax = [{:.2f}] Top Index = [{}]\n ".format(yMax-(ite1)*dy, ite1))
print("\nNo. of elements = [{}] \n".format(ite-ite1+1))
boundary_info.write("\n[Wall]\n")
boundary_info.write("\nyMin = [{:.2f}] Bottom Index = [{}]\n ".format(yMax-(ite)*dy, ite))
boundary_info.write("\nyMax = [{:.2f}] Top Index = [{}]\n ".format(yMax-(ite1)*dy, ite1))
boundary_info.write("\nNo. of elements = [{}] \n".format(ite-ite1+1))
if(flag_SlidingWall_WestBoundary==1):
print("\nSliding Wall Velocity {}\n".format(D[2][ite1+1, 0]))
boundary_info.write("\nSliding Wall Velocity {}\n".format(D[2][ite1+1, 0]))
ite=ite1+1
break
ite1=ite1-1
if(D[0][ite, 0]==1):
ite1 = ite-1
while ite1>-1:
if(D[0][ite1, 0]!=1):
print("\n[Inlet]\n")
print("\nyMin = [{:.2f}] Bottom Index = [{}]\n ".format(yMax-(ite)*dy, ite))
print("\nyMax = [{:.2f}] Top Index = [{}]\n ".format(yMax-(ite1+1)*dy, ite1+1))
print("\nNo. of elements = [{}] \n".format(ite-ite1))
print("\nInlet U-Velocity {}\n".format(D[1][ite1+1, 0]))
boundary_info.write("\n[Inlet]\n")
boundary_info.write("\nyMin = [{:.2f}] Bottom Index = [{}]\n ".format(yMax-(ite)*dy, ite))
boundary_info.write("\nyMax = [{:.2f}] Top Index = [{}]\n ".format(yMax-(ite1+1)*dy, ite1+1))
boundary_info.write("\nNo. of elements = [{}] \n".format(ite-ite1))
boundary_info.write("\nInlet U-Velocity {}\n".format(D[1][ite1+1, 0]))
ite=ite1+1
break
if(ite1==0):
print("\n[Inlet]\n")
print("\nyMin = [{:.2f}] Bottom Index = [{}]\n ".format(yMax-(ite)*dy, ite))
print("\nyMax = [{:.2f}] Top Index = [{}]\n ".format(yMax-(ite1)*dy, ite1))
print("\nNo. of elements = [{}] \n".format(ite-ite1+1))
print("\nInlet U-Velocity {}\n".format(D[1][ite1+1, 0]))
boundary_info.write("\n[Inlet]\n")
boundary_info.write("\nyMin = [{:.2f}] Bottom Index = [{}]\n ".format(yMax-(ite)*dy, ite))
boundary_info.write("\nyMax = [{:.2f}] Top Index = [{}]\n ".format(yMax-(ite1)*dy, ite1))
boundary_info.write("\nNo. of elements = [{}] \n".format(ite-ite1+1))
boundary_info.write("\nInlet U-Velocity {}\n".format(D[1][ite1+1, 0]))
ite=ite1+1
break
ite1=ite1-1
if(D[0][ite, 0]==-1):
ite1=ite-1
while ite1>-1:
if(D[0][ite1, 0]!=-1):
print("\n[Outlet]\n")
print("\nyMin = [{:.2f}] Bottom Index = [{}]\n ".format(yMax-(ite)*dy, ite))
print("\nyMax = [{:.2f}] Top Index = [{}]\n ".format(yMax-(ite1+1)*dy, ite1+1))
print("\nNo. of elements = [{}] \n".format(ite-ite1))
print("\nOutlet Pressure {}\n".format(D[3][ite1+1, 0]))
boundary_info.write("\n[Outlet]\n")
boundary_info.write("\nyMin = [{:.2f}] Bottom Index = [{}]\n ".format(yMax-(ite)*dy, ite))
boundary_info.write("\nyMax = [{:.2f}] Top Index = [{}]\n ".format(yMax-(ite1+1)*dy, ite1+1))
boundary_info.write("\nNo. of elements = [{}] \n".format(ite-ite1))
boundary_info.write("\nOutlet Pressure {}\n".format(D[3][ite1+1, 0]))
ite=ite1+1
break
if(ite1==0):
print("\n[Outlet]\n")
print("\nyMin = [{:.2f}] Bottom Index = [{}]\n ".format(yMax-(ite)*dy, ite))
print("\nyMax = [{:.2f}] Top Index = [{}]\n ".format(yMax-(ite1)*dy, ite1))
print("\nNo. of elements = [{}] \n".format(ite-ite1+1))
print("\nOutlet Pressure {}\n".format(D[3][ite1+1, 0]))
boundary_info.write("\n[Outlet]\n")
boundary_info.write("\nyMin = [{:.2f}] Bottom Index = [{}]\n ".format(yMax-(ite)*dy, ite))
boundary_info.write("\nyMax = [{:.2f}] Top Index = [{}]\n ".format(yMax-(ite1)*dy, ite1))
boundary_info.write("\nNo. of elements = [{}] \n".format(ite-ite1+1))
boundary_info.write("\nOutlet Pressure {}\n".format(D[3][ite1+1, 0]))
ite=ite1+1
break
ite1=ite1-1
ite=ite-1
flag = input("\nFinished with West Boundary Condition ?? [y/n]\n")
if(flag=='y'):
flag_WestBoundary = 1
print("\n-------------------------------------------------\n")
boundary_info.write("\n-------------------------------------------------\n")
# Eastern Boundary Defining Loop
while (flag_EastBoundary < 1):
flag_Inlet_EastBoundary = 0
flag_Outlet_EastBoundary = 0
flag_SlidingWall_EastBoundary = 0
flag = input("\nInlet in East Boundary ?? [y/n]\n")
if(flag=='y'):
flag_Inlet_EastBoundary = 1
flag = input("\nSliding Wall in East Boundary ?? [y/n]\n")
if(flag=='y'):
flag_SlidingWall_EastBoundary = 1
flag = input("\nOutlet in East Boundary ?? [y/n]\n")
if(flag=='y'):
flag_Outlet_EastBoundary = 1
# Initialising Boundary Markers
east_Boundary = np.zeros((3, 6))
east_Boundary[1, 0] = 1
east_Boundary[2, 0] = -1
# Temporary Wall
east_Boundary[0, 1] = yMin
east_Boundary[0, 2] = yMax
D[0][:, -1] = 0
D[1][:, -1] = 0
D[2][:, -1] = 0
# East Boundary Sliding U-velocity
if(flag_SlidingWall_EastBoundary==1):
east_Boundary[0, 4] = float(input("Enter V-Velocity for Sliding Wall in East Boundary\n"))
D[2][1:ny, -1] = east_Boundary[0, 4]
# Inlet definition in East Boundary
if(flag_Inlet_EastBoundary==1):
east_Boundary[1, 1] = float(input("Enter ymin for Inlet in East Boundary\n"))
eastInlet_bottom_index = int((yMax-east_Boundary[1, 1])/dy)
if(eastInlet_bottom_index==ny):
eastInlet_bottom_index-=1
east_Boundary[1, 2] = float(input("Enter ymax for Inlet in East Boundary\n"))
eastInlet_top_index = int((yMax-east_Boundary[1, 2])/dy)
if(eastInlet_top_index==0):
eastInlet_top_index+=1
east_Boundary[1, 3] = float(input("Enter U-Velocity for Inlet in East Boundary\n"))
# Boundary Marker
D[0][eastInlet_top_index:eastInlet_bottom_index+1, -1] = 1.0
# U-vel User Initialised
D[1][eastInlet_top_index:eastInlet_bottom_index+1, -1] = east_Boundary[1, 3]
# V-vel Initialised to 0
D[2][eastInlet_top_index:eastInlet_bottom_index+1, -1] = 0.0
# Outlet definition in East Boundary
if(flag_Outlet_EastBoundary==1):
flag_Outlet = 1
east_Boundary[2, 1] = float(input("Enter ymin for Outlet in East Boundary\n"))
eastOutlet_bottom_index = int((yMax-east_Boundary[2, 1])/dy)
if(eastOutlet_bottom_index==ny):
eastOutlet_bottom_index-=1
east_Boundary[2, 2] = float(input("Enter ymax for Outlet in East Boundary\n"))
eastOutlet_top_index = int((yMax-east_Boundary[2, 2])/dy)
if(eastOutlet_top_index==0):
eastOutlet_top_index+=1
east_Boundary[2, 5] = float(input("Enter Pressure for Outlet in East Boundary\n"))
# Boundary Marker
D[0][eastOutlet_top_index:eastOutlet_bottom_index+1, -1] = -1.0
# U-vel Uninitialised
D[1][eastOutlet_top_index:eastOutlet_bottom_index+1, -1] = -1000.5
# V-vel Uninitialised
D[2][eastOutlet_top_index:eastOutlet_bottom_index+1, -1] = -1000.5
# Pressure User Initialised
D[3][eastOutlet_top_index:eastOutlet_bottom_index+1, -1] = east_Boundary[2, 5]
print("\n\n------------------EAST BOUNDARY-----------------\n")
boundary_info.write("\n\n------------------EAST BOUNDARY-----------------\n")
ite=ny
while ite>-1:
if(D[0][ite, -1]==0):
ite1=ite-1
if(ite1==-1 and D[0][ite+1, -1]!=0):
print("\n[Wall]\n")
print("\nyMin = [{:.2f}] Bottom Index = [{}]\n ".format(yMax-(ite)*dy, ite))
print("\nyMax = [{:.2f}] Top Index = [{}]\n ".format(yMax-(ite1+1)*dy, ite1+1))
print("\nNo. of elements = [{}] \n".format(ite-ite1))
boundary_info.write("\n[Wall]\n")
boundary_info.write("\nyMin = [{:.2f}] Bottom Index = [{}]\n ".format(yMax-(ite)*dy, ite))
boundary_info.write("\nyMax = [{:.2f}] Top Index = [{}]\n ".format(yMax-(ite1+1)*dy, ite1+1))
boundary_info.write("\nNo. of elements = [{}] \n".format(ite-ite1))
while ite1>-1:
if(D[0][ite1, -1]!=0):
print("\n[Wall]\n")
print("\nyMin = [{:.2f}] Bottom Index = [{}]\n ".format(yMax-(ite)*dy, ite))
print("\nyMax = [{:.2f}] Top Index = [{}]\n ".format(yMax-(ite1+1)*dy, ite1+1))
print("\nNo. of elements = [{}]\n".format(ite-ite1))
boundary_info.write("\n[Wall]\n")
boundary_info.write("\nyMin = [{:.2f}] Bottom Index = [{}]\n ".format(yMax-(ite)*dy, ite))
boundary_info.write("\nyMax = [{:.2f}] Top Index = [{}]\n ".format(yMax-(ite1+1)*dy, ite1+1))
boundary_info.write("\nNo. of elements = [{}]\n".format(ite-ite1))
if(flag_SlidingWall_EastBoundary==1):
print("\nSliding Wall Velocity {}\n".format(D[2][ite1+1, -1]))
boundary_info.write("\nSliding Wall Velocity {}\n".format(D[2][ite1+1, -1]))
ite=ite1+1
break
if(ite1==0):
print("\n[Wall]\n")
print("\nyMin = [{:.2f}] Bottom Index = [{}]\n ".format(yMax-(ite)*dy, ite))
print("\nyMax = [{:.2f}] Top Index = [{}]\n ".format(yMax-(ite1)*dy, ite1))
print("\nNo. of elements = [{}]\n".format(ite-ite1+1))
boundary_info.write("\n[Wall]\n")
boundary_info.write("\nyMin = [{:.2f}] Bottom Index = [{}]\n ".format(yMax-(ite)*dy, ite))
boundary_info.write("\nyMax = [{:.2f}] Top Index = [{}]\n ".format(yMax-(ite1)*dy, ite1))
boundary_info.write("\nNo. of elements = [{}]\n".format(ite-ite1+1))
if(flag_SlidingWall_EastBoundary==1):
print("\nSliding Wall Velocity {}\n".format(D[2][ite1+1, -1]))
boundary_info.write("\nSliding Wall Velocity {}\n".format(D[2][ite1+1, -1]))
ite=ite1+1
break
ite1=ite1-1
if(D[0][ite, -1]==1):
ite1 = ite-1
while ite1>-1:
if(D[0][ite1, -1]!=1):
print("\n[Inlet]\n")
print("\nyMin = [{:.2f}] Bottom Index = [{}]\n ".format(yMax-(ite)*dy, ite))
print("\nyMax = [{:.2f}] Top Index = [{}]\n ".format(yMax-(ite1+1)*dy, ite1+1))
print("\nNo. of elements = [{}] \n".format(ite-ite1))
print("\nInlet U-Velocity {}\n".format(D[1][ite1+1, -1]))
boundary_info.write("\n[Inlet]\n")
boundary_info.write("\nyMin = [{:.2f}] Bottom Index = [{}]\n ".format(yMax-(ite)*dy, ite))
boundary_info.write("\nyMax = [{:.2f}] Top Index = [{}]\n ".format(yMax-(ite1+1)*dy, ite1+1))
boundary_info.write("\nNo. of elements = [{}] \n".format(ite-ite1))
boundary_info.write("\nInlet U-Velocity {}\n".format(D[1][ite1+1, -1]))
ite=ite1+1
break
if(ite1==0):
print("\n[Inlet]\n")
print("\nyMin = [{:.2f}] Bottom Index = [{}]\n ".format(yMax-(ite)*dy, ite))
print("\nyMax = [{:.2f}] Top Index = [{}]\n ".format(yMax-(ite1)*dy, ite1))
print("\nNo. of elements = [{}] \n".format(ite-ite1+1))
print("\nInlet U-Velocity {}\n".format(D[1][ite1+1, -1]))
boundary_info.write("\n[Inlet]\n")
boundary_info.write("\nyMin = [{:.2f}] Bottom Index = [{}]\n ".format(yMax-(ite)*dy, ite))
boundary_info.write("\nyMax = [{:.2f}] Top Index = [{}]\n ".format(yMax-(ite1)*dy, ite1))
boundary_info.write("\nNo. of elements = [{}] \n".format(ite-ite1+1))
boundary_info.write("\nInlet U-Velocity {}\n".format(D[1][ite1+1, -1]))
ite=ite1+1
break
ite1=ite1-1
if(D[0][ite, -1]==-1):
ite1=ite-1
while ite1>-1:
if(D[0][ite1, -1]!=-1):
print("\n[Outlet]\n")
print("\nyMin = [{:.2f}] Bottom Index = [{}]\n ".format(yMax-(ite)*dy, ite))
print("\nyMax = [{:.2f}] Top Index = [{}]\n ".format(yMax-(ite1+1)*dy, ite1+1))
print("\nNo. of elements = [{}] \n".format(ite-ite1))
print("\nOutlet Pressure {}\n".format(D[3][ite1+1, -1]))
boundary_info.write("\n[Outlet]\n")
boundary_info.write("\nyMin = [{:.2f}] Bottom Index = [{}]\n ".format(yMax-(ite)*dy, ite))
boundary_info.write("\nyMax = [{:.2f}] Top Index = [{}]\n ".format(yMax-(ite1+1)*dy, ite1+1))
boundary_info.write("\nNo. of elements = [{}] \n".format(ite-ite1))
boundary_info.write("\nOutlet Pressure {}\n".format(D[3][ite1+1, -1]))
ite=ite1+1
break
if(ite1==0):
print("\n[Outlet]\n")
print("\nyMin = [{:.2f}] Bottom Index = [{}]\n ".format(yMax-(ite)*dy, ite))
print("\nyMax = [{:.2f}] Top Index = [{}]\n ".format(yMax-(ite1)*dy, ite1))
print("\nNo. of elements = [{}] \n".format(ite-ite1+1))
print("\nOutlet Pressure {}\n".format(D[3][ite1+1, -1]))
boundary_info.write("\n[Outlet]\n")
boundary_info.write("\nyMin = [{:.2f}] Bottom Index = [{}]\n ".format(yMax-(ite)*dy, ite))
boundary_info.write("\nyMax = [{:.2f}] Top Index = [{}]\n ".format(yMax-(ite1)*dy, ite1))
boundary_info.write("\nNo. of elements = [{}] \n".format(ite-ite1+1))
boundary_info.write("\nOutlet Pressure {}\n".format(D[3][ite1+1, -1]))
ite=ite1+1
break
ite1=ite1-1
ite=ite-1
flag = input("Finished with East Boundary Condition ?? [y/n]")
if(flag=='y'):
flag_EastBoundary = 1
print("\n-------------------------------------------------\n")
boundary_info.write("\n-------------------------------------------------\n")
flag_Boundary = 1
return D, xMin, xMax, yMin, yMax, dx, dy, nx, ny, south_Boundary, north_Boundary, west_Boundary, east_Boundary, flag_Outlet
def obstacle(D, xMin, xMax, yMin, yMax, nx, ny, dx, dy):
package_dir = os.path.dirname(os.path.realpath(__file__))
flag_obstacle = 'n'
while flag_obstacle!='y':
scal_x = float(input("\nEnter Scaling factor for X-direction\n"))
scal_y = float(input("\nEnter Scaling factor for Y-direction\n"))
traf_x = float(input("\nOffset in X\n"))
traf_y = float(input("\nOffset in Y\n")) + dy/10
rot = float(input("\nEnter anti-clockwise rotation angle\n"))
flag_geometryCase = input("\n1. Enter address of .txt file\n2. Use default address\n")
if(flag_geometryCase==1):
geometryCase = input("\nEnter address of .txt file\nExample: /home/user/folder/Case_1/Points.txt\n")
else:
geometryCase = package_dir+"/Points.txt"
file1 = open(geometryCase, 'r')
Lines = file1.readlines()
P=[]
N=0
for line in Lines:
N+=1
P.append(np.fromstring(line.strip(), dtype=float, sep=' '))
N+=1
P.append(np.fromstring(Lines[0].strip(), dtype=float, sep=' '))
Lx = xMax-xMin
Ly = yMax-yMin
Points_x=[]
Points_y=[]
for i in range (0, N):
Points_x.append(np.sqrt((P[i][0]*scal_x)**2 + (P[i][1]*scal_y)**2)*np.cos(rot*np.pi/180+np.arctan((P[i][1])/(P[i][0])))+traf_x)
Points_y.append(np.sqrt((P[i][0]*scal_x)**2 + (P[i][1]*scal_y)**2)*np.sin(rot*np.pi/180+np.arctan(P[i][1]/P[i][0]))+traf_y)
fig = plt.subplots(figsize =(Lx*5, Ly*5))
plt.xlim(xMin, xMax)
plt.ylim(yMin, yMax)
plt.fill(Points_x, Points_y, color = 'black')
plt.show()
flag_obstacle = (input("\nConitnue [Y/N] ??\n"))
if(flag_obstacle!='y'):
continue
for i in range(0, ny+1):
for j in range (0, nx+1):
X = xMin + j*dx
Y = yMax - i*dy
count=0
for ite in range (0, N-1):
X1 = Points_x[ite]
X2 = Points_x[ite+1]
Y1 = Points_y[ite]
Y2 = Points_y[ite+1]
if(Y1<=Y<=Y2 or Y2<=Y<=Y1):
if(X1<=X2):
if(X<=(X2+(Y-Y2)*(X1-X2)/(Y1-Y2))):
count +=1
if(X2<=X1):
if(X<=(X1+(Y-Y1)*(X2-X1)/(Y2-Y1))):
count +=1
if(count==1):
D[0][i, j] = 0
D[1][i, j] = 0
D[2][i, j] = 0
D[3][i, j] = -100000
return D, P, Points_x, Points_y
def obstacleMath(D, xMin, xMax, yMin, yMax, nx, ny, dx, dy):
flag_obstacle = 'n'
while flag_obstacle!='y':
scal_x = float(input("\nEnter Scaling factor for X-direction\n"))
scal_y = float(input("\nEnter Scaling factor for Y-direction\n"))
traf_x = float(input("\nOffset in X\n"))
traf_y = float(input("\nOffset in Y\n")) + dy/10
rot = int(input("\nEnter anti-clockwise rotation angle\n"))
P=[]
asq = float(input("\nEnter Horizontal axis length for Elliptical Geometry\n"))
bsq = float(input("\nEnter Vertical axis length for Elliptical Geometry\n"))
a=asq**2
b=bsq**2
N=0
for theta in range (0, 360):
if(theta%2==0):
temp1=a*b/(b+a*np.tan(theta*np.pi/180)**2)
if(0<=theta<=90):
temp = [np.abs(np.sqrt(temp1)), np.abs(np.tan(theta*np.pi/180))*(np.abs(np.sqrt(temp1)))]
elif(90<=theta<=180):
temp = [-np.abs(np.sqrt(temp1)), np.abs(np.tan(theta*np.pi/180))*(np.abs(np.sqrt(temp1)))]
elif(180<=theta<=270):
temp = [-np.abs(np.sqrt(temp1)), -np.abs(np.tan(theta*np.pi/180))*(np.abs(np.sqrt(temp1)))]
elif(270<=theta<=360):
temp = [np.abs(np.sqrt(temp1)), -np.abs(np.tan(theta*np.pi/180))*(np.abs(np.sqrt(temp1)))]
P.append(temp)
N+=1
P.append(P[0])
Lx = xMax-xMin
Ly = yMax-yMin
Points_x=[]
Points_y=[]
for i in range(0, 181):
if(0<=(i+rot/2)<=45):
Points_x.append(np.abs(np.sqrt((P[i][0]*scal_x)**2 + (P[i][1]*scal_y)**2)*np.cos(rot*np.pi/180+np.arctan(P[i][1]/P[i][0])))+traf_x)
Points_y.append(np.abs(np.sqrt((P[i][0]*scal_x)**2 + (P[i][1]*scal_y)**2)*np.sin(rot*np.pi/180+np.arctan(P[i][1]/P[i][0])))+traf_y)
elif(45<=(i+rot/2)<=90):
Points_x.append(-np.abs(np.sqrt((P[i][0]*scal_x)**2 + (P[i][1]*scal_y)**2)*np.cos(rot*np.pi/180+np.arctan(P[i][1]/P[i][0])))+traf_x)
Points_y.append(np.abs(np.sqrt((P[i][0]*scal_x)**2 + (P[i][1]*scal_y)**2)*np.sin(rot*np.pi/180+np.arctan(P[i][1]/P[i][0])))+traf_y)
elif(90<=(i+rot/2)<=135):
Points_x.append(-np.abs(np.sqrt((P[i][0]*scal_x)**2 + (P[i][1]*scal_y)**2)*np.cos(rot*np.pi/180+np.arctan(P[i][1]/P[i][0])))+traf_x)
Points_y.append(-np.abs(np.sqrt((P[i][0]*scal_x)**2 + (P[i][1]*scal_y)**2)*np.sin(rot*np.pi/180+np.arctan(P[i][1]/P[i][0])))+traf_y)
elif(135<=(i+rot/2)<=180):
Points_x.append(np.abs(np.sqrt((P[i][0]*scal_x)**2 + (P[i][1]*scal_y)**2)*np.cos(rot*np.pi/180+np.arctan(P[i][1]/P[i][0])))+traf_x)
Points_y.append(-np.abs(np.sqrt((P[i][0]*scal_x)**2 + (P[i][1]*scal_y)**2)*np.sin(rot*np.pi/180+np.arctan(P[i][1]/P[i][0])))+traf_y)
elif(180<=(i+rot/2)<=225):
Points_x.append(np.abs(np.sqrt((P[i][0]*scal_x)**2 + (P[i][1]*scal_y)**2)*np.cos(rot*np.pi/180+np.arctan(P[i][1]/P[i][0])))+traf_x)
Points_y.append(np.abs(np.sqrt((P[i][0]*scal_x)**2 + (P[i][1]*scal_y)**2)*np.sin(rot*np.pi/180+np.arctan(P[i][1]/P[i][0])))+traf_y)
elif(-45<=(i+rot/2)<=0):
Points_x.append(np.abs(np.sqrt((P[i][0]*scal_x)**2 + (P[i][1]*scal_y)**2)*np.cos(rot*np.pi/180+np.arctan(P[i][1]/P[i][0])))+traf_x)
Points_y.append(-np.abs(np.sqrt((P[i][0]*scal_x)**2 + (P[i][1]*scal_y)**2)*np.sin(rot*np.pi/180+np.arctan(P[i][1]/P[i][0])))+traf_y)
elif(-90<=(i+rot/2)<=-45):
Points_x.append(-np.abs(np.sqrt((P[i][0]*scal_x)**2 + (P[i][1]*scal_y)**2)*np.cos(rot*np.pi/180+np.arctan(P[i][1]/P[i][0])))+traf_x)
Points_y.append(-np.abs(np.sqrt((P[i][0]*scal_x)**2 + (P[i][1]*scal_y)**2)*np.sin(rot*np.pi/180+np.arctan(P[i][1]/P[i][0])))+traf_y)
elif(-135<=(i+rot/2)<=-90):
Points_x.append(-np.abs(np.sqrt((P[i][0]*scal_x)**2 + (P[i][1]*scal_y)**2)*np.cos(rot*np.pi/180+np.arctan(P[i][1]/P[i][0])))+traf_x)
Points_y.append(np.abs(np.sqrt((P[i][0]*scal_x)**2 + (P[i][1]*scal_y)**2)*np.sin(rot*np.pi/180+np.arctan(P[i][1]/P[i][0])))+traf_y)
elif(-180<=(i+rot/2)<=-135):
Points_x.append(np.abs(np.sqrt((P[i][0]*scal_x)**2 + (P[i][1]*scal_y)**2)*np.cos(rot*np.pi/180+np.arctan(P[i][1]/P[i][0])))+traf_x)
Points_y.append(np.abs(np.sqrt((P[i][0]*scal_x)**2 + (P[i][1]*scal_y)**2)*np.sin(rot*np.pi/180+np.arctan(P[i][1]/P[i][0])))+traf_y)
fig = plt.subplots(figsize =(Lx*5, Ly*5))
plt.xlim(xMin, xMax)
plt.ylim(yMin, yMax)
plt.fill(Points_x, Points_y, color='black')
plt.show()
flag_obstacle = (input("\nConitnue [Y/N] ??\n"))
if(flag_obstacle!='y'):
continue
for i in range(0, ny+1):
for j in range (0, nx+1):
X = xMin + j*dx
Y = yMax - i*dy
count=0
for ite in range (0, 180):
X1 = Points_x[ite]
X2 = Points_x[ite+1]
Y1 = Points_y[ite]
Y2 = Points_y[ite+1]
if(Y1<Y<=Y2 or Y2<=Y<Y1):
if(X<=X1 or X<=X2):
count +=1
if(count==1):
D[0][i, j] = 0
D[1][i, j] = 0
D[2][i, j] = 0
D[3][i, j] = -100000
return D, P, Points_x, Points_y | 2D-Panel-CFD | /2D_Panel-CFD-0.0.1.tar.gz/2D_Panel-CFD-0.0.1/src/RUN_package/boundaryDomain.py | boundaryDomain.py |
import numpy as np
def U_coeffUpwind(Uxo, Uyo, Po, Uue, Uuw, Vun, Vus, D, rho, visc, urf_UV, nx, ny, dx, dy):
Aup = np.zeros((ny, nx+1))
Aue = np.zeros((ny, nx+1))
Auw = np.zeros((ny, nx+1))
Aun = np.zeros((ny, nx+1))
Aus = np.zeros((ny, nx+1))
Bup = np.zeros((ny, nx+1))
vyx = visc*dy/dx
vxy = visc*dx/dy
for i in range (0, ny):
for j in range (0, nx+1):
if(j!=0 and j!=nx):
Aup[i, j] = rho*dy*(max(Uue[i, j], 0) + max(-Uuw[i, j], 0)) +\
rho*dx*(max(Vun[i, j], 0) + max(-Vus[i, j], 0)) +\
2*(vyx+vxy)
Aue[i, j] = rho*dy*max(-Uue[i, j], 0) + vyx
Auw[i, j] = rho*dy*max(Uuw[i, j], 0) + vyx
Aun[i, j] = rho*dx*max(-Vun[i, j], 0) + vxy
Aus[i, j] = rho*dx*max(Vus[i, j], 0) + vxy
Bup[i, j] = dy*(Po[i, j-1]-Po[i, j])
# South Boundary Outlet
if(D[0][i, j] == 2 and D[0][i+1, j] == -1):
Aup[i, j] += -vxy - rho*dx*max(Vus[i, j], 0)
Aus[i, j] = 0
# North Boundary Outlet
if(D[0][i, j] == -1 and D[0][i+1, j] == 2):
Aup[i, j] += -vxy - rho*dx*max(-Vun[i, j], 0)
Aue[i, j] = 0
# South Boundary Wall
if(D[0][i, j] == 2 and D[0][i+1, j] == 0):
Aup[i, j] += 2*vxy
Aun[i, j] += vxy/3
Aus[i, j] = 0
Bup[i, j] += vxy*8*D[1][i+1, j]/3
# North Boundary Wall
if(D[0][i, j] == 0 and D[0][i+1, j] == 2):
Aup[i, j] += 2*vxy
Aus[i, j] += vxy/3
Aun[i, j] = 0
Bup[i, j] += vxy*8*D[1][i, j]/3
# South Boundary Inlet
if(D[0][i, j] == 2 and D[0][i+1, j] == 1):
Aup[i, j] += 2*vxy + 2*rho*dx*max(Vus[i, j], 0)
Aun[i, j] += vxy/3 + rho*dx*max(Vus[i, j], 0)/3
Aus[i, j] = 0
Bup[i, j] += vxy*8*D[1][i+1, j]/3 + 8*rho*dx*D[1][i+1, j]*max(Vus[i, j], 0)/3
# North Boundary Inlet
if(D[0][i, j] == 1 and D[0][i+1, j] == 2):
Aup[i, j] += 2*vxy + 2*rho*dx*max(-Vun[i, j], 0)
Aus[i, j] += vxy/3 + rho*dx*max(-Vun[i, j], 0)/3
Aun[i, j] = 0
Bup[i, j] += vxy*8*D[1][i, j]/3 + 8*rho*dx*D[1][i, j]*max(-Vun[i, j], 0)/3
# Data Point on Boundary
if(D[0][i, j]!=2 and D[0][i+1, j]!=2):
Aup[i, j] = 1
Aue[i, j] = 0
Auw[i, j] = 0
Aun[i, j] = 0
Aus[i, j] = 0
Bup[i, j] = Uxo[i, j]
# Outlet Boundary
if(D[0][i, j]==-1 and D[0][i+1, j]==-1 and j==nx):
Aup[i, j] = 1
Aue[i, j] = 0
Auw[i, j] = 1
Aun[i, j] = 0
Aus[i, j] = 0
Bup[i, j] = 0
if(D[0][i, j]==-1 and D[0][i+1, j]==-1 and j==0):
Aup[i, j] = 1
Aue[i, j] = 1
Auw[i, j] = 0
Aun[i, j] = 0
Aus[i, j] = 0
Bup[i, j] = 0
# Under-Relaxation Factor
if(D[0][i, j]==2 or D[0][i+1, j]==2):
Aup[i, j] = Aup[i, j]/urf_UV
Bup[i, j] += (1-urf_UV)*Aup[i, j]*Uxo[i, j]
# Matrix Creation
M_Bup = Bup.flatten()
M_Au = np.zeros(((ny)*(nx+1), (ny)*(nx+1)))
ite=0
for i in range (0, ny):
for j in range (0, nx+1):
M_Au[ite, ite] = Aup[i, j]
if ((ite+1)%(nx+1)!=0):
M_Au[ite, ite+1] = -Aue[i, j]
if((ite%(nx+1)!=0) and (ite!=0)):
M_Au[ite, ite-1] = -Auw[i, j]
if (ite<(ny-1)*(nx+1)):
M_Au[ite, ite+nx+1] = -Aus[i, j]
if(ite>nx):
M_Au[ite, ite-nx-1] = -Aun[i, j]
ite+=1
return M_Au, M_Bup, Aup, Aue, Auw, Aun, Aus, Bup
def V_coeffUpwind(Uxo, Uyo, Po, Uve, Uvw, Vvn, Vvs, D, rho, visc, urf_UV, nx, ny, dx, dy):
Avp = np.zeros((ny+1, nx))
Ave = np.zeros((ny+1, nx))
Avw = np.zeros((ny+1, nx))
Avn = np.zeros((ny+1, nx))
Avs = np.zeros((ny+1, nx))
Bvp = np.zeros((ny+1, nx))
vyx = visc*dy/dx
vxy = visc*dx/dy
for i in range (0, ny+1):
for j in range (0, nx):
if(i!=0 and i!=ny):
Avp[i, j] = rho*dy*(max(Uve[i, j], 0) + max(-Uvw[i, j], 0)) +\
rho*dx*(max(Vvn[i, j], 0) + max(-Vvs[i, j], 0)) +\
2*(vxy+vyx)
Ave[i, j] = rho*dy*max(-Uve[i, j], 0) + vyx
Avw[i, j] = rho*dy*max(Uvw[i, j], 0) + vyx
Avn[i, j] = rho*dx*max(-Vvn[i, j], 0) + vxy
Avs[i, j] = rho*dx*max(Vvs[i, j], 0) + vxy
Bvp[i, j] = dx*(Po[i, j]-Po[i-1, j])
# West Boundary Outlet
if(D[0][i, j] == -1 and D[0][i, j+1] == 2):
Avp[i, j] += -vyx - rho*dy*max(Uvw[i, j], 0)
Avw[i, j] = 0
# East Boundary Outlet
if(D[0][i, j] == 2 and D[0][i, j+1] == -1):
Avp[i, j] += -vyx - rho*dy*max(-Uve[i, j], 0)
Ave[i, j] = 0
# West Boundary Wall
if(D[0][i, j] == 0 and D[0][i, j+1] == 2):
Avp[i, j] += 2*vyx
Ave[i, j] += vyx/3
Avw[i, j] = 0
Bvp[i, j] += vyx*8*D[2][i, j]/3
# East Boundary Wall
if(D[0][i, j] == 2 and D[0][i, j+1] == 0):
Avp[i, j] += 2*vyx
Avw[i, j] += vyx/3
Ave[i, j] = 0
Bvp[i, j] += vyx*8*D[2][i, j+1]/3
# West Boundary Inlet
if(D[0][i, j] == 1 and D[0][i, j+1] == 2):
Avp[i, j] += 2*vyx + 2*rho*dy*max(Uvw[i, j], 0)
Ave[i, j] += vyx/3 + rho*dy*max(Uvw[i, j], 0)/3
Avw[i, j] = 0
Bvp[i, j] += vyx*8*D[2][i, j]/3 + 8*rho*dy*D[2][i, j]*max(Uvw[i, j], 0)/3
# East Boundary Inlet
if(D[0][i, j] == 2 and D[0][i, j+1] == 1):
Avp[i, j] += 2*vyx + 2*rho*dy*max(-Uve[i, j], 0)
Avw[i, j] += vyx/3 + rho*dy*max(-Uve[i, j], 0)/3
Ave[i, j] = 0
Bvp[i, j] += vyx*8*D[2][i, j+1]/3 + 8*rho*dy*D[2][i, j+1]*max(-Uve[i, j], 0)/3
# Data Point on Boundary
if(D[0][i, j]!=2 and D[0][i, j+1]!=2):
Avp[i, j] = 1
Ave[i, j] = 0
Avw[i, j] = 0
Avn[i, j] = 0
Avs[i, j] = 0
Bvp[i, j] = Uyo[i, j]
# Outlet Boundary
if(D[0][i, j]==-1 and D[0][i, j+1]==-1 and i==0):
Avp[i, j] = 1
Ave[i, j] = 0
Avw[i, j] = 0
Avn[i, j] = 0
Avs[i, j] = 1
Bvp[i, j] = 0
if(D[0][i, j]==-1 and D[0][i, j+1]==-1 and i==ny):
Avp[i, j] = 1
Ave[i, j] = 0
Avw[i, j] = 0
Avn[i, j] = 1
Avs[i, j] = 0
Bvp[i, j] = 0
# Under-Relaxation Factor
if(D[0][i, j]==2 or D[0][i, j+1]==2):
Avp[i, j] = Avp[i, j]/urf_UV
Bvp[i, j] += (1-urf_UV)*Avp[i, j]*Uyo[i, j]
# Matrix Creation
M_Bvp = Bvp.flatten()
M_Av = np.zeros(((ny+1)*(nx), (ny+1)*(nx)))
ite=0
for i in range (0, ny+1):
for j in range (0, nx):
M_Av[ite, ite] = Avp[i, j]
if ((ite+1)%(nx)!=0):
M_Av[ite, ite+1] = -Ave[i, j]
if((ite%(nx)!=0) and (ite!=0)):
M_Av[ite, ite-1] = -Avw[i, j]
if (ite<(ny)*(nx)):
M_Av[ite, ite+nx] = -Avs[i, j]
if(ite>nx-1):
M_Av[ite, ite-nx] = -Avn[i, j]
ite+=1
return M_Av, M_Bvp, Avp, Ave, Avw, Avn, Avs, Bvp | 2D-Panel-CFD | /2D_Panel-CFD-0.0.1.tar.gz/2D_Panel-CFD-0.0.1/src/RUN_package/UV_coeff.py | UV_coeff.py |
#%%
import numpy as np
import matplotlib as mpl
from matplotlib import cm
from matplotlib import pyplot as plt
from scipy import linalg
import sys, os
mpl.rcParams.update({'font.size': 15})
caseFolder = input("\nEnter address of testCase ...\nExample: /home/user/folder/Case_1\n")
try:
os.mkdir(caseFolder)
except:
print("\nFolder already there!!\n")
flag_EmptyFolder = input("\nEmpty Folder ?? [y/n]\n")
if(flag_EmptyFolder=='y'):
filelist = [ f for f in os.listdir(caseFolder) ]
for f in filelist:
os.remove(os.path.join(caseFolder, f))
else:
sys.exit()
package_dir = os.path.dirname(os.path.realpath(__file__))
print(package_dir)
from RUN_package.boundaryDomain import domain, obstacle, obstacleMath
from RUN_package.initialConditions import initialCond
from RUN_package.faceInterpolation import U_face, V_face
from RUN_package.UV_coeff import U_coeffUpwind, V_coeffUpwind
from RUN_package.P_coeff import P_coeffSIMPLE
from RUN_package.correction import correctionSIMPLE
from RUN_package.postProcessing import postProcessor
D, \
xMin, xMax, yMin, yMax, \
dx, dy, nx, ny, \
south_Boundary, north_Boundary, west_Boundary, east_Boundary, \
flag_Outlet =\
domain(caseFolder)
flag_obstacleGeometry = int(input("\n1. Use .txt file as Address \n2. Add elliptical obstacle\n3. No Obstacle Geometry\n"))
if(flag_obstacleGeometry==1):
D, Points, Points_x, Points_y = obstacle(D, xMin, xMax, yMin, yMax, nx, ny, dx, dy)
elif(flag_obstacleGeometry==2):
D, Points, Points_x, Points_y = obstacleMath(D, xMin, xMax, yMin, yMax, nx, ny, dx, dy)
Lx = xMax-xMin
Ly = yMax-yMin
# Grid - Plot
xg = np.linspace(xMin, xMax, num=nx+2)
yg = np.linspace(yMax, yMin, num=ny+2)
Xg, Yg = np.meshgrid(xg, yg)
fig, ax = plt.subplots(figsize =(Lx*6*1.4, Ly*6))
plt.xlabel('x [m]')
plt.ylabel('y [m]')
cmap = plt.get_cmap('tab10')
ylorbr = cm.get_cmap('tab10', 4)
norm = mpl.colors.Normalize(vmin=-1.5, vmax=2.5)
col = plt.pcolormesh(xg, yg, D[0], cmap = ylorbr, norm=norm)
cbar = plt.colorbar(col)
cbar.set_label('Grid\n-1.0 - Outlet 0.0 - Wall 1.0 - Inlet 2.0 - Wall')
plt.axis('scaled')
# show plot
plt.show()
flag_continue = input("\nContinue or Abort ?? [y/n]\n")
if(flag_continue=='n'):
sys.exit()
urf_UV = float(input("\nEnter Under-relaxation factor for UV\n"))
urf_P = float(input("\nEnter Under-relaxation factor for P\n"))
rho = float(input("\nEnter Density of fluid [kg/m^3]\n"))
visc = float(input("\nEnter Viscosity of Fluid [Pa*s]\n"))
max_iter = int(input("\nEnter Max Iteration for this SIM\n"))
MaxRes_Mass = float(input("\nEnter Min Mass Imbalance Residual for this SIM\n"))
Uxo, Uyo, Po = initialCond(D, nx, ny, flag_Outlet)
Ux_values=[]
Ux_values.append(Uxo)
Uy_values=[]
Uy_values.append(Uyo)
P_values=[]
P_values.append(Po)
Res_U_values=[]
Res_V_values=[]
Res_Mass_values=[]
logRes_Mass=0.0
n=0
iter_ = np.linspace(0, max_iter, max_iter)
temp = np.linspace(0, max_iter, max_iter)
fig, ax = plt.subplots(figsize =(Lx*6*1.4, Ly*6))
plt.xlabel('x [m]')
plt.ylabel('y [m]')
while n<max_iter and logRes_Mass>MaxRes_Mass:
Uue, Uuw, Vun, Vus = U_face(Uxo, Uyo, D, nx, ny)
Uve, Uvw, Vvn, Vvs = V_face(Uxo, Uyo, D, nx, ny)
M_Au, M_Bup, Aup, Aue, Auw, Aun, Aus, Bup = \
U_coeffUpwind(Uxo, Uyo, Po, Uue, Uuw, Vun, Vus, D, rho, visc, urf_UV, nx, ny, dx, dy)
M_Av, M_Bvp, Avp, Ave, Avw, Avn, Avs, Bvp = \
V_coeffUpwind(Uxo, Uyo, Po, Uve, Uvw, Vvn, Vvs, D, rho, visc, urf_UV, nx, ny, dx, dy)
# Momentum Predictor
M_Uxs = linalg.solve(M_Au, M_Bup)
M_Uys = linalg.solve(M_Av, M_Bvp)
Uxs = np.reshape(M_Uxs, (ny, nx+1))
Uys = np.reshape(M_Uys, (ny+1, nx))
M_Ap, M_Bpp, App, Ape, Apw, Apn, Aps, Bpp = P_coeffSIMPLE(Aup, Avp, Uxs, Uys, D, nx, ny, dx, dy)
if(flag_Outlet==0):
M_Ap[int(ny*nx/4), :] = 0
M_Ap[int(ny*nx/4), int(ny*nx/4)] = 1
M_Bpp[int(ny*nx/4)] = 0
# Pressure Correction values
M_Pc = linalg.solve(M_Ap, M_Bpp)
Pc = np.reshape(M_Pc, (ny, nx))
# Residuals
Res_U = M_Au.dot(Uxo.flatten())-M_Bup
Res_V = M_Av.dot(Uyo.flatten())-M_Bvp
logRes_U = np.log(np.sum(np.abs(Res_U)))
logRes_V = np.log(np.sum(np.abs(Res_V)))
logRes_Mass = np.log(np.sum(np.abs(Bpp)))
# Momentum & Pressure Correction step
Ux, Uy, P = correctionSIMPLE(Uxs, Uys, Po, Pc, D, Aup, Avp, urf_UV, urf_P, nx, ny, dx, dy)
# Writing Results
Ux_values.append(Ux.copy())
Uy_values.append(Uy.copy())
P_values.append(P.copy())
Res_U_values.append(logRes_U)
Res_V_values.append(logRes_V)
Res_Mass_values.append(logRes_Mass)
# Prepping for next step
Uxo = Ux.copy()
Uyo = Uy.copy()
Po = P.copy()
iter_ = np.linspace(0, n+1, n+1)
ax.set_xlim(-0.5, n+2)
max_Res=0
min_Res=0
for k in range (0, n+1):
if(max_Res<Res_U_values[k]):
max_Res=Res_U_values[k]
if(max_Res<Res_V_values[k]):
max_Res=Res_U_values[k]
if(max_Res<Res_Mass_values[k]):
max_Res=Res_Mass_values[k]
if(min_Res>Res_U_values[k]):
min_Res=Res_U_values[k]
if(min_Res>Res_V_values[k]):
min_Res=Res_U_values[k]
if(min_Res>Res_Mass_values[k]):
min_Res=Res_Mass_values[k]
ax.set_ylim(min_Res-1, max_Res+1)
plt.plot(iter_, Res_U_values, color='red', label='Ux-Mom Residual')
plt.plot(iter_, Res_V_values, color='blue', label='Uy-Mom Residual')
plt.plot(iter_, Res_Mass_values, color='green', label='Mass Residual')
plt.legend()
plt.show()
n+=1
if(n<max_iter):
max_iter=n
max_iter
uxq_values = []
uyq_values = []
for n in range (0, max_iter):
ux1 = Ux_values[n].copy()
uxq = np.zeros((2*ny+1, 2*nx+1))
for i in range (0, ny):
for j in range (0, nx+1):
uxq[2*i+1, 2*j] = ux1[i, j]
for j in range (0, nx+1):
uxq[0, 2*j] = D[1][0, j]
uxq[-1, 2*j] = D[1][-1, j]
for i in range (0, ny-1):
for j in range (0, nx+1):
uxq[2*i+2, 2*j] = (uxq[2*i+1, 2*j] + uxq[2*i+3, 2*j])/2
for i in range (0, 2*ny):
for j in range (0, nx):
uxq[i, 2*j+1] = (uxq[i, 2*j] + uxq[i, 2*j+2])/2
uy1 = Uy_values[n].copy()
uyq = np.zeros((2*ny+1, 2*nx+1))
for i in range (0, ny+1):
for j in range (0, nx):
uyq[2*i, 2*j+1] = uy1[i, j]
for i in range (0, ny+1):
uyq[2*i, 0] = D[2][i, 0]
uyq[2*i, -1] = D[2][i, -1]
for i in range (0, ny+1):
for j in range (0, nx-1):
uyq[2*i, 2*j+2] = (uyq[2*i, 2*j+1] + uyq[2*i, 2*j+3])/2
for i in range (0, ny):
for j in range (0, 2*nx):
uyq[2*i+1, j] = (uyq[2*i, j] + uyq[2*i+2, j])/2
uxq_values.append(uxq)
uyq_values.append(uyq)
#%%
flag_Validation = input("Validation check for Lid-driven Cavity Re-[100, 1000, 5000] ?? [y/n]")
if(flag_Validation=='y'):
mpl.rcParams.update({'font.size': 10})
ghia_u100 = np.array([1, 0.84123, 0.78871, 0.73722, 0.68717, 0.23151, 0.00332, -0.13641, -0.20581, -0.2109, -0.15662, -0.1015, -0.06434, -0.04775, -0.04192, -0.03717, 0])
ghia_u1000 = np.array([1, 0.65928, 0.57492, 0.51117, 0.46604, 0.33304, 0.18719, 0.05702, -0.0608, -0.10648, -0.27805, -0.38289, -0.2973, -0.2222, -0.20196, -0.18109, 0])
ghia_u5000 = np.array([1, 0.48223, 0.4612, 0.45992, 0.46036, 0.33556, 0.20087, 0.08183, -0.03039, -0.07404, -0.22855, -0.3305, -0.40435, -0.43643, -0.42901, -0.41165, 0])
ghia_y = np.array([1, 0.9766, 0.9688, 0.9609, 0.9531, 0.8516, 0.7344, 0.6172, 0.5, 0.4531, 0.2813, 0.1719, 0.1016, 0.0703, 0.0625, 0.0547, 0])
yq = np.linspace(yMax, yMin, num=2*ny+1)
Re = rho*D[1][0, int(nx/2)]*Lx/visc
if(Re==100):
plt.scatter(ghia_u100, ghia_y, c='green', label='Ghia etal.')
plt.title('Reynolds No. 100 Mesh -[{}X{}]'.format(nx, ny))
if(Re==1000):
plt.scatter(ghia_u1000, ghia_y, c='green', label='Ghia etal.')
plt.title('Reynolds No. 1000 Mesh -[{}X{}]'.format(nx, ny))
if(Re==5000):
plt.scatter(ghia_u5000, ghia_y, c='green', label='Ghia etal.')
plt.title('Reynolds No. 5000 Mesh -[{}X{}]'.format(nx, ny))
plt.plot(uxq_values[max_iter-1][:, nx], yq, label ='Centreline X-Velocity')
plt.legend()
plt.show()
flag_Validation = input("Validation check for Fully developed flow ?? [y/n]")
if(flag_Validation=='y'):
mpl.rcParams.update({'font.size': 10})
Re = rho*D[1][int(ny/2), 0]*Ly/visc
yq = np.linspace(yMax, yMin, num=2*ny+1)
plt.plot(uxq_values[max_iter-1][:, int(8*(2*nx+1)/10)], yq, label ='Centreline X-Velocity - at X={}'.format(xMin+int(8*(2*nx+1)/10)*dx/2))
# plt.set_xlabel('Y')
plt.title('Fully Developed Flow Re = {}'.format(Re))
plt.legend()
plt.show()
xq = np.linspace(xMin, xMax, num=2*nx+1)
plt.plot(uxq_values[max_iter-1][ny, :], xq, label ='Centreline X-Velocity - at Y={}'.format(yMax-ny*dy/2))
# plt.set_xlabel('X')
plt.title('Fully Developed Flow Re = {}'.format(Re))
plt.legend()
plt.show()
#%%
postProcessor(P_values, uxq_values, uyq_values, \
xMin, xMax, yMin, yMax, nx, ny, dx, dy, Points_x, Points_y, \
caseFolder, Res_U_values, Res_V_values, Res_Mass_values, \
max_iter, min_Res, max_Res) | 2D-Panel-CFD | /2D_Panel-CFD-0.0.1.tar.gz/2D_Panel-CFD-0.0.1/src/RUN_package/RUN-spyder.py | RUN-spyder.py |
import numpy as np
def P_coeffSIMPLE(Aup, Avp, Uxs, Uys, D, nx, ny, dx, dy):
App =np.zeros((ny, nx))
Ape =np.zeros((ny, nx))
Apw =np.zeros((ny, nx))
Apn =np.zeros((ny, nx))
Aps =np.zeros((ny, nx))
Bpp =np.zeros((ny, nx))
for j in range (0, nx):
for i in range (0, ny):
if (D[0][i, j+1]!=2 and D[0][i+1, j+1]!=2):
Ape[i, j] = 0
else:
Ape[i, j] = dy**2/Aup[i, j+1]
if (D[0][i, j]!=2 and D[0][i+1, j]!=2):
Apw[i, j] = 0
else:
Apw[i, j] = dy**2/Aup[i, j]
if (D[0][i, j]!=2 and D[0][i, j+1]!=2):
Apn[i, j] = 0
else:
Apn[i, j] = dx**2/Avp[i, j]
if (D[0][i+1, j]!=2 and D[0][i+1, j+1]!=2):
Aps[i, j] = 0
else:
Aps[i, j] = dx**2/Avp[i+1, j]
App[i, j] = Ape[i, j] + Apw[i, j] + Apn[i, j] + Aps[i, j]
Bpp[i, j] = dy*(Uxs[i, j]-Uxs[i, j+1]) + \
dx*(Uys[i+1, j]-Uys[i, j])
if(D[0][i, j]==-1 or D[0][i+1, j]==-1 or D[0][i, j+1]==-1 or D[0][i+1, j+1]==-1):
App[i, j] = 1
Ape[i, j] = 0
Apw[i, j] = 0
Apn[i, j] = 0
Aps[i, j] = 0
Bpp[i, j] = 0
if(D[0][i, j]!=2 and D[0][i+1, j]!=2 and D[0][i, j+1]!=2 and D[0][i+1, j+1]!=2):
App[i, j] = 1
Ape[i, j] = 0
Apw[i, j] = 0
Apn[i, j] = 0
Aps[i, j] = 0
Bpp[i, j] = 0
M_Bpp = Bpp.flatten()
M_A = np.zeros((ny*nx, ny*nx))
ite=0
for i in range (0, ny):
for j in range (0, nx):
M_A[ite, ite] = App[i, j]
if ((ite+1)%(nx)!=0):
M_A[ite, ite+1] = -Ape[i, j]
if(ite%(nx)!=0):
M_A[ite, ite-1] = -Apw[i, j]
if (ite<(ny-1)*(nx)):
M_A[ite, ite+nx] = -Aps[i, j]
if(ite>nx-1):
M_A[ite, ite-nx] = -Apn[i, j]
ite+=1
return M_A, M_Bpp, App, Ape, Apw, Apn, Aps, Bpp | 2D-Panel-CFD | /2D_Panel-CFD-0.0.1.tar.gz/2D_Panel-CFD-0.0.1/src/RUN_package/P_coeff.py | P_coeff.py |
#%%
import numpy as np
import matplotlib as mpl
from matplotlib import cm
from matplotlib import pyplot as plt
from scipy import linalg
import sys, os
mpl.rcParams.update({'font.size': 15})
caseFolder = input("\nEnter address of testCase ...\nExample: /home/user/folder/Case_1\n")
try:
os.mkdir(caseFolder)
except:
print("\nFolder already there!!\n")
flag_EmptyFolder = input("\nEmpty Folder ?? [y/n]\n")
if(flag_EmptyFolder=='y'):
filelist = [ f for f in os.listdir(caseFolder) ]
for f in filelist:
os.remove(os.path.join(caseFolder, f))
else:
sys.exit()
package_dir = os.path.dirname(os.path.realpath(__file__))
print(package_dir)
from RUN_package.boundaryDomain import domain, obstacle, obstacleMath
from RUN_package.initialConditions import initialCond
from RUN_package.faceInterpolation import U_face, V_face
from RUN_package.UV_coeff import U_coeffUpwind, V_coeffUpwind
from RUN_package.P_coeff import P_coeffSIMPLE
from RUN_package.correction import correctionSIMPLE
from RUN_package.postProcessing import postProcessor
D, \
xMin, xMax, yMin, yMax, \
dx, dy, nx, ny, \
south_Boundary, north_Boundary, west_Boundary, east_Boundary, \
flag_Outlet =\
domain(caseFolder)
flag_obstacleGeometry = int(input("\n1. Use .txt file as Address \n2. Add elliptical obstacle\n3. No Obstacle Geometry\n"))
if(flag_obstacleGeometry==1):
D, Points, Points_x, Points_y = obstacle(D, xMin, xMax, yMin, yMax, nx, ny, dx, dy)
elif(flag_obstacleGeometry==2):
D, Points, Points_x, Points_y = obstacleMath(D, xMin, xMax, yMin, yMax, nx, ny, dx, dy)
Lx = xMax-xMin
Ly = yMax-yMin
# Grid - Plot
xg = np.linspace(xMin, xMax, num=nx+2)
yg = np.linspace(yMax, yMin, num=ny+2)
Xg, Yg = np.meshgrid(xg, yg)
fig, ax = plt.subplots(figsize =(Lx*6*1.4, Ly*6))
plt.xlabel('x [m]')
plt.ylabel('y [m]')
cmap = plt.get_cmap('tab10')
ylorbr = cm.get_cmap('tab10', 4)
norm = mpl.colors.Normalize(vmin=-1.5, vmax=2.5)
col = plt.pcolormesh(xg, yg, D[0], cmap = ylorbr, norm=norm)
cbar = plt.colorbar(col)
cbar.set_label('Grid\n-1.0 - Outlet 0.0 - Wall 1.0 - Inlet 2.0 - Wall')
plt.axis('scaled')
# show plot
plt.show()
flag_continue = input("\nContinue or Abort ?? [y/n]\n")
if(flag_continue=='n'):
sys.exit()
urf_UV = float(input("\nEnter Under-relaxation factor for UV\n"))
urf_P = float(input("\nEnter Under-relaxation factor for P\n"))
rho = float(input("\nEnter Density of fluid [kg/m^3]\n"))
visc = float(input("\nEnter Viscosity of Fluid [Pa*s]\n"))
max_iter = int(input("\nEnter Max Iteration for this SIM\n"))
MaxRes_Mass = float(input("\nEnter Min Mass Imbalance Residual for this SIM\n"))
Uxo, Uyo, Po = initialCond(D, nx, ny, flag_Outlet)
Ux_values=[]
Ux_values.append(Uxo)
Uy_values=[]
Uy_values.append(Uyo)
P_values=[]
P_values.append(Po)
Res_U_values=[]
Res_V_values=[]
Res_Mass_values=[]
logRes_Mass=0.0
n=0
iter_ = np.linspace(0, max_iter, max_iter)
temp = np.linspace(0, max_iter, max_iter)
plt.ion()
fig = plt.figure()
ax = fig.add_subplot(111)
line1, = ax.plot(iter_, temp, color = 'red', label = 'Ux-Mom Residual')
line2, = ax.plot(iter_, temp, color = 'blue', label = 'Uy-Mom Residual')
line3, = ax.plot(iter_, temp, color = 'green', label = 'Mass Residual')
ax.legend()
while n<max_iter and logRes_Mass>MaxRes_Mass:
Uue, Uuw, Vun, Vus = U_face(Uxo, Uyo, D, nx, ny)
Uve, Uvw, Vvn, Vvs = V_face(Uxo, Uyo, D, nx, ny)
M_Au, M_Bup, Aup, Aue, Auw, Aun, Aus, Bup = \
U_coeffUpwind(Uxo, Uyo, Po, Uue, Uuw, Vun, Vus, D, rho, visc, urf_UV, nx, ny, dx, dy)
M_Av, M_Bvp, Avp, Ave, Avw, Avn, Avs, Bvp = \
V_coeffUpwind(Uxo, Uyo, Po, Uve, Uvw, Vvn, Vvs, D, rho, visc, urf_UV, nx, ny, dx, dy)
# Momentum Predictor
M_Uxs = linalg.solve(M_Au, M_Bup)
M_Uys = linalg.solve(M_Av, M_Bvp)
Uxs = np.reshape(M_Uxs, (ny, nx+1))
Uys = np.reshape(M_Uys, (ny+1, nx))
M_Ap, M_Bpp, App, Ape, Apw, Apn, Aps, Bpp = P_coeffSIMPLE(Aup, Avp, Uxs, Uys, D, nx, ny, dx, dy)
if(flag_Outlet==0):
M_Ap[int(ny*nx/4), :] = 0
M_Ap[int(ny*nx/4), int(ny*nx/4)] = 1
M_Bpp[int(ny*nx/4)] = 0
# Pressure Correction values
M_Pc = linalg.solve(M_Ap, M_Bpp)
Pc = np.reshape(M_Pc, (ny, nx))
# Residuals
Res_U = M_Au.dot(Uxo.flatten())-M_Bup
Res_V = M_Av.dot(Uyo.flatten())-M_Bvp
logRes_U = np.log(np.sum(np.abs(Res_U)))
logRes_V = np.log(np.sum(np.abs(Res_V)))
logRes_Mass = np.log(np.sum(np.abs(Bpp)))
# Momentum & Pressure Correction step
Ux, Uy, P = correctionSIMPLE(Uxs, Uys, Po, Pc, D, Aup, Avp, urf_UV, urf_P, nx, ny, dx, dy)
# Writing Results
Ux_values.append(Ux.copy())
Uy_values.append(Uy.copy())
P_values.append(P.copy())
Res_U_values.append(logRes_U)
Res_V_values.append(logRes_V)
Res_Mass_values.append(logRes_Mass)
# Prepping for next step
Uxo = Ux.copy()
Uyo = Uy.copy()
Po = P.copy()
iter_ = np.linspace(0, n+1, n+1)
ax.set_xlim(-0.5, n+2)
max_Res=0
min_Res=0
for k in range (0, n+1):
if(max_Res<Res_U_values[k]):
max_Res=Res_U_values[k]
if(max_Res<Res_V_values[k]):
max_Res=Res_U_values[k]
if(max_Res<Res_Mass_values[k]):
max_Res=Res_Mass_values[k]
if(min_Res>Res_U_values[k]):
min_Res=Res_U_values[k]
if(min_Res>Res_V_values[k]):
min_Res=Res_U_values[k]
if(min_Res>Res_Mass_values[k]):
min_Res=Res_Mass_values[k]
ax.set_ylim(min_Res-1, max_Res+1)
line1.set_ydata(Res_U_values)
line1.set_xdata(iter_)
line2.set_ydata(Res_V_values)
line2.set_xdata(iter_)
line3.set_ydata(Res_Mass_values)
line3.set_xdata(iter_)
fig.canvas.draw()
fig.canvas.flush_events()
n+=1
if(n<max_iter):
max_iter=n
max_iter
uxq_values = []
uyq_values = []
for n in range (0, max_iter):
ux1 = Ux_values[n].copy()
uxq = np.zeros((2*ny+1, 2*nx+1))
for i in range (0, ny):
for j in range (0, nx+1):
uxq[2*i+1, 2*j] = ux1[i, j]
for j in range (0, nx+1):
uxq[0, 2*j] = D[1][0, j]
uxq[-1, 2*j] = D[1][-1, j]
for i in range (0, ny-1):
for j in range (0, nx+1):
uxq[2*i+2, 2*j] = (uxq[2*i+1, 2*j] + uxq[2*i+3, 2*j])/2
for i in range (0, 2*ny):
for j in range (0, nx):
uxq[i, 2*j+1] = (uxq[i, 2*j] + uxq[i, 2*j+2])/2
uy1 = Uy_values[n].copy()
uyq = np.zeros((2*ny+1, 2*nx+1))
for i in range (0, ny+1):
for j in range (0, nx):
uyq[2*i, 2*j+1] = uy1[i, j]
for i in range (0, ny+1):
uyq[2*i, 0] = D[2][i, 0]
uyq[2*i, -1] = D[2][i, -1]
for i in range (0, ny+1):
for j in range (0, nx-1):
uyq[2*i, 2*j+2] = (uyq[2*i, 2*j+1] + uyq[2*i, 2*j+3])/2
for i in range (0, ny):
for j in range (0, 2*nx):
uyq[2*i+1, j] = (uyq[2*i, j] + uyq[2*i+2, j])/2
uxq_values.append(uxq)
uyq_values.append(uyq)
#%%
flag_Validation = input("Validation check for Lid-driven Cavity Re-[100, 1000, 5000] ?? [y/n]")
if(flag_Validation=='y'):
mpl.rcParams.update({'font.size': 10})
ghia_u100 = np.array([1, 0.84123, 0.78871, 0.73722, 0.68717, 0.23151, 0.00332, -0.13641, -0.20581, -0.2109, -0.15662, -0.1015, -0.06434, -0.04775, -0.04192, -0.03717, 0])
ghia_u1000 = np.array([1, 0.65928, 0.57492, 0.51117, 0.46604, 0.33304, 0.18719, 0.05702, -0.0608, -0.10648, -0.27805, -0.38289, -0.2973, -0.2222, -0.20196, -0.18109, 0])
ghia_u5000 = np.array([1, 0.48223, 0.4612, 0.45992, 0.46036, 0.33556, 0.20087, 0.08183, -0.03039, -0.07404, -0.22855, -0.3305, -0.40435, -0.43643, -0.42901, -0.41165, 0])
ghia_y = np.array([1, 0.9766, 0.9688, 0.9609, 0.9531, 0.8516, 0.7344, 0.6172, 0.5, 0.4531, 0.2813, 0.1719, 0.1016, 0.0703, 0.0625, 0.0547, 0])
yq = np.linspace(yMax, yMin, num=2*ny+1)
Re = rho*D[1][0, int(nx/2)]*Lx/visc
if(Re==100):
plt.scatter(ghia_u100, ghia_y, c='green', label='Ghia etal.')
plt.title('Reynolds No. 100 Mesh -[{}X{}]'.format(nx, ny))
if(Re==1000):
plt.scatter(ghia_u1000, ghia_y, c='green', label='Ghia etal.')
plt.title('Reynolds No. 1000 Mesh -[{}X{}]'.format(nx, ny))
if(Re==5000):
plt.scatter(ghia_u5000, ghia_y, c='green', label='Ghia etal.')
plt.title('Reynolds No. 5000 Mesh -[{}X{}]'.format(nx, ny))
plt.plot(uxq_values[max_iter-1][:, nx], yq, label ='Centreline X-Velocity')
plt.legend()
plt.show()
flag_Validation = input("Validation check for Fully developed flow ?? [y/n]")
if(flag_Validation=='y'):
mpl.rcParams.update({'font.size': 10})
Re = rho*D[1][int(ny/2), 0]*Ly/visc
yq = np.linspace(yMax, yMin, num=2*ny+1)
plt.plot(uxq_values[max_iter-1][:, int(8*(2*nx+1)/10)], yq, label ='Centreline X-Velocity - at X={}'.format(xMin+int(8*(2*nx+1)/10)*dx/2))
# plt.set_xlabel('Y')
plt.title('Fully Developed Flow Re = {}'.format(Re))
plt.legend()
plt.show()
xq = np.linspace(xMin, xMax, num=2*nx+1)
plt.plot(uxq_values[max_iter-1][ny, :], xq, label ='Centreline X-Velocity - at Y={}'.format(yMax-ny*dy/2))
# plt.set_xlabel('X')
plt.title('Fully Developed Flow Re = {}'.format(Re))
plt.legend()
plt.show()
#%%
postProcessor(P_values, uxq_values, uyq_values, \
xMin, xMax, yMin, yMax, nx, ny, dx, dy, Points_x, Points_y, \
caseFolder, Res_U_values, Res_V_values, Res_Mass_values, \
max_iter, min_Res, max_Res) | 2D-Panel-CFD | /2D_Panel-CFD-0.0.1.tar.gz/2D_Panel-CFD-0.0.1/src/RUN_package/RUN.py | RUN.py |
# 2D Cellular Automaton
This project was inspired by discussions in MATH 340 Mathematical Excursions. While we visualized multiple starting indicies for 2D cellular automata in Excel, I knew a Python script would allow greater functioniality and easier usage.
I came across a respository on GitHub by Zhiming Wang titled [rule30](https://github.com/zmwangx/rule30). Nearly all the code is borrowed from there and made it unnecessary for me to start from scratch. All the functionalities from Wang's solution exist in this project, with the only additions being supporting multiple starting indicies.
# Table of Contents
1. [Installation](#Installation)
2. [Usage](#Usage)
4. [Credit](#Credit)
5. [License](License)
## Installation
`pip install 2DCellularAutomaton`
## Usage
```python
from CellularAutomaton import Automaton
rows = 100 #Any positive number
rule = 30 #From 1-256. More can be seen here https://mathworld.wolfram.com/ElementaryCellularAutomaton.html
starting_indicies = [20, 60] #Note this refers to the columns and columns = 2 * rows - 1, which is why rows - 1 yields center.
block_size = 1
automata = Automaton(rows=rows, rule=rule, starting_indicies=starting_indicies)
image = automata.image(block_size=block_size)
image.save('Rule 30 | Column 20 and 60.jpeg')
```
Output
<img src="image.jpeg" alt="drawing" width="600"/>
## Credit
1. Zhiming Wang's [rule30](https://github.com/zmwangx/rule30)
## License
MIT | 2D-cellular-automaton | /2D-cellular-automaton-0.0.1.tar.gz/2D-cellular-automaton-0.0.1/README.md | README.md |
import bitarray
from . import visualize
from typing import Literal
class Automaton(object):
"""
An elementary cellular automaton with muliple initial state support.
The number of rows, `rows`, is given at class instantiation, and the
automaton is only simulated to that depth. Horizontally, we only
keep the states of the 2 * `rows` - 1 cells centered around the
initial 1.
One can access the history matrix via `matrix` or as a numpy array
via `nparray`, or as a printable string via `string`, or as an image
via `image`.
Parameters
----------
rows : int
rule : int, optional
The Wolfram code for the rule (in the range [0, 255]). See
`Wolfram code <https://en.wikipedia.org/wiki/Wolfram_code>`_ on
Wikipedia. Default is 30.
starting_indicies: list[int], optional
Defaults to rows - 1.
method: str, optional
'New' results in faster generation while 'Old' utilizies the separate
generation methods dependent on parity of rule.
Nearly all of the code is from https://github.com/zmwangx/rule30.
"""
def __init__(self, rows: int, rule: int = 30,
starting_indicies: list[int] = [],
method: Literal['New', 'Old'] = 'New'):
if type(rows) is not int or rows < 0:
raise ValueError("The rows should be a positive integer.")
self._rows = rows
self._columns = rows * 2 - 1
if type(starting_indicies) is not list or not all(
[type(index) is int and index <= rows * 2 - 1
for index in starting_indicies]):
raise ValueError('The starting indicies should be a list of \
integers less than 2 * rows - 1.')
elif len(starting_indicies) == 0:
self._starting_indicies = [rows - 1]
else:
self._starting_indicies = starting_indicies
if type(rule) is not int or not 0 <= rule <= 255:
raise ValueError("The rule should be an integer between 0-255.")
self._rule = rule
# Unpack rule
u = [bit == '1' for bit in reversed('{:08b}'.format(rule))]
# Wolfram code is big-endian in terms of the bit position of
# each of the 2^3=8 configurations (just like how we write
# numbers in a left-to-right system); we map the ruleset to
# little-endian form, so that, for instance, the rule for 110 is
# stored in 0b011.
self._rule_unpacked = [
u[0b000], # 000
u[0b100], # 001
u[0b010], # 010
u[0b110], # 011
u[0b001], # 100
u[0b101], # 101
u[0b011], # 110
u[0b111], # 111
]
if type(method) is not str and method not in ['New', 'Old']:
raise ValueError('The method should be a string New or Old.')
self._method = method
self._matrix = []
self._generate()
def __str__(self):
return self.string()
@property
def rows(self):
"""Number of rows in the history matrix.
Returns
-------
int
"""
return self._rows
@property
def columns(self):
"""Number of columns in the history matrix.
Always equals 2 * `rows` - 1.
Returns
-------
int
"""
return self._columns
@property
def starting_indicies(self):
"""The custom starting indicies of the first row.
Returns
-------
list[int]
"""
return self._starting_indicies
@property
def rule(self):
"""Wolfram code for the rule.
Returns
-------
int
"""
return self._rule
@property
def method(self):
"""'New' refers to the `generate_both` function while
'Old' refers to `generate_even`and `generate_odd` depending
on the parity of rule.
Returns
-------
str
"""
return self._method
@property
def matrix(self):
"""The history matrix.
This is a list of `rows` rows, where each row is a bitarray of
length `columns`. See `bitarray's reference
<https://pypi.python.org/pypi/bitarray>`_ for more details.
This is the internal representation of the `Automaton` class.
Returns
-------
List[bitarray]
"""
return self._matrix
def string(self, zero='0', one='1'):
"""Returns a printable string representation of the matrix.
Parameters
----------
zero : str, optional
The character to print for a cell of value 0. Default is '0'.
one : str, optional
The character to print for a cell of value 1. Default is '1'.
Returns
-------
str
Examples
--------
>>> import rule30
>>> print(rule30.Automaton(5, rule=30).string())
000010000
000111000
001100100
011011110
110010001
"""
return '\n'.join([''.join([one if bit else zero for bit in row])
for row in self._matrix])
def image(self, block_size=1):
"""Returns an image for the matrix.
Parameters
----------
block_size : int, optional
Size in pixels of each cell (drawn as a square). Default is 1.
Returns
-------
PIL.Image.Image
"""
return visualize.image_from_matrix(self._matrix, block_size=block_size)
@staticmethod
def _zeros(length):
# Returns a zeroed little-endian bitarray of specified length
buf = bitarray.bitarray(length, endian='little')
buf.setall(0)
return buf
def _generate(self):
if self.method == 'New':
self._gen_both()
else:
self._gen_even() if self.rule % 2 == 0 else self._gen_odd()
def _gen_both(self):
rows = self._rows
columns = self._columns
rule_unpacked = self._rule_unpacked
row = self._zeros(columns)
for index in self.starting_indicies:
row[index] = 1 # Multiple starting states
self._matrix.append(row)
# Evolution
for i in range(1, rows):
lastrow = row
row = self._zeros(columns)
for j in range(0, columns):
row[j] = rule_unpacked[int.from_bytes(
lastrow[j - 1: j + 2].tobytes(), 'little')]
self._matrix.append(row)
# I found this function to be slower than _gen_both.
def _gen_even(self):
rows = self._rows
columns = self._columns
rule_unpacked = self._rule_unpacked
row = self._zeros(columns)
for index in self.starting_indicies:
row[index] = 1 # Multiple starting states
self._matrix.append(row)
# Evolution
for i in range(1, rows):
lastrow = row
row = self._zeros(columns)
max_left_end = max(min(self.starting_indicies) - i - 1, 1)
max_right_end = max(max(self.starting_indicies) + i, columns - 1)
for j in range(max_left_end, max_right_end):
if len(row[j - 1:]) < 3:
continue
row[j] = rule_unpacked[int.from_bytes(
lastrow[j - 1: j + 2].tobytes(), 'little')]
# Leftover code from previous version, unsure if still necessary.
# The left and right endpoints of the last row need special
# attention because we don't have all three neighbors from
# the previous row.
if i == rows - 1:
row[0] = rule_unpacked[lastrow[0] * 2 + lastrow[1] * 4]
row[columns - 1] = rule_unpacked[lastrow[columns - 2] +
lastrow[columns - 1] * 2]
self._matrix.append(row)
# I have not figured out how to approach this like with _gen_even.
# I found this function to also perform slower than _gen_both.
def _gen_odd(self):
rows = self._rows
columns = self._columns
rule_unpacked = self._rule_unpacked
# In order to compute the states of the middle (2n-1) cells on
# the n-th row, we need to start from the states of the middle
# (4n-3) cells on the first row, and step by step compute the
# middle (4n-3-2i) cells on the (i+1)-th row.
row = self._zeros(4 * rows - 3)
row[2 * rows - 1] = 1
self._matrix.append(row[rows - 1: rows - 1 + columns])
# Evolution
for i in range(1, rows):
lastrow = row
columns_to_compute = 4 * rows - 3 - 2 * i
row = self._zeros(columns_to_compute)
for j in range(columns_to_compute):
row[j] = rule_unpacked[int.from_bytes(
lastrow[j: j + 3].tobytes(), 'little')]
self._matrix.append(row[rows - 1 - i: rows - 1 - i + columns]) | 2D-cellular-automaton | /2D-cellular-automaton-0.0.1.tar.gz/2D-cellular-automaton-0.0.1/CellularAutomaton/automaton.py | automaton.py |
# 2Keys
A easy to setup second keyboard, designed for everyone.
For a full setup guide, see [here](https://github.com/Gum-Joe/2Keys/blob/master/docs/SETUP.md)
For keyboard mappings, see [here](https://github.com/Gum-Joe/2Keys/blob/master/docs/MAPPINGS.md)
### Support
Windows is supported only as the server (where the hotkeys will run) and a raspberry pi is required to run the detector.
## WARNING
This will download a copy of [AutoHotkey_H](https://hotkeyit.github.io/v2/), a DLL version of [AutoHotkey](http://autohotkey.com/)
## Building
To build the server, where hotkeys are run:
```
$ cd server
$ yarn
```
To build the detector:
```
$ cd detector
$ pip3 install -r required.txt
```
## Devices
**Server**: The device running the hotkeys sever, i.e. where the hot keys will be run
**Detecter**: Device that handles detection of key presses & which keyboard it is and sends this to the server
## Sofware used & inspiration
Inspired by LTT editor Taran's second keyboard project: [https://github.com/TaranVH/2nd-keyboard](https://github.com/TaranVH/2nd-keyboard)
2Keys uses AutoHotkey_H (a DLL version of AutoHotkey): [https://hotkeyit.github.io/v2/](https://hotkeyit.github.io/v2/)
## License
Copyright 2018 Kishan Sambhi
2Keys is free software: you can redistribute it and/or modify
it under the terms of the GNU General Public License as published by
the Free Software Foundation, either version 3 of the License, or
(at your option) any later version.
2Keys is distributed in the hope that it will be useful,
but WITHOUT ANY WARRANTY; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
GNU General Public License for more details.
You should have received a copy of the GNU General Public License
along with 2Keys. If not, see <https://www.gnu.org/licenses/>. | 2Keys | /2Keys-0.5.1.tar.gz/2Keys-0.5.1/README.md | README.md |
# This file is part of 2Keys.
# 2Keys is free software: you can redistribute it and/or modify
# it under the terms of the GNU General Public License as published by
# the Free Software Foundation, either version 3 of the License, or
# (at your option) any later version.
#
# 2Keys is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU General Public License for more details.
#
# You should have received a copy of the GNU General Public License
# along with 2Keys. If not, see <https://www.gnu.org/licenses/>.
###
# File generated by 2Keys
# DO NOT MODIFY
# Please run using sudo
# 2Keys script to register and handler systemd services for 2Keys watchers
keyboards={{ keyboards }} # Please use format (keyboard keyboard keyboard)
# Functions
register() {
echo Registering services...
for keyboard in ${keyboards[@]};
do
echo "Adding script for ${keyboard}..."
echo "Chmodding with 644..."
chmod 644 ./.2Keys/2Keys-${keyboard}.service
echo "Adding..."
systemctl enable $PWD/.2Keys/2Keys-${keyboard}.service
done
echo Reloading daemon...
systemctl daemon-reload
start
}
disable() {
echo Disabling services...
stop
for keyboard in ${keyboards[@]};
do
echo "Disabling script for ${keyboard}..."
systemctl disable 2Keys-${keyboard}.service
echo "To completely remove the scripts, remove the ./.2Keys dir."
echo "Note that you will be unable to ever reuse the 2Keys services again once this is done"
done
}
enable_scripts() {
echo Enabling services...
for keyboard in ${keyboards[@]};
do
echo "Enabling script for ${keyboard}..."
systemctl enable $PWD/.2Keys/2Keys-${keyboard}.service
done
}
start() {
echo Starting services...
for keyboard in ${keyboards[@]};
do
echo "Starting script for ${keyboard}..."
systemctl start 2Keys-${keyboard}.service
done
}
stop() {
echo Stopping services...
for keyboard in ${keyboards[@]};
do
echo "Stopping script for ${keyboard}..."
systemctl stop 2Keys-${keyboard}.service
done
}
restart() {
echo Restarting services...
for keyboard in ${keyboards[@]};
do
echo "Restarting script for ${keyboard}..."
systemctl restart 2Keys-${keyboard}.service
done
}
status() {
echo Getting statuses of services...
for keyboard in ${keyboards[@]};
do
echo "Get status of ${keyboard}..."
systemctl status 2Keys-${keyboard}.service
echo ""
done
}
# Root check
if [ "$EUID" -ne 0 ]
then echo "Please run as root"
exit 1
fi
# Parse arg
case $1 in
register)
register
;;
disable)
disable
;;
enable)
enable_scripts
;;
start)
start
;;
stop)
stop
;;
restart)
restart
;;
status)
status
;;
help)
echo Valid comands:
echo "register: Registers and enables (starts on startup) services for use"
echo "disable: Disbales services (stops them being started on startup)"
echo "enable: Enables services (used after disabled has been run)"
echo "start: Starts services"
echo "stop: Stops services"
echo "restart: Restarts services"
echo "status: Gets statuses of all services"
;;
*)
echo Invalid command.
exit 1
;;
esac | 2Keys | /2Keys-0.5.1.tar.gz/2Keys-0.5.1/twokeys/assets/register.sh | register.sh |
import sys
import requests
import json
import os
from os import path
import yaml
import colorful
from ..util.constants import SCRIPTS_ROOT, DEFAULT_PORT, MODULE_NAME
from ..util.logger import Logger
from ..util.config import load_config
from ..daemon import generate_daemon
from ..add_keyboard import add_keyboards
logger = Logger("init")
def init(**args):
logger.info("Welcome to the INIT for the detector!")
ipv4 = None
port = DEFAULT_PORT
# Use opts if we can
if args["address"] == None:
logger.info("First we need to know where to find the server")
logger.info("Enter the ipv4 address of the server below:")
ipv4 = input("")
else:
ipv4 = args["address"]
# Port arg
if args["port"] == None:
logger.info("Enter the port of the server below:")
port = input("")
else:
port = args["port"]
# Make request, get config in JSON format
logger.info("Fetching config...")
try:
config_json = requests.get("http://" + ipv4 + ":" + port + "/api/get/config")
except requests.exceptions.ConnectionError:
logger.err("Couldn't estanblish a connection to the server.")
logger.err("Please check your internet connection.")
exit() # Can't do any more
if config_json.status_code >= 400: # i.e. 404 or 500
logger.err("ERROR: Request for config unsucessful!")
logger.err("Got status code " + str(config_json.status_code) + " with response:")
logger.err(config_json.text)
logger.debug("Headers: ")
logger.debug(config_json.headers)
exit()
config = json.loads(config_json.text)
# Save config
logger.info("Saving config to " + os.getcwd() + "...")
logger.debug("Opening config file handler...")
config_file = open("config.yml", "w")
logger.debug("Writing config...")
config_file.write("# Config for 2Keys\n# ONLY FOR USE BY THE PROGRAM\n# To change the config, update it on the client and run \"2Keys config-update\" here\n" +
yaml.dump(config, default_flow_style=False))
config_file.close() # Needed so that add keyboard can read it
# Then scan for keyboards
# Since running directly from here causes issues with async not stopping etc, holding everything up
# run 2Keys add
# (essentially run in another process)
# Do for each keyboard in config.keyboards
logger.info("Running scripts to add path for keyboard input...")
# Check for --no-path-request, which implies paths are already in config
if args["no_path_request"]:
logger.warn("--no-path-request flag was given")
logger.warn("It is assumed all keyboard /dev/input paths are already given in the config.")
logger.warn("Skipping to daemon generation.")
else:
add_keyboards(config)
# Add daemons
generate_daemon(config["name"], config["keyboards"].keys()) | 2Keys | /2Keys-0.5.1.tar.gz/2Keys-0.5.1/twokeys/init/init.py | init.py |
# CLI for 2Keys
# I'm just making my own since that's easier for me to understand
import click
import sys
from ..watcher import Keyboard
from ..util import Logger, load_config, constants
from ..add_keyboard import gen_async_handler, add_keyboard
from ..init import init as init_cli
from ..sync import sync_config
from ..daemon import generate_daemon
logger = Logger("cli")
@click.group()
@click.option("--debug", is_flag=True, help="Enable debugging")
@click.option("--silent", is_flag=True, help="Don't log")
def cli(debug, silent):
return
@cli.command()
@click.option("--address", "-a", help="Specify the IPv4 address of the server")
@click.option("--port", "-p", help="Specify the port the server is running on")
@click.option("--no-path-request", is_flag=True, help="Don't run the interactive keyboard detector (assumes all /dev/input/ paths have already been put into the config on the server)")
def init(address, port, no_path_request):
init_cli(address=address, port=port, no_path_request=no_path_request)
@cli.command()
@click.option("-y", "--yes", is_flag=True, help="Don't ask for prompts")
def sync(yes):
logger.warn("This will overwrite the copy of the config on the detector. Proceed? [Y/n]")
proceed = ""
if yes:
logger.warn("-y flag was given. Proceeding...")
proceed = "y"
else:
# ASK
proceed = input("").lower()
# DO IT
if proceed == "y":
sync_config()
@cli.command()
@click.argument("keyboard", default="")
@click.option(
"--inputs-path",
"-i",
help="Provide an alternative path to use as the source of keyboard input 'files' (default: /dev/input/by-id)",
default=constants.KEYBOARDS_PATH_BASE
)
def add(keyboard, inputs_path):
add_keyboard(keyboard, gen_async_handler, inputs_path)
@cli.command()
@click.argument("keyboard")
@click.option("-n", "--no-lock", is_flag=True, help="Don't lock the keyboard")
def watch(keyboard, no_lock):
if keyboard == "":
logger.err("Please provide a keyboard to watch.")
exit()
# Keyboard specified, watch it
config = load_config()
keyboard = Keyboard(config["keyboards"][keyboard], keyboard)
if not no_lock:
try:
keyboard.lock() # Grabs keyboard
keyboard.watch_keyboard()
except (KeyboardInterrupt, SystemExit, OSError):
keyboard.unlock()
exit(0)
else:
keyboard.watch_keyboard()
# Command to generate daemons
@cli.command()
@click.argument("keyboards", nargs=-1, required=False)
def daemon_gen(keyboards):
logger.info("Generating daemon files...")
config = load_config()
keyboard_list = config["keyboards"].keys()
if keyboards != ():
# Use args instead
keyboard_list = keyboards
generate_daemon(config["name"], config["keyboards"].keys()) | 2Keys | /2Keys-0.5.1.tar.gz/2Keys-0.5.1/twokeys/cli/cli.py | cli.py |
import struct
import time
import sys
import aiofiles
import requests
import json
from evdev import InputDevice
from os import path
from ..util.constants import KEYBOARDS_PATH_BASE, KEYBOARD_EVENT_FORMAT, KEYBOARD_EVENT_SIZE, MAX_KEY_MAPS
from ..util.keyboard_map import keys as KEY_MAP
from ..util.config import load_config
from ..util.logger import Logger
logger = Logger("detect")
class Keyboard:
# keyboard: Keyboard config
# name: Name of keyboard
def __init__(self, keyboard, name):
self.config = load_config()
logger.debug("Got keyboard: " + str(keyboard))
self.keyboard = keyboard
self.name = name
# File for input that corresponds to the keyboard.
self.keyboard_path = keyboard["path"]
# Device from evdev storage
self.keyboard_device = InputDevice(self.keyboard_path)
# Array of pressed keys
# is array of booleans, with the index = key code
# i.e. if pressed_or_not[2] == true, then 2 has been pressed down. Once set to false, the key has been 'unpressed'
self.pressed_or_not = [False] * MAX_KEY_MAPS
# Current keys being pressed
self.keys = [""]
# Local stores of key mappings
self.map = KEY_MAP
# Apply mappings
if "map" in self.keyboard:
self.apply_mappings(self.keyboard["map"])
# Store hotkeys list
self.hotkeys = self.standardise_hotkeys(keyboard["hotkeys"])
# Store array of hotkeys split into chars as this makes checking easier
# current hotkey, used for when watching for an up event
self.current_hotkey_up = None
self.last_hotkey = None
# Custom mapping
# Takes in key/value of key: code and adds to map array
def apply_mappings(self, maps):
for key, code in maps.items():
logger.debug("Mapped " + key + " as (" + key + ") to code " + str(code))
self.map[code] = "(" + key + ")"
# Keyboard watcher
# TODO: Use const from evdev instead of manually checking for cleaner code and no magic numbers
def watch_keyboard(self):
logger.info("Watching for key presses on " + self.name + "...")
for event in self.keyboard_device.read_loop():
type = event.type
code = event.code
value = event.value
# We only want event type 1, as that is a key press
# If key is already pressed, ignore event provided value not 0 (key unpressed)
if (type == 1 or type == 0x1):
logger.info("Key pressed. Code %u, value %u. Mapping: %s" %
(code, value, self.map[code]))
# Set key in array
# Only done if value 1 so as to not conflict with ups
if value == 1:
self.change_key_state(code, value)
logger.debug(self.keys)
# Run alogrithm to check keys against hotkey
# Only run though if value is 0 or 1 to prevent duplicate hotkeys
if value < 2:
# Proceed with regular hotkey logic
checked_hotkey = self.check_for_hotkey(self.keys)
if checked_hotkey != False:
hotkey = self.hotkeys[checked_hotkey]
logger.info("Registered hotkey:")
logger.info(checked_hotkey)
logger.info(hotkey)
# Is is an up, down, or multi function?
if hotkey["type"] == "down" and value == 1:
self.send_hotkey(checked_hotkey, value)
elif hotkey["type"] == "up" and value == 0:
self.send_hotkey(checked_hotkey, value)
elif hotkey["type"] == "multi":
# The server handles picking the right hotkey
self.send_hotkey(checked_hotkey, value)
else:
logger.warn("Hotkey not send as it's type " + hotkey["type"])
# Set key in array
# Only done if value 0 so as to not conflict with downs
if value == 0:
self.change_key_state(code, value)
logger.debug(self.keys)
#elif type != 0 or code != 0 or value != 0:
# print("Event type %u, code %u, value %u at %d.%d" % \
# (type, code, value, tv_sec, tv_usec))
else:
# Events with code, type and value == 0 are "separator" events
print("===========================================")
# Handle change of state (down/up) of key code
# down = True
# Up (as in not pressed) = False
def change_key_state(self, code, value):
if value == 1:
# Key not yet pressed
self.pressed_or_not[code] = True
# Add to self.keys string
if isinstance(self.map[code], str):
self.keys = [combo + self.map[code] for combo in self.keys] # Add to each candidate combo
else:
# Array in use
# Add as different candidates
new_keys = []
for combo in self.keys:
for mapping in self.map[code]:
new_keys.append(combo + mapping)
self.keys = new_keys
elif value == 0:
# Key unpressed, remove
self.pressed_or_not[code] = False
# Remove from combos
if isinstance(self.map[code], str):
self.keys = [combo.replace(self.map[code], "") for combo in self.keys] # Remove from each combo
else:
# Array
for mapping in self.map[code]:
logger.debug("Checking mapping " + str(mapping))
new_keys = []
index = 0
while index < len(self.keys):
combo = self.keys[index]
if combo.find(mapping) >= 0 or combo == mapping: # Only should run if in, to avoid duplicates
new_combo = combo.replace(mapping, "")
if len(new_combo) > 0:
new_keys.append(new_combo) # Remove from each
index += 1
self.keys = new_keys
# If keys array is empty, make sure to add somewhere for next set
if len(self.keys) < 1:
self.keys = [""]
# Standardise hotkey config
# hotkey = hotkeys mappings
# Standard config:
# type: down
# func: Function
def standardise_hotkeys(self, hotkeys):
new_hotkeys = hotkeys
for key, value in new_hotkeys.items():
if isinstance(value, str):
# Only has function
new_hotkeys[key] = {
"type": "down",
"func": value
}
# Else it has to be a regular one OR a muti one as ups require type: up, muties just need an object
# The below fixes #13. where if no type was specified but a func: was the program crashes
elif "type" not in new_hotkeys[key]:
# Function is present, but no type
new_hotkeys[key]["type"] = ("down" if isinstance(value["func"], str) else "multi") # If func is str, use down, if not, use "multi"
return new_hotkeys
# Hotkey detector algorithm
# Return the key of the hotkey if hotkey
# candidate = array of hotkey strings to check
def check_for_hotkey(self, candidates):
# Check each candidate
for combo in candidates:
logger.debug("Checking hotkey in combo " + combo)
for key, mapping in self.hotkeys.items():
# Check each candidate
# Step 1: Check length. If lengths are different, hotkeys can't match
if len(key) != len(combo):
continue
# Step 2: Check if chars are equal
split_hotkey = list(key) # Split into array for easy checking
split_current_keys = list(combo)
if set(split_hotkey).issubset(set(split_current_keys)) or set(split_current_keys).issubset(set(split_hotkey)):
return key # Candidate and hotkey matches, return hotkey location
# If none are true, then this isn't the right one (function yet to return)
return False
# Hotkey sender
# Send hotkey runner command -> server
# hotkey = hotkey ref in config
# value = Value of event type (up, down) from evdev
def send_hotkey(self, hotkey, value):
logger.info("Sending hotkey %s to server..." % hotkey)
try:
data_hotkey = { "keyboard": self.name, "hotkey": hotkey, "value": value }
TYPE_JSON = {"Content-Type": "application/json"} # So the server can interpret it
requests.post("http://" + self.config["addresses"]["server"]["ipv4"] + ":" + str(self.config["addresses"]["server"]["port"]) + "/api/post/trigger", data=json.dumps(data_hotkey), headers=TYPE_JSON, timeout=2)
except requests.exceptions.ConnectionError:
logger.err("Couldn't estanblish a connection to the server.")
logger.err("Please check your internet connection.")
except requests.exceptions.Timeout:
logger.err("The request timed out")
logger.warn("This means either the server isn't running, or is busy running another hotkey.")
logger.warn("Please note the hotkey may still execute after the server has finished running the hotkeys it is currently running")
# Locks (grabs) keyboard
def lock(self):
logger.info("Locking keyboard....")
self.keyboard_device.grab()
# Unlocks (ungrabs) keyboard
def unlock(self):
logger.info("Unlocking keyboard...")
self.keyboard_device.ungrab()
# str keyboard: Keyboard file in /dev/input/by-id
class AsyncKeyboard:
# Root defaults to /dev/input/by-id
def __init__(self, keyboard, root):
# File for input that corresponds to the keyboard.
self.keyboard = path.join(root, keyboard)
# Open keyboard events file in binary mode
self.in_file = open(self.keyboard, "rb")
# Run checker
self.run = True
# IMPORTANT: Don't use non async functions in this. That includes the logger
async def keyboard_watcher(self, callback):
# Only seems to run on key press. Strange.
# Solution, as this makes it hard to stop was to add a callback to part 2
async with aiofiles.open(self.keyboard, "rb") as in_file:
event = await in_file.read(KEYBOARD_EVENT_SIZE) # Open input file
while event and self.run:
print("[ASYNC DEBUG] Key pressed on " + self.keyboard)
break;
await in_file.close()
# Stop all
if self.run:
await callback(self.keyboard)
return self.run
# Stop watching as it's no longer needed
async def stop_watch(self):
print("[DEBUG] CLASS: STOPPING " + self.keyboard)
self.run = False
return
#await self.in_file.close() | 2Keys | /2Keys-0.5.1.tar.gz/2Keys-0.5.1/twokeys/watcher/watch_keyboard.py | watch_keyboard.py |
Keyboard mappings
Copy paste from https://github.com/torvalds/linux/blob/master/include/uapi/linux/input-event-codes.h:
/*
* Keys and buttons
*
* Most of the keys/buttons are modeled after USB HUT 1.12
* (see http://www.usb.org/developers/hidpage).
* Abbreviations in the comments:
* AC - Application Control
* AL - Application Launch Button
* SC - System Control
*/
As such some c code is leftover (mostly '#define's.)
None = invalid key
An array of key mapping means multiple keys can correspond to it
Currently 249 keys are mapped
"""
from .constants import MAX_KEY_MAPS
# Applies custom tag to key names
def custom_name(name):
return "$" + name + "$"
#define KEY_(.*?) (0?[xX]?[0-9a-fA-F]+)(.*) # RegExp to convert everything in VSCode
#keys[$2] = "$1" # Replacement RegExp to convert everything in VSCode
keys = [""] * MAX_KEY_MAPS # Index corresponds to code
keys[0] = None # RESERVED
keys[1] = custom_name("ESC")
keys[2] = "1"
keys[3] = "2"
keys[4] = "3"
keys[5] = "4"
keys[6] = "5"
keys[7] = "6"
keys[8] = "7"
keys[9] = "8"
keys[10] = "9"
keys[11] = "0"
keys[12] = "-" # MINUS
keys[13] = "=" # EQUAL
keys[14] = custom_name("BACKSPACE") # BACKSPACE
keys[15] = custom_name("TAB")
keys[16] = "Q"
keys[17] = "W"
keys[18] = "E"
keys[19] = "R"
keys[20] = "T"
keys[21] = "Y"
keys[22] = "U"
keys[23] = "I"
keys[24] = "O"
keys[25] = "P"
keys[26] = ["[", "{"] # LEFTBRACE
keys[27] = ["]", "}"] # RIGHTBRACE
keys[28] = custom_name("ENTER") # ENTER
keys[29] = ["<^", "^"] # LEFTCTRL
keys[30] = "A"
keys[31] = "S"
keys[32] = "D"
keys[33] = "F"
keys[34] = "G"
keys[35] = "H"
keys[36] = "J"
keys[37] = "K"
keys[38] = "L"
keys[39] = ";" # SEMICOLON
keys[40] = "\'" # APOSTROPHE
keys[41] = "`" # GRAVE
keys[42] = ["<+", "+"] # LEFTSHIFT
keys[43] = "\\" # BACKSLASH
keys[44] = "Z"
keys[45] = "X"
keys[46] = "C"
keys[47] = "V"
keys[48] = "B"
keys[49] = "N"
keys[50] = "M"
keys[51] = "," # COMMA
keys[52] = "." # DOT
keys[53] = "/" # SLASH
keys[54] = ["+>", "+"] # RIGHTSHIFT
keys[55] = custom_name("NUM_*") # KEYPAD Asterisk
keys[56] = ["<!", "!"] # LEFTALT
keys[57] = custom_name("SPACE")
keys[58] = custom_name("CAPS") # CAPSLOCK
keys[59] = "F1"
keys[60] = "F2"
keys[61] = "F3"
keys[62] = "F4"
keys[63] = "F5"
keys[64] = "F6"
keys[65] = "F7"
keys[66] = "F8"
keys[67] = "F9"
keys[68] = "F10"
keys[69] = custom_name("NUMLOCK") # NUMLOCK
keys[70] = custom_name("SCROLLLOCK")
# For these, NUM was KP
keys[71] = "NUM7"
keys[72] = "NUM8"
keys[73] = "NUM9"
keys[74] = custom_name("NUM_-")
keys[75] = "NUM4"
keys[76] = "NUM5"
keys[77] = "NUM6"
keys[78] = custom_name("NUM_+")
keys[79] = "NUM1"
keys[80] = "NUM2"
keys[81] = "NUM3"
keys[82] = "NUM0"
keys[83] = custom_name("NUM_.")
# Not included because internationlisation is a pain & I have no idea what char to reference as
# Could be reenabled by surround ZENKAKUHANKAKU with $s and using as: (example config)
# hotkeys:
# $ZENKAKUHANKAKU$: SomeFunction
# Referenced as IC (Internationlisation confusion)
# keys[85] = "ZENKAKUHANKAKU"
keys[86] = custom_name("#") # 102ND key added to UK keyboard
keys[87] = "F11"
keys[88] = "F12"
# keys[89] = "RO" # No idea what this is, IC?
# keys[90] = "KATAKANA" # IC
# keys[91] = "HIRAGANA" # IC
# keys[92] = "HENKAN" # IC
# keys[93] = "KATAKANAHIRAGANA" # IC
# keys[94] = "MUHENKAN" # IC
# keys[95] = "KPJPCOMMA" # No idea what key this is
keys[96] = custom_name("NUM_ENTER") # KPENTER; _ used for readability
keys[97] = ["^>", "^"] # RIGHTCTRL; _ used for readability
keys[98] = custom_name("NUM_/") # KPSLASH
keys[99] = custom_name("PRINT_SCR") # Print Screen & SYSREQ. SYSREQ is used for recovery. Thus, should throw an error if the user tries to use so the system can be recovered
keys[100] = ["!>", "!"]
keys[101] = custom_name("LINE_FEED") # New line char, backwards-compatibility; _ used for readability
keys[102] = custom_name("HOME")
keys[103] = custom_name("UP")
keys[104] = custom_name("PAGE_UP") # _ used for readability
keys[105] = custom_name("LEFT")
keys[106] = custom_name("RIGHT")
keys[107] = custom_name("END")
keys[108] = custom_name("DOWN")
keys[109] = custom_name("PAGE_DOWN") # _ used for readability
keys[110] = custom_name("INSERT")
keys[111] = custom_name("DELETE")
keys[112] = custom_name("MACRO") # Macro key from old keyboards. NOT new macro keys from i.e. Corsair keyboard. Backwards-compatibility
keys[113] = custom_name("MUTE") # KEY_MIN_INTERESTING (see bottom)
keys[114] = custom_name("VOL_DOWN") # VOLUMEUP. _ used for readability
keys[115] = custom_name("VOL_UP") # VOLUMEUP
keys[116] = None # POWER; Throw error so the user can power the system down
keys[117] = custom_name("NUM_=") # Numpad equals
keys[118] = custom_name("NUM_+-") # KPPLUSMINUS
keys[119] = custom_name("PAUSE")
keys[120] = custom_name("SCALE")
keys[121] = custom_name("NUM_,") # KP Comma
# keys[122] = "HANGEUL" # IC
#define KEY_HANGUEL KEY_HANGEUL # Leftover from Linux
keys[123] = "HANJA" # IC
keys[124] = "YEN" # IC
keys[125] = [custom_name("LEFTMETA"), "#", "<#"] # IC, Windows key
keys[126] = [custom_name("RIGHTMETA"), "#", "#>"] # IC, Windows key
keys[127] = [custom_name("COMPOSE"), custom_name("CONTEXT_MENU")] # Compose two chars, useful as a mode key. Also the context menu key
keys[128] = custom_name("STOP") # Stop key, don't know what for. NOT MEDIA CONTROL
keys[129] = custom_name("AGAIN")
keys[130] = custom_name("PROPS")
keys[131] = custom_name("UNDO")
keys[132] = custom_name("FRONT")
keys[133] = custom_name("COPY")
keys[134] = custom_name("OPEN")
keys[135] = custom_name("PASTE")
keys[136] = custom_name("FIND")
keys[137] = custom_name("CUT")
keys[138] = custom_name("HELP")
keys[139] = custom_name("MENU")
keys[140] = custom_name("CALC")
keys[141] = custom_name("SETUP")
keys[142] = custom_name("SLEEP")
keys[143] = custom_name("WAKEUP")
keys[144] = custom_name("FILE")
keys[145] = custom_name("SENDFILE")
keys[146] = custom_name("DELETEFILE")
keys[147] = custom_name("XFER")
keys[148] = custom_name("PROG1")
keys[149] = custom_name("PROG2")
keys[150] = custom_name("WWW")
keys[151] = custom_name("MSDOS")
keys[152] = [custom_name("COFFEE"), custom_name("SCREEN_LOCK")]
#define KEY_SCREENLOCK KEY_COFFEE
keys[153] = custom_name("ROTATE_DISPLAY")
#define KEY_DIRECTION KEY_ROTATE_DISPLAY
keys[154] = custom_name("CYCLE_WINDOWS")
keys[155] = custom_name("MAIL")
keys[156] = custom_name("BOOKMARKS")
keys[157] = custom_name("COMPUTER")
keys[158] = custom_name("BACK")
keys[159] = custom_name("FORWARD")
keys[160] = custom_name("CLOSE_CD")
keys[161] = custom_name("EJECT_CD")
keys[162] = custom_name("EJECT_CLOSE_CD")
keys[163] = custom_name("MEDIA_NEXT") # CD
keys[164] = custom_name("MEDIA_PLAY_PAUSE") # CD
keys[165] = custom_name("MEDIA_PREVIOUS") # CD
keys[166] = custom_name("MEDIA_STOP") # CD
keys[167] = custom_name("MEDIA_RECORD") # RECORD
keys[168] = custom_name("MEDIA_REWIND") # REWIND
keys[169] = custom_name("PHONE") # Media Select Telephone
keys[170] = custom_name("ISO")
keys[171] = custom_name("CONFIG")
keys[172] = custom_name("APP_HOMEPAGE") # AC
keys[173] = custom_name("APP_REFRESH") # AC
keys[174] = custom_name("APP_EXIT") # AC
keys[175] = custom_name("MOVE")
keys[176] = custom_name("EDIT")
keys[177] = custom_name("SCROLL_UP")
keys[178] = custom_name("SCROLL_DOWN")
# keys[179] = custom_name("KPLEFTPAREN") # What is this?
# keys[180] = custom_name("KPRIGHTPAREN") # What is this?
keys[181] = custom_name("APP_NEW") # AC
keys[182] = custom_name("APP_REDO") # AC
keys[183] = "F13"
keys[184] = "F14"
keys[185] = "F15"
keys[186] = "F16"
keys[187] = "F17"
keys[188] = "F18"
keys[189] = "F19"
keys[190] = "F20"
keys[191] = "F21"
keys[192] = "F22"
keys[193] = "F23"
keys[194] = "F24"
keys[200] = custom_name("MEDIA_PLAY") # Different to MEDIA_PLAY_PAUSE, this button can only play not pause. (CDs)
keys[201] = custom_name("MEDIA_PAUSE")
# keys[202] = custom_name("PROG3") # Programmable buttons?
# keys[203] = custom_name("PROG4") # Programmable buttons?
keys[204] = custom_name("DASHBOARD")
keys[205] = custom_name("SUSPEND")
keys[206] = custom_name("CLOSE")
keys[207] = custom_name("PLAY") # How is this different to keys[200]
keys[208] = custom_name("MEDIA_FASTFORWARD")
keys[209] = custom_name("SOUND_BASSBOOST")
keys[210] = custom_name("PRINT")
keys[211] = custom_name("HP")
keys[212] = custom_name("CAMERA")
keys[213] = custom_name("SOUND")
keys[214] = custom_name("QUESTION")
keys[215] = custom_name("EMAIL")
keys[216] = custom_name("CHAT")
keys[217] = custom_name("SEARCH")
keys[218] = custom_name("CONNECT")
keys[219] = custom_name("FINANCE") # Launch Finance app
keys[220] = custom_name("SPORT")
keys[221] = custom_name("SHOP")
keys[222] = custom_name("ALTERASE")
keys[223] = custom_name("CANCEL")
keys[224] = custom_name("BRIGHTNESS_DOWN")
keys[225] = custom_name("BRIGHTNESS_UP")
keys[226] = custom_name("MEDIA")
keys[227] = custom_name("SWITCHVIDEOMODE") # "Cycle between available video outputs"
# keys[228] = custom_name("KBDILLUMTOGGLE") # What are these?
# keys[229] = custom_name("KBDILLUMDOWN") # What are these?
# keys[230] = custom_name("KBDILLUMUP") # What are these?
keys[231] = custom_name("APP_SEND") # AC
keys[232] = custom_name("APP_REPLY") # AC
keys[233] = custom_name("APP_FORWARDMAIL") # AC
keys[234] = custom_name("APP_SAVE") # AC
keys[235] = custom_name("DOCUMENTS")
keys[236] = custom_name("BATTERY")
keys[237] = custom_name("SET_BLUETOOTH") # Bluetooth on/off
keys[238] = custom_name("SET_WLAN") # Wifi on/off
keys[239] = custom_name("SET_UWB") # UWB is a radio type (like 2G, 3G, wifi)
keys[240] = custom_name("UNKNOWN") # As it says, unknown
keys[241] = custom_name("VIDEO_SOURCE_NEXT") # next video source
keys[242] = custom_name("VIDEO_SOURCE_PREV") # previous video source
keys[243] = custom_name("BRIGHTNESS_CYCLE") # "brightness up, after max is min"
keys[244] = custom_name("BRIGHTNESS_AUTO") # Auto brightness, "rely on ambient"
#define KEY_BRIGHTNESS_ZERO KEY_BRIGHTNESS_AUTO
keys[245] = custom_name("DISPLAY_OFF")
keys[246] = custom_name("SET_WWAN") # "Wireless WAN (LTE, UMTS, GSM, etc.)". KEY_WWAN
#define KEY_WIMAX KEY_WWAN # ALias for above
keys[247] = custom_name("RFKILL") # "Key that controls all radios"
keys[248] = custom_name("MICMUTE") # Microphone mute
# "Code 255 is reserved for special needs of AT keyboard driver"
# I think this is mouse stuff, so it remains commented out
#define BTN_MISC 0x100
#define BTN_0 0x100
#define BTN_1 0x101
#define BTN_2 0x102
#define BTN_3 0x103
#define BTN_4 0x104
#define BTN_5 0x105
#define BTN_6 0x106
#define BTN_7 0x107
#define BTN_8 0x108
#define BTN_9 0x109
#define BTN_MOUSE 0x110
#define BTN_LEFT 0x110
#define BTN_RIGHT 0x111
#define BTN_MIDDLE 0x112
#define BTN_SIDE 0x113
#define BTN_EXTRA 0x114
#define BTN_FORWARD 0x115
#define BTN_BACK 0x116
#define BTN_TASK 0x117
#define BTN_JOYSTICK 0x120
#define BTN_TRIGGER 0x120
#define BTN_THUMB 0x121
#define BTN_THUMB2 0x122
#define BTN_TOP 0x123
#define BTN_TOP2 0x124
#define BTN_PINKIE 0x125
#define BTN_BASE 0x126
#define BTN_BASE2 0x127
#define BTN_BASE3 0x128
#define BTN_BASE4 0x129
#define BTN_BASE5 0x12a
#define BTN_BASE6 0x12b
#define BTN_DEAD 0x12f
#define BTN_GAMEPAD 0x130
#define BTN_SOUTH 0x130
#define BTN_A BTN_SOUTH
#define BTN_EAST 0x131
#define BTN_B BTN_EAST
#define BTN_C 0x132
#define BTN_NORTH 0x133
#define BTN_X BTN_NORTH
#define BTN_WEST 0x134
#define BTN_Y BTN_WEST
#define BTN_Z 0x135
#define BTN_TL 0x136
#define BTN_TR 0x137
#define BTN_TL2 0x138
#define BTN_TR2 0x139
#define BTN_SELECT 0x13a
#define BTN_START 0x13b
#define BTN_MODE 0x13c
#define BTN_THUMBL 0x13d
#define BTN_THUMBR 0x13e
#define BTN_DIGI 0x140
#define BTN_TOOL_PEN 0x140
#define BTN_TOOL_RUBBER 0x141
#define BTN_TOOL_BRUSH 0x142
#define BTN_TOOL_PENCIL 0x143
#define BTN_TOOL_AIRBRUSH 0x144
#define BTN_TOOL_FINGER 0x145
#define BTN_TOOL_MOUSE 0x146
#define BTN_TOOL_LENS 0x147
#define BTN_TOOL_QUINTTAP 0x148 /* Five fingers on trackpad */
#define BTN_STYLUS3 0x149
#define BTN_TOUCH 0x14a
#define BTN_STYLUS 0x14b
#define BTN_STYLUS2 0x14c
#define BTN_TOOL_DOUBLETAP 0x14d
#define BTN_TOOL_TRIPLETAP 0x14e
#define BTN_TOOL_QUADTAP 0x14f /* Four fingers on trackpad */
#define BTN_WHEEL 0x150
#define BTN_GEAR_DOWN 0x150
#define BTN_GEAR_UP 0x151
# NOTE: 0x160 != decimal 160
# Below here is the rest
# These are untouched since I have no idea if people would want to use them
# plus it would take too long to adjust them all
# Thus, they are disabled
# keys[0x160] = "OK"
# keys[0x161] = "SELECT"
# keys[0x162] = "GOTO"
# keys[0x163] = "CLEAR"
# keys[0x164] = "POWER2"
# keys[0x165] = "OPTION"
# keys[0x166] = "INFO"
# keys[0x167] = "TIME"
# keys[0x168] = "VENDOR"
# keys[0x169] = "ARCHIVE"
# keys[0x16a] = "PROGRAM"
# keys[0x16b] = "CHANNEL"
# keys[0x16c] = "FAVORITES"
# keys[0x16d] = "EPG"
# keys[0x16e] = "PVR"
# keys[0x16f] = "MHP"
# keys[0x170] = "LANGUAGE"
# keys[0x171] = "TITLE"
# keys[0x172] = "SUBTITLE"
# keys[0x173] = "ANGLE"
# keys[0x174] = "ZOOM"
# keys[0x175] = "MODE"
# keys[0x176] = "KEYBOARD"
# keys[0x177] = "SCREEN"
# keys[0x178] = "PC"
# keys[0x179] = "TV"
# keys[0x17a] = "TV2"
# keys[0x17b] = "VCR"
# keys[0x17c] = "VCR2"
# keys[0x17d] = "SAT"
# keys[0x17e] = "SAT2"
# keys[0x17f] = "CD"
# keys[0x180] = "TAPE"
# keys[0x181] = "RADIO"
# keys[0x182] = "TUNER"
# keys[0x183] = "PLAYER"
# keys[0x184] = "TEXT"
# keys[0x185] = "DVD"
# keys[0x186] = "AUX"
# keys[0x187] = "MP3"
# keys[0x188] = "AUDIO"
# keys[0x189] = "VIDEO"
# keys[0x18a] = "DIRECTORY"
# keys[0x18b] = "LIST"
# keys[0x18c] = "MEMO"
# keys[0x18d] = "CALENDAR"
# keys[0x18e] = "RED"
# keys[0x18f] = "GREEN"
# keys[0x190] = "YELLOW"
# keys[0x191] = "BLUE"
# keys[0x192] = "CHANNELUP"
# keys[0x193] = "CHANNELDOWN"
# keys[0x194] = "FIRST"
# keys[0x195] = "LAST"
# keys[0x196] = "AB"
# keys[0x197] = "NEXT"
# keys[0x198] = "RESTART"
# keys[0x199] = "SLOW"
# keys[0x19a] = "SHUFFLE"
# keys[0x19b] = "BREAK"
# keys[0x19c] = "PREVIOUS"
# keys[0x19d] = "DIGITS"
# keys[0x19e] = "TEEN"
# keys[0x19f] = "TWEN"
# keys[0x1a0] = "VIDEOPHONE"
# keys[0x1a1] = "GAMES"
# keys[0x1a2] = "ZOOMIN"
# keys[0x1a3] = "ZOOMOUT"
# keys[0x1a4] = "ZOOMRESET"
# keys[0x1a5] = "WORDPROCESSOR"
# keys[0x1a6] = "EDITOR"
# keys[0x1a7] = "SPREADSHEET"
# keys[0x1a8] = "GRAPHICSEDITOR"
# keys[0x1a9] = "PRESENTATION"
# keys[0x1aa] = "DATABASE"
# keys[0x1ab] = "NEWS"
# keys[0x1ac] = "VOICEMAIL"
# keys[0x1ad] = "ADDRESSBOOK"
# keys[0x1ae] = "MESSENGER"
# keys[0x1af] = "DISPLAYTOGGLE"
#define KEY_BRIGHTNESS_TOGGLE KEY_DISPLAYTOGGLE
# keys[0x1b0] = "SPELLCHECK"
# keys[0x1b1] = "LOGOFF"
# keys[0x1b2] = "DOLLAR"
# keys[0x1b3] = "EURO"
# keys[0x1b4] = "FRAMEBACK"
# keys[0x1b5] = "FRAMEFORWARD"
# keys[0x1b6] = "CONTEXT_MENU"
# keys[0x1b7] = "MEDIA_REPEAT"
# keys[0x1b8] = "10CHANNELSUP"
# keys[0x1b9] = "10CHANNELSDOWN"
# keys[0x1ba] = "IMAGES"
# keys[0x1c0] = "DEL_EOL"
# keys[0x1c1] = "DEL_EOS"
# keys[0x1c2] = "INS_LINE"
# keys[0x1c3] = "DEL_LINE"
# keys[0x1d0] = "FN"
# keys[0x1d1] = "FN_ESC"
# keys[0x1d2] = "FN_F1"
# keys[0x1d3] = "FN_F2"
# keys[0x1d4] = "FN_F3"
# keys[0x1d5] = "FN_F4"
# keys[0x1d6] = "FN_F5"
# keys[0x1d7] = "FN_F6"
# keys[0x1d8] = "FN_F7"
# keys[0x1d9] = "FN_F8"
# keys[0x1da] = "FN_F9"
# keys[0x1db] = "FN_F10"
# keys[0x1dc] = "FN_F11"
# keys[0x1dd] = "FN_F12"
# keys[0x1de] = "FN_1"
# keys[0x1df] = "FN_2"
# keys[0x1e0] = "FN_D"
# keys[0x1e1] = "FN_E"
# keys[0x1e2] = "FN_F"
# keys[0x1e3] = "FN_S"
# keys[0x1e4] = "FN_B"
# keys[0x1f1] = "BRL_DOT1"
# keys[0x1f2] = "BRL_DOT2"
# keys[0x1f3] = "BRL_DOT3"
# keys[0x1f4] = "BRL_DOT4"
# keys[0x1f5] = "BRL_DOT5"
# keys[0x1f6] = "BRL_DOT6"
# keys[0x1f7] = "BRL_DOT7"
# keys[0x1f8] = "BRL_DOT8"
# keys[0x1f9] = "BRL_DOT9"
# keys[0x1fa] = "BRL_DOT10"
# keys[0x200] = "NUMERIC_0"
# keys[0x201] = "NUMERIC_1"
# keys[0x202] = "NUMERIC_2"
# keys[0x203] = "NUMERIC_3"
# keys[0x204] = "NUMERIC_4"
# keys[0x205] = "NUMERIC_5"
# keys[0x206] = "NUMERIC_6"
# keys[0x207] = "NUMERIC_7"
# keys[0x208] = "NUMERIC_8"
# keys[0x209] = "NUMERIC_9"
# keys[0x20a] = "NUMERIC_STAR"
# keys[0x20b] = "NUMERIC_POUND"
# keys[0x20c] = "NUMERIC_A"
# keys[0x20d] = "NUMERIC_B"
# keys[0x20e] = "NUMERIC_C"
# keys[0x20f] = "NUMERIC_D"
# keys[0x210] = "CAMERA_FOCUS"
# keys[0x211] = "WPS_BUTTON"
# keys[0x212] = "TOUCHPAD_TOGGLE"
# keys[0x213] = "TOUCHPAD_ON"
# keys[0x214] = "TOUCHPAD_OFF"
# keys[0x215] = "CAMERA_ZOOMIN"
# keys[0x216] = "CAMERA_ZOOMOUT"
# keys[0x217] = "CAMERA_UP"
# keys[0x218] = "CAMERA_DOWN"
# keys[0x219] = "CAMERA_LEFT"
# keys[0x21a] = "CAMERA_RIGHT"
# keys[0x21b] = "ATTENDANT_ON"
# keys[0x21c] = "ATTENDANT_OFF"
# keys[0x21d] = "ATTENDANT_TOGGLE"
# keys[0x21e] = "LIGHTS_TOGGLE"
#define BTN_DPAD_UP 0x220
#define BTN_DPAD_DOWN 0x221
#define BTN_DPAD_LEFT 0x222
#define BTN_DPAD_RIGHT 0x223
# keys[0x230] = "ALS_TOGGLE"
# keys[0x231] = "ROTATE_LOCK_TOGGLE"
# keys[0x240] = "BUTTONCONFIG"
# keys[0x241] = "TASKMANAGER"
# keys[0x242] = "JOURNAL"
# keys[0x243] = "CONTROLPANEL"
# keys[0x244] = "APPSELECT"
# keys[0x245] = "SCREENSAVER"
# keys[0x246] = "VOICECOMMAND"
# keys[0x247] = "ASSISTANT"
# keys[0x250] = "BRIGHTNESS_MIN"
# keys[0x251] = "BRIGHTNESS_MAX"
# keys[0x260] = "KBDINPUTASSIST_PREV"
# keys[0x261] = "KBDINPUTASSIST_NEXT"
# keys[0x262] = "KBDINPUTASSIST_PREVGROUP"
# keys[0x263] = "KBDINPUTASSIST_NEXTGROUP"
# keys[0x264] = "KBDINPUTASSIST_ACCEPT"
# keys[0x265] = "KBDINPUTASSIST_CANCEL"
# Diagonal movement keys
# keys[0x266] = "RIGHT_UP"
# keys[0x267] = "RIGHT_DOWN"
# keys[0x268] = "LEFT_UP"
# keys[0x269] = "LEFT_DOWN"
# keys[0x26a] = "ROOT_MENU"
# Show Top Menu of the Media (e.g. DVD) */
# keys[0x26b] = "MEDIA_TOP_MENU"
# keys[0x26c] = "NUMERIC_11"
# keys[0x26d] = "NUMERIC_12"
#
# "Toggle Audio Description: refers to an audio service that helps blind and
# visually impaired consumers understand the action in a program. Note: in
# some countries this is referred to as "Video Description"."
#
# keys[0x26e] = "AUDIO_DESC"
# keys[0x26f] = "3D_MODE"
# keys[0x270] = "NEXT_FAVORITE"
# keys[0x271] = "STOP_RECORD"
# keys[0x272] = "PAUSE_RECORD"
# keys[0x273] = "VOD"
# keys[0x274] = "UNMUTE"
# keys[0x275] = "FASTREVERSE"
# keys[0x276] = "SLOWREVERSE"
###
# "Control a data application associated with the currently viewed channel,
# e.g. teletext or data broadcast application (MHEG, MHP, HbbTV, etc.)"
###
# keys[0x277] = "DATA"
# keys[0x278] = "ONSCREEN_KEYBOARD"
# WHAT ARE THESE:
#define BTN_TRIGGER_HAPPY 0x2c0
#define BTN_TRIGGER_HAPPY1 0x2c0
#define BTN_TRIGGER_HAPPY2 0x2c1
#define BTN_TRIGGER_HAPPY3 0x2c2
#define BTN_TRIGGER_HAPPY4 0x2c3
#define BTN_TRIGGER_HAPPY5 0x2c4
#define BTN_TRIGGER_HAPPY6 0x2c5
#define BTN_TRIGGER_HAPPY7 0x2c6
#define BTN_TRIGGER_HAPPY8 0x2c7
#define BTN_TRIGGER_HAPPY9 0x2c8
#define BTN_TRIGGER_HAPPY10 0x2c9
#define BTN_TRIGGER_HAPPY11 0x2ca
#define BTN_TRIGGER_HAPPY12 0x2cb
#define BTN_TRIGGER_HAPPY13 0x2cc
#define BTN_TRIGGER_HAPPY14 0x2cd
#define BTN_TRIGGER_HAPPY15 0x2ce
#define BTN_TRIGGER_HAPPY16 0x2cf
#define BTN_TRIGGER_HAPPY17 0x2d0
#define BTN_TRIGGER_HAPPY18 0x2d1
#define BTN_TRIGGER_HAPPY19 0x2d2
#define BTN_TRIGGER_HAPPY20 0x2d3
#define BTN_TRIGGER_HAPPY21 0x2d4
#define BTN_TRIGGER_HAPPY22 0x2d5
#define BTN_TRIGGER_HAPPY23 0x2d6
#define BTN_TRIGGER_HAPPY24 0x2d7
#define BTN_TRIGGER_HAPPY25 0x2d8
#define BTN_TRIGGER_HAPPY26 0x2d9
#define BTN_TRIGGER_HAPPY27 0x2da
#define BTN_TRIGGER_HAPPY28 0x2db
#define BTN_TRIGGER_HAPPY29 0x2dc
#define BTN_TRIGGER_HAPPY30 0x2dd
#define BTN_TRIGGER_HAPPY31 0x2de
#define BTN_TRIGGER_HAPPY32 0x2df
#define BTN_TRIGGER_HAPPY33 0x2e0
#define BTN_TRIGGER_HAPPY34 0x2e1
#define BTN_TRIGGER_HAPPY35 0x2e2
#define BTN_TRIGGER_HAPPY36 0x2e3
#define BTN_TRIGGER_HAPPY37 0x2e4
#define BTN_TRIGGER_HAPPY38 0x2e5
#define BTN_TRIGGER_HAPPY39 0x2e6
#define BTN_TRIGGER_HAPPY40 0x2e7
# We avoid low common keys in module aliases so they don't get huge.
#define KEY_MIN_INTERESTING KEY_MUTE # From original c code
# i.e. KEY_MUTE is the minimum interesting one
# Have no idea if this is correct
# NOTE: Should be 0x300
keys[MAX_KEY_MAPS - 1] = None # MAX keys is 0x2ff
#define KEY_CNT (KEY_MAX+1) # No ides what this is | 2Keys | /2Keys-0.5.1.tar.gz/2Keys-0.5.1/twokeys/util/keyboard_map.py | keyboard_map.py |
import asyncio
from os import path, listdir, getcwd, system
import colorful
from ..util.constants import KEYBOARDS_PATH_BASE, KEYBOARD_EVENT_FORMAT, KEYBOARD_EVENT_SIZE, SCRIPTS_ROOT, MODULE_NAME
from ..util.logger import Logger
from ..util.config import load_config
from ..watcher import AsyncKeyboard as AsyncKeyboardWatcher
logger = Logger("detect")
# Function to add keyboards (s is emphasised) from config
def add_keyboards(config):
for key, value in config["keyboards"].items():
logger.info("Running script to add keyboard for keyboard " + colorful.cyan(key) + "...")
print("") # Padding
system("cd " + getcwd() + " && python3 -m " + MODULE_NAME + " add " + key)
print("") # Padding
def add_keyboard(name, gen_handler, inputs_path):
# Check if paths not given
config = load_config()
if name == "" or name not in config["keyboards"]:
logger.warn("No keyboard supplied.")
logger.warn("Detection will be ran on all keyboards.")
logger.warn("To just generate daemons, use the 'daemon-gen' command")
logger.info("Running detection on all keyboards...")
return add_keyboards(config)
logger.info("Mapping keyboard " + name)
logger.info("Scanning for keyboards...")
if not path.isdir(inputs_path): # Make sure there's something to detect
logger.err("Couldn't scan for keyboards")
logger.err("Verify you have at least one keyboard plugged in")
logger.err("and the dir " + inputs_path + " exists")
logger.err("You can specify a custom path with the --inputs-path option")
exit()
# Scan
# From https://stackoverflow.com/questions/3207219/how-do-i-list-all-files-of-a-directory
keyboards = listdir(inputs_path)
logger.debug("Keyboards:")
logger.debug(keyboards)
logger.info("Press a button on the keyboard you want to map to register it.")
# Then watch all keyboards and ask for one to be pressed
keyboards_events = [AsyncKeyboardWatcher(keyboard_path, inputs_path) for keyboard_path in keyboards] # Keyboard watch classes for each input
handler = gen_handler(keyboards_events, name) # The handler needs access to keyboards_events, which it won't on exe in the watcher, as well as keyboard name
# Run
jobs = [keyboards_events[i].keyboard_watcher(handler) for i in range(0, len(keyboards))] # Create jobs list
loop = asyncio.get_event_loop()
loop.run_until_complete(asyncio.wait(jobs)) | 2Keys | /2Keys-0.5.1.tar.gz/2Keys-0.5.1/twokeys/add_keyboard/add_keyboard.py | add_keyboard.py |
from .add_keyboard import add_keyboard
import sys
import os
import signal
import aiofiles
from ..util import Logger
import yaml
from .sync_keyboard_path import update_server_keyboard_path
logger = Logger("add")
PID = os.getpid()
# IMPORTANT: Don't use non async functions in this. That includes the logger
# EXCEPTIONS ARE NOT CAUGHT
def gen_async_handler(keyboards, keyboard_name):
async def handler(keyboard):
print("[DEBUG] STOPPING WATCH")
# Stop each keyboard object one by one, then write config
for keyboard_stop in keyboards:
print("[DEBUG] ROOT: STOPPING " + keyboard_stop.keyboard)
await keyboard_stop.stop_watch()
# Write config
logger.info("Writing keyboard " + keyboard + " as " + keyboard_name)
logger.debug("Opening config...")
# 1: Open current file for updating
async with aiofiles.open(os.getcwd() + "/config.yml", mode="r") as config_file:
logger.debug("ASYNC FILE OPS") # DEBUG: signal start of async file ops, so as to help detect where program breaks
config_contents = await config_file.read() # Read config
logger.debug("Contents:\n" + config_contents)
# Parse it into python obj
config = yaml.load(config_contents, Loader=yaml.FullLoader)
logger.debug("Parsed contents: " + str(config))
config["keyboards"][keyboard_name]["path"] = keyboard # Update keyboard with path in /dev/input
logger.debug("Writing config...")
# r+ appends, so we have to create a new stream so we cam write
async with aiofiles.open("config.yml", mode="w") as config_write:
await config_write.write("# Config for 2Keys\n# ONLY FOR USE BY THE PROGRAM\n# To change the config, update it on the client and run \"2Keys config-update\" here\n" +
yaml.dump(config, default_flow_style=False)) # Write it
await config_write.close() # Close so other programs can use
logger.info("Config writen.")
logger.info("Updating path on server....")
await update_server_keyboard_path(keyboard_name, keyboard)
os.kill(PID, signal.SIGTERM) # Exit() does't work, so we have to self kill the script
exit() # So only one ^C is needed to end the program
return
return handler | 2Keys | /2Keys-0.5.1.tar.gz/2Keys-0.5.1/twokeys/add_keyboard/async_handler.py | async_handler.py |
import sys
import os
import stat
import pystache
from ..util.logger import Logger
from ..util.constants import DAEMON_TEMPLATE_PATH, SCRIPTS_ROOT, LOCAL_ROOT, DAEMON_TEMPLATE_SCRIPT_PATH
logger = Logger("daemon")
# Generates a systemd unit file
# Name: Name of 2Keys project
# Keyboards: Array of keyboard names
def generate_daemon(name, keyboards):
logger.info("Creating systemd unit scripts...")
template = open(DAEMON_TEMPLATE_PATH, "r").read() # Open template
for keyboard in keyboards:
script = pystache.render(template, {
"name": name,
"index_path": "2Keys",
"keyboard": keyboard,
"detector_path": SCRIPTS_ROOT,
"version": str(sys.version_info[0]) + "." + str(sys.version_info[1]),
"pwd": os.getcwd()
})
if not os.path.exists(LOCAL_ROOT):
logger.info("Making local root ./.2Keys...")
os.makedirs(LOCAL_ROOT)
UNIT_FILE_NAME = "2Keys-%s.service" % keyboard
logger.info("Creating unit file {}...".format(UNIT_FILE_NAME))
unitFile = open(LOCAL_ROOT + "/" + UNIT_FILE_NAME, "w")
logger.info("Writing...")
unitFile.write(script)
logger.info("Writing a shell script to manage the unit files (services/daemons)...")
shTemplate = open(DAEMON_TEMPLATE_SCRIPT_PATH, "r").read()
keyboard_string = "("
# Create array of keyboards
for keyboard in keyboards:
keyboard_string += keyboard + " "
# End array
keyboard_string = keyboard_string[0:-1] + ")"
# Render mustache template
shScript = pystache.render(shTemplate, {
"keyboards": keyboard_string
})
shScriptFile = open(LOCAL_ROOT + "/register.sh", "w")
shScriptFile.write(shScript)
logger.info("Making executable...")
# From https://stackoverflow.com/questions/12791997/how-do-you-do-a-simple-chmod-x-from-within-python
st = os.stat(LOCAL_ROOT + "/register.sh")
os.chmod(LOCAL_ROOT + "/register.sh", st.st_mode | stat.S_IEXEC)
logger.info("")
logger.info("Generated unit files to start 2Keys on startup!")
logger.info("To install the services for use, please run:")
logger.info(" sudo bash ./.2Keys/register.sh register")
logger.info("For help on how to use the script:")
logger.info(" sudo bash ./.2Keys/register.sh help") | 2Keys | /2Keys-0.5.1.tar.gz/2Keys-0.5.1/twokeys/daemon/scripts.py | scripts.py |
import enum
from typing import Tuple
USAGE = """\
i2w <x> <z>: convert coordinate in 8K radar image to the corresponding coordinate in 2b2t world
image2world <x> <z>: convert coordinate in 8K radar image to the corresponding coordinate in 2b2t world
w2i <x> <z>: convert coordinate in 2b2t world to the corresponding coordinate in 8K radar image
world2image <x> <z>: convert coordinate in 2b2t world to the corresponding coordinate in 8K radar image
q (or quit, exit): exit program
h (or help): show this help menu\
"""
class RadarImageType(enum.Enum):
RADAR_4K = (3840, 2160, 8)
RADAR_8K = (7680, 4320, 4)
def world_to_image(loc, image_type=RadarImageType.RADAR_8K) -> Tuple[int, int]:
"""
Given a coordinate in 2b2t overworld, return the corresponding pixel coordinate in radar image.
"""
x, z = loc
off_x, off_z, chunks_per_pixel = image_type.value[0] // 2, image_type.value[1] // 2, image_type.value[2]
return 3840 + x // 16 // chunks_per_pixel, 2160 + z // 16 // chunks_per_pixel
def image_to_world(loc, image_type=RadarImageType.RADAR_8K) -> Tuple[int, int]:
"""
Given a position in radar image, return the center coordinate of the corresponding range in 2b2t overworld.
"""
x, z = loc
off_x, off_z, chunks_per_pixel = image_type.value[0] // 2, image_type.value[1] // 2, image_type.value[2]
x, z = x - off_x, z - off_z
return int((x + 0.5) * 16 * chunks_per_pixel), int((z + 0.5) * 16 * chunks_per_pixel)
def main():
""" REPL """
while True:
try:
inp = input('> ').strip().split(' ') or None
cmd = inp[0] if len(inp) > 0 else None
if cmd == 'i2w' or cmd == 'image2world':
world_x, world_y = image_to_world((int(inp[1]), int(inp[2])))
print(f'World: ({world_x}, {world_y})')
print(f'Nether: ({world_x // 8}, {world_y // 8})')
elif cmd == 'w2i' or cmd == 'world2image':
print(world_to_image((int(inp[1]), int(inp[2]))))
elif cmd == 'q' or cmd == 'quit' or cmd == 'exit':
break
elif cmd == 'h' or cmd == 'help':
print(USAGE)
elif not cmd:
pass
else:
print('Invalid command. Run \'help\' or \'h\' for usage description.')
except (ValueError, IndexError):
print('Invalid command. Type `help` or `h` for help.')
except KeyboardInterrupt:
print()
pass # Ignore Ctrl-C event
if __name__ == '__main__':
main() | 2b2t | /2b2t-0.3.0.tar.gz/2b2t-0.3.0/bbtt/coord/__main__.py | __main__.py |
# 2b2t.py
This package handles all things 2b2t!
## Installation
You can install the package with pip.
```
pip install 2b2t
```
or
```
pip3 install 2b2t
```
Or you can install with GitHub
1. Clone the repo
2. Run setup.py
3. Install Python3 if you don't already have it
4. Install requests and colorama if you don't already have it
```
git clone https://github.com/BGP0/2b2t.py.git
python3 setup.py build install
pip3 install requests
pip3 install colorama
```
## Usage
[Docs](https://github.com/BGP0/2b2t.py/wiki)
## Credits
Orginal Software made by BGP
Thanks to SkilzMastr for turning the checker into a package!
| 2b2t.py | /2b2t.py-1.7.6.tar.gz/2b2t.py-1.7.6/README.md | README.md |
import requests
from colorama import Fore, init
import threading
class listCheck():
def __init__(self, usernames, printOut=False):
self.usernames = usernames
self.printOut = printOut
init(convert=True)
def listBased(self, usernameList):
for username in usernameList:
self.check(username)
def check(self, username):
data = f'ign={username}'
headers = {
'Content-Type': 'application/x-www-form-urlencoded'
}
request = requests.request('POST', "https://donate.2b2t.org/category/738999", data=data, headers=headers)
if self.printOut:
if 'rate limited' in request.text:
print(Fore.LIGHTMAGENTA_EX + f"YOU'VE BEEN RATELIMITED!! :(")
elif 'not a valid' in request.text:
print(Fore.LIGHTRED_EX + f"{username} is not a valid username")
elif 'Unable' in request.text:
print(Fore.LIGHTRED_EX + f"Unable to find a player with the username: {input}")
elif 'banned' not in request.text:
print(Fore.LIGHTRED_EX + f"{username} is not currently banned")
else:
print(Fore.LIGHTGREEN_EX + f"{username} is currently banned")
else:
if 'rate limited' in request.text:
return 0
elif 'not a valid' in request.text:
return 1
elif 'Unable' in request.text:
return 2
elif 'banned' not in request.text:
return False
else:
return True
def l1(self):
for i in range(len(self.lines1)):
self.check(self.lines1[i])
def l2(self):
for i in range(len(self.lines2)):
self.check(self.lines2[i])
def start(self):
if self.usernames is list:
self.listBased(self.usernames)
else:
self.lines = [item.replace("\n", "") for item in open(self.usernames, 'r').readlines()]
self.lines1 = self.lines[:len(self.lines)//2]
self.lines2 = self.lines[len(self.lines)//2:]
self.threads = []
self.t1 = threading.Thread(target=self.l1)
self.t2 = threading.Thread(target=self.l2)
self.threads.append(self.t1)
self.threads.append(self.t2)
self.t1.start()
self.t2.start()
print(('\nFinished loading all threads.\n').center(119))
for x in self.threads:
x.join()
input(Fore.RESET + 'Finished Checking!') | 2b2t.py | /2b2t.py-1.7.6.tar.gz/2b2t.py-1.7.6/py2b/listCheck.py | listCheck.py |
# Python Module for 2Captcha API
The easiest way to quickly integrate [2Captcha] captcha solving service into your code to automate solving of any types of captcha.
- [Python Module for 2Captcha API](#python-module-for-2captcha-api)
- [Installation](#installation)
- [Configuration](#configuration)
- [TwoCaptcha instance options](#twocaptcha-instance-options)
- [Solve captcha](#solve-captcha)
- [Captcha options](#captcha-options)
- [Normal Captcha](#normal-captcha)
- [Text Captcha](#text-captcha)
- [ReCaptcha v2](#recaptcha-v2)
- [ReCaptcha v3](#recaptcha-v3)
- [FunCaptcha](#funcaptcha)
- [GeeTest](#geetest)
- [hCaptcha](#hcaptcha)
- [GeeTest v4](#geetest-v4)
- [Lemin Cropped Captcha](#lemin-cropped-captcha)
- [Cloudflare Turnstile](#cloudflare-turnstile)
- [Amazon WAF](#amazon-waf)
- [KeyCaptcha](#keycaptcha)
- [Capy](#capy)
- [Grid](#grid)
- [Canvas](#canvas)
- [ClickCaptcha](#clickcaptcha)
- [Rotate](#rotate)
- [Other methods](#other-methods)
- [send / getResult](#send--getresult)
- [balance](#balance)
- [report](#report)
- [Error handling](#error-handling)
- [Proxies](#proxies)
- [Async calls](#async-calls)
## Installation
This package can be installed with Pip:
```pip3 install 2captcha-python```
## Configuration
TwoCaptcha instance can be created like this:
```python
from twocaptcha import TwoCaptcha
solver = TwoCaptcha('YOUR_API_KEY')
```
Also there are few options that can be configured:
```python
config = {
'server': '2captcha.com',
'apiKey': 'YOUR_API_KEY',
'softId': 123,
'callback': 'https://your.site/result-receiver',
'defaultTimeout': 120,
'recaptchaTimeout': 600,
'pollingInterval': 10,
}
solver = TwoCaptcha(**config)
```
### TwoCaptcha instance options
| Option | Default value | Description |
| ---------------- | -------------- | -------------------------------------------------------------------------------------------------------------------------------------------------- |
| server | `2captcha.com` | API server. You can set it to `rucaptcha.com` if your account is registered there |
| softId | - | your software ID obtained after publishing in [2captcha sofware catalog] |
| callback | - | URL of your web-sever that receives the captcha recognition result. The URl should be first registered in [pingback settings] of your account |
| defaultTimeout | 120 | Polling timeout in seconds for all captcha types except ReCaptcha. Defines how long the module tries to get the answer from `res.php` API endpoint |
| recaptchaTimeout | 600 | Polling timeout for ReCaptcha in seconds. Defines how long the module tries to get the answer from `res.php` API endpoint |
| pollingInterval | 10 | Interval in seconds between requests to `res.php` API endpoint, setting values less than 5 seconds is not recommended |
> **IMPORTANT:** once `callback` is defined for `TwoCaptcha` instance, all methods return only the captcha ID and DO NOT poll the API to get the result. The result will be sent to the callback URL.
To get the answer manually use [getResult method](#send--getresult)
## Solve captcha
When you submit any image-based captcha use can provide additional options to help 2captcha workers to solve it properly.
### Captcha options
| Option | Default Value | Description |
| ------------- | ------------- | -------------------------------------------------------------------------------------------------- |
| numeric | 0 | Defines if captcha contains numeric or other symbols [see more info in the API docs][post options] |
| minLen | 0 | minimal answer lenght |
| maxLen | 0 | maximum answer length |
| phrase | 0 | defines if the answer contains multiple words or not |
| caseSensitive | 0 | defines if the answer is case sensitive |
| calc | 0 | defines captcha requires calculation |
| lang | - | defines the captcha language, see the [list of supported languages] |
| hintImg | - | an image with hint shown to workers with the captcha |
| hintText | - | hint or task text shown to workers with the captcha |
Below you can find basic examples for every captcha type. Check out [examples directory] to find more examples with all available options.
### Normal Captcha
To bypass a normal captcha (distorted text on image) use the following method. This method also can be used to recognize any text on the image.
```python
result = solver.normal('path/to/captcha.jpg', param1=..., ...)
# OR
result = solver.normal('https://site-with-captcha.com/path/to/captcha.jpg', param1=..., ...)
```
### Text Captcha
This method can be used to bypass a captcha that requires to answer a question provided in clear text.
```python
result = solver.text('If tomorrow is Saturday, what day is today?', param1=..., ...)
```
### ReCaptcha v2
Use this method to solve ReCaptcha V2 and obtain a token to bypass the protection.
```python
result = solver.recaptcha(sitekey='6Le-wvkSVVABCPBMRTvw0Q4Muexq1bi0DJwx_mJ-',
url='https://mysite.com/page/with/recaptcha',
param1=..., ...)
```
### ReCaptcha v3
This method provides ReCaptcha V3 solver and returns a token.
```python
result = solver.recaptcha(sitekey='6Le-wvkSVVABCPBMRTvw0Q4Muexq1bi0DJwx_mJ-',
url='https://mysite.com/page/with/recaptcha',
version='v3',
param1=..., ...)
```
### FunCaptcha
FunCaptcha (Arkoselabs) solving method. Returns a token.
```python
result = solver.funcaptcha(sitekey='6Le-wvkSVVABCPBMRTvw0Q4Muexq1bi0DJwx_mJ-',
url='https://mysite.com/page/with/funcaptcha',
param1=..., ...)
```
### GeeTest
Method to solve GeeTest puzzle captcha. Returns a set of tokens as JSON.
```python
result = solver.geetest(gt='f1ab2cdefa3456789012345b6c78d90e',
challenge='12345678abc90123d45678ef90123a456b',
url='https://www.site.com/page/',
param1=..., ...)
```
### hCaptcha
Use this method to solve hCaptcha challenge. Returns a token to bypass captcha.
```python
result = solver.hcaptcha(sitekey='10000000-ffff-ffff-ffff-000000000001',
url='https://www.site.com/page/',
param1=..., ...)
```
### GeeTest v4
Use this method to solve GeeTest v4. Returns the response in JSON.
```python
result = solver.geetest_v4(captcha_id='e392e1d7fd421dc63325744d5a2b9c73',
url='https://www.site.com/page/',
param1=..., ...)
```
### Lemin Cropped Captcha
Use this method to solve hCaptcha challenge. Returns JSON with answer containing the following values: answer, challenge_id.
```python
result = solver.lemin(captcha_id='CROPPED_1abcd2f_a1234b567c890d12ef3a456bc78d901d',
div_id='lemin-cropped-captcha',
url='https://www.site.com/page/',
param1=..., ...)
```
### Cloudflare Turnstile
Use this method to solve Cloudflare Turnstile. Returns JSON with the token.
```python
result = solver.turnstile(sitekey='0x1AAAAAAAAkg0s2VIOD34y5',
url='http://mysite.com/',
param1=..., ...)
```
### Amazon WAF
Use this method to solve Amazon WAF Captcha also known as AWS WAF Captcha is a part of Intelligent threat mitigation for Amazon AWS. Returns JSON with the token.
```python
result = solver.amazon_waf(sitekey='0x1AAAAAAAAkg0s2VIOD34y5',
iv='CgAHbCe2GgAAAAAj',
context='9BUgmlm48F92WUoqv97a49ZuEJJ50TCk9MVr3C7WMtQ0X6flVbufM4n8mjFLmbLVAPgaQ1Jydeaja94iAS49ljb+sUNLoukWedAQZKrlY4RdbOOzvcFqmD/ZepQFS9N5w15Exr4VwnVq+HIxTsDJwRviElWCdzKDebN/mk8/eX2n7qJi5G3Riq0tdQw9+C4diFZU5E97RSeahejOAAJTDqduqW6uLw9NsjJBkDRBlRjxjn5CaMMo5pYOxYbGrM8Un1JH5DMOLeXbq1xWbC17YSEoM1cRFfTgOoc+VpCe36Ai9Kc='
url='https://non-existent-example.execute-api.us-east-1.amazonaws.com/latest'
param1=..., ...)
```
### KeyCaptcha
Token-based method to solve KeyCaptcha.
```python
result = solver.keycaptcha(s_s_c_user_id=10,
s_s_c_session_id='493e52c37c10c2bcdf4a00cbc9ccd1e8',
s_s_c_web_server_sign='9006dc725760858e4c0715b835472f22-pz-',
s_s_c_web_server_sign2='2ca3abe86d90c6142d5571db98af6714',
url='https://www.keycaptcha.ru/demo-magnetic/',
param1=..., ...)
```
### Capy
Token-based method to bypass Capy puzzle captcha.
```python
result = solver.capy(sitekey='PUZZLE_Abc1dEFghIJKLM2no34P56q7rStu8v',
url='http://mysite.com/',
api_server='https://jp.api.capy.me/',
param1=..., ...)
```
### Grid
Grid method is originally called Old ReCaptcha V2 method. The method can be used to bypass any type of captcha where you can apply a grid on image and need to click specific grid boxes. Returns numbers of boxes.
```python
result = solver.grid('path/to/captcha.jpg', param1=..., ...)
```
### Canvas
Canvas method can be used when you need to draw a line around an object on image. Returns a set of points' coordinates to draw a polygon.
```python
result = solver.canvas('path/to/captcha.jpg', param1=..., ...)
```
### ClickCaptcha
ClickCaptcha method returns coordinates of points on captcha image. Can be used if you need to click on particular points on the image.
```python
result = solver.coordinates('path/to/captcha.jpg', param1=..., ...)
```
### Rotate
This method can be used to solve a captcha that asks to rotate an object. Mostly used to bypass FunCaptcha. Returns the rotation angle.
```python
result = solver.rotate('path/to/captcha.jpg', param1=..., ...)
```
## Other methods
### send / getResult
These methods can be used for manual captcha submission and answer polling.
```python
import time
. . . . .
id = solver.send(file='path/to/captcha.jpg')
time.sleep(20)
code = solver.get_result(id)
```
### balance
Use this method to get your account's balance
```python
balance = solver.balance()
```
### report
Use this method to report good or bad captcha answer.
```python
solver.report(id, True) # captcha solved correctly
solver.report(id, False) # captcha solved incorrectly
```
### Error handling
In case of an error, the captcha solver throws an exception. It's important to properly handle these cases. We recommend using `try except` to handle exceptions.
```python
try:
result = solver.text('If tomorrow is Saturday, what day is today?')
except ValidationException as e:
# invalid parameters passed
print(e)
except NetworkException as e:
# network error occurred
print(e)
except ApiException as e:
# api respond with error
print(e)
except TimeoutException as e:
# captcha is not solved so far
print(e)
```
### Proxies
You can pass your proxy as an additional argument for methods: recaptcha, funcaptcha and geetest. The proxy will be forwarded to the API to solve the captcha.
```python
proxy={
'type': 'HTTPS',
'uri': 'login:password@IP_address:PORT'
}
```
### Async calls
You can also make async calls with [asyncio], for example:
```python
import asyncio
import concurrent.futures
from twocaptcha import TwoCaptcha
captcha_result = await captchaSolver(image)
async def captchaSolver(image):
loop = asyncio.get_running_loop()
with concurrent.future.ThreadPoolExecutor() as pool:
result = await loop.run_in_executor(pool, lambda: TwoCaptcha(API_KEY).normal(image))
return result
```
[2Captcha]: https://2captcha.com/
[2captcha sofware catalog]: https://2captcha.com/software
[pingback settings]: https://2captcha.com/setting/pingback
[post options]: https://2captcha.com/2captcha-api#normal_post
[list of supported languages]: https://2captcha.com/2captcha-api#language
[examples directory]: /examples
[asyncio]: https://docs.python.org/3/library/asyncio.html
| 2captcha-python | /2captcha-python-1.2.1.tar.gz/2captcha-python-1.2.1/README.md | README.md |
import os, sys
import time
import requests
from base64 import b64encode
try:
from .api import ApiClient
except ImportError:
from api import ApiClient
class SolverExceptions(Exception):
pass
class ValidationException(SolverExceptions):
pass
class NetworkException(SolverExceptions):
pass
class ApiException(SolverExceptions):
pass
class TimeoutException(SolverExceptions):
pass
class TwoCaptcha():
def __init__(self,
apiKey,
softId=None,
callback=None,
defaultTimeout=120,
recaptchaTimeout=600,
pollingInterval=10,
server = '2captcha.com'):
self.API_KEY = apiKey
self.soft_id = softId
self.callback = callback
self.default_timeout = defaultTimeout
self.recaptcha_timeout = recaptchaTimeout
self.polling_interval = pollingInterval
self.api_client = ApiClient(post_url = str(server))
self.max_files = 9
self.exceptions = SolverExceptions
def normal(self, file, **kwargs):
'''
Wrapper for solving normal captcha (image)
Required:
file (image, base64, or url)
Optional params:
phrase
numeric
minLen
maxLen
phrase
caseSensitive
calc
lang
hintText
hintImg
softId
callback
proxy = {'type': 'HTTPS', 'uri': 'login:password@IP_address:PORT'})
'''
method = self.get_method(file)
result = self.solve(**method, **kwargs)
return result
def text(self, text, **kwargs):
'''
Wrapper for solving text captcha
Required:
text
Optional params:
lang
softId
callback
'''
result = self.solve(text=text, method='post', **kwargs)
return result
def recaptcha(self, sitekey, url, version='v2', enterprise=0, **kwargs):
'''
Wrapper for solving recaptcha (v2, v3)
Required:
sitekey
url
Optional params:
invisible
version
enterprise
action
score
softId
callback
proxy = {'type': 'HTTPS', 'uri': 'login:password@IP_address:PORT'})
'''
params = {
'googlekey': sitekey,
'url': url,
'method': 'userrecaptcha',
'version': version,
'enterprise': enterprise,
**kwargs,
}
result = self.solve(timeout=self.recaptcha_timeout, **params)
return result
def funcaptcha(self, sitekey, url, **kwargs):
'''
Wrapper for solving funcaptcha
Required:
sitekey
url
Optional params:
surl
userAgent
softId
callback
proxy = {'type': 'HTTPS', 'uri': 'login:password@IP_address:PORT'})
**{'data[key]': 'anyStringValue'}
'''
result = self.solve(publickey=sitekey,
url=url,
method='funcaptcha',
**kwargs)
return result
def geetest(self, gt, challenge, url, **kwargs):
'''
Wrapper for solving geetest captcha
Required:
gt
challenge
url
Optional params:
apiServer
softId
callback
proxy = {'type': 'HTTPS', 'uri': 'login:password@IP_address:PORT'})
'''
result = self.solve(gt=gt,
challenge=challenge,
url=url,
method='geetest',
**kwargs)
return result
def hcaptcha(self, sitekey, url, **kwargs):
'''
Wrapper for solving hcaptcha
Required:
sitekey
url
Optional params:
invisible
data
softId
callback
proxy = {'type': 'HTTPS', 'uri': 'login:password@IP_address:PORT'})
'''
result = self.solve(sitekey=sitekey,
url=url,
method='hcaptcha',
**kwargs)
return result
def keycaptcha(self, s_s_c_user_id, s_s_c_session_id,
s_s_c_web_server_sign, s_s_c_web_server_sign2, url,
**kwargs):
'''
Wrapper for solving
Required:
s_s_c_user_id
s_s_c_session_id
s_s_c_web_server_sign
s_s_c_web_server_sign2
url
Optional params:
softId
callback
proxy = {'type': 'HTTPS', 'uri': 'login:password@IP_address:PORT'})
'''
params = {
's_s_c_user_id': s_s_c_user_id,
's_s_c_session_id': s_s_c_session_id,
's_s_c_web_server_sign': s_s_c_web_server_sign,
's_s_c_web_server_sign2': s_s_c_web_server_sign2,
'url': url,
'method': 'keycaptcha',
**kwargs,
}
result = self.solve(**params)
return result
def capy(self, sitekey, url, **kwargs):
'''
Wrapper for solving capy
Required:
sitekey
url
Optional params:
softId
callback
proxy = {'type': 'HTTPS', 'uri': 'login:password@IP_address:PORT'})
'''
result = self.solve(captchakey=sitekey,
url=url,
method='capy',
**kwargs)
return result
def grid(self, file, **kwargs):
'''
Wrapper for solving grid captcha (image)
Required:
file (image or base64)
Optional params:
rows
cols
previousId
canSkip
lang
hintImg
hintText
softId
callback
proxy = {'type': 'HTTPS', 'uri': 'login:password@IP_address:PORT'})
'''
method = self.get_method(file)
params = {
'recaptcha': 1,
**method,
**kwargs,
}
result = self.solve(**params)
return result
def canvas(self, file, **kwargs):
'''
Wrapper for solving canvas captcha (image)
Required:
file (image or base64)
Optional params:
previousId
canSkip
lang
hintImg
hintText
softId
callback
proxy = {'type': 'HTTPS', 'uri': 'login:password@IP_address:PORT'})
'''
if not ('hintText' in kwargs or 'hintImg' in kwargs):
raise ValidationException(
'parameters required: hintText and/or hintImg')
method = self.get_method(file)
params = {
'recaptcha': 1,
'canvas': 1,
**method,
**kwargs,
}
result = self.solve(**params)
return result
def coordinates(self, file, **kwargs):
'''
Wrapper for solving coordinates captcha (image)
Required:
file (image or base64)
Optional params:
hintImg
hintText
lang
softId
callback
proxy = {'type': 'HTTPS', 'uri': 'login:password@IP_address:PORT'})
'''
method = self.get_method(file)
params = {
'coordinatescaptcha': 1,
**method,
**kwargs,
}
result = self.solve(**params)
return result
def rotate(self, files, **kwargs):
'''
Wrapper for solving rotate captcha (image)
Required:
files (images)
Optional params:
angle
lang
hintImg
hintText
softId
callback
proxy = {'type': 'HTTPS', 'uri': 'login:password@IP_address:PORT'})
'''
if isinstance(files, str):
file = self.get_method(files)['file']
result = self.solve(file=file, method='rotatecaptcha', **kwargs)
return result
elif isinstance(files, dict):
files = list(files.values())
files = self.extract_files(files)
result = self.solve(files=files, method='rotatecaptcha', **kwargs)
return result
def geetest_v4(self, captcha_id, url, **kwargs):
'''
Wrapper for solving geetest_v4 captcha
Required:
captcha_id
url
Optional params:
'''
result = self.solve(captcha_id=captcha_id,
url=url,
method='geetest_v4',
**kwargs)
return result
def lemin(self, captcha_id, div_id, url, **kwargs):
'''
Wrapper for solving Lemin Cropped Captcha
Required:
captcha_id
div_id
url
Optional params:
'''
result = self.solve(captcha_id=captcha_id,
div_id=div_id,
url=url,
method='lemin',
**kwargs)
return result
def turnstile(self, sitekey, url, **kwargs):
'''
Wrapper for solving Cloudflare Turnstile
Required:
sitekey
url
Optional params:
action
data
'''
result = self.solve(sitekey=sitekey,
url=url,
method='turnstile',
**kwargs)
return result
def amazon_waf(self, sitekey, iv, context, url, **kwargs):
'''
Wrapper for solving Amazon WAF
Required:
sitekey
iv
context
url
Optional params:
'''
result = self.solve(sitekey=sitekey,
iv=iv,
context=context,
url=url,
method='amazon_waf',
**kwargs)
return result
def solve(self, timeout=0, polling_interval=0, **kwargs):
'''
sends captcha, receives result
Parameters
----------
timeout : float
polling_interval : int
**kwargs : all captcha params
Returns
-------
result : string
'''
id_ = self.send(**kwargs)
result = {'captchaId': id_}
if self.callback is None:
timeout = float(timeout or self.default_timeout)
sleep = int(polling_interval or self.polling_interval)
code = self.wait_result(id_, timeout, sleep)
result.update({'code': code})
return result
def wait_result(self, id_, timeout, polling_interval):
max_wait = time.time() + timeout
while time.time() < max_wait:
try:
return self.get_result(id_)
except NetworkException:
time.sleep(polling_interval)
raise TimeoutException(f'timeout {timeout} exceeded')
def get_method(self, file):
if not file:
raise ValidationException('File required')
if not '.' in file and len(file) > 50:
return {'method': 'base64', 'body': file}
if file.startswith('http'):
img_resp = requests.get(file)
if img_resp.status_code != 200:
raise ValidationException(f'File could not be downloaded from url: {file}')
return {'method': 'base64', 'body': b64encode(img_resp.content).decode('utf-8')}
if not os.path.exists(file):
raise ValidationException(f'File not found: {file}')
return {'method': 'post', 'file': file}
def send(self, **kwargs):
params = self.default_params(kwargs)
params = self.rename_params(params)
params, files = self.check_hint_img(params)
response = self.api_client.in_(files=files, **params)
if not response.startswith('OK|'):
raise ApiException(f'cannot recognize response {response}')
return response[3:]
def get_result(self, id_):
response = self.api_client.res(key=self.API_KEY, action='get', id=id_)
if response == 'CAPCHA_NOT_READY':
raise NetworkException
if not response.startswith('OK|'):
raise ApiException(f'cannot recognize response {response}')
return response[3:]
def balance(self):
'''
get my balance
Returns
-------
balance : float
'''
response = self.api_client.res(key=self.API_KEY, action='getbalance')
return float(response)
def report(self, id_, correct):
'''
report of solved captcha: good/bad
Parameters
----------
id_ : captcha ID
correct : True/False
Returns
-------
None.
'''
rep = 'reportgood' if correct else 'reportbad'
self.api_client.res(key=self.API_KEY, action=rep, id=id_)
return
def rename_params(self, params):
replace = {
'caseSensitive': 'regsense',
'minLen': 'min_len',
'maxLen': 'max_len',
'minLength': 'min_len',
'maxLength': 'max_len',
'hintText': 'textinstructions',
'hintImg': 'imginstructions',
'url': 'pageurl',
'score': 'min_score',
'text': 'textcaptcha',
'rows': 'recaptcharows',
'cols': 'recaptchacols',
'previousId': 'previousID',
'canSkip': 'can_no_answer',
'apiServer': 'api_server',
'softId': 'soft_id',
'callback': 'pingback',
'datas': 'data-s',
}
new_params = {
v: params.pop(k)
for k, v in replace.items() if k in params
}
proxy = params.pop('proxy', '')
proxy and new_params.update({
'proxy': proxy['uri'],
'proxytype': proxy['type']
})
new_params.update(params)
return new_params
def default_params(self, params):
params.update({'key': self.API_KEY})
callback = params.pop('callback', self.callback)
soft_id = params.pop('softId', self.soft_id)
if callback: params.update({'callback': callback})
if soft_id: params.update({'softId': soft_id})
self.has_callback = bool(callback)
return params
def extract_files(self, files):
if len(files) > self.max_files:
raise ValidationException(
f'Too many files (max: {self.max_files})')
not_exists = [f for f in files if not (os.path.exists(f))]
if not_exists:
raise ValidationException(f'File not found: {not_exists}')
files = {f'file_{e+1}': f for e, f in enumerate(files)}
return files
def check_hint_img(self, params):
hint = params.pop('imginstructions', None)
files = params.pop('files', {})
if not hint:
return params, files
if not '.' in hint and len(hint) > 50:
return params, files
if not os.path.exists(hint):
raise ValidationException(f'File not found: {hint}')
if not files:
files = {'file': params.pop('file', {})}
files.update({'imginstructions': hint})
return params, files
if __name__ == '__main__':
key = sys.argv[1]
sol = TwoCaptcha(key) | 2captcha-python | /2captcha-python-1.2.1.tar.gz/2captcha-python-1.2.1/twocaptcha/solver.py | solver.py |
import requests
class NetworkException(Exception):
pass
class ApiException(Exception):
pass
class ApiClient():
def __init__(self, post_url = '2captcha.com'):
self.post_url = post_url
def in_(self, files={}, **kwargs):
'''
sends POST-request (files and/or params) to solve captcha
Parameters
----------
files : TYPE, optional
DESCRIPTION. The default is {}.
**kwargs : TYPE
DESCRIPTION.
Raises
------
NetworkException
DESCRIPTION.
ApiException
DESCRIPTION.
Returns
-------
resp : TYPE
DESCRIPTION.
'''
try:
current_url = 'https://'+self.post_url+'/in.php'
if files:
files = {key: open(path, 'rb') for key, path in files.items()}
resp = requests.post(current_url,
data=kwargs,
files=files)
[f.close() for f in files.values()]
elif 'file' in kwargs:
with open(kwargs.pop('file'), 'rb') as f:
resp = requests.post(current_url,
data=kwargs,
files={'file': f})
else:
resp = requests.post(current_url,
data=kwargs)
except requests.RequestException as e:
raise NetworkException(e)
if resp.status_code != 200:
raise NetworkException(f'bad response: {resp.status_code}')
resp = resp.content.decode('utf-8')
if 'ERROR' in resp:
raise ApiException(resp)
return resp
def res(self, **kwargs):
'''
sends additional GET-requests (solved captcha, balance, report etc.)
Parameters
----------
**kwargs : TYPE
DESCRIPTION.
Raises
------
NetworkException
DESCRIPTION.
ApiException
DESCRIPTION.
Returns
-------
resp : TYPE
DESCRIPTION.
'''
try:
current_url_out = 'https://'+self.post_url+'/res.php'
resp = requests.get(current_url_out, params=kwargs)
if resp.status_code != 200:
raise NetworkException(f'bad response: {resp.status_code}')
resp = resp.content.decode('utf-8')
if 'ERROR' in resp:
raise ApiException(resp)
except requests.RequestException as e:
raise NetworkException(e)
return resp | 2captcha-python | /2captcha-python-1.2.1.tar.gz/2captcha-python-1.2.1/twocaptcha/api.py | api.py |
# 2ch-downloader
Download all files of 2ch.hk thread.
## Installation
``` sh
pip install 2ch-downloader
```
## Usage
```
usage: 2ch-downloader [-h] [-d DIRECTORY] [--max-directory-name-length LENGTH] URL
positional arguments:
URL Thread url
options:
-h, --help show this help message and exit
-d DIRECTORY, --directory DIRECTORY
Download directory
--max-directory-name-length LENGTH
Max thread directory name length, 128 by default
```
| 2ch-downloader | /2ch-downloader-0.0.3.tar.gz/2ch-downloader-0.0.3/README.md | README.md |
__prog__ = "2ch-downloader"
__desc__ = "Download all files of 2ch.hk thread."
__version__ = "0.0.3"
import argparse
import html
import json
import os
import re
import sys
from dataclasses import dataclass
from pathlib import Path
import requests
@dataclass
class File:
name: str
url: str
size: int
id: str
def download_thread_media(url: str, path: Path, max_directory_name_length: int) -> None:
BASE_URL = "https://2ch.hk"
api_url = url.replace(".html", ".json")
response = json.loads(requests.get(api_url).text)
thread = response["threads"][0]
board = response["board"]["name"]
thread_id = int(response["current_thread"])
thread_name = html.unescape(thread["posts"][0]["subject"])
directory_name = f'{board} {thread_id} {thread_name.replace("/", "_")}'
print(f"Thread {directory_name}")
if len(directory_name) > max_directory_name_length:
directory_name = directory_name[:max_directory_name_length]
path = path / directory_name
os.makedirs(path, exist_ok=True)
os.chdir(path)
files: list[File] = []
for post in thread["posts"]:
if post["files"]:
for file in post["files"]:
basename, extension = os.path.splitext(file["name"])
files.append(
File(
file["fullname"] or f'Sticker{extension}',
BASE_URL + file["path"],
file["size"],
basename,
)
)
for file in files:
download_file(file)
def download_file(file: File) -> None:
filename = f"{file.id} {file.name}"
# Иногда ни размер файла, ни его хеш, отдаваемые api, не соответствуют действительности
# Проверять их бессмысленно
if os.path.exists(filename):
print(f"'{filename}' has already been downloaded", file=sys.stderr)
else:
print(f"Downloading '{filename}' ({file.size} KB)", file=sys.stderr)
r = requests.get(file.url)
with open(filename, "wb") as f:
f.write(r.content)
def thread_url(url: str) -> str:
thread_url_regex = re.compile(
r"(?:https?:\/\/)?2ch.hk\/[a-z]+\/res\/[0-9]+.html", flags=re.I
)
if not thread_url_regex.match(url):
raise ValueError("Provided url is not a 2ch.hk thread url")
return url
def main():
parser = argparse.ArgumentParser(prog=__prog__, description=__desc__)
parser.add_argument(
"url",
metavar="URL",
type=thread_url,
help="Thread url",
)
parser.add_argument(
"-d",
"--directory",
type=str,
default=".",
help="Download directory",
)
parser.add_argument(
"--max-directory-name-length",
metavar="LENGTH",
type=int,
default=128,
help="Max thread directory name length, 128 by default",
)
args = parser.parse_args()
download_thread_media(
args.url, Path(args.directory), args.max_directory_name_length
)
if __name__ == "__main__":
main() | 2ch-downloader | /2ch-downloader-0.0.3.tar.gz/2ch-downloader-0.0.3/_2ch_downloader.py | _2ch_downloader.py |
# 2D Utils
[](https://github.com/erlete/2dutils/actions/workflows/python-publish.yml)
A collection of 2D utilities for coordinate representation and manipulation.
## Features
The following features are currently implemented:
* `Coordinate2D` - A custom 2D coordinate representation based on the built-in `tuple` class, but with extended functionality.
* `Circumcenter` - A class for calculating the circumcenter and circumradius of three coordinates.
## Installation
### macOS/UNIX
```bash
git clone https://github.com/erlete/2d-utils
```
**Note: this package is not yet available on PyPI, but it will be really soon.**
## Usage
Once the package has been installed, its modules can be easily imported into custom programs via the `import` statement.
```python
from coordinate import Coordinate2D
from circumcenter import Circumcenter
```
| 2dutils | /2dutils-1.2.0.tar.gz/2dutils-1.2.0/README.md | README.md |
from itertools import combinations
from math import sqrt
from coordinate import Coordinate2D
class Circumcenter:
"""Class for triangle calculation.
This class is used to calculate the circumcenter of a triangle, given its
three vertices as 2D coordinates.
Args:
a (Coordinate2D): First vertex of the triangle.
b (Coordinate2D): Second vertex of the triangle.
c (Coordinate2D): Third vertex of the triangle.
Attributes:
a (Coordinate2D): First vertex of the triangle.
b (Coordinate2D): Second vertex of the triangle.
c (Coordinate2D): Third vertex of the triangle.
circumcenter (Coordinate2D): The circumcenter of the triangle.
radius (float): The radius of the circumcircle of the triangle.
"""
def __init__(self, a: Coordinate2D,
b: Coordinate2D, c: Coordinate2D) -> None:
# The initial setting variable allows value validation via attribute
# setters but prevents automatic recalculation of the circumcenter
# and circumradius.
self._initial_setting = False
self._a = a
self._b = b
self._c = c
self._initial_setting = True
self._calculate() # Initial calculation:
@property
def a(self) -> Coordinate2D:
"""First vertex of the triangle.
Returns:
Coordinate2D: First vertex of the triangle.
Note:
If the value of the vertex is changed, the circumcenter and radius
of the triangle are recalculated.
"""
return self._a
@a.setter
def a(self, value: Coordinate2D) -> None:
"""First vertex of the triangle.
Args:
value (Coordinate2D): First vertex of the triangle.
Raises:
TypeError: If the value is not a Coordinate2D object.
Note:
If the value of the vertex is changed, the circumcenter and radius
of the triangle are recalculated.
"""
if not isinstance(value, Coordinate2D):
raise TypeError("a must be a Coordinate2D instance")
self._a = value
if not self._initial_setting:
self._calculate()
@property
def b(self) -> Coordinate2D:
"""Second vertex of the triangle.
Returns:
Coordinate2D: Second vertex of the triangle.
Note:
If the value of the vertex is changed, the circumcenter and radius
of the triangle are recalculated.
"""
return self._b
@b.setter
def b(self, value: Coordinate2D) -> None:
"""Second vertex of the triangle.
Args:
value (Coordinate2D): Second vertex of the triangle.
Raises:
TypeError: If the value is not a Coordinate2D object.
Note:
If the value of the vertex is changed, the circumcenter and radius
of the triangle are recalculated.
"""
if not isinstance(value, Coordinate2D):
raise TypeError("b must be a Coordinate2D instance")
self._b = value
if not self._initial_setting:
self._calculate()
@property
def c(self) -> Coordinate2D:
"""Third vertex of the triangle.
Returns:
Coordinate2D: Third vertex of the triangle.
Note:
If the value of the vertex is changed, the circumcenter and radius
of the triangle are recalculated.
"""
return self._c
@c.setter
def c(self, value: Coordinate2D) -> None:
"""Third vertex of the triangle.
Args:
value (Coordinate2D): Third vertex of the triangle.
Raises:
TypeError: If the value is not a Coordinate2D object.
Note:
If the value of the vertex is changed, the circumcenter and radius
of the triangle are recalculated.
"""
if not isinstance(value, Coordinate2D):
raise TypeError("c must be a Coordinate2D instance")
self._c = value
if not self._initial_setting:
self._calculate()
@property
def circumcenter(self) -> Coordinate2D:
"""Circumcenter of the triangle.
Returns:
Coordinate2D: The circumcenter of the triangle.
"""
return self._circumcenter
@property
def circumradius(self) -> float:
"""Radius of the circumcircle of the triangle.
Returns:
float: The radius of the circumcircle of the triangle.
"""
return self._circumradius
def _ensure_non_collinear(self) -> None:
"""Ensures that the triangle is not collinear.
This method computes the determinant of the matrix formed by the
coordinates of the triangle's vertices. If the determinant is zero,
the triangle is collinear, and the circumcenter and circumradius
cannot be calculated.
Raises:
ValueError: If the triangle is collinear.
"""
if not (self._a.x * (self._b.y - self._c.y)
+ self._b.x * (self._c.y - self._a.y)
+ self._c.x * (self._a.y - self._b.y)):
raise ValueError("The triangle is collinear")
def _calculate(self) -> None:
"""Calculates the circumcenter and radius of the triangle."""
self._ensure_non_collinear()
if any(v[1] - v[0] for v in combinations(
(self._a, self._b, self._c), 2)):
# Vertical alignment prevention:
if not (self._b - self._a).x:
self._a, self._c = self._c, self._a
if not (self._c - self._a).x:
self._a, self._b = self._b, self._a
# Segment displacement:
displacement = {
"ab": Coordinate2D(
self.b.x - self.a.x,
self.b.y - self.a.y
),
"ac": Coordinate2D(
self.c.x - self.a.x,
self.c.y - self.a.y
)
}
# Unitary vectors:
unitary = {
"ab": Coordinate2D(
displacement["ab"].y / sqrt(
displacement["ab"].x ** 2 + displacement["ab"].y ** 2
),
-displacement["ab"].x / sqrt(
displacement["ab"].x ** 2 + displacement["ab"].y ** 2
)
),
"ac": Coordinate2D(
displacement["ac"].y / sqrt(
displacement["ac"].x ** 2 + displacement["ac"].y ** 2
),
-displacement["ac"].x / sqrt(
displacement["ac"].x ** 2 + displacement["ac"].y ** 2
)
)
}
# V-vector for vertical intersection:
vertical = {
"ab": Coordinate2D(
unitary["ab"].x / unitary["ab"].y,
1
),
"ac": Coordinate2D(
-(unitary["ac"].x / unitary["ac"].y),
1
)
}
# Midpoint setting:
midpoint = {
"ab": Coordinate2D(
displacement["ab"].x / 2 + self.a.x,
displacement["ab"].y / 2 + self.a.y
),
"ac": Coordinate2D(
displacement["ac"].x / 2 + self.a.x,
displacement["ac"].y / 2 + self.a.y
)
}
# Midpoint height equivalence:
intersection = Coordinate2D(
midpoint["ab"].x + (
(midpoint["ac"].y - midpoint["ab"].y) / unitary["ab"].y
) * unitary["ab"].x,
midpoint["ac"].y
)
# Circumcenter calculation:
self._circumcenter = Coordinate2D(
intersection.x + (
(midpoint["ac"].x - intersection.x)
/ (vertical["ab"].x + vertical["ac"].x)
) * vertical["ab"].x,
intersection.y + (
midpoint["ac"].x - intersection.x
) / (
vertical["ab"].x + vertical["ac"].x
)
)
# Circumradius calculation:
self._circumradius = sqrt(
(self.a.x - self._circumcenter.x) ** 2
+ (self.a.y - self._circumcenter.y) ** 2
) | 2dutils | /2dutils-1.2.0.tar.gz/2dutils-1.2.0/src/circumcenter.py | circumcenter.py |
from __future__ import annotations
from math import ceil, floor, trunc
from typing import Generator
class Coordinate2D:
"""Represents a pair of real-value, two dimensional coordinates.
This class represents a two-dimensional coordinate. Its main purpose is to
serve as a normalized conversion format for data going in and out of the
scripts contained in the project.
Parameters:
-----------
- x: float
The x-coordinate.
- y: float
The y-coordinate.
"""
NUMERICAL = (int, float)
SEQUENTIAL = (tuple, list, set)
def __init__(self, x_value: float, y_value: float) -> None:
self.x: float = float(x_value)
self.y: float = float(y_value)
@property
def x(self) -> float:
return self._x
@x.setter
def x(self, value) -> None:
if not isinstance(value, (int, float)):
raise TypeError("x must be an int or float")
self._x = value
@property
def y(self) -> float:
return self._y
@y.setter
def y(self, value) -> None:
if not isinstance(value, (int, float)):
raise TypeError("y must be an int or float")
self._y = value
def __add__(self, value) -> Coordinate2D:
if not isinstance(value, Coordinate2D):
raise TypeError("value must be a Coordinate2D.")
return Coordinate2D(self._x + value.x, self._y + value.y)
def __sub__(self, value) -> Coordinate2D:
if not isinstance(value, Coordinate2D):
raise TypeError("value must be a Coordinate2D.")
return Coordinate2D(self._x - value.x, self._y - value.y)
def __mul__(self, value) -> Coordinate2D:
if isinstance(value, Coordinate2D):
return Coordinate2D(self._x * value.x, self._y * value.y)
elif isinstance(value, self.NUMERICAL):
return Coordinate2D(self._x * value, self._y * value)
elif isinstance(value, self.SEQUENTIAL) and len(value) >= 2:
return Coordinate2D(self._x * value[0], self._y * value[1])
raise TypeError("value must be a Coordinate2D, a numerical type, or a"
"sequential type with at least 2 elements.")
def __truediv__(self, value) -> Coordinate2D:
if not isinstance(value, Coordinate2D):
raise TypeError("value must be a Coordinate2D.")
return Coordinate2D(self._x / value, self._y / value)
def __floordiv__(self, value) -> Coordinate2D:
if not isinstance(value, Coordinate2D):
raise TypeError("value must be a Coordinate2D.")
return Coordinate2D(self._x // value, self._y // value)
def __mod__(self, value) -> Coordinate2D:
if not isinstance(value, Coordinate2D):
raise TypeError("value must be a Coordinate2D.")
return Coordinate2D(self._x % value, self._y % value)
def __pow__(self, value) -> Coordinate2D:
if not isinstance(value, Coordinate2D):
raise TypeError("value must be a Coordinate2D.")
return Coordinate2D(self._x ** value, self._y ** value)
def __neg__(self) -> Coordinate2D:
return Coordinate2D(-self._x, -self._y)
def __pos__(self) -> Coordinate2D:
return Coordinate2D(+self._x, +self._y)
def __abs__(self) -> Coordinate2D:
return Coordinate2D(abs(self._x), abs(self._y))
def __invert__(self) -> Coordinate2D:
return Coordinate2D(~self._x, ~self._y)
def __round__(self, n: int = 0) -> Coordinate2D:
if not isinstance(n, int):
raise TypeError("n must be an int.")
return Coordinate2D(round(self._x, n), round(self._y, n))
def __floor__(self) -> Coordinate2D:
return Coordinate2D(floor(self._x), floor(self._y))
def __ceil__(self) -> Coordinate2D:
return Coordinate2D(ceil(self._x), ceil(self._y))
def __trunc__(self) -> Coordinate2D:
return Coordinate2D(trunc(self._x), trunc(self._y))
def __lt__(self, value) -> bool:
if not isinstance(value, Coordinate2D):
raise TypeError("value must be a Coordinate.")
return self._x < value.x and self._y < value.y
def __le__(self, value) -> bool:
if not isinstance(value, Coordinate2D):
raise TypeError("value must be a Coordinate.")
return self._x <= value.x and self._y <= value.y
def __gt__(self, value) -> bool:
if not isinstance(value, Coordinate2D):
raise TypeError("value must be a Coordinate.")
return self._x > value.x and self._y > value.y
def __ge__(self, value) -> bool:
if not isinstance(value, Coordinate2D):
raise TypeError("value must be a Coordinate.")
return self._x >= value.x and self._y >= value.y
def __eq__(self, value) -> bool:
if not isinstance(value, Coordinate2D):
raise TypeError("value must be a Coordinate.")
return self._x == value.x and self._y == value.y
def __ne__(self, value) -> bool:
if not isinstance(value, Coordinate2D):
raise TypeError("value must be a Coordinate.")
return self._x != value.x or self._y != value.y
def __str__(self) -> str:
return f"Coordinate2D({self._x}, {self._y})"
def __repr__(self) -> str:
return f"Coordinate2D({self._x}, {self._y})"
def __hash__(self) -> int:
return hash((self._x, self._y))
def __bool__(self) -> bool:
return self._x != 0 or self._y != 0
def __len__(self) -> int:
return 2
def __getitem__(self, index) -> float:
if index == 0:
return self._x
elif index == 1:
return self._y
else:
raise IndexError("index must be 0 or 1")
def __setitem__(self, index, value) -> None:
if index == 0:
self._x = value
elif index == 1:
self._y = value
else:
raise IndexError("index must be 0 or 1")
def __iter__(self) -> Generator[float, None, None]:
yield self._x
yield self._y
def __reversed__(self) -> Generator[float, None, None]:
yield self._y
yield self._x
def __getattr__(self, name) -> float:
if name == 'x':
return self._x
elif name == 'y':
return self._y
else:
raise AttributeError(f"{name} is not an attribute")
def __delattr__(self, name) -> None:
if name == 'x':
del self._x
elif name == 'y':
del self._y
else:
raise AttributeError(f"{name} is not an attribute")
def __getstate__(self) -> tuple[float, float]:
return self._x, self._y
def __setstate__(self, state) -> None:
self._x, self._y = state
def __reduce__(self) -> tuple[Coordinate2D, tuple[float, float]]:
return self.__class__, (self._x, self._y)
def __copy__(self) -> Coordinate2D:
return self.__class__(self._x, self._y)
def __bytes__(self) -> bytes:
return bytes(self._x) + bytes(self._y)
def __int__(self) -> int:
return int(self._x) + int(self._y)
def __float__(self) -> float:
return float(self._x) + float(self._y)
def __complex__(self) -> complex:
return complex(self._x) + complex(self._y)
def __oct__(self) -> str:
return oct(self._x) + oct(self._y)
def __hex__(self) -> str:
return hex(self._x) + hex(self._y)
def __index__(self) -> float:
return self._x + self._y
def __coerce__(self, value) -> tuple[float, float]:
if not isinstance(value, Coordinate2D):
raise TypeError("value must be a Coordinate.")
return self._x, value.x
def __format__(self, format_spec) -> str:
return format(self._x, format_spec) + format(self._y, format_spec) | 2dutils | /2dutils-1.2.0.tar.gz/2dutils-1.2.0/src/coordinate.py | coordinate.py |
# 2dwavesim
This is a project that simulates waves on 2D plates/rooms. Given boundaries (or walls) and points where oscillation will be forced, this will simulate the resulting wavemodes!
Currently it supports setting the attenuation properties of individual boundaries, multiple forcing points based on either data or a function, and any wall shape you want. It also supports variable time and space steps and spans (as long as you keep numerically stable!), as well as custom wavespeed and attenuation on the material.

TODO:
- add tests
- frequency-dependant wall transmission values
- 3D??
## Usage
There are two main Classes:
`Room(ds, width, height,*, walls=[], physics_params={})`
<ul>
This creates an instance of a `Room` class, which contains any walls or sources of the system.
`ds` is a float which defines the unit of distance between two grid points.
`width` and `height` and floats which define the dimentions of the grid. If they are not exact multiples of `ds`, then they are upper bounds on the number of points above the nearest multiple.
`walls` is a list of `Wall` objects. This can be optionally set after construction as well.
`physics_params` is a dict with structure `{'wavespeed':float, 'attenuation':float}`. Wavespeed represents the speed of the propigating wave on the room's medium, and attenuation represents the attenuation factor of waves on the medium. By defaut, wavespeed is assumed to be 343 units/s and attenuation is assumed to be $2^{-2}$ units
$^{-1}$.
**`Room.add_source_func(loc, func)`**
<ul>
Add a source based on a function.
`loc` is the room-specific coordinate of the source. Note: unless `ds=1`, this is not the same as the list indices of the point in the room.
`func` is a function taking a float (the time) and outputing a float (the value of the wave at that time). This should be something like `lambda t: np.sin(t)`, for example.
</ul>
**`Room.add_source_data(loc, data)`**
<ul>
Add a source based on a list of values. Careful! Make sure you use a `dt` value that matches the table data, as an entry of the data list will be used on every time tick. For example, if you make the data table represent the value of the wave every 200ms, then be sure to set `dt` to 200ms as well when you run the simulation. If there are less points in the list of values than there are time steps, then a value of 0 is used for all time steps past the last data point.
`loc` is the room-specific coordinate of the source. Note: unless `ds=1`, this is not the same as the list indices of the point in the room.
`data` is a list of floats (the value of the wave at that time). This should be something like `np.sin(np.arange(0, 10, 0.2))`, for example.
</ul>
**`Room.add_walls(walls)`**
<ul>
Add walls to the system after constructing the Room object.
`walls` is a list of `Wall` objects to add the the system.
</ul>
**`Room.create_mask()`**
<ul>
Create a mask for the values of the room based on the currently set walls. This is automatically done when running the simulation, but it can be run beforehand if you want to plot the mask for visualization.
</ul>
**`Room.get_mask()`**
<ul>
Returns a 2D array of the wall mask as currently calculated.
</ul>
**`Room.run(dt, t_final)`**
<ul>
Calculate the wave propagation from the set sources and using the set walls. This will simulate from `t=0` to `t_final` at `dt` time steps. If `t_final` isn't an exact multiple of `dt`, then it acts like an upper bound.
`dt` a float giving the time step for the simulation. A smaller value means more time resolution. WARNING: Numerical stability will almost certainly be lost if this is not set to satisfy the [CFL Condition](https://en.wikipedia.org/wiki/Courant%E2%80%93Friedrichs%E2%80%93Lewy_condition), namely $\frac{u*dt}{ds} \leq C_{max}$ where $u$ is the `wavespeed` and $c_{max}$ is approximately 1 for the numerical method being used.
`t_final` a float giving an upper limit for the amount of time to be simulated. A higher value will take more time to simulate, and will likely just repeat the steady state after a certain point in time...
</ul>
</ul>
`Wall(endpoint1, endpoint2, transmission)`
<ul>
This creates an instance of a `Wall` class, which contains the wall's endpoints and transmission factor.
`endpoint1` and `endpoint2` are tuple2 or 2-list2 of floats giving the position of each end of the wall in the room-specific coordinates. Note: unless `ds=1`, this is not the same as the list indices of the point in the room.
`transmission` is a float in [0,1] which defines the proportion of wave amplitude able to penetrate the wall. If 0, then all energy is reflected back inwards, and if 1 then the wall "isn't there".
</ul>
## Visualization
The `visualization` module contains a few functions for visualizing results, or processing results into an easily displayed format.
**`animate(data, *, filepath='', frame_space=10, walls=[])`**
<ul>
Automatically animate the given data using `matplotlib.animation.ArtistAnimation`. The animation file can optionally be saved to a file.
`data` is a 3D array of waveform over time, which is the output from running the simulation.
`filename` is the name and path of output file. Leave this blank to not save. Output formats are those supported by `matplotlib.animation.ArtistAnimation`, which is at least ".gif" and ".webp".
`frame_space` is the temporal resolution of resulting animation. Make sure this isn't too small!
`walls` is to optionally include the walls in the animation. They won't be visible if this isn't included.
</ul>
**`get_steady_state_index(data, *, sample_points, rms_tolerance=0.1, window_size=0.1)`**
<ul>
This function calculates the windowed RMS of the given points over time. This data is compared to the RMS value at the end of the simulation. Then the latest time index where all point RMS's are within a tolerance to the final RMS is taken as the time index where steady-state is reached.
`data` is a 3D array of waveform over time, which is the output from running the simulation.
`sample_points` is a list of points in the room which will be checked for RMS.
`rms_tolerance` is a float in [0, 1] defining the limit on the amount the RMS is allowed to change from the final value and still be considered steady-state.
`window_size` is a float in [0, 1] defining the percent of total points to consider in the window.
</ul>
**`get_standing_waves(data, *, steady_state_kwargs=None)`**
<ul>
This function calculates when the steady state begins, and returns a 2D array which is the average of the absolute value of all of the rooms points across all steady state times.
`data` is a 3D array of waveform over time, which is the output from running the simulation.
`steady_state_kwargs` is a dict of the keyword arguments to pass to `get_steady_state_index`. If `None`, then the default parameters and a sample point at the middle of the room are used.
</ul>
| 2dwavesim | /2dwavesim-1.0.0.tar.gz/2dwavesim-1.0.0/README.md | README.md |
# 2dxPlay
Play konami 2dx files from the command line.
Install with pip and run it from anywhere.
`pip install 2dxPlay`
### Usage:
```
2dxplay infile.2dx
```
If the file contains more than one wav, it will play all tracks sequentially.
Press Ctrl+C to pause and enter a track number to jump to, or `q` to quit.
###Tip: Play any 2dx file just by double clicking on it:
Right click on a 2dx file, choose Open With > Look for another app on this PC.
Navigate to your python installation where 2dxPlay.exe is installed. (I use python 3.7, so pip
installed it in `%appdata%\Local\Programs\Python\Python37\Scripts`).
Check the "Always use this app" box and profit.
 | 2dxPlay | /2dxPlay-1.0.1.tar.gz/2dxPlay-1.0.1/README.md | README.md |
import os
os.environ['PYGAME_HIDE_SUPPORT_PROMPT'] = "hide"
from os import path
from pygame import mixer
from twodxPlay.twodx import Twodx
import sys, traceback
from datetime import timedelta
import struct
import argparse
version='V1.0.1'
parser = argparse.ArgumentParser(description='A command line app for playing and extracting konami 2dx audio files.')
parser.add_argument('infile')
parser.add_argument('-e', '--extract', nargs='?', const='all', help='Extract the wav files. Optionally specify a number, a comma-separated list of numbers or a range such as 10-20.'
' Don\'t use spaces, and NOTE that extracted wav files will start at 0.wav. Example: 2dxplay -e 1,4-8,10 infile.2dx')
parser.add_argument('-o', '--output-dir', default='output', help='Directory to extract files to. Will be created if not exists. Default is ./output/')
if len(sys.argv) <= 1:
print(f"2dxPlay {version}")
parser.print_help()
sys.exit(1)
args = parser.parse_args()
mixer.init()
infile = args.infile
outdir = args.output_dir
try:
twodx_file = Twodx(path.abspath(infile))
except struct.error:
traceback.print_exc()
print("Probably not a 2dx file.")
sys.exit(1)
except:
traceback.print_exc()
sys.exit(1)
current_track = 1
# function from bemaniutils/afputils.
def parse_intlist(data):
ints = []
for chunk in data.split(","):
chunk = chunk.strip()
if '-' in chunk:
start, end = chunk.split('-', 1)
start_int = int(start.strip())
end_int = int(end.strip())
ints.extend(range(start_int, end_int + 1))
else:
ints.append(int(chunk))
return sorted(set(ints))
def extract(data=None):
if outdir != '.' and not os.path.exists(outdir):
os.mkdir(outdir)
if args.extract == 'all':
for i in range(twodx_file.file_count):
wav = twodx_file.load(i)
print(f"Writing {i+1}/{twodx_file.file_count} wavs to {outdir}")
with open(f"{outdir}/{i}.wav", 'wb') as f:
f.write(wav.getbuffer())
else:
if data:
ints = parse_intlist(data)
else:
ints = parse_intlist(args.extract)
for i in ints:
wav = twodx_file.load(i-1)
print(f"Writing {i-1}.wav to {outdir}")
with open(f"{args.output_dir}/{i-1}.wav", 'wb') as f:
f.write(wav.getbuffer())
def main():
if args.extract:
extract()
sys.exit()
global current_track
try:
while current_track-1 < twodx_file.file_count:
wav = twodx_file.load(current_track-1)
sound = mixer.Sound(wav)
channel = sound.play()
length = timedelta(seconds=int(sound.get_length()))
print(f"Playing: {infile} Track {current_track}/{twodx_file.file_count} Length: {length}")
while channel.get_busy():
pass
current_track += 1
except KeyboardInterrupt:
mixer.stop()
while "the answer is invalid":
try:
reply = str(input('Enter track number to jump to, e [track(s) to extract], or q to quit: ')).lower().strip()
except (EOFError, KeyboardInterrupt):
sys.exit(0)
if reply == 'q':
sys.exit(0)
if reply[0] == 'e':
ints = reply.split()
if len(reply) > 1:
extract(reply.split()[1])
else:
continue
if reply.isdigit() and int(reply) in range(1, twodx_file.file_count + 1):
current_track = int(reply)
main()
except Exception:
traceback.print_exc()
sys.exit(1)
if __name__ == "__main__":
main() | 2dxPlay | /2dxPlay-1.0.1.tar.gz/2dxPlay-1.0.1/twodxPlay/main.py | main.py |
import base64
import hashlib
import hmac
class OTP(object):
def __init__(self, s, digits=6, digest=hashlib.sha1):
"""
@param [String] secret in the form of base32
@option options digits [Integer] (6)
Number of integers in the OTP
Google Authenticate only supports 6 currently
@option options digest [Callable] (hashlib.sha1)
Digest used in the HMAC
Google Authenticate only supports 'sha1' currently
@returns [OTP] OTP instantiation
"""
self.digits = digits
self.digest = digest
self.secret = s
def generate_otp(self, input):
"""
@param [Integer] input the number used seed the HMAC
Usually either the counter, or the computed integer
based on the Unix timestamp
"""
hmac_hash = hmac.new(
self.byte_secret(),
self.int_to_bytestring(input),
self.digest,
).digest()
hmac_hash = bytearray(hmac_hash)
offset = hmac_hash[-1] & 0xf
code = ((hmac_hash[offset] & 0x7f) << 24 |
(hmac_hash[offset + 1] & 0xff) << 16 |
(hmac_hash[offset + 2] & 0xff) << 8 |
(hmac_hash[offset + 3] & 0xff))
str_code = str(code % 10 ** self.digits)
while len(str_code) < self.digits:
str_code = '0' + str_code
return str_code
def byte_secret(self):
missing_padding = len(self.secret) % 8
if missing_padding != 0:
self.secret += '=' * (8 - missing_padding)
return base64.b32decode(self.secret, casefold=True)
@staticmethod
def int_to_bytestring(i, padding=8):
"""
Turns an integer to the OATH specified
bytestring, which is fed to the HMAC
along with the secret
"""
result = bytearray()
while i != 0:
result.append(i & 0xFF)
i >>= 8
# It's necessary to convert the final result from bytearray to bytes because
# the hmac functions in python 2.6 and 3.3 don't work with bytearray
return bytes(bytearray(reversed(result)).rjust(padding, b'\0')) | 2fa | /2fa-0.0.1.tar.gz/2fa-0.0.1/pyotp/otp.py | otp.py |
import datetime
import time
from pyotp import utils
from pyotp.otp import OTP
class TOTP(OTP):
def __init__(self, *args, **kwargs):
"""
@option options [Integer] interval (30) the time interval in seconds
for OTP This defaults to 30 which is standard.
"""
self.interval = kwargs.pop('interval', 30)
super(TOTP, self).__init__(*args, **kwargs)
def at(self, for_time, counter_offset=0):
"""
Accepts either a Unix timestamp integer or a Time object.
Time objects will be adjusted to UTC automatically
@param [Time/Integer] time the time to generate an OTP for
@param [Integer] counter_offset an amount of ticks to add to the time counter
"""
if not isinstance(for_time, datetime.datetime):
for_time = datetime.datetime.fromtimestamp(int(for_time))
return self.generate_otp(self.timecode(for_time) + counter_offset)
def now(self):
"""
Generate the current time OTP
@return [Integer] the OTP as an integer
"""
return self.generate_otp(self.timecode(datetime.datetime.now()))
def verify(self, otp, for_time=None, valid_window=0):
"""
Verifies the OTP passed in against the current time OTP
@param [String/Integer] otp the OTP to check against
@param [Integer] valid_window extends the validity to this many counter ticks before and after the current one
"""
if for_time is None:
for_time = datetime.datetime.now()
if valid_window:
for i in range(-valid_window, valid_window + 1):
if utils.strings_equal(str(otp), str(self.at(for_time, i))):
return True
return False
return utils.strings_equal(str(otp), str(self.at(for_time)))
def provisioning_uri(self, name, issuer_name=None):
"""
Returns the provisioning URI for the OTP
This can then be encoded in a QR Code and used
to provision the Google Authenticator app
@param [String] name of the account
@return [String] provisioning uri
"""
return utils.build_uri(self.secret, name, issuer_name=issuer_name)
def timecode(self, for_time):
i = time.mktime(for_time.timetuple())
return int(i / self.interval) | 2fa | /2fa-0.0.1.tar.gz/2fa-0.0.1/pyotp/totp.py | totp.py |
from __future__ import print_function, unicode_literals, division, absolute_import
import unicodedata
try:
from urllib.parse import quote
except ImportError:
from urllib import quote
def build_uri(secret, name, initial_count=None, issuer_name=None):
"""
Returns the provisioning URI for the OTP; works for either TOTP or HOTP.
This can then be encoded in a QR Code and used to provision the Google
Authenticator app.
For module-internal use.
See also:
http://code.google.com/p/google-authenticator/wiki/KeyUriFormat
@param [String] the hotp/totp secret used to generate the URI
@param [String] name of the account
@param [Integer] initial_count starting counter value, defaults to None.
If none, the OTP type will be assumed as TOTP.
@param [String] the name of the OTP issuer; this will be the
organization title of the OTP entry in Authenticator
@return [String] provisioning uri
"""
# initial_count may be 0 as a valid param
is_initial_count_present = (initial_count is not None)
otp_type = 'hotp' if is_initial_count_present else 'totp'
base = 'otpauth://%s/' % otp_type
if issuer_name:
issuer_name = quote(issuer_name)
base += '%s:' % issuer_name
uri = '%(base)s%(name)s?secret=%(secret)s' % {
'name': quote(name, safe='@'),
'secret': secret,
'base': base,
}
if is_initial_count_present:
uri += '&counter=%s' % initial_count
if issuer_name:
uri += '&issuer=%s' % issuer_name
return uri
def strings_equal(s1, s2):
"""
Timing-attack resistant string comparison.
Normal comparison using == will short-circuit on the first mismatching
character. This avoids that by scanning the whole string, though we
still reveal to a timing attack whether the strings are the same
length.
"""
s1 = unicodedata.normalize('NFKC', s1)
s2 = unicodedata.normalize('NFKC', s2)
try:
# Python 3.3+ and 2.7.7+ include a timing-attack-resistant
# comparison function, which is probably more reliable than ours.
# Use it if available.
from hmac import compare_digest
return compare_digest(s1, s2)
except ImportError:
pass
if len(s1) != len(s2):
return False
differences = 0
for c1, c2 in zip(s1, s2):
differences |= ord(c1) ^ ord(c2)
return differences == 0 | 2fa | /2fa-0.0.1.tar.gz/2fa-0.0.1/pyotp/utils.py | utils.py |
import time
import threading
import urwid
import argparse
import random
from feeds.display import FeedsTreeBrowser
from feeds.utils import *
from feeds.input_mod import *
f = None
def insert_feeds(feeds):
global q, f
try:
all_src = []
while True:
s = q.get(timeout=7)
content = s.preview()
all_src.append(s)
# added new check for None
if content is None:
continue
f.init_data(content)
q.queue.clear()
except Exception as e:
pass
finally:
f.done = True
def _start():
global q, f, progress
# cli
parser = argparse.ArgumentParser(prog='2feeds')
group = parser.add_mutually_exclusive_group()
group.add_argument('-a', '--add',
action='store_true', help='add a new feed')
group.add_argument('-l', '--list', action='store_true',
help='list saved feeds')
group.add_argument('-r', '--remove', action='store_true',
help='remove a feed from your collection')
group.add_argument('-c', '--clear', action='store_true',
help='clear feed collection and/or clear viewing cache')
args = parser.parse_args()
cli = False
if args.add or args.list or args.remove or args.clear:
cli = True
if args.add:
addLink()
elif args.list:
listSavedSites()
elif args.remove:
delete_feed()
elif args.clear:
clear_data()
if cli:
exit()
feeds = getParam('url')
random.shuffle(feeds)
feeds = tuple(feeds)
if len(feeds) == 0:
print('add some feeds using -a option.')
exit(-1)
# get feeds
progress = set(getProgress())
request_feeds = threading.Thread(
target=get_feeds_from_url, args=(feeds, progress,))
request_feeds.start()
# init display
f = FeedsTreeBrowser()
# grab feeds from queue
add_feed = threading.Thread(target=insert_feeds, args=(feeds,))
add_feed.start()
# run event loop
links = f.main()
if links != []:
saveProgress(links)
request_feeds.join()
add_feed.join()
if __name__ == "__main__":
_start() | 2feeds | /2feeds-1.0.210-py3-none-any.whl/feeds/main.py | main.py |
import os
import re
import json
import time
DIR_BASE = os.path.dirname(os.path.abspath(__file__))
dir_path = DIR_BASE + "/.app_data"
udata_path = dir_path + "/udata.dat"
ucache_path = dir_path + "/ucache.dat"
regex = re.compile(
r'^(?:http|ftp)s?://'
)
def checkFiles():
if not os.path.isdir(dir_path):
os.makedirs(dir_path)
if not os.path.exists(udata_path):
with open(udata_path, 'w') as f:
f.write('[]')
if not os.path.exists(ucache_path):
with open(ucache_path, 'w') as f:
f.write('[]')
def addLink():
checkFiles()
links = getOldLinks()
link_set = set([x['url'] for x in links])
link = input('Paste RSS link here: ')
if re.match(regex, link) is None:
print("Not a valid url")
exit()
title = input('Site-name: ')
data = {
'url': link,
'title': title,
'time': time.time()
}
if data['url'] in link_set:
print('Already added, exiting')
exit()
links.append(data)
with open(udata_path, 'w') as f:
json.dump(links, f)
def getOldLinks():
with open(udata_path, 'r') as f:
content = f.read()
data = json.loads(content)
return data
def listSavedSites():
sites = getParam('title')
if len(sites) == 0:
print('Empty collection. Please update')
exit()
for x in sites:
print(' x ' + x)
def getParam(param):
checkFiles()
links_dict = getOldLinks()
result = []
for x in links_dict:
result.append(x[param])
return result
def updateParam(param, item):
checkFiles()
links_dict = getOldLinks()
if len(links_dict) == 0:
print('Nothing to update')
exit()
valid = []
for x in links_dict:
if item != x[param]:
valid.append(x)
with open(udata_path, 'w') as f:
json.dump(valid, f)
def delete_feed():
listSavedSites()
item = input('Enter site-name: ')
updateParam('title', item)
def getProgress():
checkFiles()
with open(ucache_path, 'r') as f:
content = f.read()
data = json.loads(content)
return data
def saveProgress(url):
checkFiles()
links = getProgress()
links.extend(url)
with open(ucache_path, 'w') as f:
json.dump(list(set(links)), f)
def clear_data():
choice = input('would you like to clear viewing cache? (y/n): ')
if choice in ('y', "Y", "yes", "Yes"):
with open(ucache_path, 'w') as f:
f.write('[]')
choice = input('would you like to clear feed sources? (y/n): ')
if choice in ('y', "Y", "yes", "Yes"):
with open(udata_path, 'w') as f:
f.write('[]') | 2feeds | /2feeds-1.0.210-py3-none-any.whl/feeds/input_mod.py | input_mod.py |
import queue
import threading
import feedparser
from html.parser import HTMLParser
q = queue.Queue()
class MLStripper(HTMLParser):
def __init__(self):
self.reset()
self.strict = False
self.convert_charrefs = True
self.fed = []
def handle_data(self, d):
self.fed.append(d)
def get_data(self):
return ''.join(self.fed)
def strip_tags(html):
s = MLStripper()
s.feed(html)
return s.get_data().strip()
class RSS_parser:
def __init__(self, url, progress):
self.url = url
self.stories = []
self.init(progress)
self.pos = 0
def __len__(self):
return len(self.stories)
def init(self, progress):
p = feedparser.parse(self.url)
exception = p.get('bozo_exception', None)
if exception is not None:
exit(-1) # exit if no internet connection; no attempt to reconnect
for i in range(len(p.entries)):
url = p.entries[i].link
if url not in progress:
self.stories.append(
{"name": strip_tags(p.entries[i].title), "url": url, "children": [{"name": strip_tags(p.entries[i].description)}]})
def preview(self):
# added return
if self.pos > len(self.stories):
return None
count = 0
end = len(self.stories) if self.pos + \
3 > len(self.stories) else self.pos + 3
for i in range(self.pos, end):
count += 1
result = self.stories[self.pos:end]
self.pos += count
return result
# add dependency ?
def get_feeds_from_url(urls, progress):
for url in urls:
try:
r = RSS_parser(url, progress)
if len(r) == 0:
continue
q.put(r)
except Exception as e:
pass # is it okay to pass ?
def check_if_empty(all_feeds):
if len(list(filter(lambda x: x.pos >= len(x.stories), all_feeds))) == len(
all_feeds
):
return True
return False | 2feeds | /2feeds-1.0.210-py3-none-any.whl/feeds/utils.py | utils.py |
import urwid
import time
import threading
import webbrowser
state = {
'expanded': {},
'isFlagged': {}
}
class FeedTreeWidget(urwid.TreeWidget):
""" Display widget for leaf nodes """
unexpanded_icon = urwid.AttrMap(urwid.TreeWidget.unexpanded_icon,
'dirmark')
expanded_icon = urwid.AttrMap(urwid.TreeWidget.expanded_icon,
'dirmark')
indent_cols = 4
def __init__(self, node, is_expanded=None, is_flagged=None):
self.__super.__init__(node)
if self.get_node()._parent is not None:
self.keypress(None, "-")
if is_expanded is not None:
if is_expanded:
self.keypress(None, "+")
else:
self.keypress(None, "-")
self._w = urwid.AttrWrap(self._w, None)
if is_flagged is not None:
self.flagged = is_flagged
else:
self.flagged = False
self.update_w()
def update_w(self):
"""Update the attributes of self.widget based on self.flagged.
"""
if self.flagged:
self._w.attr = 'flagged focus' # 'flagged'
self._w.focus_attr = 'flagged focus'
else:
self._w.attr = 'body'
self._w.focus_attr = 'focus'
def get_display_text(self):
return self.get_node().get_value()['name']
def get_feed_url(self):
return self.get_node().get_value()['url']
def selectable(self):
"""
Allow selection of non-leaf nodes so children may be (un)expanded
"""
return not self.is_leaf and self.get_node()._parent is not None
def mouse_event(self, size, event, button, col, row, focus):
if self.is_leaf or event != 'mouse press' or button != 1 or self.get_node()._parent is None:
return False
if row == 0 and col == self.get_indent_cols():
self.expanded = not self.expanded
self.update_expanded_icon()
return True
return False
def unhandled_keys(self, size, key):
"""
Override this method to intercept keystrokes in subclasses.
Default behavior: Toggle flagged on space, ignore other keys.
"""
if key in (" ", "enter"):
if key == "enter":
self.flagged = True
else:
self.flagged = not self.flagged
self.update_w()
return key
def keypress(self, size, key):
"""Handle expand & collapse requests (non-leaf nodes)"""
if self.is_leaf:
return key
if key in ("+", "right"):
self.expanded = True
self.update_expanded_icon()
elif key in ("-", "left"):
self.expanded = False
self.update_expanded_icon()
elif key in (" ", "enter"):
return self.unhandled_keys(size, key)
elif self._w.selectable():
return self.__super.keypress(size, key)
else:
return key
def load_inner_widget(self):
if self.is_leaf:
return urwid.AttrMap(urwid.Text(self.get_display_text()), 'summary')
else:
return urwid.Text(self.get_display_text())
def get_indented_widget(self):
widget = self.get_inner_widget()
div = urwid.Divider()
if not self.is_leaf:
widget = urwid.Columns([('fixed', 1,
[self.unexpanded_icon, self.expanded_icon][self.expanded]),
widget], dividechars=1)
else:
widget = urwid.Pile([widget, div])
indent_cols = self.get_indent_cols()
return urwid.Padding(widget,
width=('relative', 100), left=indent_cols, right=indent_cols)
class Node(urwid.TreeNode):
""" Data storage object for leaf nodes """
def load_widget(self):
return FeedTreeWidget(self)
class FeedParentNode(urwid.ParentNode):
""" Data storage object for interior/parent nodes """
def __init__(self, value, parent=None, key=None, depth=None):
urwid.TreeNode.__init__(
self, value, parent=parent, key=key, depth=depth)
self._child_keys = None
self._children = {}
self.is_expanded = None
self.is_flagged = None
def load_widget(self):
return FeedTreeWidget(self, self.is_expanded, self.is_flagged)
def load_child_keys(self):
data = self.get_value()
return range(len(data['children']))
def load_child_node(self, key):
"""Return either an Node or ParentNode"""
childdata = self.get_value()['children'][key]
childdepth = self.get_depth() + 1
if 'children' in childdata:
childclass = FeedParentNode
else:
childclass = Node
return childclass(childdata, parent=self, key=key, depth=childdepth)
def append_child_node(self, data):
self._value["children"].extend(data["children"])
saved_links = []
class CustomTree(urwid.TreeListBox):
def unhandled_input(self, size, input):
"""Handle macro-navigation keys"""
if input == 'left':
return input
elif input == '-':
self.collapse_focus_parent(size)
elif input == "enter":
widget, pos = self.body.get_focus()
url = widget.get_feed_url()
webbrowser.open(url)
self.add_link(url)
elif input == " ":
widget, pos = self.body.get_focus()
url = widget.get_feed_url()
self.add_or_remove_link(url)
else:
return input
def add_link(self, url):
if url not in saved_links:
saved_links.append(url)
def add_or_remove_link(self, url):
if url not in saved_links:
saved_links.append(url)
else:
saved_links.pop(saved_links.index(url))
class FeedsTreeBrowser:
palette = [
('body', 'white', 'black'),
('dirmark', 'light red', 'black', 'bold'),
# ('title', 'default,bold', 'default', 'bold'),
('key', 'light cyan', 'black'),
('focus', 'light gray', 'dark blue', 'standout'),
('title', 'default,bold', 'default', 'bold'),
('summary', 'black', 'light gray'),
# ('flagged', 'black', 'dark green', ('bold', 'underline')),
('flagged', 'black', 'light green'),
('flagged focus', 'black', 'dark cyan',
('bold', 'standout', 'underline')),
# ('foot', 'light gray', 'black'),
# ('flag', 'dark gray', 'light gray'),
# ('error', 'dark red', 'light gray'),
]
footer_text = [
('title', ""), "",
('key', "UP"), ",", ('key', "DOWN"), ",",
# ('key', "PAGE UP"), ",", ('key', "PAGE DOWN"),
" ",
('key', "+"), ",",
('key', "-"), ",",
('key', "RIGHT"), ",",
('key', "LEFT"), " ",
('key', "ENTER"), ",",
('key', "SPACE"), " ",
('key', "Q"),
]
def __init__(self, data=None):
self.header = urwid.Text("2Feeds", align="left")
self.footer = urwid.AttrMap(urwid.Text(self.footer_text),
'foot')
self.start = True
self.done = False
def init_data(self, data):
if self.start:
self.topnode = FeedParentNode(None)
d_ = {"name": "/", "children": []}
d_["children"].extend(data)
self.topnode._value = d_
self.tw = urwid.TreeWalker(self.topnode)
self.listbox = CustomTree(self.tw)
self.start = False
else:
d_ = self.topnode._value
d_["children"].extend(data)
self.topnode = FeedParentNode(d_)
def generate_frame(self):
return urwid.Padding(urwid.Frame(
urwid.AttrMap(self.listbox, 'body'),
header=urwid.AttrMap(self.header, 'title'),
footer=self.footer), left=1, right=2)
def store_state(self):
node = self.loop.widget.base_widget.body.base_widget._body.focus
parent = node.get_parent()
keys = parent.get_child_keys()
for key in keys:
state['expanded'][key] = parent.get_child_widget(key).expanded
state['isFlagged'][key] = parent.get_child_widget(key).flagged
def restore_state(self):
keys = self.topnode.get_child_keys()
prev_len = len(state['expanded'])
for i, key in enumerate(keys):
_ = self.topnode.get_child_node(key)
if i < prev_len:
self.topnode._children[key].is_expanded = state['expanded'][key]
self.topnode._children[key].is_flagged = state['isFlagged'][key]
def refresh(self, _loop, _data):
"""
Redraw screen;
TODO: is there a better way?
"""
global stop_adding
node = self.tw.get_focus()[1]
key = node.get_key()
self.store_state()
self.restore_state()
self.tw = urwid.TreeWalker(self.topnode.get_child_node(key))
# self.loop.widget.base_widget.body.base_widget._body = self.tw
self.listbox = CustomTree(self.tw)
self.loop.widget.base_widget.body = self.listbox
if not self.done:
self.loop.set_alarm_in(1, self.refresh)
def main(self):
"""Run the program."""
try:
count = 0
while not hasattr(self, 'listbox'):
time.sleep(1)
count += 1
if count == 15:
# print('time exceeded')
raise Exception('time exceeded')
continue
self.frame = urwid.AttrMap(self.generate_frame(), 'body')
self.loop = urwid.MainLoop(self.frame, self.palette,
unhandled_input=self.unhandled_input)
self.loop.set_alarm_in(1, self.refresh)
self.loop.run()
except Exception as e:
err_msg = """error occured:\n- check internet connection\n- add more sources"""
print(err_msg)
exit(-1)
finally:
return saved_links
def unhandled_input(self, k):
if k in ('q', 'Q'):
raise urwid.ExitMainLoop() | 2feeds | /2feeds-1.0.210-py3-none-any.whl/feeds/display.py | display.py |
import inspect
try:
import urlparse
except ImportError: # Python 3
from urllib import parse as urlparse
try:
from urllib import urlencode
except ImportError: # Python 3
from urllib.parse import urlencode
import requests
from dgis.exceptions import DgisError
from dgis.utils import force_text
__all__ = ('bind_api',)
def __init__(self, api):
self.api = api
def execute(self, *args, **kwargs):
# Build GET parameters for query
parameters = {}
# Both positional
for idx, arg in enumerate(args):
if arg is None:
continue
try:
parameters[self.allowed_param[idx]] = force_text(arg)
except IndexError:
raise ValueError('Too many parameters supplied')
# And keyword parameters
for key, arg in kwargs.items():
if arg is None:
continue
if key in parameters:
raise ValueError('Multiple values for parameter %s supplied' % key)
parameters[key] = force_text(arg)
parameters.update({
'key': self.api.key,
'version': self.api.version,
'output': 'json',
})
url = urlparse.urlunparse(['http', self.api.host, self.path, None, urlencode(parameters), None])
response = requests.get(url).json
if inspect.ismethod(response):
response = response()
if response['response_code'] != '200':
raise DgisError(int(response['response_code']), response['error_message'], response['error_code'])
# Register view if required
if self.register_views and self.api.register_views:
if requests.get(response['register_bc_url']).text == '0':
raise DgisError(404, 'View registration cannot be processed', 'registerViewFailed')
return response
def bind_api(**config):
properties = {
'path': config['path'],
'method': config.get('method', 'GET'),
'allowed_param': config['allowed_param'],
'register_views': config.get('register_views', False),
'__init__': __init__,
'execute': execute,
}
cls = type('API%sMethod' % config['path'].title().replace('/', ''), (object,), properties)
def _call(api, *args, **kwargs):
return cls(api).execute(*args, **kwargs)
return _call | 2gis | /2gis-1.3.1.tar.gz/2gis-1.3.1/dgis/binder.py | binder.py |
from dgis.binder import bind_api
__version__ = '1.3.1'
class API(object):
"""2GIS API"""
def __init__(self, key, host='catalog.api.2gis.ru', version='1.3', register_views=True):
"""
Parameters::
key : user API key
host : base URL for queries
version : API version for working
register_views : send information to stats server about a firm profile viewing
"""
self.key = key
self.host = host
self.version = version
self.register_views = register_views
"""Projects lists
http://api.2gis.ru/doc/firms/list/project-list/
"""
project_list = bind_api(
path='/project/list',
allowed_param=[],
)
"""List of the project cities
http://api.2gis.ru/doc/firms/list/city-list/
"""
city_list = bind_api(
path='/city/list',
allowed_param=['where', 'project_id']
)
def rubricator(self, **kwargs):
"""Rubrics search
http://api.2gis.ru/doc/firms/list/rubricator/
"""
# `show_children' parameter must be an integer
kwargs['show_children'] = int(kwargs.pop('show_children', False))
return self._rubricator(**kwargs)
_rubricator = bind_api(
path='/rubricator',
allowed_param=['where', 'id', 'parent_id', 'show_children', 'sort'],
)
def search(self, **kwargs):
"""Firms search
http://api.2gis.ru/doc/firms/searches/search/
"""
point = kwargs.pop('point', False)
if point:
kwargs['point'] = '%s,%s' % (point[0], point[1])
bound = kwargs.pop('bound', False)
if bound:
kwargs['bound[point1]'] = bound[0]
kwargs['bound[point2]'] = bound[1]
filters = kwargs.pop('filters', False)
if filters:
for k, v in filters.items():
kwargs['filters[%s]' % k] = v
return self._search(**kwargs)
_search = bind_api(
path='/search',
allowed_param=['what', 'where', 'point', 'radius', 'bound', 'page', 'pagesize', 'sort', 'filters'],
)
def search_in_rubric(self, **kwargs):
"""Firms search in rubric
http://api.2gis.ru/doc/firms/searches/searchinrubric/
"""
point = kwargs.pop('point', False)
if point:
kwargs['point'] = '%s,%s' % point
bound = kwargs.pop('bound', False)
if bound:
kwargs['bound[point1]'] = bound[0]
kwargs['bound[point2]'] = bound[1]
filters = kwargs.pop('filters', False)
if filters:
for k, v in filters.items():
kwargs['filters[%s]' % k] = v
return self._search_in_rubric(**kwargs)
_search_in_rubric = bind_api(
path='/searchinrubric',
allowed_param=['what', 'where', 'point', 'radius', 'bound', 'page', 'pagesize', 'sort', 'filters'],
)
"""Firm filials
http://api.2gis.ru/doc/firms/searches/firmsbyfilialid/
"""
firms_by_filial_id = bind_api(
path='/firmsByFilialId',
allowed_param=['firmid', 'page' 'pagesize'],
)
"""Adverts search
http://api.2gis.ru/doc/firms/searches/adssearch/
"""
ads_search = bind_api(
path='/ads/search',
allowed_param=['what', 'where', 'format', 'page', 'pagesize'],
)
"""Firm profile
http://api.2gis.ru/doc/firms/profiles/profile/
"""
profile = bind_api(
path='/profile',
allowed_param=['id', 'hash'],
register_views=True,
)
def geo_search(self, **kwargs):
"""Geo search
http://api.2gis.ru/doc/geo/search/
"""
if 'types' in kwargs:
kwargs['types'] = ','.join(kwargs['types'])
bound = kwargs.pop('bound', False)
if bound:
kwargs['bound[point1]'] = bound[0]
kwargs['bound[point2]'] = bound[1]
return self._geo_search(**kwargs)
_geo_search = bind_api(
path='/geo/search',
allowed_param=['q', 'types', 'radius', 'limit', 'project', 'bound', 'format'],
)
"""Information about a geo object"""
geo_get = bind_api(
path='/geo/search',
allowed_param=['id', ],
) | 2gis | /2gis-1.3.1.tar.gz/2gis-1.3.1/dgis/__init__.py | __init__.py |
# Python 2ip Module
**2ip** allows you to make requests to the 2ip.me API to retrieve provider/geographic information for IP addresses. Requests are (optionally, on by default) cached to prevent unnecessary API lookups when possible.
## Installation
Install the module from PyPI:
```bash
python3 -m pip install 2ip
```
## Methods
The following methods are available.
### TwoIP (Initialisation)
When initialising the 2ip module the following parameters may be specified:
* *Optional* `key`: The API key to use for lookups. If no API key defined the API lookups will use the rate limited free API.
### geo
The geographic lookup method accepts the following parameters:
* *Required* `ip`: The IP address to lookup.
* *Optional* `format` {**json**,xml}: The output format for the request. `json` will return a dict and `xml` will return a string.
* *Optional* `force` {True,**False**}: Force an API lookup even if there is a cache entry.
* *Optional* `cache` {**True**,False}: Allow the lookup result to be cached.
### provider
The provider lookup method accepts the following parameters:
* *Required* `ip`: The IP address to lookup.
* *Optional* `format` {**json**,xml}: The output format for the request. `json` will return a dict and `xml` will return a string.
* *Optional* `force` {True,**False**}: Force an API lookup even if there is a cache entry.
* *Optional* `cache` {**True**,False}: Allow the lookup result to be cached.
## Examples
Some example scripts are included in the [examples](examples/) directory.
### Provider API
Retrieve provider information for the IP address `192.0.2.0` as a `dict`:
```python
>>> from twoip import TwoIP
>>> twoip = TwoIP(key = None)
>>> twoip.provider(ip = '192.0.2.0')
{'ip': '192.0.2.0',
'ip_range_end': '3221226239',
'ip_range_start': '3221225984',
'mask': '24',
'name_ripe': 'Reserved AS',
'name_rus': '',
'route': '192.0.2.0'}
```
Retrieve provider information for the IP address `192.0.2.0` as a XML string:
```python
>>> from twoip import TwoIP
>>> twoip = TwoIP(key = None)
>>> twoip.provider(ip = '192.0.2.0', format = 'xml')
'<?xml version="1.0" encoding="UTF-8"?>\n<provider_api><ip>192.0.2.0</ip><name_ripe>Reserved AS</name_ripe><name_rus></name_rus><ip_range_start>3221225984</ip_range_start><ip_range_end>3221226239</ip_range_end><route>192.0.2.0</route><mask>24</mask></provider_api>'
```
### Geographic API
Retrieve geographic information for the IP address `8.8.8.8` as a `dict`:
```python
>>> from twoip import TwoIP
>>> twoip = TwoIP(key = None)
>>> twoip.geo(ip = '8.8.8.8')
{'city': 'Mountain view',
'country': 'United states of america',
'country_code': 'US',
'country_rus': 'США',
'country_ua': 'США',
'ip': '8.8.8.8',
'latitude': '37.405992',
'longitude': '-122.078515',
'region': 'California',
'region_rus': 'Калифорния',
'region_ua': 'Каліфорнія',
'time_zone': '-08:00',
'zip_code': '94043'}
```
Retrieve geographic information for the IP address `8.8.8.8` as a XML string:
```python
>>> from twoip import TwoIP
>>> twoip = TwoIP(key = None)
>>> twoip.geo(ip = '8.8.8.8', format = 'xml')
'<?xml version="1.0" encoding="UTF-8"?>\n<geo_api><ip>8.8.8.8</ip><country_code>US</country_code><country>United states of america</country><country_rus>США</country_rus><country_ua>США</country_ua><region>California</region><region_rus>Калифорния</region_rus><region_ua>Каліфорнія</region_ua><city>Mountain view</city><latitude>37.405992</latitude><longitude>-122.078515</longitude><zip_code>94043</zip_code><time_zone>-08:00</time_zone></geo_api>'
```
## Roadmap/Todo
- [ ] Support for email API
- [ ] Support for MAC address API
- [ ] Support for hosting API
- [x] Option to retrieve data as XML
- [ ] Unit tests
- [x] Deduplicate handler to retrieve information from API
| 2ip | /2ip-0.0.2.tar.gz/2ip-0.0.2/README.md | README.md |
import logging
import requests
from ipaddress import ip_address
from typing import Optional, Union
class TwoIP(object):
"""
2ip.me API client
"""
# The API URL
api = 'https://api.2ip.me'
# List of available API endpoints
endpoints = {
'ip': {
'geo': {
'format': [ 'json', 'xml' ],
'description': 'Geographic information for IP addresses',
},
'provider': {
'format': [ 'json', 'xml' ],
'description': 'Provider information for IP addresses',
},
},
}
def __init__(self, key: Optional[str] = None) -> object:
"""Set up new 2ip.me API client.
Parameters
----------
key : str, optional
Optional API key to use for requests to the API
Returns
-------
TwoIP : object
The TwoIP API object
Examples
--------
>>> twoip = TwoIP(key = None)
"""
# Set API key
self.__key = key
# Create empty cache
self.__cache = {}
for endpoint_type in self.endpoints:
for type in self.endpoints[endpoint_type]:
self.__cache[type] = {}
for format in self.endpoints[endpoint_type][type]['format']:
self.__cache[type][format] = {}
# Debugging
if key:
logging.debug(f'Setup of new TwoIP object')
else:
logging.debug(f'Setup of new TwoIP object (No API key used - rate limits will apply)')
@staticmethod
def __api_request(url: str, params: Optional[dict] = None) -> object:
"""Send request to the 2ip API and return the requests object.
Parameters
----------
url : str
The URL to send request to
params : dict, Optional
Optional GET parameters to add to the request
Returns
-------
object
The requests object
Raises
------
RuntimeError
If the API request fails
"""
# Send the request
try:
req = requests.request(method = 'GET', url = url, params = params)
except Exception as e:
raise RuntimeError('API request failed') from e
# Make sure API request didn't result in rate limit
if req.status_code == 429:
raise RuntimeError(f'API has reached rate limit; retry in an hour or use an API key')
# Make sure response code is fine
if req.status_code != 200:
raise RuntimeError(f'Received unexpected response code "{req.status_code}" from API')
# Return the requests object
return req
def __api_param(self, params: dict) -> dict:
"""Add API key to the list of parameters for the API request if available.
Parameters
----------
params : dict
The parameters to add API key to
Returns
-------
dict
The built parameter list
"""
# Add API key if available
if self.__key:
params['key'] = self.__key
# Return params
return params
@staticmethod
def __test_ip(ip: str) -> bool:
"""Test if the IP address provided is really an IP address.
Parameters
----------
params : ip
The IP address to test
Returns
-------
bool
True if the string provided is an IP address or False if it is not an IP address
"""
logging.debug(f'Testing if "{ip}" is a valid IP address')
# Check for exception when creating an ip_address object
try:
ip_address(ip)
except Exception as e:
logging.error(f'Could not validate string "{ip}" as an IP address: {e}')
return False
# IP provided is valid
return True
def __execute_ip(self, ip: str, format: str, type: str) -> Union[dict, str]:
"""Execute an API lookup for an IP address based provider (geographic information or provider information)
Parameters
----------
ip : str
The IP address to lookup
format : {'json', 'xml'}
The format for which results should be returned - json results will be returned as a dict
type : {'geo','provider'}
The API type to lookup ('geo' for geographic information or 'provider' for provider information)
Returns
-------
dict or str
The lookup result in either dict format (for JSON lookups) or string format (for XML lookups)
"""
# Make sure there is an endpoint for the requested type
if not self.endpoints['ip'][type]:
raise ValueError(f'Invalid lookup type "{type}" requested')
# Make sure that the format selected is valid
if format not in self.endpoints['ip'][type]['format']:
raise ValueError(f'The "{type}" API endpoint does not support the requested format "{format}"')
# Build the API URL that the request will be made to
url = f'{self.api}/{type}.{format}'
# Set the parameters for the request (this will add the API key if available)
params = self.__api_param({'ip': ip})
# Send API request
try:
response = self.__api_request(url = url, params = params)
except Exception as e:
logging.error('Failed to send API request: {e}')
raise RuntimeError('Could not run API lookup') from e
# Make sure content type of the response is expected
if response.headers.get('Content-Type') != f'application/{format}':
if response.headers.get('Content-Type'):
raise RuntimeError(f'Expecting Content-Type header "application/{format}" but got "{req.headers.get("Content-Type")}"')
else:
raise RuntimeError(f'API request did not provide any Content-Type header')
# If the format requests is XML, return it immediately otherwise attempt to parse the json response into a dict
if format == 'xml':
return response.text
elif format == 'json':
try:
json = response.json()
except Exception as e:
raise RuntimeError('Exception parsing response from API as json') from e
return json
else:
raise RuntimeError(f'No output handler configured for the output format "{format}"')
def geo(self, ip: str, format: str = 'json', force: bool = False, cache: bool = True) -> Union[dict, str]:
"""Perform a Geographic information lookup to the 2ip API.
By default, lookups will be cached. Subsequent lookups will use the cache (unless disabled) to prevent excessive API requests.
Parameters
----------
ip : str
The IP address to lookup
format : {'xml','json'}
The requested output format - JSON will result in a dict otherwise XML will result in a str
force : bool, default = False
Force request to be sent to the API even if the lookup has already been cached
cache: bool, default = True
Allow caching of the request (cache can still be bypassed by using the force parameter)
Returns
-------
dict or str
The response from API in either a dictionary (if the format is set to json) or a string (if the format is set to XML)
Raises
------
ValueError
Invalid IP address or invalid output format requested
RuntimeError
API request failure
Examples
--------
>>> twoip = TwoIP(key = '12345678')
>>> twoip.geo(ip = '192.0.2.0')
{'ip': '192.0.2.0', 'country_code': '-', 'country': '-', 'country_rus': 'Неизвестно', 'country_ua': 'Невідомо',
'region': '-', 'region_rus': 'Неизвестно', 'region_ua': 'Невідомо', 'city': '-', 'city_rus': 'Неизвестно',
'city_ua': 'Невідомо', 'zip_code': '-', 'time_zone': '-'}
"""
# Make sure IP address provided is valid
if self.__test_ip(ip = ip):
logging.debug(f'2ip geo API lookup request for IP "{ip}" (force: {force})')
else:
raise ValueError(f'Could not run geo lookup for invalid IP address "{ip}"')
# If the lookup is cached and this is not a forced lookup, return the cached result
if ip in self.__cache['geo'][format]:
logging.debug(f'Cached API entry available for IP "{ip}"')
if force:
logging.debug('Cached entry will be ignored for lookup; proceeding with API request')
else:
logging.debug('Cached entry will be returned; use the force parameter to force API request')
return self.__cache['geo'][format][ip]
# Run the request
response = self.__execute_ip(ip = ip, format = format, type = 'geo')
# Cache the response if allowed
if cache:
self.__cache['geo'][format][ip] = response
# Return response
return response
def provider(self, ip: str, format: str = 'json', force: bool = False, cache: bool = True) -> Union[dict, str]:
"""Perform a provider information lookup to the 2ip API.
By default, lookups will be cached. Subsequent lookups will use the cache (unless disabled) to prevent excessive API requests.
Parameters
----------
ip : str
The IP address to lookup
format : {'xml','json'}
The requested output format - JSON will result in a dict otherwise XML will result in a str
force : bool, default = False
Force request to be sent to the API even if the lookup has already been cached
cache: bool, default = True
Allow caching of the request (cache can still be bypassed by using the force parameter)
Returns
-------
dict or str
The response from API in either a dictionary (if the format is set to json) or a string (if the format is set to XML)
Raises
------
ValueError
Invalid IP address or invalid output format requested
RuntimeError
API request failure
Examples
--------
>>> twoip = TwoIP(key = '12345678')
>>> twoip.provider(ip = '192.0.2.0')
{'ip': '192.0.2.0', 'name_ripe': 'Reserved AS', 'name_rus': '', 'ip_range_start': '3221225984', 'ip_range_end': '3221226239',
'route': '192.0.2.0', 'mask': '24'}
"""
# Make sure IP address provided is valid
if self.__test_ip(ip = ip):
logging.debug(f'2ip provider API lookup request for IP "{ip}" (force: {force})')
else:
raise ValueError(f'Could not run provider lookup for invalid IP address "{ip}"')
# If the lookup is cached and this is not a forced lookup, return the cached result
if ip in self.__cache['provider'][format]:
logging.debug(f'Cached API entry available for IP "{ip}"')
if force:
logging.debug('Cached entry will be ignored for lookup; proceeding with API request')
else:
logging.debug('Cached entry will be returned; use the force parameter to force API request')
return self.__cache['provider'][format][ip]
# Run the request
response = self.__execute_ip(ip = ip, format = format, type = 'provider')
# Cache the response if allowed
if cache:
self.__cache['provider'][format][ip] = response
# Return response
return response | 2ip | /2ip-0.0.2.tar.gz/2ip-0.0.2/twoip/twoip.py | twoip.py |
##########
2lazy2rest
##########
A simple way to produce short-to-medium document using *reStructuredText*
Multi-format themes
Render the same document in HTML, ODT, PDF keeping the main visual identity
Unified interface
- Tired of switching between rst2* tools having different arguments or behavior ?
- Would like to not lose *code-blocks* or some rendering options switching the output format ?
This tool try to address this
Make your own theme
TODO: templates will be customizable easily (say, probably colors only)
How to use it
#############
Dependencies
============
You'll need **rst2pdf** to use all the features, other rst2* tools are coming from docutils.
Using
=====
.. code-block:: console
mkrst [-h] [--html] [--pdf] [--odt] [--theme THEME]
[--themes-dir THEMES_DIR]
FILE
optional arguments:
-h, --help show this help message and exit
--html Generate HTML output
--pdf Generate PDF output
--odt Generate ODT output
--theme THEME Use a different theme
--themes-dir THEMES_DIR
Change the folder searched for theme
.. code-block:: console
popo:~/2lazy2rest% ./mkrst test_page.rst --html --pdf
Using ./themes/default
html: test_page.html
pdf: test_page.pdf
Customizing
===========
Make a copy of ``themes/default``, edit to your needs the copy and use the **--theme** option with the name of your copy, that's All !
Example
-------
.. code-block:: console
popo:~/2lazy2rest% cp -r themes/default themes/red
popo:~/2lazy2rest% sed -si 's/#FEFEFE/red/g' themes/red/html/stylesheet.css
popo:~/2lazy2rest% ./mkrst test_page.rst --html --theme red
Issues
######
- ODT style is unfinished
- PDF & HTML still needs more ReST coverage
- No skin generation from template yet
| 2lazy2rest | /2lazy2rest-0.2.2.tar.gz/2lazy2rest-0.2.2/README.rst | README.rst |
from __future__ import absolute_import, print_function
import argparse
import importlib.util
import os
import subprocess
import mkrst_themes
THEMES_DIR = os.path.dirname(mkrst_themes.__file__)
THEME = "default"
SUPPORTED_FORMATS = "html pdf odt".split()
DEBUG = "DEBUG" in os.environ
class CommandExecutor:
GENERAL_OPTIONS = ["--syntax-highlight=short", "--smart-quotes=yes"]
def __init__(self, opts: argparse.Namespace):
self._src: str = opts.input
self._verbose: bool = opts.verbose
self._extensions: str = opts.ext
def set_theme(self, path: str, name: str) -> None:
if os.path.exists(name):
self.theme_dir = os.path.abspath(name)
else:
self.theme_dir = os.path.join(path, name)
print("Using %s" % self.theme_dir)
def _list_resources(self, origin: str, extension: str) -> list[str]:
r = []
for fname in os.listdir(os.path.join(self.theme_dir, origin)):
if fname.endswith(extension):
r.append(os.path.join(self.theme_dir, origin, fname))
r.sort()
if DEBUG:
print("List(*%s) = %s" % (extension, ", ".join(r)))
return r
def get_output(self, fmt: str) -> str:
return self._src.rsplit(".", 1)[0] + "." + fmt
def __execute(self, args: list[str]) -> None | tuple[bytes, bytes]:
if DEBUG:
print("Executing %s" % " ".join(args))
proc = subprocess.Popen(args, stdout=subprocess.PIPE, stderr=subprocess.PIPE)
out, err = proc.communicate()
def show_result() -> None:
print("\nStdout:")
print(out.decode("utf-8").strip())
print("\nStderr:")
print(err.decode("utf-8").strip())
if proc.returncode == 0:
if self._verbose:
show_result()
return (out, err)
else:
print(" PROBLEM! ".center(80, "#"))
show_result()
return None
def make_pdf(self) -> None:
out = self.get_output("pdf")
opts = [o for o in self.GENERAL_OPTIONS if "syntax" not in o]
opts += ["--font-path", os.path.join(self.theme_dir, "fonts")]
args = [
"rst2pdf",
"--fit-background-mode=scale",
"--use-floating-images",
"--real-footnotes",
"-e",
"preprocess",
] + opts
for style in self._list_resources("pdf", ".yaml"):
args.extend(("-s", style))
if self._extensions:
args.extend(("-e", self._extensions))
args.extend(("-o", out, self._src))
if self.__execute(args):
print(out)
def make_odt(self) -> None:
out = self.get_output("odt")
css = self._list_resources("odt", "odt")[0]
args = ["rst2odt"] + self.GENERAL_OPTIONS + ["--add-syntax-highlighting", "--stylesheet", css, self._src, out]
if self.__execute(args):
print(out)
def make_html(self) -> None:
out = self.get_output("html")
css = ",".join(self._list_resources("html", "css"))
args = ["rst2html"] + self.GENERAL_OPTIONS + ["--embed-stylesheet", "--stylesheet-path", css, self._src, out]
if self.__execute(args):
print(out)
# post process
links = os.path.join(self.theme_dir, "html", "links.html")
if os.path.exists(links):
data = open(out).read()
data = data.replace("<html>", "<html>\n" + open(links).read())
open(out, "w").write(data)
def main() -> None:
parser = argparse.ArgumentParser()
parser.add_argument("-v", "--verbose", action="store_true", help="Display commands' output")
parser.add_argument(
"input",
type=str,
metavar="FILE",
nargs="?",
help="The .rst or .txt file to convert",
)
for fmt in SUPPORTED_FORMATS:
parser.add_argument("--%s" % fmt, action="store_true", help="Generate %s output" % fmt.upper())
parser.add_argument("-t", "--theme", type=str, default=THEME, help="Use a different theme")
parser.add_argument(
"--themes-dir",
type=str,
default=THEMES_DIR,
help="Change the folder searched for theme",
)
parser.add_argument("--ext", type=str, default="", help="Load docutils extensions (coma separated)")
parser.add_argument("-l", "--list", action="store_true", help="List available themes")
args = parser.parse_args()
cmd = CommandExecutor(args)
cmd.set_theme(args.themes_dir, args.theme)
if args.list:
for t in os.listdir(THEMES_DIR):
if t[0] not in "_." and os.path.isdir(os.path.join(THEMES_DIR, t)):
print(t)
elif not any(getattr(args, fmt) for fmt in SUPPORTED_FORMATS):
print("No action ! Give one or more --(%s)" % "|".join(SUPPORTED_FORMATS))
else:
if not args.input:
print("No input file !")
else:
for fmt in SUPPORTED_FORMATS:
if getattr(args, fmt):
fn = getattr(cmd, "make_%s" % fmt)
print(" %5s: " % fmt, end="")
fn()
if __name__ == "__main__":
main() | 2lazy2rest | /2lazy2rest-0.2.2.tar.gz/2lazy2rest-0.2.2/mkrst/__init__.py | __init__.py |
This is a continually-updated market data analysis machine learning transformer model intended for summarizing & interpreting market & economic data.
To use:
1. gh repo clone sscla1/2op_ai
2. install requirements.txt wherever running your application
3. import statement: from 2op_ai import ai_pipeline
4. use case for summarizer & interpreter
def summarize_text(self, text):
"""
Summarizes the provided text.
Args:
text (str): The text to be summarized.
Returns:
str: The summarized text.
"""
try:
# Perform text summarization using AI pipeline
summary = self.ai_pipeline(text, max_length=600)[0]['summary_text']
return summary
except Exception as e:
logging.error(f"An error occurred when trying to summarize text: {e}")
return ""
def generate_interpretation(self, summary):
"""
Generates interpretation for the summary using the AI pipeline.
Args:
summary (str): The summary to generate interpretation for.
Returns:
str or None: The generated interpretation or None if an error occurs.
"""
logging.info("Attempting to generate interpretation.")
try:
input_text = "Explain the provided data, its affect on the market, and specific sectors, in layman's terms, in 300 words or less: " + summary
interpretation = self.ai(input_text, max_length=300, do_sample=True)[0]['generated_text']
return interpretation
except Exception as e:
logging.error(f"An error occurred when trying to generate interpretation: {e}")
return None | 2op-ai | /2op_ai-0.1.tar.gz/2op_ai-0.1/README.md | README.md |
import logging
import os
import pandas as pd
import nbformat
from nbconvert import PythonExporter
import chardet
import pdb
class DataProcessor:
def __init__(self, data_dir):
# Initialize the data directory and logger
self.data_dir = data_dir
self.logger = self.setup_logging()
def setup_logging(self):
# Set up logging with both console and file handlers
logger = logging.getLogger(__name__)
logger.setLevel(logging.INFO)
formatter = logging.Formatter('%(asctime)s - %(levelname)s - %(message)s')
stream_handler = logging.StreamHandler()
stream_handler.setFormatter(formatter)
logger.addHandler(stream_handler)
log_file_path = os.path.join(self.data_dir, "data_processing.log")
file_handler = logging.FileHandler(log_file_path)
file_handler.setFormatter(formatter)
logger.addHandler(file_handler)
return logger
def export_notebook_to_script(self, notebook_path, script_path):
# Export Jupyter Notebook to a Python script
nb = nbformat.read(notebook_path, as_version=nbformat.NO_CONVERT)
exporter = PythonExporter()
python_code, _ = exporter.from_notebook_node(nb)
with open(script_path, "w") as f:
f.write(python_code)
def load_csv(self, file_path, encoding='utf-8', dtype=None):
# Load CSV file with error handling for different encodings
try:
with open(file_path, 'rb') as f:
result = chardet.detect(f.read())
file_encoding = result['encoding']
return pd.read_csv(file_path, encoding=file_encoding, dtype=dtype, low_memory=False)
except UnicodeDecodeError:
self.logger.warning(f"UnicodeDecodeError encountered while loading CSV '{file_path}' with encoding '{encoding}'. Trying 'latin-1' encoding.")
try:
return pd.read_csv(file_path, encoding='latin-1', dtype=dtype, low_memory=False)
except Exception as e:
self.logger.error(f"Error loading CSV '{file_path}' with 'latin-1' encoding: {e}")
return None
except Exception as e:
self.logger.error(f"Error loading CSV '{file_path}': {e}")
return None
def load_excel(self, file_path, encoding='utf-8'):
# Load Excel file with error handling for different encodings
try:
return pd.read_excel(file_path, engine='openpyxl')
except UnicodeDecodeError:
self.logger.warning(f"UnicodeDecodeError encountered while loading Excel '{file_path}' with encoding '{encoding}'. Trying 'latin-1' encoding.")
try:
return pd.read_excel(file_path, engine='openpyxl', encoding='latin-1')
except Exception as e:
self.logger.error(f"Error loading Excel '{file_path}' with 'latin-1' encoding: {e}")
return None
except Exception as e:
self.logger.error(f"Error loading Excel '{file_path}': {e}")
return None
def process_sroie_files(self, directory):
# Process SROIE files from the given directory
sroie_data = []
for root, _, files in os.walk(directory):
for file in files:
if file.endswith('.txt'):
file_path = os.path.join(root, file)
try:
with open(file_path, 'rb') as txt_file:
raw_data = txt_file.read()
detected_encoding = chardet.detect(raw_data)['encoding']
text_data = raw_data.decode(detected_encoding)
sroie_data.append(text_data)
except Exception as e:
self.logger.error(f"Error reading '{file_path}': {e}")
sroie_df = pd.DataFrame({'data': sroie_data})
return sroie_df
def process_data(self):
# Process various data files and concatenate them into a single dataset
try:
breakpoint()
root_dir = os.path.join(self.data_dir)
# Load files
currency_data = self.load_csv(os.path.join(root_dir, 'currency.csv'))
fortune500_data = self.load_csv(os.path.join(root_dir, 'fortune500.csv'))
msft_data = self.load_csv(os.path.join(root_dir, 'msft.csv'))
snp_data = self.load_csv(os.path.join(root_dir, 's&p.csv'))
balancesheets_data = self.load_csv(os.path.join(root_dir, 'bank_fail', 'balancesheets.csv'))
banklist_data = self.load_csv(os.path.join(root_dir, 'bank_fail', 'banklist.csv'))
final_datasets_dir = os.path.join(root_dir, 'income', 'final_datasets')
final_datasets_files = [
'Final_32_region_deciles_1958-2015.csv',
'Final_Global_Income_Distribution.csv',
'Final_Historical_data_ISO.csv'
]
final_datasets_data = pd.concat(
[pd.read_csv(os.path.join(final_datasets_dir, file)) for file in final_datasets_files]
)
input_data_dir = os.path.join(root_dir, 'income', 'input_data')
input_data_files = [
'WDI_GINI_data.csv'
]
input_data = pd.concat(
[pd.read_csv(os.path.join(input_data_dir, file)) for file in input_data_files]
)
mapping_files_dir = os.path.join(root_dir, 'income', 'mapping_files')
mapping_files_files = [
'GCAM_region_names.csv',
'WIDER_mapping_file.csv'
]
mapping_files = pd.concat(
[pd.read_csv(os.path.join(mapping_files_dir, file)) for file in mapping_files_files]
)
train_dataset = pd.concat([currency_data, fortune500_data, msft_data, snp_data,
balancesheets_data, banklist_data, final_datasets_data,
input_data, mapping_files])
etfs_dir = os.path.join(root_dir, 'nasdaq', 'etfs')
etfs_files = [file for file in os.listdir(etfs_dir) if file.endswith('.csv')]
etfs_data = pd.concat([pd.read_csv(os.path.join(etfs_dir, file)) for file in etfs_files])
stocks_dir = os.path.join(root_dir, 'nasdaq', 'stocks')
stocks_files = [file for file in os.listdir(stocks_dir) if file.endswith('.csv')]
stocks_data = pd.concat([pd.read_csv(os.path.join(stocks_dir, file)) for file in stocks_files])
symbols_valid_meta_data = self.load_csv(os.path.join(root_dir, 'nasdaq', 'symbols_valid_meta.csv'))
train_dataset = pd.concat([train_dataset, etfs_data, stocks_data, symbols_valid_meta_data])
news_dir = os.path.join(root_dir, 'news')
news_files = [file for file in os.listdir(news_dir) if file.endswith('.csv')]
news_data = pd.concat([pd.read_csv(os.path.join(news_dir, file)) for file in news_files])
train_dataset = pd.concat([train_dataset, news_data])
prediction_dir = os.path.join(root_dir, 'prediction')
prediction_files = [file for file in os.listdir(prediction_dir) if file.endswith('.csv')]
prediction_data = pd.concat([pd.read_csv(os.path.join(prediction_dir, file)) for file in prediction_files])
train_dataset = pd.concat([train_dataset, prediction_data])
sroie_dir = os.path.join(root_dir, 'sroie')
sroie_df = self.process_sroie_files(sroie_dir)
breakpoint()
train_dataset = pd.concat([train_dataset, sroie_df]) # Concatenate sroie_df with the train_dataset
self.logger.info("Data processing completed successfully.")
return train_dataset
except Exception as e:
self.logger.error(f"An error occurred during data processing: {e}")
return None
if __name__ == "__main__":
script_dir = os.path.dirname(os.path.abspath(__file__))
project_dir = os.path.dirname(script_dir)
data_dir = os.path.join(project_dir, "training_data")
data_processor = DataProcessor(data_dir)
currency_data = data_processor.load_csv(os.path.join(data_processor.data_dir, 'currency.csv'))
print("Currency Data:", currency_data)
fortune500_data = data_processor.load_csv(os.path.join(data_processor.data_dir, 'fortune500.csv'))
print("Fortune500 Data:", fortune500_data)
msft_data = data_processor.load_csv(os.path.join(data_processor.data_dir, 'msft.csv'))
print("MSFT Data:", msft_data)
snp_data = data_processor.load_csv(os.path.join(data_processor.data_dir, 's&p.csv'))
print("S&P Data:", snp_data)
breakpoint()
trained_dataset = data_processor.process_data()
if trained_dataset is not None:
print("Data processing completed successfully.")
trained_dataset.to_csv(os.path.join(data_dir, "trained_dataset.csv"), index=False)
else:
print("Data processing failed.") | 2op-ai | /2op_ai-0.1.tar.gz/2op_ai-0.1/src/processing.py | processing.py |
import os
import logging
import torch
from transformers import AutoTokenizer, AutoModelForCausalLM
from processing import DataProcessor
from finetuning import finetune_model
def main():
# Get the directory of the currently executing script
script_dir = os.path.dirname(os.path.abspath(__file__))
logging.basicConfig(level=logging.INFO, format="%(asctime)s - %(levelname)s - %(message)s")
try:
# Instantiate the tokenizer and model
tokenizer = AutoTokenizer.from_pretrained("Salesforce/xgen-7b-8k-base", trust_remote_code=True)
model = AutoModelForCausalLM.from_pretrained("Salesforce/xgen-7b-8k-base", torch_dtype=torch.bfloat16)
# Set up fine-tuning parameters
num_epochs = 10
batch_size = 16
learning_rate = 1e-5
# Construct the project directory path using the script's directory
project_dir = os.path.dirname(script_dir)
data_dir = os.path.join(project_dir, "training_data")
print("Instantiating DataProcessor...")
data_processor = DataProcessor(data_dir)
print("Processing data...")
trained_dataset = data_processor.process_data() # Only train dataset for fine-tuning
if trained_dataset is not None:
# Calculate validation dataset (val_dataset) for fine-tuning
val_dataset = trained_dataset.sample(frac=0.2, random_state=42)
trained_dataset.to_csv(os.path.join(data_dir, "trained_dataset.csv"), index=False)
else:
logging.error("Data processing failed.")
return
# Fine-tune the model using the imported function
print("Fine-tuning model...")
finetune_model(model, trained_dataset, val_dataset, num_epochs, batch_size, learning_rate)
logging.info("Fine-tuning completed successfully.")
# Example usage for aipipeline, summarizer, and interpreter
from transformers import pipeline, TextGenerationPipeline
# Create an AI pipeline
aipipeline = pipeline("text-generation", model=model, tokenizer=tokenizer)
# Summarizer
summarizer = TextGenerationPipeline(model=model, tokenizer=tokenizer)
# Interpreter
def interpret(text):
inputs = tokenizer(text, return_tensors="pt")
outputs = model(**inputs)
predicted_class = torch.argmax(outputs.logits, dim=-1).item()
return predicted_class
except Exception as e:
logging.error(f"An error occurred: {e}")
raise
if __name__ == "__main__":
main() | 2op-ai | /2op_ai-0.1.tar.gz/2op_ai-0.1/src/main.py | main.py |
import logging
import os
import torch
from transformers import AutoTokenizer, AutoModelForCausalLM
from processing import DataProcessor
logging.basicConfig(level=logging.INFO, format="%(asctime)s - %(levelname)s - %(message)s")
def finetune_model(model, train_dataset, val_dataset, num_epochs, batch_size, learning_rate):
try:
# Instantiate the tokenizer with the unk_token set
tokenizer = AutoTokenizer.from_pretrained(
"Salesforce/xgen-7b-8k-base",
unk_token="<unk>", # Set the unk_token
trust_remote_code=True
)
# Enable token tracing
tokenizer._tokenizer.enable_tracing()
# Set up the optimizer and loss function, data loaders, and fine-tuning loop
optimizer = torch.optim.AdamW(model.parameters(), lr=learning_rate)
loss_fn = torch.nn.CrossEntropyLoss()
train_loader = torch.utils.data.DataLoader(train_dataset, batch_size=batch_size, shuffle=True)
val_loader = torch.utils.data.DataLoader(val_dataset, batch_size=batch_size)
# Fine-tuning loop
for epoch in range(num_epochs):
model.train()
for inputs, labels in train_loader:
optimizer.zero_grad()
# Tokenize the text using the tokenizer
tokenized_inputs = tokenizer(inputs, return_tensors="pt", padding=True, truncation=True)
outputs = model(**tokenized_inputs)
logits = outputs.logits
loss = loss_fn(logits, labels)
loss.backward()
optimizer.step()
model.eval()
with torch.no_grad():
total_loss = 0.0
total_correct = 0
total_samples = 0
for inputs, labels in val_loader:
# Tokenize the text using the tokenizer
tokenized_inputs = tokenizer(inputs, return_tensors="pt", padding=True, truncation=True)
outputs = model(**tokenized_inputs)
logits = outputs.logits
loss = loss_fn(logits, labels)
total_loss += loss.item()
_, predicted = torch.max(logits, 1)
total_correct += (predicted == labels).sum().item()
total_samples += labels.size(0)
# Calculate val_loss and val_accuracy
val_loss = total_loss / len(val_loader)
val_accuracy = total_correct / total_samples
logging.info(f"Epoch {epoch+1}: Val Loss = {val_loss:.4f}, Val Accuracy = {val_accuracy:.4f}")
# Disable token tracing
tokenizer._tokenizer.disable_tracing()
except Exception as e:
logging.error(f"An error occurred during fine-tuning: {e}")
raise
if __name__ == "__main__":
# Get the directory of the currently executing script
script_dir = os.path.dirname(os.path.abspath(__file__))
# Construct the project directory path using the script's directory
project_dir = os.path.dirname(script_dir)
data_dir = os.path.join(project_dir, "training_data") # Construct the data directory path
# Instantiate the DataProcessor class
data_processor = DataProcessor(data_dir)
# Load and process the datasets
trained_dataset = data_processor.process_data()
if trained_dataset is not None:
# Calculate batch size and learning rate
batch_size = 32
num_samples = len(trained_dataset)
num_epochs = 5
learning_rate = 1e-4
# Calculate validation dataset (val_dataset) for fine-tuning
val_dataset = trained_dataset.sample(frac=0.2, random_state=42)
# Instantiate the model for fine-tuning
model = AutoModelForCausalLM.from_pretrained("Salesforce/xgen-7b-8k-base")
# Fine-tune the model using the train_dataset and val_dataset
finetune_model(model, trained_dataset, val_dataset, num_epochs, batch_size, learning_rate)
logging.info("Fine-tuning completed successfully.")
else:
logging.error("Data processing failed.") | 2op-ai | /2op_ai-0.1.tar.gz/2op_ai-0.1/src/finetuning.py | finetuning.py |
__version__ = '0.1.0'
import os
import sys
import re
# TODO use ast
# TODO use 2to3
# TODO use pyflakes
def classify_string(s):
"Returns 2 or 3 based on a string s of Python source code, defaulting to 3"
first_line = s.splitlines()[0]
if '#!' in first_line:
if 'python3' in first_line:
return 3
elif 'python2' in first_line:
return 2
if 'yield from' in s:
return 3
if 'Error, e' in s:
return 2
if 'xrange' in s:
return 2
if (re.search(r'''u"[^"]*"[.]decode''', s) or
re.search(r'''u'[^']*'[.]decode''', s) or
re.search(r'''b"[^"]*"[.]encode''', s) or
re.search(r'''b'[^']*'[.]encode''', s)):
return 2
return 3
def has_pycache(filename):
"Returns bool for if there exists a __pycache__ directory next to a file"
neighbors = os.listdir(os.path.dirname(os.path.abspath(filename)))
if '__pycache__' in neighbors:
return True
else:
return False
def pycache_classify(filename):
"""Returns 3, 2, or None for prediction from __pycache__ directory
prefers 3 if both found, returns None if neither found"""
dirname, name = os.path.split(os.path.abspath(filename))
neighbors = os.listdir(dirname)
if '__pycache__' in neighbors:
classifications = [
n[-6:-5]
for n in os.listdir(os.path.join(dirname, '__pycache__'))
if n.split('.', 1)[0] == name.split('.', 1)[0]]
if '3' in classifications:
return 3
elif '2' in classifications:
return 2
return None
def classify(filename):
r = pycache_classify(filename)
if r:
return r
return classify_string(open(filename).read())
def main(error_if_not=None):
"""Writes 2 or 3 and a newline to stdout, uses exit code for error_if_not"""
r = classify(sys.argv[-1])
print r
if error_if_not is not None:
sys.exit(0 if r == error_if_not else r)
if __name__ == '__main__':
main() | 2or3 | /2or3-0.1.0.tar.gz/2or3-0.1.0/lib2or3/__init__.py | __init__.py |
from tuprolog import logger
# noinspection PyUnresolvedReferences
import jpype
# noinspection PyUnresolvedReferences
import jpype.imports
# noinspection PyUnresolvedReferences
import java.io as _jio
# noinspection PyUnresolvedReferences
import java.nio.file as _jnio_file
from io import IOBase, SEEK_SET
from typing import Union
InputStream = _jio.InputStream
Reader = _jio.Reader
InputStreamReader = _jio.InputStreamReader
BufferedReader = _jio.BufferedReader
Files = _jnio_file.Files
Paths = _jnio_file.Paths
@jpype.JImplementationFor("java.io.InputStream")
class _JvmInputStream:
def __jclass_init__(cls):
IOBase.register(cls)
@jpype.JOverride
def close(self):
self._closed = True
self.close_()
@property
def closed(self):
if hasattr(self, '_closed'):
return self._closed
else:
return False
def fileno(self):
raise OSError("This IOBase is backed by an instance of <java.io.InputStream>")
def flush(self):
pass
def isatty(self):
return False
def readable(self):
return True
def seekable(self):
return False
def writable(self):
return False
def _buffer(self):
if not hasattr(self, '_buffered'):
self._buffered = BufferedReader(InputStreamReader(self))
def readline(self, size=-1):
self._buffer()
self._buffered.readLine()
def readlines(self, hint=-1):
self._buffer()
stream = self._buffered.read.lines()
if hint >= 0:
stream = stream.limit(hint)
return list(stream)
def seek(self, offset: int, whence=SEEK_SET):
raise OSError("This IOBase is backed by an instance of <java.io.InputStream>")
def tell(self):
raise OSError("This IOBase is backed by an instance of <java.io.InputStream>")
def truncate(self, size=None):
raise OSError("This IOBase is backed by an instance of <java.io.InputStream>")
def writelines(self):
raise OSError("This IOBase is backed by an instance of <java.io.InputStream>")
def open_file(path: str) -> InputStream:
return Files.newInputStream(Paths.get(path))
def ensure_input_steam(input: Union[str, InputStream]) -> InputStream:
if isinstance(input, str):
return open_file(input)
elif isinstance(input, InputStream):
return input
else:
raise TypeError("Invalid type: " + type(input))
logger.debug("Configure JVM-specific IO extensions") | 2ppy | /2ppy-0.4.0-py3-none-any.whl/tuprolog/jvmioutils.py | jvmioutils.py |
from tuprolog import logger
# noinspection PyUnresolvedReferences
import jpype
# noinspection PyUnresolvedReferences
import jpype.imports
# noinspection PyProtectedMember
from _jpype import _JObject as JObjectClass
# noinspection PyUnresolvedReferences
import java.util as _jutils
# noinspection PyUnresolvedReferences
import java.lang as _jlang
# noinspection PyUnresolvedReferences
import kotlin as _kotlin
# noinspection PyUnresolvedReferences
import kotlin.sequences as _ksequences
# noinspection PyUnresolvedReferences
import it.unibo.tuprolog.utils as _tuprolog_utils
from typing import Iterable as PyIterable
from typing import Iterator as PyIterator
from typing import Mapping, MutableMapping, Callable, Any
from .jvmioutils import *
Arrays = _jutils.Arrays
ArrayList = _jutils.ArrayList
Iterator = _jutils.Iterator
Map = _jutils.Map
NoSuchElementException = _jutils.NoSuchElementException
Iterable = _jlang.Iterable
JavaSystem = _jlang.System
Object = _jlang.Object
Pair = _kotlin.Pair
Triple = _kotlin.Triple
Sequence = _ksequences.Sequence
SequencesKt = _ksequences.SequencesKt
PyUtils = _tuprolog_utils.PyUtils
def protect_iterable(iterable: Iterable) -> Iterable:
return PyUtils.iterable(iterable)
@jpype.JImplements("java.util.Iterator", deferred=True)
class _IteratorAdapter(object):
def __init__(self, iterator):
assert isinstance(iterator, PyIterator)
self._iterator = iterator
self._queue = None
@jpype.JOverride
def hasNext(self):
if self._queue is None:
try:
self._queue = [next(self._iterator)]
return True
except StopIteration:
return False
elif len(self._queue) > 0:
return True
else:
try:
self._queue.append(next(self._iterator))
return True
except StopIteration:
return False
@jpype.JOverride
def next(self):
if self.hasNext():
return self._queue.pop(0)
else:
raise NoSuchElementException()
@jpype.JImplements("java.lang.Iterable", deferred=True)
class _IterableAdapter(object):
def __init__(self, iterable):
assert isinstance(iterable, PyIterable)
self._iterable = iterable
@jpype.JOverride
def iterator(self):
return _IteratorAdapter(iter(self._iterable))
def kpair(items: PyIterable) -> Pair:
if isinstance(items, Pair):
return items
i = iter(items)
first = next(i)
second = next(i)
return Pair(first, second)
@jpype.JConversion("kotlin.Pair", instanceof=PyIterable, excludes=str)
def _kt_pair_covert(jcls, obj):
return kpair(obj)
def ktriple(items: PyIterable) -> Triple:
if isinstance(items, Triple):
return items
i = iter(items)
first = next(i)
second = next(i)
third = next(i)
return Triple(first, second, third)
@jpype.JConversion("kotlin.Triple", instanceof=PyIterable, excludes=str)
def _kt_triple_covert(jcls, obj):
return ktriple(obj)
def jlist(iterable: PyIterable) -> Iterable:
assert isinstance(iterable, PyIterable)
if isinstance(iterable, list):
return Arrays.asList(iterable)
lst = ArrayList()
for item in iterable:
lst.add(item)
return lst
def jiterable(iterable: PyIterable) -> Iterable:
assert isinstance(iterable, PyIterable)
return _IterableAdapter(iterable)
@jpype.JConversion("java.lang.Iterable", instanceof=PyIterable, excludes=str)
def _java_iterable_convert(jcls, obj):
return jiterable(obj)
def jarray(type, rank: int = 1):
return jpype.JArray(type, rank)
def jiterator(iterator: PyIterator) -> Iterator:
assert isinstance(iterator, PyIterator)
return _IteratorAdapter(iterator)
def jmap(dictionary: Mapping) -> Map:
assert isinstance(dictionary, Mapping)
return Map@dictionary
def _java_obj_repr(java_object: Object) -> str:
return str(java_object.toString())
# replaces the default __repr__ implementation for java objects, making them use _java_obj_repr
JObjectClass.__repr__ = _java_obj_repr
@jpype.JImplementationFor("kotlin.sequences.Sequence")
class _KtSequence:
def __jclass_init__(self):
PyIterable.register(self)
def __iter__(self):
return protect_iterable(self).iterator()
def ksequence(iterable: PyIterable) -> Sequence:
return SequencesKt.asSequence(jiterable(iterable))
@jpype.JConversion("kotlin.sequences.Sequence", instanceof=PyIterable, excludes=str)
def _kt_sequence_convert(jcls, obj):
return ksequence(obj)
@jpype.JImplementationFor("java.util.stream.Stream")
class _JvmStream:
def __jclass_init__(self):
PyIterable.register(self)
def __iter__(self):
return self.iterator()
@jpype.JImplementationFor("java.lang.Comparable")
class _JvmComparable:
def __jclass_init__(self):
pass
def __lt__(self, other):
return self.compareTo(other) < 0
def __gt__(self, other):
return self.compareTo(other) > 0
def __le__(self, other):
return self.compareTo(other) <= 0
def __ge__(self, other):
return self.compareTo(other) >= 0
@jpype.JImplementationFor("java.lang.Throwable")
class _JvmThrowable:
def __jclass_init__(self):
pass
@property
def message(self):
return self.getMessage()
@property
def localized_message(self):
return self.getLocalizedMessage()
@property
def cause(self):
return self.getCause()
class _KtFunction(Callable):
def __init__(self, arity: int, function: Callable):
self._function = function
self._arity = arity
def invoke(self, *args):
assert len(args) == self._arity
return self._function(*args)
def __call__(self, *args):
return self.invoke(*args)
_kt_function_classes: MutableMapping[int, Any] = dict()
def kfunction(arity: int):
if arity not in _kt_function_classes:
@jpype.JImplements("kotlin.jvm.functions.Function" + str(arity), deferred=True)
class _KtFunctionN(_KtFunction):
def __init__(self, f):
super().__init__(arity, f)
@jpype.JOverride
def invoke(self, *args):
return super().invoke(*args)
_kt_function_classes[arity] = _KtFunctionN
return _kt_function_classes[arity]
logger.debug("Configure JVM-specific extensions") | 2ppy | /2ppy-0.4.0-py3-none-any.whl/tuprolog/jvmutils.py | jvmutils.py |
from collections import Mapping
from itertools import chain
from tuprolog import logger
# noinspection PyUnresolvedReferences
import jpype
from tuprolog.pyutils import iterable_or_varargs
@jpype.JImplementationFor("it.unibo.tuprolog.solve.SolverFactory")
class _KtSolverFactory:
def __jclass_init__(self):
pass
@jpype.JImplementationFor("it.unibo.tuprolog.solve.SolveOptions")
class _KtSolveOptions:
_static_keys = {'lazy', 'is_lazy', 'eager', 'is_eager', 'timeout', 'limit'}
def __jclass_init__(self):
Mapping.register(self)
@property
def is_lazy(self):
return self.isLazy()
@property
def is_eager(self):
return self.isEager()
@property
def timeout(self):
return self.getTimeout()
@property
def limit(self):
return self.getLimit()
@property
def options(self):
return self.getOptions()
def __len__(self):
return 4 + len(self.options)
def __iter__(self):
return chain(_KtSolveOptions._static_keys, self.options)
def __contains__(self, item):
return item in _KtSolveOptions._static_keys
def __getitem__(self, item, default=None):
if item in {'lazy', 'is_lazy'}:
return self.is_lazy
elif item in {'eager', 'is_eager'}:
return self.is_eager
elif item == 'timeout':
return self.timeout
elif item == 'limit':
return self.limit
elif item in self.options:
return self.options[item]
elif default is not None:
return default
return KeyError(f"No such option: {item}")
@jpype.JImplementationFor("it.unibo.tuprolog.solve.ExecutionContextAware")
class _KtExecutionContextAware:
def __jclass_init__(self):
pass
@property
def libraries(self):
return self.getLibraries()
@property
def flags(self):
return self.getFlags()
@property
def static_kb(self):
return self.getStaticKb()
@property
def dynamic_kb(self):
return self.getDynamicKb()
@property
def operators(self):
return self.getOperators()
@property
def input_channels(self):
return self.getInputChannles()
@property
def output_channels(self):
return self.getOutputChannles()
@property
def standard_output(self):
return self.getStandardOutput()
@property
def standard_input(self):
return self.getStandardInput()
@property
def standard_error(self):
return self.getStandardError()
@property
def warnings(self):
return self.getWarnings()
@jpype.JImplementationFor("it.unibo.tuprolog.solve.Solver")
class _KtSolver:
def __jclass_init__(self):
pass
@jpype.JOverride
def solve(self, goal, options=None):
if options is None:
return self.solve_(goal)
elif options.is_eager:
return self.solveList(goal, options)
else:
return self.solve_(goal, options)
@jpype.JOverride
def solve_once(self, goal, options=None):
if options is None:
return self.solveOnce(goal)
else:
return self.solveOnce(goal, options)
@jpype.JImplementationFor("it.unibo.tuprolog.solve.MutableSolver")
class _KtMutableSolver:
def __jclass_init__(self):
pass
def load_library(self, library):
return self.loadLibrary(library)
def unload_library(self, library):
return self.unloadLibrary(library)
def set_libraries(self, libraries):
return self.setLibraries(libraries)
def load_static_kb(self, theory):
return self.loadStaticKb(theory)
def load_static_clauses(self, *clauses):
return iterable_or_varargs(clauses, lambda cs: self.loadStaticClauses(cs))
def append_static_kb(self, theory):
return self.appendStaticKb(theory)
def reset_static_kb(self):
return self.resetStaticKb()
def load_dynamic_kb(self, theory):
return self.loadDynamicKb(theory)
def load_dynamic_clauses(self, *clauses):
return iterable_or_varargs(clauses, lambda cs: self.loadDynamicClauses(cs))
def append_dynamic_kb(self, theory):
return self.appendDynamicKb(theory)
def reset_dynamic_kb(self):
return self.resetDynamicKb()
def assert_a(self, clause):
return self.assertA(clause)
def assert_z(self, clause):
return self.assertZ(clause)
def retract_all(self, clause):
return self.retractAll(clause)
def set_flag(self, *args):
return self.setFlag(*args)
@property
def standard_output(self):
return self.getStandardOutput()
@property
def standard_input(self):
return self.getStandardInput()
@property
def standard_error(self):
return self.getStandardError()
@property
def warnings(self):
return self.getWarnings()
@standard_input.setter
def standard_input(self, channel):
return self.setStandardInput(channel)
@standard_output.setter
def standard_output(self, channel):
return self.setStandardOutput(channel)
@standard_error.setter
def standard_error(self, channel):
return self.setStandardError(channel)
@warnings.setter
def warnings(self, channel):
return self.setWarnings(channel)
@jpype.JImplementationFor("it.unibo.tuprolog.solve.Solution")
class _KtSolution:
def __jclass_init__(self):
pass
@property
def is_yes(self):
return self.isYes()
@property
def is_no(self):
return self.isNo()
@property
def is_halt(self):
return self.isHalt()
@property
def substitution(self):
return self.getSubstitution()
@property
def exception(self):
return self.getExecption()
@property
def solved_query(self):
return self.getSolvedQuery()
def clean_up(self):
return self.cleanUp()
def value_of(self, variable):
return self.valueOf(variable)
logger.debug("Configure Kotlin adapters for types in it.unibo.tuprolog.solve.*") | 2ppy | /2ppy-0.4.0-py3-none-any.whl/tuprolog/solve/_ktadapt.py | _ktadapt.py |
from tuprolog import logger
# noinspection PyUnresolvedReferences
import jpype.imports
from ._ktadapt import *
# noinspection PyUnresolvedReferences
import it.unibo.tuprolog.solve as _solve
from tuprolog.core import Indicator, Struct, Term, Substitution, EMPTY_UNIFIER, TermFormatter
from tuprolog.solve.exception import ResolutionException
from tuprolog.jvmutils import jlist, jmap, JavaSystem
from functools import singledispatch
from typing import Iterable, Mapping, Any
ExecutionContext = _solve.ExecutionContext
ExecutionContextAware = _solve.ExecutionContextAware
MutableSolver = _solve.MutableSolver
Signature = _solve.Signature
Solution = _solve.Solution
SolutionFormatter = _solve.SolutionFormatter
SolveOptions = _solve.SolveOptions
Solver = _solve.Solver
SolverFactory = _solve.SolverFactory
Time = _solve.Time
@singledispatch
def signature(name: str, arity: int, vararg: bool = False) -> Signature:
return Signature(name, arity, vararg)
@signature.register
def _signature_from_indicator(indicator: Indicator) -> Signature:
return Signature.fromIndicator(indicator)
@signature.register
def _signature_from_term(term: Term) -> Signature:
return Signature.fromSignatureTerm(term)
@singledispatch
def yes_solution(
signature: Signature,
arguments: Iterable[Term],
substitution: Substitution.Unifier = EMPTY_UNIFIER
) -> Solution.Yes:
return Solution.yes(signature, jlist(arguments), substitution)
@yes_solution.register
def _yes_solution_from_query(query: Struct, substitution: Substitution.Unifier = EMPTY_UNIFIER) -> Solution.Yes:
return Solution.yes(query, substitution)
@singledispatch
def no_solution(signature: Signature, arguments: Iterable[Term]) -> Solution.No:
return Solution.no(signature, jlist(arguments))
@no_solution.register
def _no_solution_from_query(query: Struct) -> Solution.No:
return Solution.no(query)
@singledispatch
def halt_solution(signature: Signature, arguments: Iterable[Term], exception: ResolutionException) -> Solution.Halt:
return Solution.halt(signature, jlist(arguments), exception)
@halt_solution.register
def _halt_solution_from_query(query: Struct, exception: ResolutionException) -> Solution.No:
return Solution.halt(query, exception)
def current_time_instant() -> int:
return JavaSystem.currentTimeMillis()
def solution_formatter(term_formatter: TermFormatter = TermFormatter.prettyExpressions()) -> SolutionFormatter:
return SolutionFormatter.of(term_formatter)
MAX_TIMEOUT: int = SolveOptions.MAX_TIMEOUT
ALL_SOLUTIONS: int = SolveOptions.ALL_SOLUTIONS
def solve_options(
lazy: bool = True,
timeout: int = MAX_TIMEOUT,
limit: int = ALL_SOLUTIONS,
custom: Mapping[str, Any] = dict(),
**kwargs: Any
) -> SolveOptions:
opts = dict(kwargs)
for key in custom:
opts[key] = custom[key]
temp = SolveOptions.of(lazy, timeout, limit)
if len(opts) > 0:
return temp.setOptions(jmap(opts))
return temp
logger.debug("Loaded JVM classes from it.unibo.tuprolog.solve.*") | 2ppy | /2ppy-0.4.0-py3-none-any.whl/tuprolog/solve/__init__.py | __init__.py |
from tuprolog import logger
# noinspection PyUnresolvedReferences
import jpype.imports
# noinspection PyUnresolvedReferences
import it.unibo.tuprolog.solve.stdlib.primitive as _primitive
Abolish = _primitive.Abolish.INSTANCE
Arg = _primitive.Arg.INSTANCE
ArithmeticEqual = _primitive.ArithmeticEqual.INSTANCE
ArithmeticGreaterThan = _primitive.ArithmeticGreaterThan.INSTANCE
ArithmeticGreaterThanOrEqualTo = _primitive.ArithmeticGreaterThanOrEqualTo.INSTANCE
ArithmeticLowerThan = _primitive.ArithmeticLowerThan.INSTANCE
ArithmeticLowerThanOrEqualTo = _primitive.ArithmeticLowerThanOrEqualTo.INSTANCE
ArithmeticNotEqual = _primitive.ArithmeticNotEqual.INSTANCE
Assert = _primitive.Assert.INSTANCE
AssertA = _primitive.AssertA.INSTANCE
AssertZ = _primitive.AssertZ.INSTANCE
Atom = _primitive.Atom.INSTANCE
AtomChars = _primitive.AtomChars.INSTANCE
AtomCodes = _primitive.AtomCodes.INSTANCE
AtomConcat = _primitive.AtomConcat.INSTANCE
AtomLength = _primitive.AtomLength.INSTANCE
Atomic = _primitive.Atomic.INSTANCE
BagOf = _primitive.BagOf.INSTANCE
Between = _primitive.Between.INSTANCE
Callable = _primitive.Callable.INSTANCE
CharCode = _primitive.CharCode.INSTANCE
Clause = _primitive.Clause.INSTANCE
Compound = _primitive.Compound.INSTANCE
CopyTerm = _primitive.CopyTerm.INSTANCE
CurrentFlag = _primitive.CurrentFlag.INSTANCE
CurrentOp = _primitive.CurrentOp.INSTANCE
EnsureExecutable = _primitive.EnsureExecutable.INSTANCE
FindAll = _primitive.FindAll.INSTANCE
Float = _primitive.Float.INSTANCE
Functor = _primitive.Functor.INSTANCE
GetDurable = _primitive.GetDurable.INSTANCE
GetEphemeral = _primitive.GetEphemeral.INSTANCE
GetPersistent = _primitive.GetPersistent.INSTANCE
Ground = _primitive.Ground.INSTANCE
Halt = _primitive.Halt.INSTANCE
Halt1 = _primitive.Halt1.INSTANCE
Integer = _primitive.Integer.INSTANCE
Is = _primitive.Is.INSTANCE
Natural = _primitive.Natural.INSTANCE
NewLine = _primitive.NewLine.INSTANCE
NonVar = _primitive.NonVar.INSTANCE
NotUnifiableWith = _primitive.NotUnifiableWith.INSTANCE
Number = _primitive.Number.INSTANCE
NumberChars = _primitive.NumberChars.INSTANCE
NumberCodes = _primitive.NumberCodes.INSTANCE
Op = _primitive.Op.INSTANCE
Repeat = _primitive.Repeat.INSTANCE
Retract = _primitive.Retract.INSTANCE
RetractAll = _primitive.RetractAll.INSTANCE
Reverse = _primitive.Reverse.INSTANCE
SetDurable = _primitive.SetDurable.INSTANCE
SetEphemeral = _primitive.SetEphemeral.INSTANCE
SetFlag = _primitive.SetFlag.INSTANCE
SetOf = _primitive.SetOf.INSTANCE
SetPersistent = _primitive.SetPersistent.INSTANCE
Sleep = _primitive.Sleep.INSTANCE
SubAtom = _primitive.SubAtom.INSTANCE
TermGreaterThan = _primitive.TermGreaterThan.INSTANCE
TermGreaterThanOrEqualTo = _primitive.TermGreaterThanOrEqualTo.INSTANCE
TermIdentical = _primitive.TermIdentical.INSTANCE
TermLowerThan = _primitive.TermLowerThan.INSTANCE
TermLowerThanOrEqualTo = _primitive.TermLowerThanOrEqualTo.INSTANCE
TermNotIdentical = _primitive.TermNotIdentical.INSTANCE
TermNotSame = _primitive.TermNotSame.INSTANCE
TermSame = _primitive.TermSame.INSTANCE
UnifiesWith = _primitive.UnifiesWith.INSTANCE
Univ = _primitive.Univ.INSTANCE
Var = _primitive.Var.INSTANCE
Write = _primitive.Write.INSTANCE
logger.debug("Loaded JVM classes from it.unibo.tuprolog.solve.stdlib.primitive.*") | 2ppy | /2ppy-0.4.0-py3-none-any.whl/tuprolog/solve/stdlib/primitive/__init__.py | __init__.py |
from tuprolog import logger
# noinspection PyUnresolvedReferences
import jpype.imports
# noinspection PyUnresolvedReferences
import it.unibo.tuprolog.solve.channel as _channel
# noinspection PyUnresolvedReferences
import kotlin.jvm.functions as _kfunctions
from tuprolog.jvmutils import jmap, kfunction
from typing import Callable, Mapping
from functools import singledispatch
Channel = _channel.Channel
ChannelStore = _channel.ChannelStore
InputChannel = _channel.InputChannel
InputStore = _channel.InputStore
OutputChannel = _channel.OutputChannel
OutputStore = _channel.OutputStore
Listener = _kfunctions.Function1
def std_in() -> InputChannel:
return InputChannel.stdIn()
@singledispatch
def input_channel(generator: Callable, availability_checker: Callable = None) -> InputChannel:
if availability_checker is None:
return InputChannel.of(kfunction(0)(generator))
else:
return InputChannel.of(kfunction(0)(generator), kfunction(0)(availability_checker))
@input_channel.register
def input_channel_from_string(string: str) -> InputChannel:
return InputChannel.of(string)
def input_store(stdin: InputChannel = None, channels: Mapping[str, InputChannel] = None) -> InputStore:
if stdin is None and channels is None:
return InputStore.fromStandard()
elif channels is None:
return InputStore.fromStandard(stdin)
else:
cs = {InputStore.STDIN: stdin or std_in()}
for k, v in channels:
cs[k] = v
return InputStore.of(jmap(cs))
def std_out() -> OutputChannel:
return OutputChannel.stdOut()
def std_err() -> OutputChannel:
return OutputChannel.stdErr()
def warn() -> OutputChannel:
return OutputChannel.warn()
def output_channel(consumer: Callable) -> OutputChannel:
return OutputChannel.of(kfunction(1)(consumer))
def output_store(
stdout: OutputChannel = None,
stderr: OutputChannel = None,
warnings: OutputChannel = None,
channels: Mapping[str, OutputChannel] = None
) -> OutputStore:
if all((channel is None for channel in (stdout, stderr, warnings))):
return OutputStore.fromStandard()
elif channels is None:
return OutputStore.fromStandard(
stdout or std_out(),
stderr or std_err(),
warnings or warn()
)
else:
cs = {
OutputStore.STDOUT: stdout or std_out(),
OutputStore.STDERR: stderr or std_err()
}
for k, v in channels:
cs[k] = v
return OutputStore.of(jmap(cs), warnings or warn())
def listener(consumer: Callable) -> Listener:
return kfunction(1)(consumer)
logger.debug("Loaded JVM classes from it.unibo.tuprolog.solve.channel.*") | 2ppy | /2ppy-0.4.0-py3-none-any.whl/tuprolog/solve/channel/__init__.py | __init__.py |
from tuprolog import logger
# noinspection PyUnresolvedReferences
import jpype.imports
# noinspection PyUnresolvedReferences
import it.unibo.tuprolog.solve.flags as _flags
from tuprolog.core import Term
from tuprolog.jvmutils import kpair, jmap, jarray, Pair
from typing import Iterable, Union, Mapping
DoubleQuotes = _flags.DoubleQuotes
FlagStore = _flags.FlagStore
LastCallOptimization = _flags.LastCallOptimization
MaxArity = _flags.MaxArity
NotableFlag = _flags.NotableFlag
Unknown = _flags.Unknown
Flag = Pair
EMPTY_FLAG_STORE: FlagStore = FlagStore.EMPTY
DEFAULT_FLAG_STORE: FlagStore = FlagStore.DEFAULT
DoubleQuotes: NotableFlag = DoubleQuotes.INSTANCE
LastCallOptimization: NotableFlag = LastCallOptimization.INSTANCE
MaxArity: NotableFlag = MaxArity.INSTANCE
Unknown: NotableFlag = Unknown.INSTANCE
def flag(first: Union[str, NotableFlag, Iterable], value: Term = None) -> Flag:
if isinstance(first, NotableFlag):
if value is None:
return first.toPair()
else:
return first.to(value)
elif isinstance(first, str):
if value is None:
raise ValueError("Argument value is None")
return Flag(first, value)
elif isinstance(first, Iterable) and value is None:
return kpair(first)
else:
raise ValueError("Argument first is not iterable nor str")
def flag_store(*flags: Union[NotableFlag, Flag, Iterable, FlagStore], **kwargs: Mapping[str, Term]):
normal_flags = []
notable_flags = []
other_stores = []
for f in flags:
if isinstance(f, NotableFlag):
notable_flags.append(f)
elif isinstance(f, Flag):
normal_flags.append(f)
elif isinstance(f, FlagStore):
other_stores.append(f)
else:
normal_flags.append(flag(f))
store1 = FlagStore.of(jarray(NotableFlag)@notable_flags)
store2 = FlagStore.of(jarray(Flag)@normal_flags)
store3 = FlagStore.of(jmap(kwargs))
store = store1.plus(store2).plus(store3)
for s in other_stores:
store = store.plus(s)
return store
logger.debug("Loaded JVM classes from it.unibo.tuprolog.solve.flags.*") | 2ppy | /2ppy-0.4.0-py3-none-any.whl/tuprolog/solve/flags/__init__.py | __init__.py |
from tuprolog import logger
# noinspection PyUnresolvedReferences
import jpype.imports
from tuprolog.core import Term, Clause, Integer
from tuprolog.solve import ExecutionContext, Signature, Solution, current_time_instant, MAX_TIMEOUT
# from tuprolog.solve.sideffcts import SideEffect
from tuprolog.pyutils import iterable_or_varargs
from tuprolog.jvmutils import jlist
from typing import List, Iterable, Callable
# noinspection PyUnresolvedReferences
import it.unibo.tuprolog.solve.primitive as _primitive
Primitive = _primitive.Primitive
Solve = _primitive.Solve
PrimitiveWrapper = _primitive.PrimitiveWrapper
SolveRequest = Solve.Request
SolveResponse = Solve.Request
@jpype.JImplements(Primitive)
class AbstractPrimitive(object):
@jpype.JOverride
def solve(self, request: SolveRequest) -> Iterable[SolveResponse]:
raise NotImplementedError()
def primitive(callable: Callable[[SolveResponse], Iterable[SolveResponse]]) -> Primitive:
class CallableToPrimitiveAdapter(AbstractPrimitive):
def solve(self, request: SolveRequest) -> Iterable[SolveResponse]:
return callable(request)
return CallableToPrimitiveAdapter()
def solve_request(
signature: Signature,
arguments: List[Term],
context: ExecutionContext,
issuing_instant: int = current_time_instant(),
max_duration: int = MAX_TIMEOUT
) -> SolveRequest:
return SolveRequest(signature, arguments, context, issuing_instant, max_duration)
def solve_response(solution: Solution, *side_effects) -> SolveResponse:
return iterable_or_varargs(side_effects, lambda ses: SolveResponse(solution, None, jlist(ses)))
def check_term_is_recursively_callable(request: SolveRequest, term: Term):
return PrimitiveWrapper.checkTermIsRecursivelyCallable(request, term)
def ensuring_all_arguments_are_instantiated(request: SolveRequest) -> SolveRequest:
return PrimitiveWrapper.ensuringAllArgumentsAreInstantiated(request)
def ensuring_procedure_has_permission(request: SolveRequest, signature: Signature, operation) -> SolveRequest:
return PrimitiveWrapper.ensuringProcedureHasPermission(request, signature, operation)
def ensuring_clause_procedure_has_permission(request: SolveRequest, clause: Clause, operation) -> SolveRequest:
return PrimitiveWrapper.ensuringClauseProcedureHasPermission(request, clause, operation)
def ensuring_argument_is_well_formed_indicator(request: SolveRequest, index: int) -> SolveRequest:
return PrimitiveWrapper.ensuringArgumentIsWellFormedIndicator(request, index)
def not_implemented(request: SolveRequest, message: str) -> SolveResponse:
return PrimitiveWrapper.notImplemented(request, message)
def not_supported(request: SolveRequest, message: str) -> SolveResponse:
return PrimitiveWrapper.notSupported(request, message)
def ensuring_argument_is_well_formed_clause(request: SolveRequest, index: int) -> SolveRequest:
return PrimitiveWrapper.ensuringArgumentIsWellFormedClause(request, index)
def ensuring_argument_is_instantiated(request: SolveRequest, index: int) -> SolveRequest:
return PrimitiveWrapper.ensuringArgumentIsInstantiated(request, index)
def ensuring_argument_is_numeric(request: SolveRequest, index: int) -> SolveRequest:
return PrimitiveWrapper.ensuringArgumentIsNumeric(request, index)
def ensuring_argument_is_struct(request: SolveRequest, index: int) -> SolveRequest:
return PrimitiveWrapper.ensuringArgumentIsStruct(request, index)
def ensuring_argument_is_callable(request: SolveRequest, index: int) -> SolveRequest:
return PrimitiveWrapper.ensuringArgumentIsCallable(request, index)
def ensuring_argument_is_variable(request: SolveRequest, index: int) -> SolveRequest:
return PrimitiveWrapper.ensuringArgumentIsVariable(request, index)
def ensuring_argument_is_compound(request: SolveRequest, index: int) -> SolveRequest:
return PrimitiveWrapper.ensuringArgumentIsCompound(request, index)
def ensuring_argument_is_atom(request: SolveRequest, index: int) -> SolveRequest:
return PrimitiveWrapper.ensuringArgumentIsCompound(request, index)
def ensuring_argument_is_constant(request: SolveRequest, index: int) -> SolveRequest:
return PrimitiveWrapper.ensuringArgumentIsConstant(request, index)
def ensuring_argument_is_ground(request: SolveRequest, index: int) -> SolveRequest:
return PrimitiveWrapper.ensuringArgumentIsGround(request, index)
def ensuring_argument_is_char(request: SolveRequest, index: int) -> SolveRequest:
return PrimitiveWrapper.ensuringArgumentIsChar(request, index)
def ensuring_argument_is_specifier(request: SolveRequest, index: int) -> SolveRequest:
return PrimitiveWrapper.ensuringArgumentIsSpecifier(request, index)
def ensuring_argument_is_integer(request: SolveRequest, index: int) -> SolveRequest:
return PrimitiveWrapper.ensuringArgumentIsInteger(request, index)
def ensuring_argument_is_list(request: SolveRequest, index: int) -> SolveRequest:
return PrimitiveWrapper.ensuringArgumentIsList(request, index)
def ensuring_argument_is_arity(request: SolveRequest, index: int) -> SolveRequest:
return PrimitiveWrapper.ensuringArgumentIsArity(request, index)
def ensuring_argument_is_non_negative_integer(request: SolveRequest, index: int) -> SolveRequest:
return PrimitiveWrapper.ensuringArgumentIsNonNegativeInteger(request, index)
def is_character_code(integer: Integer) -> bool:
return PrimitiveWrapper.isCharacterCode(integer)
def ensuring_term_is_char_code(request: SolveRequest, term: Term) -> SolveRequest:
return PrimitiveWrapper.ensuringTermIsCharCode(request, term)
def ensuring_term_is_well_formed_list(request: SolveRequest, term: Term) -> SolveRequest:
return PrimitiveWrapper.ensuringTermIsWellFormedList(request, term)
def ensuring_argument_is_char_code(request: SolveRequest, index: int) -> SolveRequest:
return PrimitiveWrapper.ensuringArgumentIsCharCode(request, index)
logger.debug("Loaded JVM classes from it.unibo.tuprolog.solve.primitive.*") | 2ppy | /2ppy-0.4.0-py3-none-any.whl/tuprolog/solve/primitive/__init__.py | __init__.py |
from typing import Iterable, Union
from tuprolog import logger
# noinspection PyUnresolvedReferences
import jpype.imports
# noinspection PyUnresolvedReferences
import it.unibo.tuprolog.solve.exception.error as errors
from tuprolog.core import Term
from tuprolog.solve import ExecutionContext, Signature
DomainError = errors.DomainError
Domain = DomainError.Expected
DOMAIN_wATOM_PROPERTY = Domain.ATOM_PROPERTY
DOMAIN_BUFFERING_MODE = Domain.BUFFERING_MODE
DOMAIN_CHARACTER_CODE_LIST = Domain.CHARACTER_CODE_LIST
DOMAIN_CLOSE_OPTION = Domain.CLOSE_OPTION
DOMAIN_DATE_TIME = Domain.DATE_TIME
DOMAIN_EOF_ACTION = Domain.EOF_ACTION
DOMAIN_FLAG_VALUE = Domain.FLAG_VALUE
DOMAIN_FORMAT_CONTROL_SEQUENCE = Domain.FORMAT_CONTROL_SEQUENCE
DOMAIN_IO_MODE = Domain.IO_MODE
DOMAIN_WELL_FORMED_LIST = Domain.WELL_FORMED_LIST
DOMAIN_NON_EMPTY_LIST = Domain.NON_EMPTY_LIST
DOMAIN_NOT_LESS_THAN_ZERO = Domain.NOT_LESS_THAN_ZERO
DOMAIN_OPERATOR_PRIORITY = Domain.OPERATOR_PRIORITY
DOMAIN_OPERATOR_SPECIFIER = Domain.OPERATOR_SPECIFIER
DOMAIN_ORDER = Domain.ORDER
DOMAIN_OS_FILE_PERMISSION = Domain.OS_FILE_PERMISSION
DOMAIN_OS_FILE_PROPERTY = Domain.OS_FILE_PROPERTY
DOMAIN_OS_PATH = Domain.OS_PATH
DOMAIN_PREDICATE_PROPERTY = Domain.PREDICATE_PROPERTY
DOMAIN_FLAG = Domain.FLAG
DOMAIN_READ_OPTION = Domain.READ_OPTION
DOMAIN_SELECTABLE_ITEM = Domain.SELECTABLE_ITEM
DOMAIN_SOCKET_ADDRESS = Domain.SOCKET_ADDRESS
DOMAIN_SOCKET_DOMAIN = Domain.SOCKET_DOMAIN
DOMAIN_SOURCE_SINK = Domain.SOURCE_SINK
DOMAIN_STREAM = Domain.STREAM
DOMAIN_STREAM_OPTION = Domain.STREAM_OPTION
DOMAIN_STREAM_OR_ALIAS = Domain.STREAM_OR_ALIAS
DOMAIN_STREAM_POSITION = Domain.STREAM_POSITION
DOMAIN_STREAM_PROPERTY = Domain.STREAM_PROPERTY
DOMAIN_STREAM_SEEK_METHOD = Domain.STREAM_SEEK_METHOD
DOMAIN_STREAM_TYPE = Domain.STREAM_TYPE
DOMAIN_TERM_STREAM_OR_ALIAS = Domain.TERM_STREAM_OR_ALIAS
DOMAIN_VAR_BINDING_OPTION = Domain.VAR_BINDING_OPTION
DOMAIN_WRITE_OPTION = Domain.WRITE_OPTION
DOMAIN_CLAUSE = Domain.CLAUSE
DOMAIN_RULE = Domain.RULE
DOMAIN_FACT = Domain.FACT
DIRECTIVE = Domain.DIRECTIVE
def domain_error_for_flag_values(
context: ExecutionContext,
procedure: Signature,
flag_values: Iterable[Term],
actual: Term,
index: int = None
) -> DomainError:
return DomainError.forFlagValues(context, procedure, flag_values, actual, index)
def domain_error_for_argument(
context: ExecutionContext,
procedure: Signature,
expected: Domain,
actual: Term,
index: int = None
) -> DomainError:
return DomainError.forArgument(context, procedure, expected, actual, index)
def domain_error_for_term(
context: ExecutionContext,
expected: Domain,
actual_value: Term,
) -> DomainError:
return DomainError.forTerm(context, expected, actual_value)
def domain_error_for_goal(
context: ExecutionContext,
procedure: Signature,
expected: Domain,
actual: Term,
) -> DomainError:
return DomainError.forGoal(context, procedure, expected, actual)
def domain(name: Union[str, Term]) -> Domain:
if isinstance(name, str):
return Domain.of(name)
else:
return Domain.fromTerm(name)
logger.debug("Loaded JVM classes from it.unibo.tuprolog.solve.exception.error.DomainError.*") | 2ppy | /2ppy-0.4.0-py3-none-any.whl/tuprolog/solve/exception/error/domain/__init__.py | __init__.py |
from typing import Union
# noinspection PyUnresolvedReferences
import jpype.imports
from tuprolog import logger
from tuprolog.core import Term
from tuprolog.solve import ExecutionContext, Signature
# noinspection PyUnresolvedReferences
import it.unibo.tuprolog.solve.exception.error as errors
PermissionError = errors.PermissionError
Operation = PermissionError.Operation
Permission = PermissionError.Permission
OPERATION_ACCESS = Operation.ACCESS
OPERATION_ADD_ALIAS = Operation.ADD_ALIAS
OPERATION_CLOSE = Operation.CLOSE
OPERATION_CREATE = Operation.CREATE
OPERATION_INPUT = Operation.INPUT
OPERATION_INVOKE = Operation.INVOKE
OPERATION_MODIFY = Operation.MODIFY
OPERATION_OPEN = Operation.OPEN
OPERATION_OUTPUT = Operation.OUTPUT
OPERATION_REPOSITION = Operation.REPOSITION
PERMISSION_BINARY_STREAM = Permission.BINARY_STREAM
PERMISSION_FLAG = Permission.FLAG
PERMISSION_OPERATOR = Permission.OPERATOR
PERMISSION_PAST_END_OF_STREAM = Permission.PAST_END_OF_STREAM
PERMISSION_PRIVATE_PROCEDURE = Permission.PRIVATE_PROCEDURE
PERMISSION_SOURCE_SINK = Permission.SOURCE_SINK
PERMISSION_STATIC_PROCEDURE = Permission.STATIC_PROCEDURE
PERMISSION_OOP_METHOD = Permission.OOP_METHOD
PERMISSION_STREAM = Permission.STREAM
PERMISSION_TEXT_STREAM = Permission.TEXT_STREAM
def permission_error(
context: ExecutionContext,
procedure: Signature,
operation: Operation,
permission: Permission,
culprit: Term
) -> PermissionError:
return PermissionError.of(context, procedure, operation, permission, culprit)
def operation(name: Union[str, Term]) -> Operation:
if isinstance(name, str):
return Operation.of(name)
else:
return Operation.fromTerm(name)
def permission(name: Union[str, Term]) -> Permission:
if isinstance(name, str):
return Permission.of(name)
else:
return Permission.fromTerm(name)
logger.debug("Loaded JVM classes from it.unibo.tuprolog.solve.exception.error.PermissionError.*") | 2ppy | /2ppy-0.4.0-py3-none-any.whl/tuprolog/solve/exception/error/permission/__init__.py | __init__.py |
from typing import Union
# noinspection PyUnresolvedReferences
import jpype.imports
from tuprolog import logger
from tuprolog.core import Term
from tuprolog.solve import ExecutionContext, Signature
# noinspection PyUnresolvedReferences
import it.unibo.tuprolog.solve.exception.error as errors
TypeError = errors.TypeError
Type = TypeError.Expected
TYPE_ATOM = Type.ATOM
TYPE_ATOMIC = Type.ATOMIC
TYPE_BOOLEAN = Type.BOOLEAN
TYPE_BYTE = Type.BYTE
TYPE_CALLABLE = Type.CALLABLE
TYPE_CHARACTER = Type.CHARACTER
TYPE_COMPOUND = Type.COMPOUND
TYPE_DEALIASING_EXPRESSION = Type.DEALIASING_EXPRESSION
TYPE_EVALUABLE = Type.EVALUABLE
TYPE_FLOAT = Type.FLOAT
TYPE_INTEGER = Type.INTEGER
TYPE_IN_CHARACTER = Type.IN_CHARACTER
TYPE_LIST = Type.LIST
TYPE_NUMBER = Type.NUMBER
TYPE_OBJECT_REFERENCE = Type.OBJECT_REFERENCE
TYPE_PAIR = Type.PAIR
TYPE_PREDICATE_INDICATOR = Type.PREDICATE_INDICATOR
TYPE_REFERENCE = Type.REFERENCE
TYPE_TYPE_REFERENCE = Type.TYPE_REFERENCE
TYPE_URL = Type.URL
TYPE_VARIABLE = Type.VARIABLE
def type_error_for_flag_values(
context: ExecutionContext,
procedure: Signature,
expected: Type,
actual: Term,
message: str
) -> TypeError:
return TypeError.of(context, procedure, expected, actual, message)
def type_error_for_argument_list(
context: ExecutionContext,
procedure: Signature,
expected: Type,
actual: Term,
index: int = None
) -> TypeError:
return TypeError.forArgumentList(context, procedure, expected, actual, index)
def type_error_for_argument(
context: ExecutionContext,
procedure: Signature,
expected: Type,
actual: Term,
index: int = None
) -> TypeError:
return TypeError.forArgument(context, procedure, expected, actual, index)
def type_error_for_term(
context: ExecutionContext,
expected: Type,
actual_value: Term,
) -> TypeError:
return TypeError.forTerm(context, expected, actual_value)
def type_error_for_goal(
context: ExecutionContext,
procedure: Signature,
expected: Type,
actual: Term,
) -> TypeError:
return TypeError.forGoal(context, procedure, expected, actual)
def type(name: Union[str, Term]) -> Type:
if isinstance(name, str):
return TypeError.valueOf(name)
else:
return TypeError.fromTerm(name)
logger.debug("Loaded JVM classes from it.unibo.tuprolog.solve.exception.error.TypeError.*") | 2ppy | /2ppy-0.4.0-py3-none-any.whl/tuprolog/solve/exception/error/type/__init__.py | __init__.py |
from functools import reduce
from tuprolog import logger
# noinspection PyUnresolvedReferences
import jpype.imports
from tuprolog.core.operators import OperatorSet, operator_set
from tuprolog.theory import Theory, theory
from tuprolog.solve import Signature
from tuprolog.solve.primitive import Primitive
from tuprolog.solve.function import LogicFunction
from tuprolog.jvmutils import jiterable
from typing import Union, Mapping, Iterable
# noinspection PyUnresolvedReferences
import it.unibo.tuprolog.solve.library as _library
Library = _library.Library
Libraries = _library.Libraries
LibraryGroup = _library.LibraryGroup
AliasedLibrary = _library.AliasedLibrary
def library(
alias: str = None,
primitives: Mapping[Signature, Primitive] = dict(),
theory: Theory = theory(),
operators: OperatorSet = operator_set(),
functions: Mapping[Signature, LogicFunction] = dict(),
) -> Union[Library, AliasedLibrary]:
if alias is None:
return Library.unaliased(primitives, theory, operators, functions)
else:
return Library.aliased(alias, primitives, theory, operators, functions)
def aliased(alias: str, library: Library) -> AliasedLibrary:
return Library.of(alias, library)
def libraries(
*libs: Union[Libraries, AliasedLibrary, Iterable[Union[Libraries, AliasedLibrary]]],
**kwargs: Library
) -> Libraries:
all_libraries = []
aliased_libs = []
queue = list(libs)
while len(queue) > 0:
current = queue.pop()
if isinstance(current, Libraries):
all_libraries.append(current)
elif isinstance(current, AliasedLibrary):
aliased_libs.append(current)
else:
queue.extend(current)
for alias in kwargs:
aliased_libs.append(aliased(alias, kwargs[alias]))
first = Libraries.of(jiterable(aliased_libs))
return reduce(lambda a, b: a.plus(b), all_libraries, first)
logger.debug("Loaded JVM classes from it.unibo.tuprolog.solve.library.*") | 2ppy | /2ppy-0.4.0-py3-none-any.whl/tuprolog/solve/library/__init__.py | __init__.py |
from functools import reduce
from tuprolog import logger
# noinspection PyUnresolvedReferences
import jpype
# noinspection PyUnresolvedReferences
import jpype.imports
# noinspection PyUnresolvedReferences
from it.unibo.tuprolog.solve.sideffects import SideEffect
# noinspection PyUnresolvedReferences
from it.unibo.tuprolog.solve.sideffects import SideEffectFactory
# noinspection PyUnresolvedReferences
from it.unibo.tuprolog.solve.sideffects import SideEffectsBuilder
# noinspection PyUnresolvedReferences
from tuprolog.core import Clause, Term
# noinspection PyUnresolvedReferences
from tuprolog.core.operators import Operator, OperatorSet
# noinspection PyUnresolvedReferences
from tuprolog.solve.library import Library, AliasedLibrary, Libraries, libraries as new_libraries
# noinspection PyUnresolvedReferences
from tuprolog.solve.channel import InputChannel, OutputChannel
from tuprolog.pyutils import iterable_or_varargs, dict_or_keyword_args
from tuprolog.jvmutils import jlist, jmap
from typing import Mapping, Union, Iterable, Any
def _forward_iterable_or_varargs(callable, args, *callable_args):
return iterable_or_varargs(args, lambda xs: callable(jlist(xs), *callable_args))
def _forward_dict_or_keywords(callable, dict, kwargs, *callable_args):
return dict_or_keyword_args(dict, kwargs, lambda ds: callable(jmap(ds), *callable_args))
def reset_static_kb(*clauses: Union[Clause, Iterable[Clause]]) -> SideEffect.ResetStaticKb:
return _forward_iterable_or_varargs(SideEffect.ResetStaticKb, clauses)
def add_static_clauses(*clauses: Union[Clause, Iterable[Clause]]) -> SideEffect.AddStaticClauses:
return _forward_iterable_or_varargs(SideEffect.AddStaticClauses, clauses)
def remove_static_clauses(*clauses: Union[Clause, Iterable[Clause]]) -> SideEffect.RemoveStaticClauses:
return _forward_iterable_or_varargs(SideEffect.RemoveStaticClauses, clauses)
def reset_dynamic_kb(*clauses: Union[Clause, Iterable[Clause]]) -> SideEffect.ResetDynamicKb:
return _forward_iterable_or_varargs(SideEffect.ResetDynamicKb, clauses)
def add_dynamic_clauses(*clauses: Union[Clause, Iterable[Clause]]) -> SideEffect.AddDynamicClauses:
return _forward_iterable_or_varargs(SideEffect.AddDynamicClauses, clauses)
def remove_dynamic_clauses(*clauses: Union[Clause, Iterable[Clause]]) -> SideEffect.RemoveDynamicClauses:
return _forward_iterable_or_varargs(SideEffect.RemoveDynamicClauses, clauses)
def set_flags(flags: Mapping[str, Term] = {}, **kwargs: Term) -> SideEffect.SetFlags:
return _forward_dict_or_keywords(SideEffect.SetFlags, flags, kwargs)
def reset_flags(flags: Mapping[str, Term] = {}, **kwargs: Term) -> SideEffect.ResetFlags:
return _forward_dict_or_keywords(SideEffect.ResetFlags, flags, kwargs)
def clear_flags(*flag_names: str) -> SideEffect.ClearFlags:
return _forward_iterable_or_varargs(SideEffect.ClearFlags, flag_names)
def load_library(alias: str, library: Library) -> SideEffect.LoadLibrary:
return SideEffect.LoadLibrary(alias, library)
def unload_libraries(*aliases: str) -> SideEffect.UnloadLibraries:
return _forward_iterable_or_varargs(SideEffect.UnloadLibraries, aliases)
def update_library(alias: str, library: Library) -> SideEffect.UpdateLibrary:
return SideEffect.UpdateLibrary(alias, library)
def add_libraries(*libraries: Union[Libraries, AliasedLibrary]) -> SideEffect.AddLibraries:
return SideEffect.AddLibraries(new_libraries(libraries))
def reset_libraries(*libraries: Union[Libraries, AliasedLibrary]) -> SideEffect.ResetLibraries:
return SideEffect.ResetLibraries(new_libraries(libraries))
def set_operators(*operators: Union[Operator, Iterable[Operator]]) -> SideEffect.SetOperators:
return _forward_iterable_or_varargs(SideEffect.SetOperators, operators)
def reset_operators(*operators: Union[Operator, Iterable[Operator]]) -> SideEffect.ResetOperators:
return _forward_iterable_or_varargs(SideEffect.ResetOperators, operators)
def remove_operators(*operators: Union[Operator, Iterable[Operator]]) -> SideEffect.RemoveOperators:
return _forward_iterable_or_varargs(SideEffect.RemoveOperators, operators)
def open_input_channels(
channels: Mapping[str, InputChannel] = {},
**kwargs: InputChannel
) -> SideEffect.OpenInputChannels:
return _forward_dict_or_keywords(SideEffect.OpenInputChannels, channels, kwargs)
def reset_input_channels(
channels: Mapping[str, InputChannel] = {},
**kwargs: InputChannel
) -> SideEffect.ResetInputChannels:
return _forward_dict_or_keywords(SideEffect.ResetInputChannels, channels, kwargs)
def close_input_channels(*names: Union[str, Iterable[str]]) -> SideEffect.CloseInputChannels:
return _forward_iterable_or_varargs(SideEffect.CloseInputChannels, names)
def open_output_channels(
channels: Mapping[str, OutputChannel] = {},
**kwargs: OutputChannel
) -> SideEffect.OpenOutputChannels:
return _forward_dict_or_keywords(SideEffect.OpenOutputChannels, channels, kwargs)
def reset_output_channels(
channels: Mapping[str, OutputChannel] = {},
**kwargs: OutputChannel
) -> SideEffect.ResetOutputChannels:
return _forward_dict_or_keywords(SideEffect.ResetOutputChannels, channels, kwargs)
def close_output_channels(*names: Union[str, Iterable[str]]) -> SideEffect.CloseOutputChannels:
return _forward_iterable_or_varargs(SideEffect.CloseOutputChannels, names)
def set_persistent_data(
data: Mapping[str, Any] = {},
**kwargs: Any
) -> SideEffect.SetPersistentData:
return _forward_dict_or_keywords(SideEffect.SetPersistentData, data, kwargs, False)
def reset_persistent_data(
data: Mapping[str, Any] = {},
**kwargs: Any
) -> SideEffect.SetPersistentData:
return _forward_dict_or_keywords(SideEffect.SetPersistentData, data, kwargs, True)
def set_durable_data(
data: Mapping[str, Any] = {},
**kwargs: Any
) -> SideEffect.SetDurableData:
return _forward_dict_or_keywords(SideEffect.SetDurableData, data, kwargs, False)
def reset_durable_data(
data: Mapping[str, Any] = {},
**kwargs: Any
) -> SideEffect.SetDurableData:
return _forward_dict_or_keywords(SideEffect.SetDurableData, data, kwargs, True)
def set_ephemeral_data(
data: Mapping[str, Any] = {},
**kwargs: Any
) -> SideEffect.SetEphemeralData:
return _forward_dict_or_keywords(SideEffect.SetEphemeralData, data, kwargs, False)
def reset_ephemeral_data(
data: Mapping[str, Any] = {},
**kwargs: Any
) -> SideEffect.SetEphemeralData:
return _forward_dict_or_keywords(SideEffect.SetEphemeralData, data, kwargs, True)
logger.debug("Loaded JVM classes from it.unibo.tuprolog.solve.sideffects.*") | 2ppy | /2ppy-0.4.0-py3-none-any.whl/tuprolog/solve/sideffcts/__init__.py | __init__.py |
from tuprolog import logger
import jpype
from tuprolog.core import indicator as new_indicator
from tuprolog.jvmutils import jiterable, protect_iterable
from tuprolog.pyutils import iterable_or_varargs
from typing import Sized
@jpype.JImplementationFor("it.unibo.tuprolog.theory.Theory")
class _KtTheory:
def __jclass_init__(cls):
Sized.register(cls)
@property
def is_mutable(self):
return self.isMutable()
def to_mutable_theory(self):
return self.toMutableTheory()
def to_immutable_theory(self):
return self.toImmutableTheory()
@jpype.JOverride
def getClauses(self):
return protect_iterable(self.getClauses_())
@property
def clauses(self):
return self.getClauses()
@jpype.JOverride
def getRules(self):
return protect_iterable(self.getRules_())
@property
def rules(self):
return self.getRules()
@jpype.JOverride
def getDirectives(self):
return protect_iterable(self.getDirectives_())
@property
def directives(self):
return self.getDirectives()
def __len__(self):
return self.getSize()
def __add__(self, other):
return self.plus(other)
def __contains__(self, item):
return self.contains(item)
def __getitem__(self, item):
return self.get(item)
def _assert(self, method, clause, *clauses):
if len(clauses) == 0:
return method(clause)
return iterable_or_varargs((clause,) + clauses, lambda cs: method(jiterable(cs)))
def assert_a(self, clause, *clauses):
self._assert(self.assertA, clause, *clauses)
def assert_z(self, clause, *clauses):
self._assert(self.assertZ, clause, *clauses)
@jpype.JOverride
def retract(self, clause, *clauses):
if len(clauses) == 0:
return self.retract_(clause)
return iterable_or_varargs((clause,) + clauses, lambda cs: self.retract_(jiterable(cs)))
def retract_all(self, clause):
return self.retractAll(clause)
@jpype.JOverride
def abolish(self, name, arity=None, indicator=None):
if name is not None:
if arity is not None:
return self.abolish_(new_indicator(name, arity))
else:
return self.abolish_(name)
elif indicator is not None:
return self.abolish_(indicator)
raise ValueError("You should provide at least either a name-arity couple or an indicator")
@jpype.JOverride
def equals(self, other, use_var_complete_name=True):
return self.equals_(other, use_var_complete_name)
def __eq__(self, other):
return self.equals(other, use_var_complete_name=True)
def to_string(self, as_prolog_text=False):
return self.toString(as_prolog_text)
@jpype.JImplementationFor("it.unibo.tuprolog.theory.RetractResult")
class _KtRetractResult:
def __jclass_init__(cls):
pass
@property
def is_success(self):
return self.isSuccess()
@property
def is_failure(self):
return self.isFailure()
@property
def theory(self):
return self.getTheory()
@property
def clauses(self):
return self.getClauses()
@property
def first_clause(self):
return self.getFirstClause()
logger.debug("Configure Kotlin adapters for types in it.unibo.tuprolog.theory.*") | 2ppy | /2ppy-0.4.0-py3-none-any.whl/tuprolog/theory/_ktadapt.py | _ktadapt.py |
from decimal import Decimal
from tuprolog import logger
import jpype
from ._ktmath import big_integer, big_decimal, BigInteger, BigDecimal
from tuprolog.utils import *
from typing import Sized, Callable
from tuprolog.jvmutils import jiterable
from tuprolog.pyutils import iterable_or_varargs
from tuprolog.jvmutils import kfunction
from ._ktmath import *
@jpype.JImplementationFor("it.unibo.tuprolog.core.Term")
class _KtTerm:
def __jclass_init__(cls):
pass
def __getitem__(self, item, *items):
return self.get(item, *items)
@property
def variables(self):
return self.getVariables()
def structurally_equals(self, other):
return self.structurallyEquals(other)
@jpype.JOverride()
def equals(self, other, use_var_complete_name: bool = True):
return self.equals_(other, use_var_complete_name)
@property
def is_var(self):
return self.isVar()
@property
def is_ground(self):
return self.isGround()
@property
def is_struct(self):
return self.isStruct()
@property
def is_truth(self):
return self.isTruth()
@property
def is_recursive(self):
return self.isRecursive()
@property
def is_atom(self):
return self.isAtom()
@property
def is_constant(self):
return self.isConstant()
@property
def is_number(self):
return self.isNumber()
@property
def is_integer(self):
return self.isInteger()
@property
def is_real(self):
return self.isReal()
@property
def is_list(self):
return self.isList()
@property
def is_tuple(self):
return self.isTuple()
@property
def is_block(self):
return self.isBlock()
@property
def is_clause(self):
return self.isClause()
@property
def is_rule(self):
return self.isRule()
@property
def is_fact(self):
return self.isFact()
@property
def is_directive(self):
return self.isDirective()
@property
def is_cons(self):
return self.isCons()
@property
def is_empty_list(self):
return self.isEmptyList()
@property
def is_true(self):
return self.isTrue()
@property
def is_fail(self):
return self.isFail()
@property
def is_indicator(self):
return self.isIndicator()
def fresh_copy(self, scope=None):
if scope is None:
return self.freshCopy()
else:
return self.freshCopy(scope)
@jpype.JImplementationFor("it.unibo.tuprolog.core.Struct")
class _KtStruct:
def __jclass_init__(cls):
pass
@property
def functor(self):
return self.getFunctor()
@property
def args(self):
return self.getArgs()
@property
def argsSequence(self):
return self.getArgsSequence()
@property
def arity(self):
return self.getArity()
@property
def is_functor_well_formed(self):
return self.isFunctorWellFormed()
@property
def indicator(self):
return self.getIndicator()
def get_arg_at(self, index):
return self.getArgAt(index)
def add_last(self, argument):
return self.addLast(argument)
def add_first(self, argument):
return self.addFirst(argument)
def insert_at(self, index, argument):
return self.addFirst(index, argument)
def set_functor(self, functor):
return self.setFunctor(functor)
def set_args(self, *args):
return iterable_or_varargs(args, lambda xs: self.setArgs(jiterable(args)))
@jpype.JImplementationFor("it.unibo.tuprolog.core.Var")
class _KtVar:
def __jclass_init__(cls):
pass
@property
def is_anonymous(self):
return self.isAnonymous()
@property
def name(self):
return self.getName()
@property
def id(self):
return self.getId()
@property
def complete_name(self):
return self.getCompleteName()
@property
def is_name_well_formed(self):
return self.isNameWellFormed()
@jpype.JImplementationFor("it.unibo.tuprolog.core.Constant")
class _KtConstant:
def __jclass_init__(cls):
pass
@property
def value(self):
return self.getValue()
@jpype.JImplementationFor("it.unibo.tuprolog.core.Numeric")
class _KtNumeric:
def __jclass_init__(cls):
pass
@property
def int_value(self):
return python_integer(self.getIntValue())
@property
def decimal_value(self):
return python_decimal(self.getDecimalValue())
def to_int(self):
return self.int_value
def to_float(self):
return float(self.decimal_value)
@jpype.JImplementationFor("it.unibo.tuprolog.core.Integer")
class _KtInteger:
def __jclass_init__(cls):
pass
@property
def value(self):
return self.int_value
@jpype.JImplementationFor("it.unibo.tuprolog.core.Real")
class _KtReal:
def __jclass_init__(cls):
pass
@property
def value(self):
return self.decimal_value
@jpype.JImplementationFor("it.unibo.tuprolog.core.Recursive")
class _KtRecursive:
def __jclass_init__(cls):
Sized.register(cls)
@property
def lazily_unfolded(self):
return self.getUnfoldedSequence()
@property
def unfolded(self):
return self.getUnfoldedList()
@property
def __len__(self):
return self.getSize()
def to_iterable(self):
return self.toSequence()
def to_list(self):
return self.toList()
def to_array(self):
return self.toArray()
@jpype.JImplementationFor("it.unibo.tuprolog.core.List")
class _KtList:
def __jclass_init__(cls):
pass
@property
def is_well_formed(self):
return self.isWellFormed()
@property
def last(self):
return self.getLast()
def estimated_length(self):
return self.getEstimatedLength()
@jpype.JImplementationFor("it.unibo.tuprolog.core.Cons")
class _KtCons:
def __jclass_init__(cls):
pass
@property
def head(self):
return self.getHead()
@property
def tail(self):
return self.getTail()
@jpype.JImplementationFor("it.unibo.tuprolog.core.Tuple")
class _KtTuple:
def __jclass_init__(cls):
pass
@property
def left(self):
return self.getLeft()
@property
def right(self):
return self.getRight()
@jpype.JImplementationFor("it.unibo.tuprolog.core.Clause")
class _KtClause:
def __jclass_init__(cls):
pass
@property
def head(self):
return self.getHead()
@property
def body(self):
return self.getBody()
@property
def is_well_formed(self):
return self.isWellFormed()
@property
def body_items(self):
return self.getBodyItems()
@property
def body_size(self):
return self.getBodySize()
def get_body_item(self, index):
return self.getBodyItem(index)
def set_head(self, head):
return self.setHead(head)
def set_body(self, term):
return self.setBody(term)
def set_head_functor(self, functor):
return self.setHeadFunctor(functor)
def set_head_args(self, *args):
return iterable_or_varargs(args, lambda xs: self.setHeadArgs(jiterable(args)))
def insert_head_arg(self, index, argument):
return self.setHeadArg(index, argument)
def add_first_head_arg(self, argument):
return self.addFirstHeadArg(argument)
def add_last_head_arg(self, argument):
return self.addLastHeadArg(argument)
def append_head_arg(self, argument):
return self.appendHeadArg(argument)
def set_body_items(self, *args):
return iterable_or_varargs(args, lambda xs: self.setBodyItems(jiterable(args)))
def insert_body_item(self, index, item):
return self.insertBodyItem(index, item)
def add_first_body_item(self, item):
return self.addFirstBodyItem(item)
def add_last_body_item(self, item):
return self.addLastBodyItem(item)
def append_body_item(self, item):
return self.appendBodyItem(item)
@jpype.JImplementationFor("it.unibo.tuprolog.core.Rule")
class _KtRule:
def __jclass_init__(cls):
pass
@property
def head_args(self):
return self.getHeadArgs()
@property
def head_arity(self):
return self.getHeadArity()
def get_head_arg(self, index):
return self.getHeadArg(index)
@jpype.JImplementationFor("it.unibo.tuprolog.core.TermConvertible")
class _KtTermConvertible:
def __jclass_init__(cls):
pass
def to_term(self):
return self.toTerm()
@jpype.JImplementationFor("it.unibo.tuprolog.core.Substitution")
class _KtSubstitution:
def __jclass_init__(cls):
pass
@property
def is_success(self):
return self.isSuccess()
@property
def is_failed(self):
return self.isFailed()
def apply_to(self, term):
return self.applyTo(term)
def get_original(self, variable):
return self.getOriginal(variable)
def get_by_name(self, name):
return self.getByName(name)
def __add__(self, other):
return self.plus(other)
def __sub__(self, other):
return self.minus(other)
@jpype.JOverride
def filter(self, filter):
if isinstance(filter, Callable):
return self.filter_(kfunction(1)(filter))
else:
return self.filter_(filter)
@jpype.JImplementationFor("it.unibo.tuprolog.core.Scope")
class _KtScope:
def __jclass_init__(cls):
pass
def __contains__(self, item):
return self.contains(item)
def __getitem__(self, item):
return self.get(item)
@property
def variables(self):
return self.getVariables()
@property
def fail(self):
return self.getFail()
@property
def empty_list(self):
return self.getEmptyList()
@property
def empty_block(self):
return self.getEmptyBlock()
def atom(self, string):
return self.atomOf(string)
def block(self, *terms):
return iterable_or_varargs(terms, lambda ts: self.blockOf(jiterable(ts)))
def clause(self, head=None, *body):
return iterable_or_varargs(body, lambda bs: self.clauseOf(head, jiterable(bs)))
def cons(self, head, tail=None):
if tail is None:
return self.listOf(head)
else:
return self.consOf(head, tail)
def directive(self, *goals):
return iterable_or_varargs(goals, lambda gs: self.directiveOf(jiterable(gs)))
def fact(self, head):
return self.factOf(head)
def indicator(self, name, arity):
return self.indicatorOf(name, arity)
def integer(self, value):
return self.intOf(big_integer(value))
def real(self, value):
return self.intOf(big_integer(value))
def numeric(self, value):
if isinstance(value, str):
return self.numOf(value)
if isinstance(value, int) or isinstance(value, BigInteger):
return self.intOf(value)
return self.realOf(value)
def rule(self, head, *body):
return iterable_or_varargs(body, lambda bs: self.ruleOf(head, jiterable(bs)))
def struct(self, functor, *args):
return iterable_or_varargs(args, lambda xs: self.structOf(functor, jiterable(xs)))
def truth(self, value):
return self.truthOf(value)
def var(self, name):
return self.varOf(name)
def list(self, *items):
return iterable_or_varargs(items, lambda xs: self.listOf(jiterable(xs)))
def list_from(self, items, last=None):
return self.listFrom(jiterable(items), last)
def tuple(self, first, second, *others):
return iterable_or_varargs(others, lambda os: self.tupleOf(jiterable([first, second] + list(os))))
@property
def anonymous(self):
return self.getAnonymous()
@property
def whatever(self):
return self.getWhatever()
logger.debug("Configure Kotlin adapters for types in it.unibo.tuprolog.core.*") | 2ppy | /2ppy-0.4.0-py3-none-any.whl/tuprolog/core/_ktadapt.py | _ktadapt.py |
from typing import Union
import decimal
import jpype.imports
# noinspection PyUnresolvedReferences
import org.gciatto.kt.math as _ktmath
# noinspection PyUnresolvedReferences
import java.lang as _java_lang
from math import ceil
from tuprolog.pyutils import and_then
BigInteger = _ktmath.BigInteger
BigDecimal = _ktmath.BigDecimal
MathContext = _ktmath.MathContext
RoundingMode = _ktmath.RoundingMode
_MAX_LONG = _java_lang.Long.MAX_VALUE
_MIN_LONG = _java_lang.Long.MIN_VALUE
BIG_INTEGER_MAX_LONG = BigInteger.of(_MAX_LONG)
BIG_INTEGER_MIN_LONG = BigInteger.of(_MIN_LONG)
BIG_INTEGER_ZERO = BigInteger.ZERO
BIG_INTEGER_TEN = BigInteger.TEN
BIG_INTEGER_ONE = BigInteger.ONE
BIG_INTEGER_NEGATIVE_ONE = BigInteger.NEGATIVE_ONE
BIG_INTEGER_TWO = BigInteger.TWO
BIG_DECIMAL_ZERO = BigDecimal.ZERO
BIG_DECIMAL_ONE = BigDecimal.ONE
BIG_DECIMAL_ONE_HALF = BigDecimal.ONE_HALF
BIG_DECIMAL_ONE_TENTH = BigDecimal.ONE_TENTH
BIG_DECIMAL_E = BigDecimal.E
BIG_DECIMAL_PI = BigDecimal.PI
def _int_to_bytes(n: int) -> bytes:
try:
size = n.bit_length() // 8 + 1
return n.to_bytes(size, 'big', signed=True)
except OverflowError as e:
e.args = ["%s: %d, whose size is %d" % (e.args[0], n, size)]
raise e
def big_integer(value: Union[str, int], radix: int = None) -> BigInteger:
if radix is not None:
assert isinstance(value, str)
return BigInteger.of(value, radix)
if isinstance(value, str):
return BigInteger.of(jpype.JString @ value)
assert isinstance(value, int)
bs = _int_to_bytes(value)
return BigInteger(jpype.JArray(jpype.JByte) @ bs, 0, len(bs))
def python_integer(value: BigInteger) -> int:
return int.from_bytes(value.toByteArray(), byteorder='big', signed=True)
_python_rounding_modes = {decimal.ROUND_DOWN, decimal.ROUND_HALF_UP, decimal.ROUND_HALF_EVEN, decimal.ROUND_CEILING,
decimal.ROUND_FLOOR, decimal.ROUND_UP, decimal.ROUND_HALF_DOWN, decimal.ROUND_05UP}
def jvm_rounding_mode(mode):
if mode == decimal.ROUND_DOWN:
return RoundingMode.DOWN
elif mode == decimal.ROUND_UP:
return RoundingMode.UP
elif mode == decimal.ROUND_FLOOR:
return RoundingMode.FLOOR
elif mode == decimal.ROUND_CEILING:
return RoundingMode.CEILING
elif mode == decimal.ROUND_HALF_UP:
return RoundingMode.HALF_UP
elif mode == decimal.ROUND_HALF_EVEN:
return RoundingMode.HALF_EVEN
elif mode == decimal.ROUND_HALF_DOWN:
return RoundingMode.HALF_DOWN
elif mode == decimal.ROUND_05UP:
raise ValueError(f"Rounding mode {decimal.ROUND_05UP} has no Java correspondence")
else:
raise ValueError(f"Not a rounding mode {mode}")
@and_then(lambda bd: bd.stripTrailingZeros())
def big_decimal(value: Union[str, int, float, decimal.Decimal], precision=0,
rounding=RoundingMode.HALF_UP) -> BigDecimal:
if precision is None:
precision = decimal.getcontext().prec
assert isinstance(precision, int)
if rounding is None:
rounding = jvm_rounding_mode(decimal.getcontext().rounding)
elif rounding in _python_rounding_modes:
rounding = jvm_rounding_mode(rounding)
assert isinstance(rounding, RoundingMode)
context = MathContext(precision, rounding)
if isinstance(value, str):
return BigDecimal.of(value, context)
if isinstance(value, decimal.Decimal):
return BigDecimal.of(jpype.JString @ str(value), context)
if isinstance(value, BigInteger):
return BigDecimal.of(value, context)
if isinstance(value, int):
return BigDecimal.of(big_integer(value), context)
assert isinstance(value, float)
return BigDecimal.of(jpype.JDouble @ value, context)
def python_decimal(value: BigDecimal) -> decimal.Decimal:
return decimal.Decimal(str(value)) | 2ppy | /2ppy-0.4.0-py3-none-any.whl/tuprolog/core/_ktmath.py | _ktmath.py |
from tuprolog import logger
import jpype
from tuprolog.core import TermVisitor
@jpype.JImplements(TermVisitor)
class AbstractTermVisitor(object):
@jpype.JOverride
def defaultValue(self, term):
raise NotImplementedError()
@jpype.JOverride
def visitTerm(self, term):
return TermVisitor.DefaultImpls.visitTerm(self, term)
@jpype.JOverride
def visitVar(self, term):
return TermVisitor.DefaultImpls.visitVar(self, term)
@jpype.JOverride
def visitConstant(self, term):
return TermVisitor.DefaultImpls.visitConstant(self, term)
@jpype.JOverride
def visitStruct(self, term):
return TermVisitor.DefaultImpls.visitStruct(self, term)
@jpype.JOverride
def visitCollection(self, term):
return TermVisitor.DefaultImpls.visitCollection(self, term)
@jpype.JOverride
def visitAtom(self, term):
return TermVisitor.DefaultImpls.visitAtom(self, term)
@jpype.JOverride
def visitTruth(self, term):
return TermVisitor.DefaultImpls.visitTruth(self, term)
@jpype.JOverride
def visitNumeric(self, term):
return TermVisitor.DefaultImpls.visitNumeric(self, term)
@jpype.JOverride
def visitInteger(self, term):
return TermVisitor.DefaultImpls.visitInteger(self, term)
@jpype.JOverride
def visitReal(self, term):
return TermVisitor.DefaultImpls.visitReal(self, term)
@jpype.JOverride
def visitBlock(self, term):
return TermVisitor.DefaultImpls.visitBlock(self, term)
@jpype.JOverride
def visitEmpty(self, term):
return TermVisitor.DefaultImpls.visitEmpty(self, term)
@jpype.JOverride
def visitEmptyBlock(self, term):
return TermVisitor.DefaultImpls.visitEmptyBlock(self, term)
@jpype.JOverride
def visitList(self, term):
return TermVisitor.DefaultImpls.visitList(self, term)
@jpype.JOverride
def visitCons(self, term):
return TermVisitor.DefaultImpls.visitCons(self, term)
@jpype.JOverride
def visitEmptyList(self, term):
return TermVisitor.DefaultImpls.visitEmptyList(self, term)
@jpype.JOverride
def visitTuple(self, term):
return TermVisitor.DefaultImpls.visitTuple(self, term)
@jpype.JOverride
def visitIndicator(self, term):
return TermVisitor.DefaultImpls.visitIndicator(self, term)
@jpype.JOverride
def visitClause(self, term):
return TermVisitor.DefaultImpls.visitClause(self, term)
@jpype.JOverride
def visitRule(self, term):
return TermVisitor.DefaultImpls.visitRule(self, term)
@jpype.JOverride
def visitFact(self, term):
return TermVisitor.DefaultImpls.visitFact(self, term)
@jpype.JOverride
def visitDirective(self, term):
return TermVisitor.DefaultImpls.visitDirective(self, term)
logger.debug("Loaded compatibility layer for JVM subtypes of " + str(TermVisitor.class_.getName())) | 2ppy | /2ppy-0.4.0-py3-none-any.whl/tuprolog/core/visitors.py | visitors.py |
from tuprolog import logger
# noinspection PyUnresolvedReferences
import jpype
import jpype.imports
from tuprolog.core import Formatter, TermFormatter
from tuprolog.core.visitors import AbstractTermVisitor
@jpype.JImplements(Formatter)
class AbstractFormatter(object):
@jpype.JOverride
def format(self, term):
raise NotImplementedError()
@jpype.JImplements(TermFormatter)
class AbstractTermFormatter(AbstractFormatter, AbstractTermVisitor):
@jpype.JOverride
def format(self, term):
return term.accept(self)
@jpype.JOverride
def defaultValue(self, term):
return super(AbstractTermVisitor).defaultValue(self, term)
@jpype.JOverride
def visitTerm(self, term):
return super(AbstractTermVisitor, self).visitTerm(term)
@jpype.JOverride
def visitVar(self, term):
return super(AbstractTermVisitor, self).visitVar(term)
@jpype.JOverride
def visitConstant(self, term):
return super(AbstractTermVisitor, self).visitConstant(term)
@jpype.JOverride
def visitStruct(self, term):
return super(AbstractTermVisitor, self).visitStruct(term)
@jpype.JOverride
def visitCollection(self, term):
return super(AbstractTermVisitor, self).visitCollection(term)
@jpype.JOverride
def visitAtom(self, term):
return super(AbstractTermVisitor, self).visitAtom(term)
@jpype.JOverride
def visitTruth(self, term):
return super(AbstractTermVisitor, self).visitTruth(term)
@jpype.JOverride
def visitNumeric(self, term):
return super(AbstractTermVisitor, self).visitNumeric(term)
@jpype.JOverride
def visitInteger(self, term):
return super(AbstractTermVisitor, self).visitInteger(term)
@jpype.JOverride
def visitReal(self, term):
return super(AbstractTermVisitor, self).visitReal(term)
@jpype.JOverride
def visitBlock(self, term):
return super(AbstractTermVisitor, self).visitBlock(term)
@jpype.JOverride
def visitEmpty(self, term):
return super(AbstractTermVisitor, self).visitEmpty(term)
@jpype.JOverride
def visitEmptyBlock(self, term):
return super(AbstractTermVisitor, self).visitEmptyBlock(term)
@jpype.JOverride
def visitList(self, term):
return super(AbstractTermVisitor, self).visitList(term)
@jpype.JOverride
def visitCons(self, term):
return super(AbstractTermVisitor, self).visitCons(term)
@jpype.JOverride
def visitEmptyList(self, term):
return super(AbstractTermVisitor, self).visitEmptyList(term)
@jpype.JOverride
def visitTuple(self, term):
return super(AbstractTermVisitor, self).visitTuple(term)
@jpype.JOverride
def visitIndicator(self, term):
return super(AbstractTermVisitor, self).visitIndicator(term)
@jpype.JOverride
def visitClause(self, term):
return super(AbstractTermVisitor, self).visitClause(term)
@jpype.JOverride
def visitRule(self, term):
return super(AbstractTermVisitor, self).visitRule(term)
@jpype.JOverride
def visitFact(self, term):
return super(AbstractTermVisitor, self).visitFact(term)
@jpype.JOverride
def visitDirective(self, term):
return super(AbstractTermVisitor, self).visitDirective(term)
logger.debug("Loaded compatibility layer for JVM subtypes of " + str(TermFormatter.class_.getName())) | 2ppy | /2ppy-0.4.0-py3-none-any.whl/tuprolog/core/formatters.py | formatters.py |
from decimal import Decimal
from tuprolog import logger
import jpype
import jpype.imports
from ._ktadapt import *
# noinspection PyUnresolvedReferences
import it.unibo.tuprolog.core as _core
from tuprolog.pyutils import iterable_or_varargs
from tuprolog.jvmutils import jiterable, jmap
from typing import Iterable, Dict, Tuple as PyTuple
Atom = _core.Atom
Block = _core.Block
Clause = _core.Clause
Cons = _core.Cons
Constant = _core.Constant
Directive = _core.Directive
Empty = _core.Empty
EmptyBlock = _core.EmptyBlock
EmptyList = _core.EmptyList
Fact = _core.Fact
Formatter = _core.Formatter
Indicator = _core.Indicator
Integer = _core.Integer
List = _core.List
Numeric = _core.Numeric
Real = _core.Real
Recursive = _core.Recursive
Rule = _core.Rule
Scope = _core.Scope
Struct = _core.Struct
Substitution = _core.Substitution
Term = _core.Term
TermComparator = _core.TermComparator
TermConvertible = _core.TermConvertible
TermFormatter = _core.TermFormatter
Terms = _core.Terms
TermVisitor = _core.TermVisitor
Truth = _core.Truth
Tuple = _core.Tuple
Var = _core.Var
@jpype.JImplements(TermConvertible)
class AbstractTermConvertible(object):
@jpype.JOverride
def toTerm(self):
raise NotImplementedError()
def atom(string: str) -> Atom:
return Atom.of(string)
def block(*terms: Union[Term, Iterable[Term]]) -> Block:
return iterable_or_varargs(terms, lambda ts: Block.of(jiterable(ts)))
def clause(head: Term = None, *body: Union[Term, Iterable[Term]]):
return iterable_or_varargs(body, lambda bs: Clause.of(head, jiterable(bs)))
def empty_logic_list() -> EmptyList:
return EmptyList.getInstance()
def cons(head: Term, tail: Term = None) -> Cons:
if tail is None:
return Cons.singleton(head)
else:
return Cons.of(head, tail)
def directive(*goals: Union[Term, Iterable[Term]]) -> Directive:
return iterable_or_varargs(goals, lambda gs: Directive.of(jiterable(gs)))
def empty_block() -> Block:
return EmptyBlock.getInstance()
def fact(struct: Union[Term, Iterable[Term]]) -> Fact:
return Fact.of(struct)
def indicator(name: Union[str, Term], arity: Union[int, Term]) -> Indicator:
return Indicator.of(name, arity)
def integer(value: Union[int, BigInteger, str]) -> Integer:
return Integer.of(big_integer(value))
def real(value: Union[float, BigDecimal, str, Decimal]) -> Real:
return Real.of(big_decimal(value))
def numeric(value: Union[int, BigInteger, BigDecimal, str, float, Decimal]) -> Numeric:
if isinstance(value, str):
return Numeric.of(jpype.JString @ value)
if isinstance(value, int) or isinstance(value, BigInteger):
return integer(value)
return real(value)
def rule(head: Struct, *body: Union[Term, Iterable[Term]]) -> Rule:
return iterable_or_varargs(body, lambda bs: Rule.of(head, jiterable(bs)))
def struct(functor: str, *args: Union[Term, Iterable[Term]]) -> Struct:
return iterable_or_varargs(args, lambda xs: Struct.of(functor, jiterable(xs)))
def truth(boolean: bool) -> Truth:
return Truth.of(boolean)
TRUE = Truth.TRUE
FALSE = Truth.FALSE
FAIL = Truth.FAIL
def logic_list(*items: Union[Term, Iterable[Term]]) -> List:
return iterable_or_varargs(items, lambda xs: List.of(jiterable(xs)))
def logic_list_from(items: Iterable[Term], last: Term = None) -> List:
return List.from_(jiterable(items), last)
def logic_tuple(first: Term, second: Term, *others: Union[Term, Iterable[Term]]) -> Tuple:
return iterable_or_varargs(others, lambda os: Tuple.of(jiterable([first, second] + list(os))))
def var(name: str) -> Var:
return Var.of(name)
def unifier(assignments: Dict[Var, Term] = {}) -> Substitution.Unifier:
return Substitution.unifier(jmap(assignments))
def substitution(assignments: Dict[Var, Term] = {}) -> Substitution:
return Substitution.of(jmap(assignments))
def failed() -> Substitution.Fail:
return Substitution.failed()
EMPTY_UNIFIER: Substitution.Unifier = substitution()
FAILED_SUBSTITUTION: Substitution.Fail = failed()
def scope(*variables: Union[Var, str]) -> Scope:
if len(variables) == 0:
return Scope.empty()
vars = [var(v) if isinstance(v, str) else v for v in variables]
return Scope.of(jpype.JArray(Var) @ vars)
def variables(*names: str) -> PyTuple[Var]:
assert len(names) > 0
return tuple((var(n) for n in names))
logger.debug("Loaded JVM classes from it.unibo.tuprolog.core.*") | 2ppy | /2ppy-0.4.0-py3-none-any.whl/tuprolog/core/__init__.py | __init__.py |
from tuprolog import logger
import jpype
from tuprolog.jvmioutils import ensure_input_steam
@jpype.JImplementationFor("it.unibo.tuprolog.core.parsing.TermParser")
class _KtTermParser:
def __jclass_init__(cls):
pass
@property
def default_operator_set(self):
return self.getDefaultOperatorSet()
def _parse(self, method, input, operators):
if operators is None:
return method(input)
else:
return method(input, operators)
def parse(self, input, operators=None):
return self.parse_term(input, operators)
def parse_term(self, input, operators=None):
return self._parse(self.parseTerm, input, operators)
def parse_struct(self, input, operators=None):
return self._parse(self.parseStruct, input, operators)
def parse_constant(self, input, operators=None):
return self._parse(self.parseConstant, input, operators)
def parse_var(self, input, operators=None):
return self._parse(self.parseVar, input, operators)
def parse_atom(self, input, operators=None):
return self._parse(self.parseAtom, input, operators)
def parse_numeric(self, input, operators=None):
return self._parse(self.parseNumeric, input, operators)
def parse_integer(self, input, operators=None):
return self._parse(self.parseInteger, input, operators)
def parse_real(self, input, operators=None):
return self._parse(self.parseReal, input, operators)
def parse_clause(self, input, operators=None):
return self._parse(self.parseClause, input, operators)
@jpype.JImplementationFor("it.unibo.tuprolog.core.parsing.TermReader")
class _KtTermReader:
def __jclass_init__(cls):
pass
@property
def default_operator_set(self):
return self.getDefaultOperatorSet()
def _read(self, method, input, operators):
input_stream = ensure_input_steam(input)
if operators is None:
return method(input_stream)
else:
return method(input_stream, operators)
def read(self, input, operators=None):
return self.read_term(input, operators)
def read_term(self, input, operators=None):
return self._read(self.readTerm, input, operators)
def read_all(self, input, operators=None):
return self.read_terms(input, operators)
def read_terms(self, input, operators=None):
return self._read(self.readTerms, input, operators)
logger.debug("Configure Kotlin adapters for types in it.unibo.tuprolog.core.parsing.*") | 2ppy | /2ppy-0.4.0-py3-none-any.whl/tuprolog/core/parsing/_ktadapt.py | _ktadapt.py |
from tuprolog import logger
# noinspection PyUnresolvedReferences
import jpype.imports
# noinspection PyUnresolvedReferences
import it.unibo.tuprolog.core.parsing as _parsing
from tuprolog.core import Term, Struct, Constant, Var, Atom, Numeric, Integer, Real, Clause
from tuprolog.jvmutils import InputStream
from tuprolog.core.operators import OperatorSet, DEFAULT_OPERATORS
from typing import Union, Iterable
from ._ktadapt import *
TermParser = _parsing.TermParser
TermReader = _parsing.TermReader
ParseException = _parsing.ParseException
InvalidTermTypeException = _parsing.InvalidTermTypeException
def _factory(source, with_default_operators: bool = True, operators: OperatorSet = None):
if operators is None:
if with_default_operators:
return source.getWithDefaultOperators()
else:
return source.getWithNoOperator()
else:
if with_default_operators:
return source.withOperators(DEFAULT_OPERATORS.plus(operators))
else:
return source.withOperators(operators)
def term_parser(with_default_operators: bool = True, operators: OperatorSet = None) -> TermParser:
return _factory(TermParser, with_default_operators, operators)
def term_reader(with_default_operators: bool = True, operators: OperatorSet = None) -> TermParser:
return _factory(TermReader, with_default_operators, operators)
DEFAULT_TERM_PARSER = term_parser()
DEFAULT_TERM_READER = term_reader()
def parse_term(string: str, operators: OperatorSet = None) -> Term:
return DEFAULT_TERM_PARSER.parse_term(string, operators)
def parse_struct(string: str, operators: OperatorSet = None) -> Struct:
return DEFAULT_TERM_PARSER.parse_struct(string, operators)
def parse_constant(string: str, operators: OperatorSet = None) -> Constant:
return DEFAULT_TERM_PARSER.parse_constant(string, operators)
def parse_var(string: str, operators: OperatorSet = None) -> Var:
return DEFAULT_TERM_PARSER.parse_var(string, operators)
def parse_atom(string: str, operators: OperatorSet = None) -> Atom:
return DEFAULT_TERM_PARSER.parse_atom(string, operators)
def parse_numeric(string: str, operators: OperatorSet = None) -> Numeric:
return DEFAULT_TERM_PARSER.parse_numeric(string, operators)
def parse_integer(string: str, operators: OperatorSet = None) -> Integer:
return DEFAULT_TERM_PARSER.parse_integer(string, operators)
def parse_real(string: str, operators: OperatorSet = None) -> Real:
return DEFAULT_TERM_PARSER.parse_real(string, operators)
def parse_clause(string: str, operators: OperatorSet = None) -> Clause:
return DEFAULT_TERM_PARSER.parse_clause(string, operators)
def read_term(input: Union[InputStream, str], operators: OperatorSet = None) -> Term:
return DEFAULT_TERM_READER.read_term(input, operators)
def read_terms(input: Union[InputStream, str], operators: OperatorSet = None) -> Iterable[Term]:
return DEFAULT_TERM_READER.read_terms(input, operators)
logger.debug("Loaded JVM classes from it.unibo.tuprolog.core.parsing.*") | 2ppy | /2ppy-0.4.0-py3-none-any.whl/tuprolog/core/parsing/__init__.py | __init__.py |
# __ (Double Underscores, 2us)
Glueing functionals by `import __` for python!
## Install
The package is written in pure python, with no dependencies other than the Python language. Just do:
```sh
pip install 2us
```
Requires Python 3.5 or higher.
## Why this?
Python is a great language for creating convenient wrappers around native code and implementing simple, human-friendly functions.
In python, a bunch of builtin higher-order methods (which means that they accept functions as arguments) such as `map`, `filter` are available.
They enable streamed data processing on containers that focus on the processing itself,
in contrast with *noisy code* on traditional command-based languages that is heavily involved in loops.
However, you may occasionally run into the situatiion where you find that there is no standard library functions to implement in-line unpacking of tuples,
adding all numbers in a list by a constant shift, so you will have to write:
```python
map(lambda x: x + 1, some_list)
map(lambda x: x[0], some_list)
```
which seems rather dumb due to the inconvenient definition of lambda functions in python.
## Using __
Start using the package by importing `__`:
```python
import __
```
And then `__` can be used to create convenient functions that are identical to those written with `lambda`. Examples:
```python
assert sum(map(__ + 1, range(1000))) == sum(map(lambda x: x + 1, range(1000)))
assert set(map(__[0], {1: 2, 4: 6}.items())) == {1, 4}
assert functools.reduce(__ + __, range(1000)) == sum(range(1000))
```
Currently there is a drawback: python do not support overriding `__contains__` returning non-boolean values, so the `in` condition should be handled separately.
```python
assert tuple(map(__.is_in([1, 2]), [3, 1, 5, 0, 2])) == (False, True, False, False, True)
assert list(map(__.contains('1'), '13')) == [True, False]
```
| 2us | /2us-0.0.1.tar.gz/2us-0.0.1/README.md | README.md |
import sys
import operator
import functools
from types import ModuleType
class LambdaProxyModule(ModuleType):
def __getattr__(self, name):
return lambda x: getattr(x, name)
def __call__(self, *args, **kwargs):
return lambda x: x(*args, **kwargs)
def __getitem__(self, key):
return operator.itemgetter(key)
def is_in(self, it):
return functools.partial(operator.contains, it)
def contains(self, other):
if self is other:
return lambda x, y: y in x
return lambda x: other in x
def __lt__(self, other):
if self is other:
return operator.__lt__
return lambda x: x < other
def __le__(self, other):
if self is other:
return operator.__le__
return lambda x: x <= other
def __eq__(self, other):
if self is other:
return operator.__eq__
return lambda x: x == other
def __ne__(self, other):
if self is other:
return operator.__ne__
return lambda x: x != other
def __gt__(self, other):
if self is other:
return operator.__gt__
return lambda x: x > other
def __ge__(self, other):
if self is other:
return operator.__ge__
return lambda x: x >= other
def __add__(self, other):
if self is other:
return operator.__add__
return lambda x: x + other
def __sub__(self, other):
if self is other:
return operator.__sub__
return lambda x: x - other
def __mul__(self, other):
if self is other:
return operator.__mul__
return lambda x: x * other
def __matmul__(self, other):
if self is other:
return operator.__matmul__
return lambda x: x @ other
def __truediv__(self, other):
if self is other:
return operator.__truediv__
return lambda x: x / other
def __floordiv__(self, other):
if self is other:
return operator.__floordiv__
return lambda x: x // other
def __mod__(self, other):
if self is other:
return operator.__mod__
return lambda x: x % other
def __pow__(self, other):
if self is other:
return operator.__pow__
return lambda x: x ** other
def __lshift__(self, other):
if self is other:
return operator.__lshift__
return lambda x: x << other
def __rshift__(self, other):
if self is other:
return operator.__rshift__
return lambda x: x >> other
def __and__(self, other):
if self is other:
return operator.__and__
return lambda x: x & other
def __xor__(self, other):
if self is other:
return operator.__xor__
return lambda x: x ^ other
def __or__(self, other):
if self is other:
return operator.__or__
return lambda x: x | other
def __radd__(self, other):
if self is other:
return lambda x, y: y + x
return lambda x: other + x
def __rsub__(self, other):
if self is other:
return lambda x, y: y - x
return lambda x: other - x
def __rmul__(self, other):
if self is other:
return lambda x, y: y * x
return lambda x: other * x
def __rmatmul__(self, other):
if self is other:
return lambda x, y: y @ x
return lambda x: other @ x
def __rtruediv__(self, other):
if self is other:
return lambda x, y: y / x
return lambda x: other / x
def __rfloordiv__(self, other):
if self is other:
return lambda x, y: y // x
return lambda x: other // x
def __rmod__(self, other):
if self is other:
return lambda x, y: y % x
return lambda x: other % x
def __rpow__(self, other):
if self is other:
return lambda x, y: y ** x
return lambda x: other ** x
def __rlshift__(self, other):
if self is other:
return lambda x, y: y << x
return lambda x: other << x
def __rrshift__(self, other):
if self is other:
return lambda x, y: y >> x
return lambda x: other >> x
def __rand__(self, other):
if self is other:
return lambda x, y: y & x
return lambda x: other & x
def __rxor__(self, other):
if self is other:
return lambda x, y: y ^ x
return lambda x: other ^ x
def __ror__(self, other):
if self is other:
return lambda x, y: y | x
return lambda x: other | x
sys.modules[__name__].__class__ = LambdaProxyModule | 2us | /2us-0.0.1.tar.gz/2us-0.0.1/__/__init__.py | __init__.py |
2vyper is an automatic verifier for smart contracts written in Vyper, based on the `Viper <http://viper.ethz.ch>`_ verification infrastructure. It is being developed at the `Programming Methodology Group <http://www.pm.inf.ethz.ch/>`_ at ETH Zurich. 2vyper was mainly developed by Robin Sierra, Christian Bräm, and Marco Eilers.
For examples of the provided specification constructs, check out the `examples folder <tests/resources/examples>`_. Note that the examples are written in Vyper 0.1.0, but 2vyper supports different versions if a version pragma is set.
A short overview of the most important specification constructs can be found `here <docs/specifications.md>`_.
For further documentation, read `our paper <https://arxiv.org/abs/2104.10274>`_ about 2vyper's specification constructs, `Robin Sierra's <https://ethz.ch/content/dam/ethz/special-interest/infk/chair-program-method/pm/documents/Education/Theses/Robin_Sierra_MA_Report.pdf>`_ and `Christian Bräm's <https://ethz.ch/content/dam/ethz/special-interest/infk/chair-program-method/pm/documents/Education/Theses/Christian%20Br%C3%A4m_MS_Report.pdf>`_ Master's theses on the tool.
Dependencies (Ubuntu Linux, MacOS)
===================================
Install Java >= 11 (64 bit) and Python >= 3.7 (64 bit).
For usage with the Viper's verification condition generation backend Carbon, you will also need to install the .NET / the Mono runtime.
Dependencies (Windows)
==========================
1. Install Java >= 11 (64 bit) and Python >= 3.7 (64 bit).
2. Install either Visual C++ Build Tools 2015 (http://go.microsoft.com/fwlink/?LinkId=691126) or Visual Studio 2015. For the latter, make sure to choose the option "Common Tools for Visual C++ 2015" in the setup (see https://blogs.msdn.microsoft.com/vcblog/2015/07/24/setup-changes-in-visual-studio-2015-affecting-c-developers/ for an explanation).
Getting Started
===============
1. Clone the 2vyper repository::
git clone https://github.com/viperproject/2vyper
cd 2vyper/
2. Create a virtual environment and activate it::
virtualenv env
source env/bin/activate
3. Install 2vyper::
pip install .
Command Line Usage
==================
To verify a specific file from the 2vyper directory, run::
2vyper [OPTIONS] path-to-file.vy
The following command line options are available::
``--verifier``
Selects the Viper backend to use for verification.
Possible options are ``silicon`` (for Symbolic Execution) and ``carbon``
(for Verification Condition Generation based on Boogie).
Default: ``silicon``.
``--viper-jar-path``
Sets the path to the required Viper binaries (``silicon.jar`` or
``carbon.jar``). Only the binary for the selected backend is
required. We recommend that you use the provided binary
packages installed by default, but you can or compile your own from
source.
Expects either a single path or a colon- (Unix) or semicolon-
(Windows) separated list of paths. Alternatively, the environment
variables ``SILICONJAR``, ``CARBONJAR`` or ``VIPERJAR`` can be set.
``--z3``
Sets the path of the Z3 executable. Alternatively, the
``Z3_EXE`` environment variable can be set.
``--boogie``
Sets the path of the Boogie executable. Required if the Carbon backend
is selected. Alternatively, the ``BOOGIE_EXE`` environment variable can be
set.
``--counterexample``
Produces a counterexample if the verification fails. Currently only works
with the default ``silicon`` backend.
``--vyper-root``
Sets the root directory for the Vyper compiler.
``--skip-vyper``
Skips type checking the given Vyper program using the Vyper compiler.
``--print-viper``
Print the generated Viper file to the command line.
To see all possible command line options, invoke ``2vyper`` without arguments.
Alternative Viper Versions
==========================
To use a custom version of the Viper infrastructure, follow the
`instructions here <https://bitbucket.org/viperproject/documentation/wiki/Home>`_. Look for
``sbt assembly`` to find instructions for packaging the required JAR files. Use the
parameters mentioned above to instruct 2vyper to use your custom Viper version.
Note that 2vyper may not always work with the most recent Viper version.
Troubleshooting
=======================
1. On Windows: During the setup, you get an error like ``Microsoft Visual C++ 14.0 is required.`` or ``Unable to fnd vcvarsall.bat``:
Python cannot find the required Visual Studio 2015 C++ installation, make sure you have either installed the Build Tools or checked the "Common Tools" option in your regular VS 2015 installation (see above).
2. While verifying a file, you get a stack trace ending with something like ``No matching overloads found``:
The version of Viper you're using does not match your version of 2vyper. Try using the the one that comes with 2vyper instead.
Build Status
============
.. image:: https://pmbuilds.inf.ethz.ch/buildStatus/icon?job=2vyper-linux-xenial&style=plastic
:alt: Build Status
:target: https://pmbuilds.inf.ethz.ch/job/2vyper-linux-xenial
| 2vyper | /2vyper-0.3.0.tar.gz/2vyper-0.3.0/README.rst | README.rst |
import os
import sys
import shutil
import twovyper.backends
def _backend_path(verifier: str):
backends = os.path.dirname(twovyper.backends.__file__)
return os.path.join(backends, f'{verifier}.jar')
def _sif_path():
backends = os.path.dirname(twovyper.backends.__file__)
return os.path.join(backends, 'silver-sif-extension.jar')
def _executable_path(cmd: str, env: str = None) -> str:
if env:
env_var = os.environ.get(env)
if env_var:
return env_var
return shutil.which(cmd, os.X_OK)
def _construct_classpath(verifier: str = 'silicon'):
""" Contstructs JAVA classpath.
First tries environment variables ``VIPERJAVAPATH``, ``SILICONJAR``
and ``CARBONJAR``. If they are undefined, then uses packaged backends.
"""
viper_java_path = os.environ.get('VIPERJAVAPATH')
silicon_jar = os.environ.get('SILICONJAR')
carbon_jar = os.environ.get('CARBONJAR')
sif_jar = os.environ.get('SIFJAR')
if viper_java_path:
return viper_java_path
if silicon_jar and verifier == 'silicon':
verifier_path = silicon_jar
elif carbon_jar and verifier == 'carbon':
verifier_path = carbon_jar
else:
verifier_path = _backend_path(verifier)
sif_path = sif_jar or _sif_path()
return os.pathsep.join([verifier_path, sif_path])
def _get_boogie_path():
""" Tries to detect path to Boogie executable.
First tries the environment variable ``BOOGIE_EXE``. If it is not
defined, it uses the system default.
"""
cmd = 'Boogie.exe' if sys.platform.startswith('win') else 'boogie'
return _executable_path(cmd, 'BOOGIE_EXE')
def _get_z3_path():
""" Tries to detect path to Z3 executable.
First tries the environment variable ``Z3_EXE``. If it is not defined,
it uses the system default (which is probably the one installed by the dependency).
"""
cmd = 'z3.exe' if sys.platform.startswith('win') else 'z3'
return _executable_path(cmd, 'Z3_EXE')
def set_classpath(v: str):
global classpath
classpath = _construct_classpath(v)
classpath = _construct_classpath()
z3_path = _get_z3_path()
boogie_path = _get_boogie_path() | 2vyper | /2vyper-0.3.0.tar.gz/2vyper-0.3.0/src/twovyper/config.py | config.py |
import argparse
import logging
import os
from time import time
from jpype import JException
########################################################################################################################
# DO NOT GLOBALLY IMPORT ANY SUBMODULE OF TWOVYPER THAT MIGHT DEPEND ON THE VYPER VERSION! #
########################################################################################################################
from twovyper import config
from twovyper import vyper
from twovyper.utils import reload_package
from twovyper.viper.jvmaccess import JVM
from twovyper.viper.typedefs import Program
from twovyper.exceptions import (
InvalidVyperException, ParseException, UnsupportedException, InvalidProgramException, ConsistencyException
)
class TwoVyper:
def __init__(self, jvm: JVM, get_model: bool = False, check_ast_inconsistencies: bool = False):
self.jvm = jvm
self.get_model = get_model
self.check_ast_inconsistencies = check_ast_inconsistencies
def translate(self, path: str, vyper_root: str = None, skip_vyper: bool = False) -> Program:
path = os.path.abspath(path)
if not os.path.isfile(path):
raise InvalidVyperException(f"Contract not found at the given location '{path}'.")
vyper.set_vyper_version(path)
from twovyper.verification import error_manager
error_manager.clear()
# Check that the file is a valid Vyper contract
if not skip_vyper:
vyper.check(path, vyper_root)
logging.debug("Start parsing.")
from twovyper.parsing import parser
vyper_program = parser.parse(path, vyper_root, name=os.path.basename(path.split('.')[0]))
logging.info("Finished parsing.")
logging.debug("Start analyzing.")
from twovyper.analysis import analyzer
for interface in vyper_program.interfaces.values():
analyzer.analyze(interface)
analyzer.analyze(vyper_program)
logging.info("Finished analyzing.")
logging.debug("Start translating.")
from twovyper.translation import translator
from twovyper.translation.translator import TranslationOptions
options = TranslationOptions(self.get_model, self.check_ast_inconsistencies)
translated = translator.translate(vyper_program, options, self.jvm)
logging.info("Finished translating.")
return translated
def verify(self, program: Program, path: str, backend: str) -> 'VerificationResult':
"""
Verifies the given Viper program
"""
logging.debug("Start verifying.")
from twovyper.verification.verifier import (
VerificationResult,
ViperVerifier, AbstractVerifier
)
verifier: AbstractVerifier = ViperVerifier[backend].value
result = verifier.verify(program, self.jvm, path, self.get_model)
logging.info("Finished verifying.")
return result
def main() -> None:
"""
Entry point for the verifier.
"""
parser = argparse.ArgumentParser()
parser.add_argument(
'vyper_file',
help='Python file to verify'
)
parser.add_argument(
'--verifier',
help='verifier to be used (carbon or silicon)',
choices=['silicon', 'carbon'],
default='silicon'
)
parser.add_argument(
'--viper-jar-path',
help='Java CLASSPATH that includes Viper class files',
default=None
)
parser.add_argument(
'--z3',
help='path to Z3 executable',
default=config.z3_path
)
parser.add_argument(
'--boogie',
help='path to Boogie executable',
default=config.boogie_path
)
parser.add_argument(
'--counterexample',
action='store_true',
help='print a counterexample if the verification fails',
)
parser.add_argument(
'--check-ast-inconsistencies',
action='store_true',
help='Check the generated Viper AST for potential inconsistencies.',
)
parser.add_argument(
'--vyper-root',
help='import root directory for the Vyper contract',
default=None
)
parser.add_argument(
'--skip-vyper',
action='store_true',
help='skip validity check of contract with the Vyper compiler'
)
parser.add_argument(
'--print-viper',
action='store_true',
help='print generated Viper program'
)
parser.add_argument(
'--write-viper-to-file',
default=None,
help='write generated Viper program to specified file'
)
parser.add_argument(
'--log',
choices=['DEBUG', 'INFO', 'WARNING', 'ERROR', 'CRITICAL'],
help='log level',
default='WARNING'
)
parser.add_argument(
'--benchmark',
type=int,
help='run verification the given number of times to benchmark performance',
default=-1
)
parser.add_argument(
'--ide-mode',
action='store_true',
help='output errors in IDE format'
)
args = parser.parse_args()
if args.viper_jar_path:
config.classpath = args.viper_jar_path
else:
config.set_classpath(args.verifier)
config.boogie_path = args.boogie
config.z3_path = args.z3
if not config.classpath:
parser.error('missing argument: --viper-jar-path')
if not config.z3_path:
parser.error('missing argument: --z3')
if args.verifier == 'carbon' and not config.boogie_path:
parser.error('missing argument: --boogie')
if args.verifier == 'carbon' and args.counterexample:
parser.error('counterexamples are only supported with the Silicon backend')
formatter = logging.Formatter()
handler = logging.StreamHandler()
handler.setFormatter(formatter)
logging.basicConfig(level=args.log, handlers=[handler])
jvm = JVM(config.classpath)
translate_and_verify(args.vyper_file, jvm, args)
def translate_and_verify(vyper_file, jvm, args, print=print):
try:
start = time()
tw = TwoVyper(jvm, args.counterexample, args.check_ast_inconsistencies)
program = tw.translate(vyper_file, args.vyper_root, args.skip_vyper)
if args.print_viper:
print(str(program))
if args.write_viper_to_file:
with open(args.write_viper_to_file, 'w') as fp:
fp.write(str(program))
backend = args.verifier
if args.benchmark >= 1:
print("Run, Total, Start, End, Time")
for i in range(args.benchmark):
start = time()
prog = tw.translate(vyper_file, args.vyper_root, args.skip_vyper)
vresult = tw.verify(prog, vyper_file, backend)
end = time()
print(f"{i}, {args.benchmark}, {start}, {end}, {end - start}")
else:
vresult = tw.verify(program, vyper_file, backend)
print(vresult.string(args.ide_mode, include_model=args.counterexample))
end = time()
duration = end - start
print(f"Verification took {duration:.2f} seconds.")
except (InvalidProgramException, UnsupportedException) as e:
print(e.error_string())
except ConsistencyException as e:
print(e.program)
print(e.message)
for error in e.errors:
print(error.toString())
except (InvalidVyperException, ParseException) as e:
print(e.message)
except JException as e:
print(e.stacktrace())
raise e
def prepare_twovyper_for_vyper_version(path: str) -> bool:
"""
Checks if the current loaded vyper version differs from the vyper version used in a vyper contract.
If it does differ, reload the whole twovyper module.
Please note that all references (e.g. via a "from ... import ...") to the twovyper module are not refreshed
automatically. Therefore, use this function with caution.
:param path: The path to a vyper contract.
:return: True iff a reload of the module was performed.
"""
version = vyper.get_vyper_version()
vyper.set_vyper_version(path)
if version != vyper.get_vyper_version():
import twovyper
reload_package(twovyper)
return True
return False
if __name__ == '__main__':
main() | 2vyper | /2vyper-0.3.0.tar.gz/2vyper-0.3.0/src/twovyper/main.py | main.py |
import importlib
import os
from contextlib import contextmanager
from types import ModuleType
from typing import Callable, Iterable, List, Optional, TypeVar
_ = object()
@contextmanager
def switch(*values):
def match(*v, where=True):
if not where or len(values) != len(v):
return False
return all(case is _ or actual == case for actual, case in zip(values, v))
yield match
T = TypeVar('T')
def first(iterable: Iterable[T]) -> Optional[T]:
return next(iter(iterable), None)
def first_index(statisfying: Callable[[T], bool], l) -> int:
return next((i for i, v in enumerate(l) if statisfying(v)), -1)
def flatten(iterables: Iterable[Iterable[T]]) -> List[T]:
return [item for subiterable in iterables for item in subiterable]
def unique(eq, iterable: Iterable[T]) -> List[T]:
unique_iterable: List[T] = []
for elem in iterable:
for uelem in unique_iterable:
if eq(elem, uelem):
break
else:
unique_iterable.append(elem)
return unique_iterable
def seq_to_list(scala_iterable):
lst = []
it = scala_iterable.iterator()
while it.hasNext():
lst.append(it.next())
return lst
def list_to_seq(lst, jvm):
seq = jvm.scala.collection.mutable.ArraySeq(len(lst))
for i, element in enumerate(lst):
seq.update(i, element)
return seq
def reload_package(package):
assert(hasattr(package, "__package__"))
fn = package.__file__
fn_dir = os.path.dirname(fn) + os.sep
module_visit = {fn}
del fn
# Reloading all modules in a depth first search matter starting from "package".
def reload_recursive_ex(module):
importlib.reload(module)
for module_child in vars(module).values():
if isinstance(module_child, ModuleType):
fn_child = getattr(module_child, "__file__", None)
if (fn_child is not None) and fn_child.startswith(fn_dir):
if fn_child not in module_visit:
module_visit.add(fn_child)
reload_recursive_ex(module_child)
reload_recursive_ex(package)
class Subscriptable(type):
"""
A metaclass to add dictionary lookup functionality to a class.
The class needs to implement the `_subscript(_)` method.
"""
def __getitem__(cls, val):
return cls._subscript(val) | 2vyper | /2vyper-0.3.0.tar.gz/2vyper-0.3.0/src/twovyper/utils.py | utils.py |
import logging
import os
import re
from subprocess import Popen, PIPE
from typing import Optional, List, Dict, TypeVar
import vvm
import vyper
from semantic_version import NpmSpec, Version
from vvm.utils.convert import to_vyper_version
from twovyper.exceptions import InvalidVyperException, UnsupportedVersionException
try:
# noinspection PyUnresolvedReferences,PyUnboundLocalVariable
_imported
except NameError:
_imported = True # means the module is being imported
else:
_imported = False # means the module is being reloaded
if _imported:
_original_get_executable = vvm.install.get_executable
def _get_executable(version=None, vvm_binary_path=None) -> str:
path = _original_get_executable(version, vvm_binary_path)
return os.path.abspath(path)
_version_pragma_pattern = re.compile(r"(?:\n|^)\s*#\s*@version\s*([^\n]*)")
_comment_pattern = re.compile(r"(?:\n|^)\s*(#|$)")
_VYPER_VERSION = to_vyper_version(vyper.__version__)
_installed_vyper_versions: Optional[List[Version]] = None
_available_vyper_versions: Optional[List[Version]] = None
_current_vyper_version: Optional[Version] = None
_cache_of_parsed_version_strings: Dict[str, NpmSpec] = {}
# noinspection PyUnboundLocalVariable
vvm.install.get_executable = _get_executable
def _parse_version_string(version_str: str) -> NpmSpec:
result = _cache_of_parsed_version_strings.get(version_str)
if result is not None:
return result
try:
result = NpmSpec(version_str)
except ValueError:
try:
version = to_vyper_version(version_str)
result = NpmSpec(str(version))
except Exception:
raise InvalidVyperException(f"Cannot parse Vyper version from pragma: {version_str}")
_cache_of_parsed_version_strings[version_str] = result
return result
def _get_vyper_pragma_spec(path: str) -> NpmSpec:
pragma_string = None
with open(path, 'r') as file:
for line in file:
pragma_match = _version_pragma_pattern.match(line)
if pragma_match is not None:
pragma_string = pragma_match.groups()[0]
pragma_string = " ".join(pragma_string.split())
break
comment_match = _comment_pattern.match(line)
if comment_match is None:
# version pragma has to comme before code
break
if pragma_string is None:
pragma_string = vyper.__version__
logging.warning(f'No version pragma found. (Using vyper version "{pragma_string}" instead.)')
return _parse_version_string(pragma_string)
def _find_vyper_version(file: str) -> str:
global _installed_vyper_versions
if _installed_vyper_versions is None:
_installed_vyper_versions = vvm.get_installed_vyper_versions()
_installed_vyper_versions.append(_VYPER_VERSION)
pragma_specs = _get_vyper_pragma_spec(file)
version = pragma_specs.select(_installed_vyper_versions)
if not version:
global _available_vyper_versions
if _available_vyper_versions is None:
_available_vyper_versions = vvm.get_installable_vyper_versions()
version = pragma_specs.select(_available_vyper_versions)
if not version:
raise InvalidVyperException(f"Invalid vyper version pragma: {pragma_specs}")
lock = vvm.install.get_process_lock(f"locked${version}")
with lock:
try:
_get_executable(version)
except vvm.exceptions.VyperNotInstalled:
vvm.install_vyper(version)
if version not in _installed_vyper_versions:
_installed_vyper_versions.append(version)
return version
def check(file: str, root=None):
"""
Checks that the file is a valid Vyper contract. If not, throws an `InvalidVyperException`.
"""
if _VYPER_VERSION == _current_vyper_version:
path_str = 'vyper'
else:
path = _get_executable(_current_vyper_version)
path_str = os.path.abspath(path)
pipes = Popen([path_str, file], stdout=PIPE, stderr=PIPE, cwd=root)
_, stderr = pipes.communicate()
if pipes.returncode != 0:
err_msg = stderr.strip().decode('utf-8')
raise InvalidVyperException(err_msg)
def get_vyper_version() -> Version:
if _current_vyper_version is None:
return _VYPER_VERSION
return _current_vyper_version
def set_vyper_version(file: str):
global _current_vyper_version
_current_vyper_version = _find_vyper_version(file)
def is_compatible_version(version: str) -> bool:
vyper_version = get_vyper_version()
specs = _parse_version_string(version)
return specs.match(vyper_version)
T = TypeVar('T')
def select_version(value_dict: Dict[str, T], default: T = None) -> T:
for version_str, value in value_dict.items():
if is_compatible_version(version_str):
return value
else:
if default is None:
raise UnsupportedVersionException()
return default | 2vyper | /2vyper-0.3.0.tar.gz/2vyper-0.3.0/src/twovyper/vyper.py | vyper.py |
import os
from twovyper.ast import ast_nodes as ast
from twovyper.ast.text import pprint
class TwoVyperException(Exception):
def __init__(self, message: str):
self.message = message
class InvalidVyperException(TwoVyperException):
"""
Exception indicating that the file is not a valid Vyper contract.
"""
def __init__(self, vyper_output: str):
message = f"Not a valid Vyper contract:\n{vyper_output}"
super().__init__(message)
class ParseException(TwoVyperException):
"""
Exception that is thrown if the contract cannot be parsed.
"""
def __init__(self, message: str):
super().__init__(message)
class TranslationException(TwoVyperException):
def __init__(self, node: ast.Node, message: str):
super().__init__(message)
self.node = node
def error_string(self) -> str:
file_name = os.path.basename(self.node.file)
line = self.node.lineno
col = self.node.col_offset
return f"{self.message} ({file_name}@{line}.{col})"
class UnsupportedException(TranslationException):
"""
Exception that is thrown when attempting to translate a Vyper element not
currently supported.
"""
def __init__(self, node: ast.Node, message: str = None):
self.node = node
if not message:
message = pprint(node)
super().__init__(node, f"Not supported: {message}")
class InvalidProgramException(TranslationException):
"""
Signals that the input program is invalid and cannot be translated.
"""
def __init__(self, node: ast.Node, reason_code: str, message: str = None):
self.node = node
self.code = 'invalid.program'
self.reason_code = reason_code
if not message:
node_msg = pprint(node, True)
message = f"Node\n{node_msg}\nnot allowed here."
super().__init__(node, f"Invalid program ({self.reason_code}): {message}")
class ConsistencyException(TwoVyperException):
"""
Exception reporting that the translated AST has a consistency error.
"""
def __init__(self, program, message: str, errors):
super().__init__(message)
self.program = program
self.errors = errors
class UnsupportedVersionException(TwoVyperException):
"""
Exception that is thrown when a unsupported Vyper version is used by a contract.
"""
def __init__(self, message: str = None):
if message is None:
from twovyper.vyper import get_vyper_version
message = f"The Vyper version ({get_vyper_version()}) used by the contract is not supported."
super().__init__(message) | 2vyper | /2vyper-0.3.0.tar.gz/2vyper-0.3.0/src/twovyper/exceptions.py | exceptions.py |
import os
from twovyper.ast import names
from twovyper.ast.nodes import VyperProgram, VyperInterface
from twovyper.exceptions import InvalidProgramException
from twovyper.utils import first
def check_symbols(program: VyperProgram):
_check_ghost_functions(program)
_check_ghost_implements(program)
_check_resources(program)
def _check_resources(program: VyperProgram):
if not isinstance(program, VyperInterface):
node = first(program.node.stmts) or program.node
for interface in program.interfaces.values():
for resource_name, resources_list in interface.resources.items():
for resource in resources_list:
if resource.file is None:
if program.resources.get(resource_name) is None:
if not program.config.has_option(names.CONFIG_ALLOCATION):
raise InvalidProgramException(node, 'alloc.not.alloc',
f'The interface "{interface.name}" uses the '
f'allocation config option. Therefore, this contract '
f'also has to enable this config option.')
raise InvalidProgramException(node, 'missing.resource',
f'The interface "{interface.name}" '
f'needs a default resource "{resource_name}" '
f'that is not present in this contract.')
continue
imported_resources = [r for r in program.resources.get(resource_name, [])
if r.file == resource.file]
if not imported_resources:
prefix_length = len(os.path.commonprefix([resource.file, program.file]))
raise InvalidProgramException(node, 'missing.resource',
f'The interface "{interface.name}" '
f'needs a resource "{resource_name}" from '
f'".{os.path.sep}{resource.file[prefix_length:]}" but it '
f'was not imported for this contract.')
imported_resources = [r for r in program.resources.get(resource_name)
if r.interface == resource.interface]
for imported_resource in imported_resources:
if resource.file != imported_resource.file:
prefix_length = len(os.path.commonprefix([resource.file, imported_resource.file]))
resource_file = resource.file[prefix_length:]
imported_resource_file = imported_resource.file[prefix_length:]
raise InvalidProgramException(node, 'duplicate.resource',
f'There are two versions of the resource '
f'"{resource_name}" defined in an interface '
f'"{imported_resource.interface}" one from '
f'[...]"{imported_resource_file}" the other from '
f'[...]"{resource_file}".')
for interface_type in program.implements:
interface = program.interfaces[interface_type.name]
for resource_name, resource in program.declared_resources.items():
if resource_name in interface.own_resources:
raise InvalidProgramException(resource.node, 'duplicate.resource',
f'A contract cannot redeclare a resource it already imports. '
f'The resource "{resource_name}" got already declared in the '
f'interface {interface.name}.')
def _check_ghost_functions(program: VyperProgram):
if not isinstance(program, VyperInterface):
node = first(program.node.stmts) or program.node
for implemented_ghost in program.ghost_function_implementations.values():
if program.ghost_functions.get(implemented_ghost.name) is None:
raise InvalidProgramException(implemented_ghost.node, 'missing.ghost',
f'This contract is implementing an unknown ghost function. '
f'None of the interfaces, this contract implements, declares a ghost '
f'function "{implemented_ghost.name}".')
for interface in program.interfaces.values():
for ghost_function_list in interface.ghost_functions.values():
for ghost_function in ghost_function_list:
imported_ghost_functions = [ghost_func
for ghost_func in program.ghost_functions.get(ghost_function.name, [])
if ghost_func.file == ghost_function.file]
if not imported_ghost_functions:
prefix_length = len(os.path.commonprefix([ghost_function.file, program.file]))
raise InvalidProgramException(node, 'missing.ghost',
f'The interface "{interface.name}" '
f'needs a ghost function "{ghost_function.name}" from '
f'".{os.path.sep}{ghost_function.file[prefix_length:]}" but it '
f'was not imported for this contract.')
imported_ghost_functions = [ghost_func
for ghost_func in program.ghost_functions.get(ghost_function.name)
if ghost_func.interface == ghost_function.interface]
for imported_ghost_function in imported_ghost_functions:
if ghost_function.file != imported_ghost_function.file:
prefix_length = len(os.path.commonprefix(
[ghost_function.file, imported_ghost_function.file]))
ghost_function_file = ghost_function.file[prefix_length:]
imported_ghost_function_file = imported_ghost_function.file[prefix_length:]
raise InvalidProgramException(node, 'duplicate.ghost',
f'There are two versions of the ghost function '
f'"{ghost_function.name}" defined in an interface '
f'"{ghost_function.interface}" one from '
f'[...]"{imported_ghost_function_file}" the other from '
f'[...]"{ghost_function_file}".')
def _check_ghost_implements(program: VyperProgram):
def check(cond, node, ghost_name, interface_name):
if not cond:
raise InvalidProgramException(node, 'ghost.not.implemented',
f'The ghost function "{ghost_name}" from the interface "{interface_name}" '
f'has not been implemented correctly.')
ghost_function_implementations = dict(program.ghost_function_implementations)
for itype in program.implements:
interface = program.interfaces[itype.name]
for ghost in interface.own_ghost_functions.values():
implementation = ghost_function_implementations.pop(ghost.name, None)
check(implementation is not None, program.node, ghost.name, itype.name)
check(implementation.name == ghost.name, implementation.node, ghost.name, itype.name)
check(len(implementation.args) == len(ghost.args), implementation.node, ghost.name, itype.name)
check(implementation.type == ghost.type, implementation.node, ghost.name, itype.name)
if len(ghost_function_implementations) > 0:
raise InvalidProgramException(first(ghost_function_implementations.values()).node, 'invalid.ghost.implemented',
f'This contract implements some ghost functions that have no declaration in '
f'any of the implemented interfaces.\n'
f'(Ghost functions without declaration: {list(ghost_function_implementations)})') | 2vyper | /2vyper-0.3.0.tar.gz/2vyper-0.3.0/src/twovyper/analysis/symbol_checker.py | symbol_checker.py |
from itertools import zip_longest
from contextlib import contextmanager
from typing import Optional, Union, Dict, Iterable
from twovyper.utils import first_index, switch, first
from twovyper.ast import ast_nodes as ast, names, types
from twovyper.ast.arithmetic import Decimal
from twovyper.ast.types import (
TypeBuilder, VyperType, MapType, ArrayType, StringType, StructType, AnyStructType,
SelfType, ContractType, InterfaceType, TupleType
)
from twovyper.ast.nodes import VyperProgram, VyperFunction, GhostFunction, VyperInterface
from twovyper.ast.text import pprint
from twovyper.ast.visitors import NodeVisitor
from twovyper.exceptions import InvalidProgramException, UnsupportedException
from twovyper.vyper import is_compatible_version, select_version
def _check(condition: bool, node: ast.Node, reason_code: str, msg: Optional[str] = None):
if not condition:
raise InvalidProgramException(node, reason_code, msg)
class TypeAnnotator(NodeVisitor):
# The type annotator annotates all expression nodes with type information.
# It works by using the visitor pattern as follows:
# - Each expression that is visited returns a list of possible types for it (ordered by
# priority) and a list of nodes which will have that type, e.g., the literal 5 will return
# the types [int128, uint256, address] when visited, and itself as the only node.
# - Statements simply return None, as they don't have types.
# - Some operations, e.g., binary operations, have as their result type the same type
# as the input type. The 'combine' method combines the input type sets for that.
# - If the required type is known at some point, 'annotate_expected' can be called with
# the expected type as the argument. It then checks that the nodes of the expected type
# and annotates all nodes.
# - If any type can be used, 'annotate' can be called to annotate all nodes with the first
# type in the list, i.e., the type with the highest priority.
def __init__(self, program: VyperProgram):
type_map = {}
for name, struct in program.structs.items():
type_map[name] = struct.type
for name, contract in program.contracts.items():
type_map[name] = contract.type
for name, interface in program.interfaces.items():
type_map[name] = interface.type
self.type_builder = TypeBuilder(type_map)
self.program = program
self.current_func: Union[VyperFunction, None] = None
self.current_loops: Dict[str, ast.For] = {}
self_type = self.program.fields.type
# Contains the possible types a variable can have
self.known_variables = {
names.BLOCK: [types.BLOCK_TYPE],
names.CHAIN: [types.CHAIN_TYPE],
names.TX: [types.TX_TYPE],
names.MSG: [types.MSG_TYPE],
names.SELF: [types.VYPER_ADDRESS, self_type],
names.LOG: [None],
names.LEMMA: [None]
}
self.known_variables.update({interface: [None] for interface in program.interfaces.keys()})
self.variables = self.known_variables.copy()
self.undecided_nodes = False
self.type_resolver = TypeResolver()
@contextmanager
def _function_scope(self, func: Union[VyperFunction, GhostFunction]):
old_func = self.current_func
self.current_func = func
old_variables = self.variables.copy()
self.variables = self.known_variables.copy()
self.variables.update({k: [v.type] for k, v in func.args.items()})
yield
self.current_func = old_func
self.variables = old_variables
@contextmanager
def _quantified_vars_scope(self):
old_variables = self.variables.copy()
yield
self.variables = old_variables
@contextmanager
def _loop_scope(self):
current_loops = self.current_loops
yield
self.current_loops = current_loops
@property
def method_name(self):
return 'visit'
def check_number_of_arguments(self, node: Union[ast.FunctionCall, ast.ReceiverCall], *expected: int,
allowed_keywords: Iterable[str] = (), required_keywords: Iterable[str] = (),
resources: int = 0):
_check(len(node.args) in expected, node, 'invalid.no.args')
for kw in node.keywords:
_check(kw.name in allowed_keywords, node, 'invalid.no.args')
for kw in required_keywords:
_check(any(k.name == kw for k in node.keywords), node, 'invalid.no.args')
if resources > 0 and self.program.config.has_option(names.CONFIG_NO_DERIVED_WEI):
_check(node.resource is not None, node, 'invalid.resource',
'The derived wei resource got disabled and cannot be used.')
if isinstance(node, ast.FunctionCall) and node.resource:
if resources == 1:
cond = not isinstance(node.resource, ast.Exchange)
elif resources == 2:
cond = isinstance(node.resource, ast.Exchange)
else:
cond = False
_check(cond, node, 'invalid.no.resources')
def annotate_program(self):
for resources in self.program.resources.values():
for resource in resources:
if (isinstance(resource.type, types.DerivedResourceType)
and isinstance(resource.type.underlying_resource, types.UnknownResourceType)):
underlying_resource, args = self._annotate_resource(resource.underlying_resource)
resource.underlying_address = underlying_resource.own_address
underlying_resource.derived_resources.append(resource)
if len(args) != 0:
_check(False, args[0], 'invalid.derived.resource',
'The underlying resource type must be declared without arguments.')
resource.type.underlying_resource = underlying_resource.type
derived_resource_member_types = resource.type.member_types.values()
underlying_resource_member_types = underlying_resource.type.member_types.values()
_check(resource.file != underlying_resource.file, resource.node, 'invalid.derived.resource',
'A resource cannot be a derived resource from a resource of the same contract.')
_check(resource.underlying_address is not None, resource.node, 'invalid.derived.resource',
'The underlying resource of a derived resource must have an address specified.')
for derived_member_type, underlying_member_type \
in zip_longest(derived_resource_member_types, underlying_resource_member_types):
_check(derived_member_type == underlying_member_type, resource.node,
'invalid.derived.resource',
'Arguments of derived resource are not matching underlying resource.')
if not isinstance(self.program, VyperInterface):
# Set analysed flag to True for all resources
for resources in self.program.resources.values():
for resource in resources:
resource.analysed = True
for function in self.program.functions.values():
with self._function_scope(function):
for pre in function.preconditions:
self.annotate_expected(pre, types.VYPER_BOOL, resolve=True)
for performs in function.performs:
self.annotate(performs, resolve=True)
self.visit(function.node)
self.resolve_type(function.node)
for name, default in function.defaults.items():
if default:
self.annotate_expected(default, function.args[name].type, resolve=True)
for post in function.postconditions:
self.annotate_expected(post, types.VYPER_BOOL, resolve=True)
for check in function.checks:
self.annotate_expected(check, types.VYPER_BOOL, resolve=True)
for loop_invariants in function.loop_invariants.values():
for loop_invariant in loop_invariants:
self.annotate_expected(loop_invariant, types.VYPER_BOOL, resolve=True)
for lemma in self.program.lemmas.values():
with self._function_scope(lemma):
for stmt in lemma.node.body:
assert isinstance(stmt, ast.ExprStmt)
self.annotate_expected(stmt.value, types.VYPER_BOOL, resolve=True)
for name, default in lemma.defaults.items():
if default:
self.annotate_expected(default, lemma.args[name].type, resolve=True)
for pre in lemma.preconditions:
self.annotate_expected(pre, types.VYPER_BOOL, resolve=True)
for ghost_function in self.program.ghost_function_implementations.values():
assert isinstance(ghost_function, GhostFunction)
with self._function_scope(ghost_function):
expression_stmt = ghost_function.node.body[0]
assert isinstance(expression_stmt, ast.ExprStmt)
self.annotate_expected(expression_stmt.value, ghost_function.type.return_type, resolve=True)
for inv in self.program.invariants:
self.annotate_expected(inv, types.VYPER_BOOL, resolve=True)
for post in self.program.general_postconditions:
self.annotate_expected(post, types.VYPER_BOOL, resolve=True)
for post in self.program.transitive_postconditions:
self.annotate_expected(post, types.VYPER_BOOL, resolve=True)
for check in self.program.general_checks:
self.annotate_expected(check, types.VYPER_BOOL, resolve=True)
if isinstance(self.program, VyperInterface):
for caller_private in self.program.caller_private:
self.annotate(caller_private, resolve=True)
def resolve_type(self, node: ast.Node):
if self.undecided_nodes:
self.type_resolver.resolve_type(node)
self.undecided_nodes = False
def generic_visit(self, node: ast.Node, *args):
assert False
def pass_through(self, node1, node=None):
vyper_types, nodes = self.visit(node1)
if node is not None:
nodes.append(node)
return vyper_types, nodes
def combine(self, node1, node2, node=None, allowed=lambda t: True):
types1, nodes1 = self.visit(node1)
types2, nodes2 = self.visit(node2)
def is_array(t):
return isinstance(t, ArrayType) and not t.is_strict
def longest_array(matching, lst):
try:
arrays = [t for t in lst if is_array(t) and t.element_type == matching.element_type]
return max(arrays, key=lambda t: t.size)
except ValueError:
raise InvalidProgramException(node if node is not None else node2, 'wrong.type')
def intersect(l1, l2):
for e in l1:
if is_array(e):
yield longest_array(e, l2)
else:
for t in l2:
if types.matches(e, t):
yield e
elif types.matches(t, e):
yield t
intersection = list(filter(allowed, intersect(types1, types2)))
if not intersection:
raise InvalidProgramException(node if node is not None else node2, 'wrong.type')
else:
if node:
return intersection, [*nodes1, *nodes2, node]
else:
return intersection, [*nodes1, *nodes2]
def annotate(self, node1, node2=None, allowed=lambda t: True, resolve=False):
if node2 is None:
combined, nodes = self.visit(node1)
else:
combined, nodes = self.combine(node1, node2, allowed=allowed)
for node in nodes:
if len(combined) > 1:
self.undecided_nodes = True
node.type = combined
else:
node.type = combined[0]
if resolve:
self.resolve_type(node1)
if node2 is not None:
self.resolve_type(node2)
def annotate_expected(self, node, expected, orelse=None, resolve=False):
"""
Checks that node has type `expected` (or matches the predicate `expected`) or, if that isn't the case,
that is has type `orelse` (or matches the predicate `orelse`).
"""
tps, nodes = self.visit(node)
def annotate_nodes(t):
node.type = t
for n in nodes:
n.type = t
if resolve:
self.resolve_type(node)
if isinstance(expected, VyperType):
if any(types.matches(t, expected) for t in tps):
annotate_nodes(expected)
return
else:
for t in tps:
if expected(t):
annotate_nodes(t)
return
if orelse:
self.annotate_expected(node, orelse)
else:
raise InvalidProgramException(node, 'wrong.type')
def visit_FunctionDef(self, node: ast.FunctionDef):
for stmt in node.body:
self.visit(stmt)
def visit_Return(self, node: ast.Return):
if node.value:
# Interfaces are allowed to just have `pass` as a body instead of a correct return,
# therefore we don't check for the correct return type.
if self.program.is_interface():
self.annotate(node.value)
else:
self.annotate_expected(node.value, self.current_func.type.return_type)
def visit_Assign(self, node: ast.Assign):
self.annotate(node.target, node.value)
def visit_AugAssign(self, node: ast.AugAssign):
self.annotate(node.target, node.value, types.is_numeric)
def visit_AnnAssign(self, node: ast.AnnAssign):
# This is a variable declaration so we add it to the type map. Since Vyper
# doesn't allow shadowing, we can just override the previous value.
variable_name = node.target.id
variable_type = self.type_builder.build(node.annotation)
self.variables[variable_name] = [variable_type]
self.annotate_expected(node.target, variable_type)
if node.value:
self.annotate_expected(node.value, variable_type)
def visit_For(self, node: ast.For):
self.annotate_expected(node.iter, lambda t: isinstance(t, ArrayType))
# Add the loop variable to the type map
var_name = node.target.id
var_type = node.iter.type.element_type
self.variables[var_name] = [var_type]
self.annotate_expected(node.target, var_type)
with self._loop_scope():
self.current_loops[var_name] = node
# Visit loop invariants
for loop_inv in self.current_func.loop_invariants.get(node, []):
self.annotate_expected(loop_inv, types.VYPER_BOOL)
# Visit body
for stmt in node.body:
self.visit(stmt)
def visit_If(self, node: ast.If):
self.annotate_expected(node.test, types.VYPER_BOOL)
for stmt in node.body + node.orelse:
self.visit(stmt)
def visit_Raise(self, node: ast.Raise):
if node.msg:
if not (isinstance(node.msg, ast.Name) and node.msg.id == names.UNREACHABLE):
self.annotate(node.msg)
def visit_Assert(self, node: ast.Assert):
self.annotate_expected(node.test, types.VYPER_BOOL)
def visit_ExprStmt(self, node: ast.ExprStmt):
self.annotate(node.value)
def visit_Log(self, node: ast.Log):
_check(isinstance(node.body, ast.FunctionCall), node, 'invalid.log')
assert isinstance(node.body, ast.FunctionCall)
for arg in node.body.args:
self.annotate(arg)
for kw in node.body.keywords:
self.annotate(kw.value)
def visit_Pass(self, node: ast.Pass):
pass
def visit_Continue(self, node: ast.Continue):
pass
def visit_Break(self, node: ast.Break):
pass
def visit_BoolOp(self, node: ast.BoolOp):
self.annotate_expected(node.left, types.VYPER_BOOL)
self.annotate_expected(node.right, types.VYPER_BOOL)
return [types.VYPER_BOOL], [node]
def visit_Not(self, node: ast.Not):
self.annotate_expected(node.operand, types.VYPER_BOOL)
return [types.VYPER_BOOL], [node]
def visit_ArithmeticOp(self, node: ast.ArithmeticOp):
return self.combine(node.left, node.right, node, types.is_numeric)
def visit_UnaryArithmeticOp(self, node: ast.UnaryArithmeticOp):
return self.pass_through(node.operand, node)
def visit_IfExpr(self, node: ast.IfExpr):
self.annotate_expected(node.test, types.VYPER_BOOL)
return self.combine(node.body, node.orelse, node)
def visit_Comparison(self, node: ast.Comparison):
self.annotate(node.left, node.right)
return [types.VYPER_BOOL], [node]
def visit_Containment(self, node: ast.Containment):
self.annotate_expected(node.list, lambda t: isinstance(t, ArrayType))
array_type = node.list.type
assert isinstance(array_type, ArrayType)
self.annotate_expected(node.value, array_type.element_type)
return [types.VYPER_BOOL], [node]
def visit_Equality(self, node: ast.Equality):
self.annotate(node.left, node.right)
return [types.VYPER_BOOL], [node]
def visit_FunctionCall(self, node: ast.FunctionCall):
name = node.name
if node.resource:
self._visit_resource(node.resource, node)
with switch(name) as case:
if case(names.CONVERT):
self.check_number_of_arguments(node, 2)
self.annotate(node.args[0])
ntype = self.type_builder.build(node.args[1])
return [ntype], [node]
elif case(names.SUCCESS):
self.check_number_of_arguments(node, 0, 1, allowed_keywords=[names.SUCCESS_IF_NOT])
if node.args:
argument = node.args[0]
assert isinstance(argument, ast.ReceiverCall)
self.annotate(argument)
_check(isinstance(argument.receiver.type, types.SelfType), argument, 'spec.success',
'Only functions defined in this contract can be called from the specification.')
return [types.VYPER_BOOL], [node]
elif case(names.REVERT):
self.check_number_of_arguments(node, 0, 1)
if node.args:
argument = node.args[0]
assert isinstance(argument, ast.ReceiverCall)
self.annotate(argument)
_check(isinstance(argument.receiver.type, types.SelfType), argument, 'spec.success',
'Only functions defined in this contract can be called from the specification.')
return [types.VYPER_BOOL], [node]
elif case(names.FORALL):
return self._visit_forall(node)
elif case(names.ACCESSIBLE):
return self._visit_accessible(node)
elif case(names.EVENT):
self.check_number_of_arguments(node, 1, 2)
return self._visit_event(node)
elif case(names.PREVIOUS):
self.check_number_of_arguments(node, 1)
self.annotate(node.args[0])
loop = self._retrieve_loop(node, names.PREVIOUS)
loop_array_type = loop.iter.type
assert isinstance(loop_array_type, ArrayType)
return [ArrayType(loop_array_type.element_type, loop_array_type.size, False)], [node]
elif case(names.LOOP_ARRAY):
self.check_number_of_arguments(node, 1)
self.annotate(node.args[0])
loop = self._retrieve_loop(node, names.LOOP_ARRAY)
loop_array_type = loop.iter.type
assert isinstance(loop_array_type, ArrayType)
return [loop_array_type], [node]
elif case(names.LOOP_ITERATION):
self.check_number_of_arguments(node, 1)
self.annotate(node.args[0])
self._retrieve_loop(node, names.LOOP_ITERATION)
return [types.VYPER_UINT256], [node]
elif case(names.CALLER):
self.check_number_of_arguments(node, 0)
return [types.VYPER_ADDRESS], [node]
elif case(names.RANGE):
self.check_number_of_arguments(node, 1, 2)
return self._visit_range(node)
elif case(names.MIN) or case(names.MAX):
self.check_number_of_arguments(node, 2)
return self.combine(node.args[0], node.args[1], node, types.is_numeric)
elif case(names.ADDMOD) or case(names.MULMOD):
self.check_number_of_arguments(node, 3)
for arg in node.args:
self.annotate_expected(arg, types.VYPER_UINT256)
return [types.VYPER_UINT256], [node]
elif case(names.SQRT):
self.check_number_of_arguments(node, 1)
self.annotate_expected(node.args[0], types.VYPER_DECIMAL)
return [types.VYPER_DECIMAL], [node]
elif case(names.SHIFT):
self.check_number_of_arguments(node, 2)
self.annotate_expected(node.args[0], types.VYPER_UINT256)
self.annotate_expected(node.args[1], types.VYPER_INT128)
elif case(names.BITWISE_NOT):
self.check_number_of_arguments(node, 1)
self.annotate_expected(node.args[0], types.VYPER_UINT256)
return [types.VYPER_UINT256], [node]
elif case(names.BITWISE_AND) or case(names.BITWISE_OR) or case(names.BITWISE_XOR):
self.check_number_of_arguments(node, 2)
self.annotate_expected(node.args[0], types.VYPER_UINT256)
self.annotate_expected(node.args[1], types.VYPER_UINT256)
return [types.VYPER_UINT256], [node]
elif case(names.OLD) or case(names.PUBLIC_OLD) or case(names.ISSUED):
self.check_number_of_arguments(node, 1)
return self.pass_through(node.args[0], node)
elif case(names.INDEPENDENT):
self.check_number_of_arguments(node, 2)
self.annotate(node.args[0])
self.annotate(node.args[1])
return [types.VYPER_BOOL], [node]
elif case(names.REORDER_INDEPENDENT):
self.check_number_of_arguments(node, 1)
self.annotate(node.args[0])
return [types.VYPER_BOOL], [node]
elif case(names.FLOOR) or case(names.CEIL):
self.check_number_of_arguments(node, 1)
self.annotate_expected(node.args[0], types.VYPER_DECIMAL)
return [types.VYPER_INT128], [node]
elif case(names.LEN):
self.check_number_of_arguments(node, 1)
self.annotate_expected(node.args[0], lambda t: isinstance(t, types.ArrayType))
return_type = select_version({'^0.2.0': types.VYPER_UINT256,
'>=0.1.0-beta.16 <0.1.0': types.VYPER_INT128})
return [return_type], [node]
elif case(names.STORAGE):
self.check_number_of_arguments(node, 1)
self.annotate_expected(node.args[0], types.VYPER_ADDRESS)
# We know that storage(self) has the self-type
first_arg = node.args[0]
if isinstance(first_arg, ast.Name) and first_arg.id == names.SELF:
return [self.program.type, AnyStructType()], [node]
# Otherwise it is just some struct, which we don't know anything about
else:
return [AnyStructType()], [node]
elif case(names.ASSERT_MODIFIABLE):
self.check_number_of_arguments(node, 1)
self.annotate_expected(node.args[0], types.VYPER_BOOL)
return [None], [node]
elif case(names.SEND):
self.check_number_of_arguments(node, 2)
self.annotate_expected(node.args[0], types.VYPER_ADDRESS)
self.annotate_expected(node.args[1], types.VYPER_WEI_VALUE)
return [None], [node]
elif case(names.SELFDESTRUCT):
self.check_number_of_arguments(node, 0, 1)
if node.args:
# The selfdestruct Vyper function
self.annotate_expected(node.args[0], types.VYPER_ADDRESS)
return [None], [node]
else:
# The selfdestruct verification function
return [types.VYPER_BOOL], [node]
elif case(names.CLEAR):
# Not type checking is needed here as clear may only occur in actual Vyper code
self.annotate(node.args[0])
return [None], [node]
elif case(names.RAW_CALL):
# No type checking is needed here as raw_call may only occur in actual Vyper code
for arg in node.args:
self.annotate(arg)
for kwarg in node.keywords:
self.annotate(kwarg.value)
idx = first_index(lambda n: n.name == names.RAW_CALL_OUTSIZE, node.keywords)
outsize_arg = node.keywords[idx].value
assert isinstance(outsize_arg, ast.Num)
size = outsize_arg.n
return [ArrayType(types.VYPER_BYTE, size, False)], [node]
elif case(names.RAW_LOG):
self.check_number_of_arguments(node, 2)
self.annotate_expected(node.args[0], ArrayType(types.VYPER_BYTES32, 4, False))
self.annotate_expected(node.args[1], types.is_bytes_array)
return [None], [node]
elif case(names.CREATE_FORWARDER_TO):
self.check_number_of_arguments(node, 1, allowed_keywords=[names.CREATE_FORWARDER_TO_VALUE])
self.annotate_expected(node.args[0], types.VYPER_ADDRESS)
if node.keywords:
self.annotate_expected(node.keywords[0].value, types.VYPER_WEI_VALUE)
return [types.VYPER_ADDRESS], [node]
elif case(names.AS_WEI_VALUE):
self.check_number_of_arguments(node, 2)
self.annotate_expected(node.args[0], types.is_integer)
unit = node.args[1]
_check(isinstance(unit, ast.Str) and
# Check that the unit exists
next((v for k, v in names.ETHER_UNITS.items() if unit.s in k), False),
node, 'invalid.unit')
return [types.VYPER_WEI_VALUE], [node]
elif case(names.AS_UNITLESS_NUMBER):
self.check_number_of_arguments(node, 1)
# We ignore units completely, therefore the type stays the same
return self.pass_through(node.args[0], node)
elif case(names.EXTRACT32):
self.check_number_of_arguments(node, 2, allowed_keywords=[names.EXTRACT32_TYPE])
self.annotate_expected(node.args[0], types.is_bytes_array)
self.annotate_expected(node.args[1], types.VYPER_INT128)
return_type = self.type_builder.build(node.keywords[0].value) if node.keywords else types.VYPER_BYTES32
return [return_type], [node]
elif case(names.CONCAT):
_check(bool(node.args), node, 'invalid.no.args')
for arg in node.args:
self.annotate_expected(arg, types.is_bytes_array)
size = sum(arg.type.size for arg in node.args)
return [ArrayType(node.args[0].type.element_type, size, False)], [node]
elif case(names.KECCAK256) or case(names.SHA256):
self.check_number_of_arguments(node, 1)
self.annotate_expected(node.args[0], types.is_bytes_array)
return [types.VYPER_BYTES32], [node]
elif case(names.BLOCKHASH):
self.check_number_of_arguments(node, 1)
self.annotate_expected(node.args[0], types.VYPER_UINT256)
return [types.VYPER_BYTES32], [node]
elif case(names.METHOD_ID):
if is_compatible_version('>=0.1.0-beta.16 <0.1.0'):
self.check_number_of_arguments(node, 2)
self.annotate_expected(node.args[0], lambda t: isinstance(t, StringType))
ntype = self.type_builder.build(node.args[1])
elif is_compatible_version('^0.2.0'):
self.check_number_of_arguments(node, 1, allowed_keywords=[names.METHOD_ID_OUTPUT_TYPE])
self.annotate_expected(node.args[0], lambda t: isinstance(t, StringType))
ntype = types.ArrayType(types.VYPER_BYTE, 4)
for keyword in node.keywords:
if keyword.name == names.METHOD_ID_OUTPUT_TYPE:
ntype = self.type_builder.build(keyword.value)
else:
assert False
is_bytes32 = ntype == types.VYPER_BYTES32
is_bytes4 = (isinstance(ntype, ArrayType) and ntype.element_type == types.VYPER_BYTE
and ntype.size == 4)
_check(is_bytes32 or is_bytes4, node, 'invalid.method_id')
return [ntype], [node]
elif case(names.ECRECOVER):
self.check_number_of_arguments(node, 4)
self.annotate_expected(node.args[0], types.VYPER_BYTES32)
self.annotate_expected(node.args[1], types.VYPER_UINT256)
self.annotate_expected(node.args[2], types.VYPER_UINT256)
self.annotate_expected(node.args[3], types.VYPER_UINT256)
return [types.VYPER_ADDRESS], [node]
elif case(names.ECADD) or case(names.ECMUL):
self.check_number_of_arguments(node, 2)
int_pair = ArrayType(types.VYPER_UINT256, 2)
self.annotate_expected(node.args[0], int_pair)
arg_type = int_pair if case(names.ECADD) else types.VYPER_UINT256
self.annotate_expected(node.args[1], arg_type)
return [int_pair], [node]
elif case(names.EMPTY):
self.check_number_of_arguments(node, 1)
ntype = self.type_builder.build(node.args[0])
return [ntype], [node]
elif case(names.IMPLIES):
self.check_number_of_arguments(node, 2)
self.annotate_expected(node.args[0], types.VYPER_BOOL)
self.annotate_expected(node.args[1], types.VYPER_BOOL)
return [types.VYPER_BOOL], [node]
elif case(names.RESULT):
self.check_number_of_arguments(node, 0, 1, allowed_keywords=[names.RESULT_DEFAULT])
if node.args:
argument = node.args[0]
_check(isinstance(argument, ast.ReceiverCall), argument, 'spec.result',
'Only functions defined in this contract can be called from the specification.')
assert isinstance(argument, ast.ReceiverCall)
self.annotate(argument)
_check(isinstance(argument.receiver.type, types.SelfType), argument, 'spec.result',
'Only functions defined in this contract can be called from the specification.')
func = self.program.functions[argument.name]
for kw in node.keywords:
self.annotate_expected(kw.value, func.type.return_type)
return [func.type.return_type], [node]
elif len(node.keywords) > 0:
_check(False, node.keywords[0].value, 'spec.result',
'The default keyword argument can only be used in combination with a pure function.')
return [self.current_func.type.return_type], [node]
elif case(names.SUM):
self.check_number_of_arguments(node, 1)
def is_numeric_map_or_array(t):
return isinstance(t, types.MapType) and types.is_numeric(t.value_type) \
or isinstance(t, types.ArrayType) and types.is_numeric(t.element_type)
self.annotate_expected(node.args[0], is_numeric_map_or_array)
if isinstance(node.args[0].type, types.MapType):
vyper_type = node.args[0].type.value_type
elif isinstance(node.args[0].type, types.ArrayType):
vyper_type = node.args[0].type.element_type
else:
assert False
return [vyper_type], [node]
elif case(names.TUPLE):
return self._visit_tuple(node.args), [node]
elif case(names.SENT) or case(names.RECEIVED) or case(names.ALLOCATED):
self.check_number_of_arguments(node, 0, 1, resources=1 if case(names.ALLOCATED) else 0)
if not node.args:
map_type = types.MapType(types.VYPER_ADDRESS, types.NON_NEGATIVE_INT)
return [map_type], [node]
else:
self.annotate_expected(node.args[0], types.VYPER_ADDRESS)
return [types.NON_NEGATIVE_INT], [node]
elif case(names.LOCKED):
self.check_number_of_arguments(node, 1)
lock = node.args[0]
_check(isinstance(lock, ast.Str) and lock.s in self.program.nonreentrant_keys(), node, 'invalid.lock')
return [types.VYPER_BOOL], [node]
elif case(names.OVERFLOW) or case(names.OUT_OF_GAS):
self.check_number_of_arguments(node, 0)
return [types.VYPER_BOOL], [node]
elif case(names.FAILED):
self.check_number_of_arguments(node, 1)
self.annotate_expected(node.args[0], types.VYPER_ADDRESS)
return [types.VYPER_BOOL], [node]
elif case(names.IMPLEMENTS):
self.check_number_of_arguments(node, 2)
address = node.args[0]
interface = node.args[1]
is_interface = (isinstance(interface, ast.Name)
and (interface.id in self.program.interfaces
or isinstance(self.program, VyperInterface) and interface.id == self.program.name))
_check(is_interface, node, 'invalid.interface')
self.annotate_expected(address, types.VYPER_ADDRESS)
return [types.VYPER_BOOL], [node]
elif case(names.INTERPRETED):
self.check_number_of_arguments(node, 1)
self.annotate_expected(node.args[0], types.VYPER_BOOL)
return [types.VYPER_BOOL], [node]
elif case(names.CONDITIONAL):
self.check_number_of_arguments(node, 2)
self.annotate_expected(node.args[0], types.VYPER_BOOL)
return self.pass_through(node.args[1], node)
elif case(names.REALLOCATE):
keywords = [names.REALLOCATE_TO, names.REALLOCATE_ACTOR]
required = [names.REALLOCATE_TO]
self.check_number_of_arguments(node, 1, allowed_keywords=keywords,
required_keywords=required, resources=1)
self.annotate_expected(node.args[0], types.VYPER_WEI_VALUE)
self.annotate_expected(node.keywords[0].value, types.VYPER_ADDRESS)
for kw in node.keywords:
self.annotate_expected(kw.value, types.VYPER_ADDRESS)
return [None], [node]
elif case(names.FOREACH):
return self._visit_foreach(node)
elif case(names.OFFER):
keywords = {
names.OFFER_TO: types.VYPER_ADDRESS,
names.OFFER_ACTOR: types.VYPER_ADDRESS,
names.OFFER_TIMES: types.VYPER_UINT256
}
required = [names.OFFER_TO, names.OFFER_TIMES]
self.check_number_of_arguments(node, 2, allowed_keywords=keywords.keys(),
required_keywords=required, resources=2)
self.annotate_expected(node.args[0], types.VYPER_WEI_VALUE)
self.annotate_expected(node.args[1], types.VYPER_WEI_VALUE)
for kw in node.keywords:
self.annotate_expected(kw.value, keywords[kw.name])
return [None], [node]
elif case(names.ALLOW_TO_DECOMPOSE):
self.check_number_of_arguments(node, 2, resources=1)
self.annotate_expected(node.args[0], types.VYPER_WEI_VALUE)
self.annotate_expected(node.args[1], types.VYPER_ADDRESS)
return [None], [node]
elif case(names.REVOKE):
keywords = {
names.REVOKE_TO: types.VYPER_ADDRESS,
names.REVOKE_ACTOR: types.VYPER_ADDRESS
}
required = [names.REVOKE_TO]
self.check_number_of_arguments(node, 2, allowed_keywords=keywords.keys(),
required_keywords=required, resources=2)
self.annotate_expected(node.args[0], types.VYPER_WEI_VALUE)
self.annotate_expected(node.args[1], types.VYPER_WEI_VALUE)
for kw in node.keywords:
self.annotate_expected(kw.value, keywords[kw.name])
return [None], [node]
elif case(names.EXCHANGE):
keywords = [names.EXCHANGE_TIMES]
self.check_number_of_arguments(node, 4, allowed_keywords=keywords,
required_keywords=keywords, resources=2)
self.annotate_expected(node.args[0], types.VYPER_WEI_VALUE)
self.annotate_expected(node.args[1], types.VYPER_WEI_VALUE)
self.annotate_expected(node.args[2], types.VYPER_ADDRESS)
self.annotate_expected(node.args[3], types.VYPER_ADDRESS)
self.annotate_expected(node.keywords[0].value, types.VYPER_UINT256)
return [None], [node]
elif case(names.CREATE):
keywords = {
names.CREATE_TO: types.VYPER_ADDRESS,
names.CREATE_ACTOR: types.VYPER_ADDRESS
}
self.check_number_of_arguments(node, 1, allowed_keywords=keywords.keys(), resources=1)
_check(node.resource is not None and node.underlying_resource is None, node, 'invalid.create',
'Only non-derived resources can be used with "create"')
self.annotate_expected(node.args[0], types.VYPER_UINT256)
for kw in node.keywords:
self.annotate_expected(kw.value, keywords[kw.name])
return [None], [node]
elif case(names.DESTROY):
keywords = [names.DESTROY_ACTOR]
self.check_number_of_arguments(node, 1, resources=1, allowed_keywords=keywords)
_check(node.resource is not None and node.underlying_resource is None, node, 'invalid.destroy',
'Only non-derived resources can be used with "destroy"')
self.annotate_expected(node.args[0], types.VYPER_UINT256)
for kw in node.keywords:
self.annotate_expected(kw.value, types.VYPER_ADDRESS)
return [None], [node]
elif case(names.RESOURCE_PAYABLE):
keywords = [names.RESOURCE_PAYABLE_ACTOR]
self.check_number_of_arguments(node, 1, resources=1, allowed_keywords=keywords)
_check(node.resource is None or node.underlying_resource is not None, node, 'invalid.payable',
'Only derived resources can be used with "payable"')
self.annotate_expected(node.args[0], types.VYPER_UINT256)
for kw in node.keywords:
self.annotate_expected(kw.value, types.VYPER_ADDRESS)
return [None], [node]
elif case(names.RESOURCE_PAYOUT):
keywords = [names.RESOURCE_PAYOUT_ACTOR]
self.check_number_of_arguments(node, 1, resources=1, allowed_keywords=keywords)
_check(node.resource is None or node.underlying_resource is not None, node, 'invalid.payout',
'Only derived resources can be used with "payout"')
self.annotate_expected(node.args[0], types.VYPER_UINT256)
for kw in node.keywords:
self.annotate_expected(kw.value, types.VYPER_ADDRESS)
return [None], [node]
elif case(names.TRUST):
keywords = [names.TRUST_ACTOR]
self.check_number_of_arguments(node, 2, allowed_keywords=keywords)
self.annotate_expected(node.args[0], types.VYPER_ADDRESS)
self.annotate_expected(node.args[1], types.VYPER_BOOL)
for kw in node.keywords:
self.annotate_expected(kw.value, types.VYPER_ADDRESS)
return [None], [node]
elif case(names.ALLOCATE_UNTRACKED):
self.check_number_of_arguments(node, 1, resources=0) # By setting resources to 0 we only allow wei.
_check(node.resource is None or node.underlying_resource is not None, node,
'invalid.allocate_untracked', 'Only derived resources can be used with "allocate_untracked"')
self.annotate_expected(node.args[0], types.VYPER_ADDRESS)
return [None], [node]
elif case(names.OFFERED):
self.check_number_of_arguments(node, 4, resources=2)
self.annotate_expected(node.args[0], types.VYPER_WEI_VALUE)
self.annotate_expected(node.args[1], types.VYPER_WEI_VALUE)
self.annotate_expected(node.args[2], types.VYPER_ADDRESS)
self.annotate_expected(node.args[3], types.VYPER_ADDRESS)
return [types.NON_NEGATIVE_INT], [node]
elif case(names.NO_OFFERS):
self.check_number_of_arguments(node, 1, resources=1)
self.annotate_expected(node.args[0], types.VYPER_ADDRESS)
return [types.VYPER_BOOL], [node]
elif case(names.ALLOWED_TO_DECOMPOSE):
self.check_number_of_arguments(node, 1, resources=1)
self.annotate_expected(node.args[0], types.VYPER_ADDRESS)
return [types.VYPER_UINT256], [node]
elif case(names.TRUSTED):
keywords = [names.TRUSTED_BY, names.TRUSTED_WHERE]
required_keywords = [names.TRUSTED_BY]
self.check_number_of_arguments(node, 1, allowed_keywords=keywords, required_keywords=required_keywords)
self.annotate_expected(node.args[0], types.VYPER_ADDRESS)
for kw in node.keywords:
self.annotate_expected(kw.value, types.VYPER_ADDRESS)
return [types.VYPER_BOOL], [node]
elif case(names.TRUST_NO_ONE):
self.check_number_of_arguments(node, 2)
self.annotate_expected(node.args[0], types.VYPER_ADDRESS)
self.annotate_expected(node.args[1], types.VYPER_ADDRESS)
return [types.VYPER_BOOL], [node]
elif len(node.args) == 1 and isinstance(node.args[0], ast.Dict):
# This is a struct initializer
self.check_number_of_arguments(node, 1)
return self._visit_struct_init(node)
elif name in self.program.contracts:
# This is a contract initializer
self.check_number_of_arguments(node, 1)
self.annotate_expected(node.args[0], types.VYPER_ADDRESS)
return [self.program.contracts[name].type], [node]
elif name in self.program.interfaces:
# This is an interface initializer
self.check_number_of_arguments(node, 1)
self.annotate_expected(node.args[0], types.VYPER_ADDRESS)
return [self.program.interfaces[name].type], [node]
elif name in self.program.ghost_functions:
possible_functions = self.program.ghost_functions[name]
if len(possible_functions) != 1:
if isinstance(self.program, VyperInterface):
function = [fun for fun in possible_functions if fun.file == self.program.file][0]
else:
implemented_interfaces = [self.program.interfaces[i.name].file for i in self.program.implements]
function = [fun for fun in possible_functions if fun.file in implemented_interfaces][0]
else:
function = possible_functions[0]
self.check_number_of_arguments(node, len(function.args) + 1)
arg_types = [types.VYPER_ADDRESS, *[arg.type for arg in function.args.values()]]
for arg_type, arg in zip(arg_types, node.args):
self.annotate_expected(arg, arg_type)
return [function.type.return_type], [node]
else:
raise UnsupportedException(node, "Unsupported function call")
def visit_ReceiverCall(self, node: ast.ReceiverCall):
receiver = node.receiver
if isinstance(receiver, ast.Name):
# A lemma call
if receiver.id == names.LEMMA:
self.annotate(receiver)
_check(not isinstance(receiver.type, (ContractType, InterfaceType)),
receiver, 'invalid.lemma.receiver',
'A receiver, with name "lemma" and with a contract- or interface-type, is not supported.')
lemma = self.program.lemmas[node.name]
self.check_number_of_arguments(node, len(lemma.args))
for arg, func_arg in zip(node.args, lemma.args.values()):
self.annotate_expected(arg, func_arg.type)
return [types.VYPER_BOOL], [node]
# A receiver ghost function call
elif receiver.id in self.program.interfaces.keys():
self.annotate(receiver)
function = self.program.interfaces[receiver.id].own_ghost_functions[node.name]
self.check_number_of_arguments(node, len(function.args) + 1)
arg_types = [types.VYPER_ADDRESS, *[arg.type for arg in function.args.values()]]
for arg_type, arg in zip(arg_types, node.args):
self.annotate_expected(arg, arg_type)
return [function.type.return_type], [node]
def expected(t):
is_self_call = isinstance(t, SelfType)
is_external_call = isinstance(t, (ContractType, InterfaceType))
is_log = t is None
return is_self_call or is_external_call or is_log
self.annotate_expected(receiver, expected)
receiver_type = receiver.type
# A self call
if isinstance(receiver_type, SelfType):
function = self.program.functions[node.name]
num_args = len(function.args)
num_defaults = len(function.defaults)
self.check_number_of_arguments(node, *range(num_args - num_defaults, num_args + 1))
for arg, func_arg in zip(node.args, function.args.values()):
self.annotate_expected(arg, func_arg.type)
return [function.type.return_type], [node]
else:
for arg in node.args:
self.annotate(arg)
for kw in node.keywords:
self.annotate(kw.value)
# A logging call
if not receiver_type:
return [None], [node]
# A contract call
elif isinstance(receiver_type, ContractType):
return [receiver_type.function_types[node.name].return_type], [node]
elif isinstance(receiver_type, InterfaceType):
interface = self.program.interfaces[receiver_type.name]
function = interface.functions[node.name]
return [function.type.return_type], [node]
else:
assert False
def _visit_resource_address(self, node: ast.Node, resource_name: str, interface: Optional[VyperInterface] = None):
self.annotate_expected(node, types.VYPER_ADDRESS)
_check(resource_name != names.UNDERLYING_WEI, node, 'invalid.resource.address',
'The underlying wei resource cannot have an address.')
is_wei = resource_name == names.WEI
ref_interface = None
if isinstance(node, ast.Name):
ref_interface = self.program
elif isinstance(node, ast.Attribute):
for field, field_type in self.program.fields.type.member_types.items():
if node.attr == field:
_check(isinstance(field_type, types.InterfaceType), node, 'invalid.resource.address')
assert isinstance(field_type, types.InterfaceType)
ref_interface = self.program.interfaces[field_type.name]
_check(is_wei or resource_name in ref_interface.own_resources, node, 'invalid.resource.address')
break
else:
_check(False, node, 'invalid.resource.address')
elif isinstance(node, ast.FunctionCall):
if node.name in self.program.ghost_functions:
if isinstance(self.program, VyperInterface):
function = self.program.ghost_functions[node.name][0]
else:
function = self.program.ghost_function_implementations.get(node.name)
if function is None:
function = self.program.ghost_functions[node.name][0]
_check(isinstance(function.type.return_type, types.InterfaceType), node, 'invalid.resource.address')
assert isinstance(function.type.return_type, types.InterfaceType)
program = first(interface for interface in self.program.interfaces.values()
if interface.file == function.file) or self.program
ref_interface = program.interfaces[function.type.return_type.name]
else:
self.check_number_of_arguments(node, 1)
ref_interface = self.program.interfaces[node.name]
_check(is_wei or resource_name in ref_interface.own_resources, node, 'invalid.resource.address')
elif isinstance(node, ast.ReceiverCall):
assert node.name in self.program.ghost_functions
assert isinstance(node.receiver, ast.Name)
interface_name = node.receiver.id
function = self.program.interfaces[interface_name].own_ghost_functions.get(node.name)
_check(isinstance(function.type.return_type, types.InterfaceType), node, 'invalid.resource.address')
assert isinstance(function.type.return_type, types.InterfaceType)
program = first(interface for interface in self.program.interfaces.values()
if interface.file == function.file) or self.program
ref_interface = program.interfaces[function.type.return_type.name]
_check(is_wei or resource_name in ref_interface.own_resources, node, 'invalid.resource.address')
else:
assert False
if not is_wei and ref_interface is not None:
if interface is not None:
if isinstance(ref_interface, VyperInterface):
_check(ref_interface.file == interface.file, node, 'invalid.resource.address')
else:
interface_names = [t.name for t in ref_interface.implements]
interfaces = [ref_interface.interfaces[name] for name in interface_names]
files = [ref_interface.file] + [i.file for i in interfaces]
_check(interface.file in files, node, 'invalid.resource.address')
elif isinstance(ref_interface, VyperInterface):
self_is_interface = isinstance(self.program, VyperInterface)
self_does_not_have_resource = resource_name not in self.program.own_resources
_check(self_is_interface or self_does_not_have_resource, node, 'invalid.resource.address')
def _annotate_resource(self, node: ast.Node, top: bool = False):
if isinstance(node, ast.Name):
resources = self.program.resources.get(node.id)
if resources is not None and len(resources) == 1:
resource = resources[0]
if self.program.config.has_option(names.CONFIG_NO_DERIVED_WEI):
_check(resource.name != names.WEI, node, 'invalid.resource')
if (top and self.program.file != resource.file
and resource.name != names.WEI
and resource.name != names.UNDERLYING_WEI):
interface = first(i for i in self.program.interfaces.values() if i.file == resource.file)
_check(any(i.name == interface.name for i in self.program.implements), node, 'invalid.resource')
_check(node.id in interface.declared_resources, node, 'invalid.resource')
else:
resource = self.program.declared_resources.get(node.id)
args = []
elif isinstance(node, ast.FunctionCall) and node.name == names.CREATOR:
resource, args = self._annotate_resource(node.args[0])
node.type = node.args[0].type
elif isinstance(node, ast.FunctionCall):
resources = self.program.resources.get(node.name)
if resources is not None and len(resources) == 1:
resource = resources[0]
else:
resource = self.program.declared_resources.get(node.name)
if self.program.config.has_option(names.CONFIG_NO_DERIVED_WEI):
_check(resource.name != names.WEI, node, 'invalid.resource')
if node.resource is not None:
assert isinstance(node.resource, ast.Expr)
self._visit_resource_address(node.resource, node.name)
resource.own_address = node.resource
elif (top and self.program.file != resource.file
and resource.name != names.WEI
and resource.name != names.UNDERLYING_WEI):
interface = first(i for i in self.program.interfaces.values() if i.file == resource.file)
_check(any(i.name == interface.name for i in self.program.implements), node, 'invalid.resource')
_check(node.name in interface.declared_resources, node, 'invalid.resource')
args = node.args
elif isinstance(node, ast.Attribute):
assert isinstance(node.value, ast.Name)
interface = self.program.interfaces[node.value.id]
if top:
_check(any(i.name == interface.name for i in self.program.implements), node, 'invalid.resource')
_check(node.attr in interface.declared_resources, node, 'invalid.resource')
resource = interface.declared_resources.get(node.attr)
args = []
elif isinstance(node, ast.ReceiverCall):
address = None
if isinstance(node.receiver, ast.Name):
interface = self.program.interfaces[node.receiver.id]
if top:
_check(any(i.name == interface.name for i in self.program.implements), node, 'invalid.resource')
_check(node.name in interface.declared_resources, node, 'invalid.resource')
elif isinstance(node.receiver, ast.Subscript):
assert isinstance(node.receiver.value, ast.Attribute)
assert isinstance(node.receiver.value.value, ast.Name)
interface_name = node.receiver.value.value.id
interface = self.program.interfaces[interface_name]
if top:
_check(node.name in interface.declared_resources, node, 'invalid.resource')
address = node.receiver.index
self._visit_resource_address(address, node.name, interface)
else:
assert False
resource = interface.declared_resources.get(node.name)
resource.own_address = address
args = node.args
elif isinstance(node, ast.Subscript):
address = node.index
resource, args = self._annotate_resource(node.value)
node.type = node.value.type
interface = None
if isinstance(node.value, ast.Name):
resource_name = node.value.id
elif isinstance(node.value, ast.Attribute):
assert isinstance(node.value.value, ast.Name)
interface = self.program.interfaces[node.value.value.id]
resource_name = node.value.attr
_check(resource_name in interface.declared_resources, node, 'invalid.resource')
else:
assert False
self._visit_resource_address(address, resource_name, interface)
resource.own_address = address
else:
assert False
if not resource:
raise InvalidProgramException(node, 'invalid.resource',
f'Could not find a resource for node:\n{pprint(node, True)}')
for arg_type, arg in zip(resource.type.member_types.values(), args):
self.annotate_expected(arg, arg_type)
node.type = resource.type
return resource, args
def _visit_resource(self, node: ast.Node, top_node: Optional[ast.FunctionCall] = None):
if isinstance(node, ast.Exchange):
self._visit_resource(node.left, top_node)
self._visit_resource(node.right, top_node)
return
resource, args = self._annotate_resource(node, True)
if len(args) != len(resource.type.member_types):
raise InvalidProgramException(node, 'invalid.resource')
if top_node is not None and top_node.underlying_resource is None:
top_node.underlying_resource = resource.underlying_resource
def _add_quantified_vars(self, var_decls: ast.Dict):
vars_types = zip(var_decls.keys, var_decls.values)
for name, type_annotation in vars_types:
var_type = self.type_builder.build(type_annotation)
self.variables[name.id] = [var_type]
name.type = var_type
def _visit_forall(self, node: ast.FunctionCall):
with self._quantified_vars_scope():
# Add the quantified variables {<x1>: <t1>, <x2>: <t2>, ...} to the
# variable map
first_arg = node.args[0]
assert isinstance(first_arg, ast.Dict)
self._add_quantified_vars(first_arg)
# Triggers can have any type
for arg in node.args[1:-1]:
self.annotate(arg)
# The quantified expression is always a Boolean
self.annotate_expected(node.args[-1], types.VYPER_BOOL)
return [types.VYPER_BOOL], [node]
def _visit_foreach(self, node: ast.FunctionCall):
with self._quantified_vars_scope():
# Add the quantified variables {<x1>: <t1>, <x2>: <t2>, ...} to the
# variable map
first_arg = node.args[0]
assert isinstance(first_arg, ast.Dict)
self._add_quantified_vars(first_arg)
# Triggers can have any type
for arg in node.args[1:-1]:
self.annotate(arg)
# Annotate the body
self.annotate(node.args[-1])
return [None], [node]
def _visit_accessible(self, node: ast.FunctionCall):
self.annotate_expected(node.args[0], types.VYPER_ADDRESS)
self.annotate_expected(node.args[1], types.VYPER_WEI_VALUE)
if len(node.args) == 3:
function_call = node.args[2]
assert isinstance(function_call, ast.ReceiverCall)
function = self.program.functions[function_call.name]
for arg, arg_type in zip(function_call.args, function.type.arg_types):
self.annotate_expected(arg, arg_type)
return [types.VYPER_BOOL], [node]
def _visit_event(self, node: ast.FunctionCall):
event = node.args[0]
assert isinstance(event, ast.FunctionCall)
event_name = event.name
event_type = self.program.events[event_name].type
for arg, arg_type in zip(event.args, event_type.arg_types):
self.annotate_expected(arg, arg_type)
if len(node.args) == 2:
self.annotate_expected(node.args[1], types.VYPER_UINT256)
return [types.VYPER_BOOL], [node]
def _visit_struct_init(self, node: ast.FunctionCall):
ntype = self.program.structs[node.name].type
struct_dict = node.args[0]
assert isinstance(struct_dict, ast.Dict)
for key, value in zip(struct_dict.keys, struct_dict.values):
value_type = ntype.member_types[key.id]
self.annotate_expected(value, value_type)
return [ntype], [node]
def _visit_range(self, node: ast.FunctionCall):
_check(len(node.args) > 0, node, 'invalid.range')
first_arg = node.args[0]
self.annotate_expected(first_arg, types.is_integer)
if len(node.args) == 1:
# A range expression of the form 'range(n)' where 'n' is a constant
assert isinstance(first_arg, ast.Num)
size = first_arg.n
elif len(node.args) == 2:
second_arg = node.args[1]
self.annotate_expected(second_arg, types.is_integer)
if isinstance(second_arg, ast.ArithmeticOp) \
and second_arg.op == ast.ArithmeticOperator.ADD \
and ast.compare_nodes(first_arg, second_arg.left):
assert isinstance(second_arg.right, ast.Num)
# A range expression of the form 'range(x, x + n)' where 'n' is a constant
size = second_arg.right.n
else:
assert isinstance(first_arg, ast.Num) and isinstance(second_arg, ast.Num)
# A range expression of the form 'range(n1, n2)' where 'n1' and 'n2 are constants
size = second_arg.n - first_arg.n
else:
raise InvalidProgramException(node, 'invalid.range')
_check(size > 0, node, 'invalid.range', 'The size of a range(...) must be greater than zero.')
return [ArrayType(types.VYPER_INT128, size, True)], [node]
def _retrieve_loop(self, node, name):
loop_var_name = node.args[0].id
loop = self.current_loops.get(loop_var_name)
_check(loop is not None, node, f"invalid.{name}")
return loop
@staticmethod
def visit_Bytes(node: ast.Bytes):
# Bytes could either be non-strict or (if it has length 32) strict
non_strict = types.ArrayType(types.VYPER_BYTE, len(node.s), False)
if len(node.s) == 32:
return [non_strict, types.VYPER_BYTES32], [node]
else:
return [non_strict], [node]
@staticmethod
def visit_Str(node: ast.Str):
string_bytes = bytes(node.s, 'utf-8')
ntype = StringType(len(string_bytes))
return [ntype], [node]
# Sets occur in trigger expressions
def visit_Set(self, node: ast.Set):
for elem in node.elements:
# Triggers can have any type
self.annotate(elem)
return [None], [node]
@staticmethod
def visit_Num(node: ast.Num):
if isinstance(node.n, int):
max_int128 = 2 ** 127 - 1
max_address = 2 ** 160 - 1
if 0 <= node.n <= max_int128:
tps = [types.VYPER_INT128,
types.VYPER_UINT256,
types.VYPER_ADDRESS,
types.VYPER_BYTES32]
elif max_int128 < node.n <= max_address:
tps = [types.VYPER_UINT256,
types.VYPER_ADDRESS,
types.VYPER_BYTES32]
elif max_address < node.n:
tps = [types.VYPER_UINT256,
types.VYPER_BYTES32]
else:
tps = [types.VYPER_INT128]
nodes = [node]
return tps, nodes
elif isinstance(node.n, Decimal):
return [types.VYPER_DECIMAL], [node]
else:
assert False
@staticmethod
def visit_Bool(node: ast.Bool):
return [types.VYPER_BOOL], [node]
def visit_Attribute(self, node: ast.Attribute):
self.annotate_expected(node.value, lambda t: isinstance(t, StructType), types.VYPER_ADDRESS)
struct_type = node.value.type if isinstance(node.value.type, StructType) else types.AddressType()
ntype = struct_type.member_types.get(node.attr)
if not ntype:
raise InvalidProgramException(node, 'invalid.storage.var')
return [ntype], [node]
def visit_Subscript(self, node: ast.Subscript):
self.annotate_expected(node.value, lambda t: isinstance(t, (ArrayType, MapType)))
if isinstance(node.value.type, MapType):
key_type = node.value.type.key_type
self.annotate_expected(node.index, key_type)
ntype = node.value.type.value_type
elif isinstance(node.value.type, ArrayType):
self.annotate_expected(node.index, types.is_integer)
ntype = node.value.type.element_type
else:
assert False
return [ntype], [node]
def visit_Name(self, node: ast.Name):
_check(node.id in self.variables, node, 'invalid.local.var')
return self.variables[node.id], [node]
def visit_List(self, node: ast.List):
# Vyper guarantees that size > 0
size = len(node.elements)
for e in node.elements:
self.annotate(e)
element_types = [e.type for e in node.elements]
possible_types = []
for element_type in element_types:
if isinstance(element_type, list):
if possible_types:
possible_types = [t for t in possible_types if t in element_type]
else:
possible_types = element_type
else:
_check(not possible_types or element_type in possible_types, node, 'invalid.element.types')
possible_types = [element_type]
break
return [types.ArrayType(element_type, size) for element_type in possible_types], [node]
def visit_Tuple(self, node: ast.Tuple):
return self._visit_tuple(node.elements), [node]
def _visit_tuple(self, elements):
length = len(elements)
for e in elements:
self.annotate(e)
element_types = [e.type if isinstance(e.type, list) else [e.type] for e in elements]
sizes = [len(element_type) for element_type in element_types]
index = [0] * length
possible_types = []
while True:
possible_type = TupleType([element_types[i][index[i]] for i in range(length)])
possible_types.append(possible_type)
for i in range(length):
index[i] = (index[i] + 1) % sizes[i]
if index[i]:
break
else:
break
return possible_types
class TypeResolver(NodeVisitor):
def resolve_type(self, node: ast.Node):
self.visit(node, None)
def visit(self, node, *args):
assert len(args) == 1
if hasattr(node, "type"):
if isinstance(node.type, list):
node.type = args[0] or node.type[0]
return super().visit(node, *args)
@staticmethod
def first_intersect(left, right):
if not isinstance(left, list):
left = [left]
if not isinstance(right, list):
right = [right]
for t in left:
if t in right:
return t
def visit_Assign(self, node: ast.Assign, _: Optional[VyperType]):
self.visit(node.target, self.first_intersect(node.target.type, node.value.type))
return self.visit(node.value, node.target.type)
def visit_AugAssign(self, node: ast.AugAssign, _: Optional[VyperType]):
self.visit(node.target, self.first_intersect(node.target.type, node.value.type))
return self.visit(node.value, node.target.type)
def visit_Comparison(self, node: ast.Comparison, _: Optional[VyperType]):
self.visit(node.left, self.first_intersect(node.left.type, node.right.type))
return self.visit(node.right, node.left.type)
def visit_Equality(self, node: ast.Equality, _: Optional[VyperType]):
self.visit(node.left, self.first_intersect(node.left.type, node.right.type))
return self.visit(node.right, node.left.type)
def visit_FunctionCall(self, node: ast.FunctionCall, _: Optional[VyperType]):
self.generic_visit(node, None)
def visit_ReceiverCall(self, node: ast.ReceiverCall, _: Optional[VyperType]):
return self.generic_visit(node, None)
def visit_Set(self, node: ast.Set, _: Optional[VyperType]):
return self.generic_visit(node, None)
def visit_List(self, node: ast.List, _: Optional[VyperType]):
return self.generic_visit(node, node.type.element_type) | 2vyper | /2vyper-0.3.0.tar.gz/2vyper-0.3.0/src/twovyper/analysis/type_annotator.py | type_annotator.py |
from contextlib import contextmanager
from enum import Enum
from itertools import chain
from typing import Optional, Union
from twovyper.utils import switch, first
from twovyper.ast import ast_nodes as ast, names
from twovyper.ast.nodes import VyperProgram, VyperFunction, VyperInterface
from twovyper.ast.visitors import NodeVisitor
from twovyper.exceptions import InvalidProgramException, UnsupportedException
def _assert(cond: bool, node: ast.Node, error_code: str, msg: Optional[str] = None):
if not cond:
raise InvalidProgramException(node, error_code, msg)
def check_structure(program: VyperProgram):
StructureChecker().check(program)
class _Context(Enum):
CODE = 'code'
INVARIANT = 'invariant'
LOOP_INVARIANT = 'loop.invariant'
CHECK = 'check'
POSTCONDITION = 'postcondition'
PRECONDITION = 'precondition'
TRANSITIVE_POSTCONDITION = 'transitive.postcondition'
CALLER_PRIVATE = 'caller.private'
GHOST_CODE = 'ghost.code'
GHOST_FUNCTION = 'ghost.function'
LEMMA = 'lemma'
GHOST_STATEMENT = 'ghost.statement'
@property
def is_specification(self):
return self not in [_Context.CODE, _Context.GHOST_FUNCTION, _Context.LEMMA]
@property
def is_postcondition(self):
return self in [_Context.POSTCONDITION, _Context.TRANSITIVE_POSTCONDITION]
class StructureChecker(NodeVisitor):
def __init__(self):
self.not_allowed = {
_Context.CODE: names.NOT_ALLOWED_BUT_IN_LOOP_INVARIANTS,
_Context.INVARIANT: names.NOT_ALLOWED_IN_INVARIANT,
_Context.LOOP_INVARIANT: names.NOT_ALLOWED_IN_LOOP_INVARIANT,
_Context.CHECK: names.NOT_ALLOWED_IN_CHECK,
_Context.POSTCONDITION: names.NOT_ALLOWED_IN_POSTCONDITION,
_Context.PRECONDITION: names.NOT_ALLOWED_IN_PRECONDITION,
_Context.TRANSITIVE_POSTCONDITION: names.NOT_ALLOWED_IN_TRANSITIVE_POSTCONDITION,
_Context.CALLER_PRIVATE: names.NOT_ALLOWED_IN_CALLER_PRIVATE,
_Context.GHOST_CODE: names.NOT_ALLOWED_IN_GHOST_CODE,
_Context.GHOST_FUNCTION: names.NOT_ALLOWED_IN_GHOST_FUNCTION,
_Context.GHOST_STATEMENT: names.NOT_ALLOWED_IN_GHOST_STATEMENT,
_Context.LEMMA: names.NOT_ALLOWED_IN_LEMMAS
}
self._inside_old = False
self._is_pure = False
self._non_pure_parent_description: Union[str, None] = None
self._visited_an_event = False
self._visited_caller_spec = False
self._num_visited_conditional = 0
self._only_one_event_allowed = False
self._function_pure_checker = _FunctionPureChecker()
self._inside_performs = False
@contextmanager
def _inside_old_scope(self):
current_inside_old = self._inside_old
self._inside_old = True
yield
self._inside_old = current_inside_old
@contextmanager
def _inside_performs_scope(self):
current_inside_performs = self._inside_performs
self._inside_performs = True
yield
self._inside_performs = current_inside_performs
@contextmanager
def _inside_pure_scope(self, node_description: str = None):
is_pure = self._is_pure
self._is_pure = True
description = self._non_pure_parent_description
self._non_pure_parent_description = node_description
yield
self._is_pure = is_pure
self._non_pure_parent_description = description
@contextmanager
def _inside_one_event_scope(self):
visited_an_event = self._visited_an_event
self._visited_an_event = False
only_one_event_allowed = self._only_one_event_allowed
self._only_one_event_allowed = True
yield
self._visited_an_event = visited_an_event
self._only_one_event_allowed = only_one_event_allowed
def check(self, program: VyperProgram):
if program.resources and not program.config.has_option(names.CONFIG_ALLOCATION):
msg = "Resources require allocation config option."
raise InvalidProgramException(first(program.node.stmts) or program.node, 'alloc.not.alloc', msg)
seen_functions = set()
for implements in program.real_implements:
interface = program.interfaces.get(implements.name)
if interface is not None:
function_names = set(function.name for function in interface.functions.values())
else:
contract = program.contracts[implements.name]
function_names = set(contract.type.function_types)
_assert(not seen_functions & function_names, program.node, 'invalid.implemented.interfaces',
f'Implemented interfaces should not have a function that shares the name with another function of '
f'another implemented interface.\n'
f'(Conflicting functions: {seen_functions & function_names})')
seen_functions.update(function_names)
for function in program.functions.values():
self.visit(function.node, _Context.CODE, program, function)
if function.is_pure():
self._function_pure_checker.check_function(function, program)
for postcondition in function.postconditions:
self.visit(postcondition, _Context.POSTCONDITION, program, function)
if function.preconditions:
_assert(not function.is_public(), function.preconditions[0],
'invalid.preconditions', 'Public functions are not allowed to have preconditions.')
for precondition in function.preconditions:
self.visit(precondition, _Context.PRECONDITION, program, function)
if function.checks:
_assert(not program.is_interface(), function.checks[0],
'invalid.checks', 'No checks are allowed in interfaces.')
for check in function.checks:
self.visit(check, _Context.CHECK, program, function)
if function.performs:
_assert(function.name != names.INIT, function.performs[0],
'invalid.performs', '__init__ does not require and must not have performs clauses.')
_assert(function.is_public(), function.performs[0],
'invalid.performs', 'Private functions are not allowed to have performs clauses.')
for performs in function.performs:
self._visit_performs(performs, program, function)
for lemma in program.lemmas.values():
for default_val in lemma.defaults.values():
_assert(default_val is None, lemma.node, 'invalid.lemma')
_assert(not lemma.postconditions, first(lemma.postconditions), 'invalid.lemma',
'No postconditions are allowed for lemmas.')
_assert(not lemma.checks, first(lemma.checks), 'invalid.lemma',
'No checks are allowed for lemmas.')
_assert(not lemma.performs, first(lemma.performs), 'invalid.lemma')
self.visit(lemma.node, _Context.LEMMA, program, lemma)
for stmt in lemma.node.body:
_assert(isinstance(stmt, ast.ExprStmt), stmt, 'invalid.lemma',
'All steps of the lemma should be expressions')
for precondition in lemma.preconditions:
self.visit(precondition, _Context.LEMMA, program, lemma)
for invariant in program.invariants:
self.visit(invariant, _Context.INVARIANT, program, None)
if program.general_checks:
_assert(not program.is_interface(), program.general_checks[0],
'invalid.checks', 'No checks are allowed in interfaces.')
for check in program.general_checks:
self.visit(check, _Context.CHECK, program, None)
for postcondition in program.general_postconditions:
self.visit(postcondition, _Context.POSTCONDITION, program, None)
if program.transitive_postconditions:
_assert(not program.is_interface(), program.transitive_postconditions[0],
'invalid.transitive.postconditions', 'No transitive postconditions are allowed in interfaces')
for postcondition in program.transitive_postconditions:
self.visit(postcondition, _Context.TRANSITIVE_POSTCONDITION, program, None)
for ghost_function in program.ghost_function_implementations.values():
self.visit(ghost_function.node, _Context.GHOST_FUNCTION, program, None)
if isinstance(program, VyperInterface):
for caller_private in program.caller_private:
self._visited_caller_spec = False
self._num_visited_conditional = 0
self.visit(caller_private, _Context.CALLER_PRIVATE, program, None)
_assert(self._visited_caller_spec, caller_private, 'invalid.caller.private',
'A caller private expression must contain "caller()"')
_assert(self._num_visited_conditional <= 1, caller_private, 'invalid.caller.private',
'A caller private expression can only contain at most one "conditional(...)"')
def visit(self, node: ast.Node, *args):
assert len(args) == 3
ctx, program, function = args
_assert(ctx != _Context.GHOST_CODE or isinstance(node, ast.AllowedInGhostCode), node, 'invalid.ghost.code')
super().visit(node, ctx, program, function)
def _visit_performs(self, node: ast.Expr, program: VyperProgram, function: VyperFunction):
with self._inside_performs_scope():
_assert(isinstance(node, ast.FunctionCall) and node.name in names.GHOST_STATEMENTS,
node, 'invalid.performs')
self.visit(node, _Context.CODE, program, function)
def visit_BoolOp(self, node: ast.BoolOp, *args):
with switch(node.op) as case:
if case(ast.BoolOperator.OR):
with self._inside_one_event_scope():
with self._inside_pure_scope('disjunctions'):
self.generic_visit(node, *args)
elif case(ast.BoolOperator.IMPLIES):
with self._inside_pure_scope('(e ==> A) as the left hand side'):
self.visit(node.left, *args)
self.visit(node.right, *args)
elif case(ast.BoolOperator.AND):
self.generic_visit(node, *args)
else:
assert False
def visit_Not(self, node: ast.Not, *args):
with self._inside_pure_scope('not expressions'):
self.generic_visit(node, *args)
def visit_Containment(self, node: ast.Containment, *args):
with self._inside_pure_scope(f"containment expressions like \"{node.op}\""):
self.generic_visit(node, *args)
def visit_Equality(self, node: ast.Equality, *args):
with self._inside_pure_scope('== expressions'):
self.generic_visit(node, *args)
def visit_IfExpr(self, node: ast.IfExpr, *args):
self.visit(node.body, *args)
self.visit(node.orelse, *args)
with self._inside_pure_scope('if expressions like (A1 if e else A2)'):
self.visit(node.test, *args)
def visit_Dict(self, node: ast.Dict, *args):
with self._inside_pure_scope('dicts'):
self.generic_visit(node, *args)
def visit_List(self, node: ast.List, *args):
with self._inside_pure_scope('lists'):
self.generic_visit(node, *args)
def visit_Tuple(self, node: ast.Tuple, *args):
with self._inside_pure_scope('tuples'):
self.generic_visit(node, *args)
def visit_For(self, node: ast.For, ctx: _Context, program: VyperProgram, function: Optional[VyperFunction]):
self.generic_visit(node, ctx, program, function)
if function:
for loop_inv in function.loop_invariants.get(node, []):
self.visit(loop_inv, _Context.LOOP_INVARIANT, program, function)
def visit_FunctionDef(self, node: ast.FunctionDef, ctx: _Context,
program: VyperProgram, function: Optional[VyperFunction]):
for stmt in node.body:
if ctx == _Context.GHOST_FUNCTION or ctx == _Context.LEMMA:
new_ctx = ctx
elif ctx == _Context.CODE:
new_ctx = _Context.GHOST_CODE if stmt.is_ghost_code else ctx
else:
assert False
self.visit(stmt, new_ctx, program, function)
@staticmethod
def visit_Name(node: ast.Name, ctx: _Context, program: VyperProgram, function: Optional[VyperFunction]):
assert program
if ctx == _Context.GHOST_FUNCTION and node.id in names.ENV_VARIABLES:
_assert(False, node, 'invalid.ghost.function')
elif ctx == _Context.CALLER_PRIVATE and node.id in names.ENV_VARIABLES:
_assert(False, node, 'invalid.caller.private')
elif ctx == _Context.INVARIANT:
_assert(node.id != names.MSG, node, 'invariant.msg')
_assert(node.id != names.BLOCK, node, 'invariant.block')
elif ctx == _Context.TRANSITIVE_POSTCONDITION:
_assert(node.id != names.MSG, node, 'postcondition.msg')
elif ctx == _Context.LEMMA:
_assert(function.node.is_lemma, node, 'invalid.lemma')
_assert(node.id != names.SELF, node, 'invalid.lemma', 'Self cannot be used in lemmas')
_assert(node.id != names.MSG, node, 'invalid.lemma', 'Msg cannot be used in lemmas')
_assert(node.id != names.BLOCK, node, 'invalid.lemma', 'Block cannot be used in lemmas')
_assert(node.id != names.TX, node, 'invalid.lemma', 'Tx cannot be used in lemmas')
def visit_FunctionCall(self, node: ast.FunctionCall, ctx: _Context,
program: VyperProgram, function: Optional[VyperFunction]):
_assert(node.name not in self.not_allowed[ctx], node, f'{ctx.value}.call')
if ctx == _Context.POSTCONDITION and function and function.name == names.INIT:
_assert(node.name != names.OLD, node, 'postcondition.init.old')
if node.name == names.PUBLIC_OLD:
if ctx == _Context.POSTCONDITION:
_assert(function and function.is_private(), node, 'postcondition.public.old')
elif ctx == _Context.PRECONDITION:
_assert(function and function.is_private(), node, 'precondition.public.old')
if function and node.name == names.EVENT:
if ctx == _Context.POSTCONDITION:
_assert(function.is_private(), node, 'postcondition.event')
elif ctx == _Context.PRECONDITION:
_assert(function.is_private(), node, 'precondition.event')
if node.name in program.ghost_functions:
_assert(ctx != _Context.LEMMA, node, 'invalid.lemma')
if len(program.ghost_functions[node.name]) > 1:
if isinstance(program, VyperInterface):
cond = node.name in program.own_ghost_functions
else:
cond = node.name in program.ghost_function_implementations
_assert(cond, node, 'duplicate.ghost',
f'Multiple interfaces declare the ghost function "{node.name}".')
if node.name == names.CALLER:
self._visited_caller_spec = True
elif node.name == names.CONDITIONAL:
self._num_visited_conditional += 1
if node.name == names.LOCKED:
_assert(not isinstance(program, VyperInterface), node, 'invalid.locked',
'Locked cannot be used in an interface.')
if node.name == names.SENT or node.name == names.RECEIVED:
_assert(not program.is_interface(), node, f'invalid.{node.name}',
f'"{node.name}" cannot be used in interfaces')
elif node.name == names.EVENT:
if ctx == _Context.PRECONDITION \
or ctx == _Context.POSTCONDITION \
or ctx == _Context.LOOP_INVARIANT:
_assert(not self._is_pure, node, 'spec.event',
"Events are only in checks pure expressions. "
f"They cannot be used in {self._non_pure_parent_description}.")
if self._only_one_event_allowed:
_assert(not self._visited_an_event, node, 'spec.event',
'In this context only one event expression is allowed.')
self._visited_an_event = True
elif node.name == names.IMPLIES:
with self._inside_pure_scope('implies(e, A) as the expression "e"'):
self.visit(node.args[0], ctx, program, function)
self.visit(node.args[1], ctx, program, function)
# Success is of the form success() or success(if_not=cond1 or cond2 or ...)
elif node.name == names.SUCCESS:
def check_success_args(arg: ast.Node):
if isinstance(arg, ast.Name):
_assert(arg.id in names.SUCCESS_CONDITIONS, arg, 'spec.success')
elif isinstance(arg, ast.BoolOp) and arg.op == ast.BoolOperator.OR:
check_success_args(arg.left)
check_success_args(arg.right)
else:
raise InvalidProgramException(arg, 'spec.success')
_assert(len(node.keywords) <= 1, node, 'spec.success')
if node.keywords:
_assert(node.keywords[0].name == names.SUCCESS_IF_NOT, node, 'spec.success')
check_success_args(node.keywords[0].value)
if len(node.keywords) == 0 and len(node.args) == 1:
argument = node.args[0]
_assert(isinstance(argument, ast.ReceiverCall), node, 'spec.success')
assert isinstance(argument, ast.ReceiverCall)
func = program.functions.get(argument.name)
_assert(func is not None, argument, 'spec.success',
'Only functions defined in this contract can be called from the specification.')
_assert(func.is_pure(), argument, 'spec.success',
'Only pure functions can be called from the specification.')
self.generic_visit(argument, ctx, program, function)
elif (ctx == _Context.INVARIANT
or ctx == _Context.TRANSITIVE_POSTCONDITION
or ctx == _Context.LOOP_INVARIANT
or ctx == _Context.GHOST_CODE
or ctx == _Context.GHOST_FUNCTION):
_assert(False, node, f'{ctx.value}.call')
return
# Accessible is of the form accessible(to, amount, self.some_func(args...))
elif node.name == names.ACCESSIBLE:
_assert(not program.is_interface(), node, 'invalid.accessible', 'Accessible cannot be used in interfaces')
_assert(not self._inside_old, node, 'spec.old.accessible')
_assert(len(node.args) == 2 or len(node.args) == 3, node, 'spec.accessible')
self.visit(node.args[0], ctx, program, function)
self.visit(node.args[1], ctx, program, function)
if len(node.args) == 3:
call = node.args[2]
_assert(isinstance(call, ast.ReceiverCall), node, 'spec.accessible')
assert isinstance(call, ast.ReceiverCall)
_assert(isinstance(call.receiver, ast.Name), node, 'spec.accessible')
assert isinstance(call.receiver, ast.Name)
_assert(call.receiver.id == names.SELF, node, 'spec.accessible')
_assert(call.name in program.functions, node, 'spec.accessible')
_assert(program.functions[call.name].is_public(), node, 'spec.accessible',
"Only public functions may be used in an accessible expression")
_assert(call.name != names.INIT, node, 'spec.accessible')
self.generic_visit(call, ctx, program, function)
return
elif node.name in [names.FORALL, names.FOREACH]:
_assert(len(node.args) >= 2 and not node.keywords, node, 'invalid.no.args')
first_arg = node.args[0]
_assert(isinstance(first_arg, ast.Dict), node.args[0], f'invalid.{node.name}')
assert isinstance(first_arg, ast.Dict)
for name in first_arg.keys:
_assert(isinstance(name, ast.Name), name, f'invalid.{node.name}')
for trigger in node.args[1:len(node.args) - 1]:
_assert(isinstance(trigger, ast.Set), trigger, f'invalid.{node.name}')
self.visit(trigger, ctx, program, function)
body = node.args[-1]
if node.name == names.FOREACH:
_assert(isinstance(body, ast.FunctionCall), body, 'invalid.foreach')
assert isinstance(body, ast.FunctionCall)
_assert(body.name in names.QUANTIFIED_GHOST_STATEMENTS, body, 'invalid.foreach')
self.visit(body, ctx, program, function)
return
elif node.name in [names.OLD, names.PUBLIC_OLD]:
with self._inside_old_scope():
self.generic_visit(node, ctx, program, function)
return
elif node.name in [names.RESOURCE_PAYABLE, names.RESOURCE_PAYOUT]:
_assert(self._inside_performs, node, f'invalid.{node.name}',
f'{node.name} is only allowed in perform clauses.')
elif node.name == names.RESULT or node.name == names.REVERT:
if len(node.args) == 1:
argument = node.args[0]
_assert(isinstance(argument, ast.ReceiverCall), node, f"spec.{node.name}")
assert isinstance(argument, ast.ReceiverCall)
func = program.functions.get(argument.name)
_assert(func is not None, argument, f"spec.{node.name}",
'Only functions defined in this contract can be called from the specification.')
_assert(func.is_pure(), argument, f"spec.{node.name}",
'Only pure functions can be called from the specification.')
self.generic_visit(argument, ctx, program, function)
if node.keywords:
self.visit_nodes([kv.value for kv in node.keywords], ctx, program, function)
if node.name == names.RESULT:
_assert(func.type.return_type is not None, argument, f"spec.{node.name}",
'Only functions with a return type can be used in a result-expression.')
return
elif (ctx == _Context.CHECK
or ctx == _Context.INVARIANT
or ctx == _Context.TRANSITIVE_POSTCONDITION
or ctx == _Context.LOOP_INVARIANT
or ctx == _Context.GHOST_CODE
or ctx == _Context.GHOST_FUNCTION):
_assert(False, node, f'{ctx.value}.call')
elif node.name == names.INDEPENDENT:
with self._inside_pure_scope('independent expressions'):
self.visit(node.args[0], ctx, program, function)
def check_allowed(arg):
if isinstance(arg, ast.FunctionCall):
is_old = len(arg.args) == 1 and arg.name in [names.OLD, names.PUBLIC_OLD]
_assert(is_old, node, 'spec.independent')
return check_allowed(arg.args[0])
if isinstance(arg, ast.Attribute):
return check_allowed(arg.value)
elif isinstance(arg, ast.Name):
allowed = [names.SELF, names.BLOCK, names.CHAIN, names.TX, *function.args]
_assert(arg.id in allowed, node, 'spec.independent')
else:
_assert(False, node, 'spec.independent')
check_allowed(node.args[1])
elif node.name == names.RAW_CALL:
if names.RAW_CALL_DELEGATE_CALL in (kw.name for kw in node.keywords):
raise UnsupportedException(node, "Delegate calls are not supported.")
elif node.name == names.PREVIOUS or node.name == names.LOOP_ARRAY or node.name == names.LOOP_ITERATION:
if len(node.args) > 0:
_assert(isinstance(node.args[0], ast.Name), node.args[0], f"invalid.{node.name}")
elif node.name == names.RANGE:
if len(node.args) == 1:
_assert(isinstance(node.args[0], ast.Num), node.args[0], 'invalid.range',
'The range operator should be of the form: range(const). '
'"const" must be a constant integer expression.')
elif len(node.args) == 2:
first_arg = node.args[0]
second_arg = node.args[1]
msg = 'The range operator should be of the form: range(const1, const2) or range(x, x + const1). '\
'"const1" and "const2" must be constant integer expressions.'
if isinstance(second_arg, ast.ArithmeticOp) \
and second_arg.op == ast.ArithmeticOperator.ADD \
and ast.compare_nodes(first_arg, second_arg.left):
_assert(isinstance(second_arg.right, ast.Num), second_arg.right, 'invalid.range', msg)
else:
_assert(isinstance(first_arg, ast.Num), first_arg, 'invalid.range', msg)
_assert(isinstance(second_arg, ast.Num), second_arg, 'invalid.range', msg)
assert isinstance(first_arg, ast.Num) and isinstance(second_arg, ast.Num)
_assert(first_arg.n < second_arg.n, node, 'invalid.range',
'The range operator should be of the form: range(const1, const2). '
'"const2" must be greater than "const1".')
if node.name in names.ALLOCATION_FUNCTIONS:
msg = "Allocation statements require allocation config option."
_assert(program.config.has_option(names.CONFIG_ALLOCATION), node, 'alloc.not.alloc', msg)
if isinstance(program, VyperInterface):
_assert(not program.config.has_option(names.CONFIG_NO_PERFORMS), node, 'alloc.not.alloc')
if self._inside_performs:
_assert(node.name not in names.ALLOCATION_SPECIFICATION_FUNCTIONS, node, 'alloc.in.alloc',
'Allocation functions are not allowed in arguments of other allocation '
'functions in a performs clause')
if node.name in names.GHOST_STATEMENTS:
msg = "Allocation statements are not allowed in constant functions."
_assert(not (function and function.is_constant()), node, 'alloc.in.constant', msg)
arg_ctx = _Context.GHOST_STATEMENT if node.name in names.GHOST_STATEMENTS and not self._inside_performs else ctx
if node.name == names.TRUST and node.keywords:
_assert(not (function and function.name != names.INIT), node, 'invalid.trusted',
f'Only the "{names.INIT}" function is allowed to have trust calls with keywords.')
if node.resource:
# Resources are only allowed in allocation functions. They can have the following structure:
# - a simple name: r
# - an exchange: r <-> s
# - a creator: creator(r)
_assert(node.name in names.ALLOCATION_FUNCTIONS, node, 'invalid.no.resources')
# All allocation functions are allowed to have a resource with an address if the function is inside of a
# performs clause. Outside of performs clauses ghost statements are not allowed to have resources with
# addresses (They can only refer to their own resources).
resources_with_address_allowed = self._inside_performs or node.name not in names.GHOST_STATEMENTS
def check_resource_address(address):
_assert(resources_with_address_allowed, address, 'invalid.resource.address')
self.generic_visit(address, ctx, program, function)
if isinstance(address, ast.Name):
_assert(address.id == names.SELF, address, 'invalid.resource.address')
elif isinstance(address, ast.Attribute):
_assert(isinstance(address.value, ast.Name), address, 'invalid.resource.address')
assert isinstance(address.value, ast.Name)
_assert(address.value.id == names.SELF, address, 'invalid.resource.address')
elif isinstance(address, ast.FunctionCall):
if address.name in program.ghost_functions:
f = program.ghost_functions.get(address.name)
_assert(f is None or len(f) == 1, address, 'invalid.resource.address',
'The ghost function is not unique.')
_assert(f is not None, address, 'invalid.resource.address')
_assert(len(address.args) >= 1, address, 'invalid.resource.address')
else:
_assert(program.config.has_option(names.CONFIG_TRUST_CASTS), address,
'invalid.resource.address', 'Using casted addresses as resource address is only allowed'
'when the "trust_casts" config is set.')
_assert(address.name in program.interfaces, address, 'invalid.resource.address')
elif isinstance(address, ast.ReceiverCall):
_assert(address.name in program.ghost_functions, address, 'invalid.resource.address')
_assert(isinstance(address.receiver, ast.Name), address, 'invalid.resource.address')
assert isinstance(address.receiver, ast.Name)
f = program.interfaces[address.receiver.id].own_ghost_functions.get(address.name)
_assert(f is not None, address, 'invalid.resource.address')
_assert(len(address.args) >= 1, address, 'invalid.resource.address')
else:
_assert(False, address, 'invalid.resource.address')
def check_resource(resource: ast.Node, top: bool, inside_subscript: bool = False):
if isinstance(resource, ast.Name):
return
elif isinstance(resource, ast.Exchange) and top and not inside_subscript:
check_resource(resource.left, False, inside_subscript)
check_resource(resource.right, False, inside_subscript)
elif isinstance(resource, ast.FunctionCall) and not inside_subscript:
if resource.name == names.CREATOR:
_assert(len(resource.args) == 1 and not resource.keywords, resource, 'invalid.resource')
check_resource(resource.args[0], False, inside_subscript)
else:
if resource.resource is not None:
check_resource_address(resource.resource)
self.generic_visit(resource, arg_ctx, program, function)
elif isinstance(resource, ast.Attribute):
_assert(isinstance(resource.value, ast.Name), resource, 'invalid.resource')
assert isinstance(resource.value, ast.Name)
interface_name = resource.value.id
interface = program.interfaces.get(interface_name)
_assert(interface is not None, resource, 'invalid.resource')
_assert(resource.attr in interface.own_resources, resource, 'invalid.resource')
elif isinstance(resource, ast.ReceiverCall) and not inside_subscript:
if isinstance(resource.receiver, ast.Name):
interface_name = resource.receiver.id
elif isinstance(resource.receiver, ast.Subscript):
_assert(isinstance(resource.receiver.value, ast.Attribute), resource, 'invalid.resource')
assert isinstance(resource.receiver.value, ast.Attribute)
_assert(isinstance(resource.receiver.value.value, ast.Name), resource, 'invalid.resource')
assert isinstance(resource.receiver.value.value, ast.Name)
interface_name = resource.receiver.value.value.id
address = resource.receiver.index
check_resource_address(address)
else:
_assert(False, resource, 'invalid.resource')
return
interface = program.interfaces.get(interface_name)
_assert(interface is not None, resource, 'invalid.resource')
_assert(resource.name in interface.own_resources, resource, 'invalid.resource')
self.generic_visit(resource, arg_ctx, program, function)
elif isinstance(resource, ast.Subscript):
address = resource.index
check_resource(resource.value, False, True)
check_resource_address(address)
else:
_assert(False, resource, 'invalid.resource')
check_resource(node.resource, True)
self.visit_nodes(chain(node.args, node.keywords), arg_ctx, program, function)
def visit_ReceiverCall(self, node: ast.ReceiverCall, ctx: _Context,
program: VyperProgram, function: Optional[VyperFunction]):
if ctx == _Context.CALLER_PRIVATE:
_assert(False, node, 'spec.call')
elif ctx.is_specification or self._inside_performs:
receiver = node.receiver
if isinstance(receiver, ast.Name):
if receiver.id == names.LEMMA:
other_lemma = program.lemmas.get(node.name)
_assert(other_lemma is not None, node, 'invalid.lemma',
f'Unknown lemma to call: {node.name}')
elif receiver.id in program.interfaces:
interface = program.interfaces[receiver.id]
_assert(node.name in interface.own_ghost_functions, node, 'spec.call',
f'Unknown ghost function to call "{node.name}" in the interface "{receiver.id}"')
else:
_assert(False, node, 'spec.call')
elif ctx == _Context.GHOST_CODE:
_assert(False, node, 'invalid.ghost.code')
else:
_assert(False, node, 'spec.call')
elif ctx == _Context.GHOST_FUNCTION:
_assert(False, node, 'invalid.ghost')
elif ctx == _Context.LEMMA:
receiver = node.receiver
_assert(isinstance(receiver, ast.Name), node, 'invalid.lemma',
'Only calls to other lemmas are allowed in lemmas.')
assert isinstance(receiver, ast.Name)
_assert(receiver.id == names.LEMMA, node, 'invalid.lemma'
'Only calls to other lemmas are allowed in lemmas.')
other_lemma = program.lemmas.get(node.name)
_assert(other_lemma is not None, node, 'invalid.lemma',
f'Unknown lemma to call: {node.name}')
_assert(other_lemma.index < function.index, node, 'invalid.lemma',
'Can only use lemmas previously defined (No recursion is allowed).')
self.generic_visit(node, ctx, program, function)
def visit_Exchange(self, node: ast.Exchange, ctx: _Context,
program: VyperProgram, function: Optional[VyperFunction]):
_assert(False, node, 'exchange.not.resource')
self.generic_visit(node, ctx, program, function)
@staticmethod
def _visit_assertion(node: Union[ast.Assert, ast.Raise], ctx: _Context):
if ctx == _Context.GHOST_CODE:
if isinstance(node, ast.Assert) and node.is_lemma:
return
_assert(node.msg and isinstance(node.msg, ast.Name), node, 'invalid.ghost.code')
assert isinstance(node.msg, ast.Name)
_assert(node.msg.id == names.UNREACHABLE, node, 'invalid.ghost.code')
def visit_Assert(self, node: ast.Assert, ctx: _Context, program: VyperProgram, function: Optional[VyperFunction]):
self._visit_assertion(node, ctx)
self.generic_visit(node, ctx, program, function)
def visit_Raise(self, node: ast.Raise, ctx: _Context, program: VyperProgram, function: Optional[VyperFunction]):
self._visit_assertion(node, ctx)
self.generic_visit(node, ctx, program, function)
class _FunctionPureChecker(NodeVisitor):
"""
Checks if a given VyperFunction is pure
"""
def __init__(self):
super().__init__()
self._ghost_allowed = False
self.max_allowed_function_index = -1
@contextmanager
def _ghost_code_allowed(self):
ghost_allowed = self._ghost_allowed
self._ghost_allowed = True
yield
self._ghost_allowed = ghost_allowed
def check_function(self, function: VyperFunction, program: VyperProgram):
# A function must be constant, private and non-payable to be valid
ghost_pure_cond = not (function.is_constant() and function.is_private() and (not function.is_payable()))
pure_decorators = [decorator for decorator in function.decorators if decorator.name == names.PURE]
if len(pure_decorators) > 1:
_assert(False, pure_decorators[0], 'invalid.pure',
'A pure function can only have exactly one pure decorator')
elif pure_decorators[0].is_ghost_code and ghost_pure_cond:
_assert(False, function.node, 'invalid.pure', 'A pure function must be constant, private and non-payable')
else:
self.max_allowed_function_index = function.index - 1
# Check checks
if function.checks:
_assert(False, function.checks[0], 'invalid.pure',
'A pure function must not have checks')
# Check performs
if function.performs:
_assert(False, function.performs[0], 'invalid.pure',
'A pure function must not have performs')
# Check preconditions
if function.preconditions:
_assert(False, function.preconditions[0], 'invalid.pure',
'A pure function must not have preconditions')
# Check postconditions
if function.postconditions:
_assert(False, function.postconditions[0], 'invalid.pure',
'A pure function must not have postconditions')
# Check loop invariants
with self._ghost_code_allowed():
for loop_invariants in function.loop_invariants.values():
self.visit_nodes(loop_invariants, program)
# Check Code
self.visit_nodes(function.node.body, program)
def visit(self, node, *args):
_assert(self._ghost_allowed or not node.is_ghost_code, node, 'invalid.pure',
'A pure function must not have ghost code statements')
return super().visit(node, *args)
def visit_Name(self, node: ast.Name, program: VyperProgram):
with switch(node.id) as case:
if case(names.MSG) \
or case(names.BLOCK) \
or case(names.TX):
_assert(False, node, 'invalid.pure', 'Pure functions are not allowed to use "msg", "block" or "tx".')
self.generic_visit(node, program)
def visit_ExprStmt(self, node: ast.ExprStmt, program: VyperProgram):
if isinstance(node.value, ast.FunctionCall) and node.value.name == names.CLEAR:
# A call to clear is an assignment
self.generic_visit(node, program)
elif isinstance(node.value, ast.Str):
# long string comment
pass
else:
# all other expressions are not valid.
_assert(False, node, 'invalid.pure', 'Pure functions are not allowed to have just an expression as a '
'statement, since expressions must not have side effects.')
def visit_FunctionCall(self, node: ast.FunctionCall, program: VyperProgram):
with switch(node.name) as case:
if (
# Vyper functions with side effects
case(names.RAW_LOG)
# Specification functions with side effects
or case(names.ACCESSIBLE)
or case(names.EVENT)
or case(names.INDEPENDENT)
or case(names.REORDER_INDEPENDENT)
):
_assert(False, node, 'invalid.pure',
f'Only functions without side effects may be used in pure functions ("{node.name}" is invalid)')
elif (
# Not supported specification functions
case(names.ALLOCATED)
or case(names.FAILED)
or case(names.IMPLEMENTS)
or case(names.ISSUED)
or case(names.LOCKED)
or case(names.OFFERED)
or case(names.OUT_OF_GAS)
or case(names.OVERFLOW)
or case(names.PUBLIC_OLD)
or case(names.RECEIVED)
or case(names.SENT)
or case(names.STORAGE)
or case(names.TRUSTED)
):
_assert(False, node, 'invalid.pure',
f'This function may not be used in pure functions ("{node.name}" is invalid)')
elif case(names.SUCCESS):
_assert(len(node.keywords) == 0, node, 'invalid.pure',
f'Only success without keywords may be used in pure functions')
self.generic_visit(node, program)
def visit_ReceiverCall(self, node: ast.ReceiverCall, program: VyperProgram):
if not isinstance(node.receiver, ast.Name):
_assert(False, node, 'invalid.receiver')
with switch(node.receiver.id) as case:
if case(names.SELF):
self.generic_visit(node, program)
_assert(program.functions[node.name].is_pure(), node, 'invalid.pure',
'Pure function may only call other pure functions')
_assert(program.functions[node.name].index <= self.max_allowed_function_index, node, 'invalid.pure',
'Only functions defined above this function can be called from here')
elif case(names.LOG):
_assert(False, node, 'invalid.pure',
'Pure function may not log events.')
else:
_assert(False, node, 'invalid.pure',
'Pure function must not call functions of another contract.')
def visit_Log(self, node: ast.Log, program: VyperProgram):
raise UnsupportedException(node, 'Pure functions that log events are not supported.') | 2vyper | /2vyper-0.3.0.tar.gz/2vyper-0.3.0/src/twovyper/analysis/structure_checker.py | structure_checker.py |
from contextlib import contextmanager
from typing import Set, Dict, List
from twovyper.ast import ast_nodes as ast, names
from twovyper.ast.nodes import VyperProgram, VyperInterface, VyperFunction
from twovyper.ast.visitors import NodeVisitor
from twovyper.analysis import heuristics
from twovyper.analysis.structure_checker import check_structure
from twovyper.analysis.symbol_checker import check_symbols
from twovyper.analysis.type_annotator import TypeAnnotator
from twovyper.exceptions import UnsupportedException
from twovyper.utils import switch
def analyze(program: VyperProgram):
"""
Checks the program for structural errors, adds type information to all program expressions
and creates an analysis for each function.
"""
if isinstance(program, VyperInterface) and program.is_stub:
return
check_symbols(program)
check_structure(program)
TypeAnnotator(program).annotate_program()
function_analyzer = _FunctionAnalyzer()
program_analyzer = _ProgramAnalyzer()
invariant_analyzer = _InvariantAnalyzer()
program.analysis = ProgramAnalysis()
for function in program.functions.values():
function.analysis = FunctionAnalysis()
function_analyzer.analyze(program, function)
# The heuristics are need for analysis, therefore do them first
heuristics.compute(program)
program_analyzer.analyze(program)
invariant_analyzer.analyze(program)
class ProgramAnalysis:
def __init__(self):
# True if and only if issued state is accessed in top-level specifications
self.uses_issued = False
# The function that is used to prove accessibility if none is given
# Is set in the heuristics computation
# May be 'None' if the heuristics is not able to determine a suitable function
self.accessible_function = None
# The invariant a tag belongs to
# Each invariant has a tag that is used in accessible so we know which invariant fails
# if we cannot prove the accessibility
self.inv_tags = {}
# Maps accessible ast.ReceiverCall nodes to their tag
self.accessible_tags = {}
# All invariants that contain allocated
self.allocated_invariants = []
class FunctionAnalysis:
def __init__(self):
# True if and only if issued state is accessed in top-level or function specifications
self.uses_issued = False
# The set of tags for which accessibility needs to be proven in the function
self.accessible_tags = set()
# The set of variable names which get changed by a loop
self.loop_used_names: Dict[str, List[str]] = {}
# True if and only if "assert e, UNREACHABLE" or "raise UNREACHABLE" is used
self.uses_unreachable = False
class _ProgramAnalyzer(NodeVisitor):
def __init__(self):
super().__init__()
def analyze(self, program: VyperProgram):
self.visit_nodes(program.invariants, program)
self.visit_nodes(program.general_postconditions, program)
self.visit_nodes(program.transitive_postconditions, program)
self.visit_nodes(program.general_checks, program)
def visit_FunctionCall(self, node: ast.FunctionCall, program: VyperProgram):
if node.name == names.ISSUED:
program.analysis.uses_issued = True
for function in program.functions.values():
function.analysis.uses_issued = True
self.generic_visit(node, program)
class _InvariantAnalyzer(NodeVisitor):
def analyze(self, program: VyperProgram):
for tag, inv in enumerate(program.invariants):
program.analysis.inv_tags[tag] = inv
self.visit(inv, program, inv, tag)
def visit_FunctionCall(self, node: ast.FunctionCall, program: VyperProgram, inv: ast.Expr, tag: int):
if node.name == names.ALLOCATED:
program.analysis.allocated_invariants.append(inv)
return
elif node.name == names.ACCESSIBLE:
program.analysis.accessible_tags[node] = tag
if len(node.args) == 3:
func_arg = node.args[2]
assert isinstance(func_arg, ast.ReceiverCall)
function_name = func_arg.name
else:
if not program.analysis.accessible_function:
msg = "No matching function for accessible could be determined."
raise UnsupportedException(node, msg)
function_name = program.analysis.accessible_function.name
program.functions[function_name].analysis.accessible_tags.add(tag)
self.generic_visit(node, program, inv, tag)
class _FunctionAnalyzer(NodeVisitor):
def __init__(self):
super().__init__()
self.inside_loop = False
self.inside_spec = False
self.used_names: Set[str] = set()
@contextmanager
def _loop_scope(self):
used_variables = self.used_names
inside_loop = self.inside_loop
self.used_names = set()
self.inside_loop = True
yield
if inside_loop:
for name in used_variables:
self.used_names.add(name)
else:
self.used_names = used_variables
self.inside_loop = inside_loop
@contextmanager
def _spec_scope(self):
inside_spec = self.inside_spec
self.inside_spec = True
yield
self.inside_spec = inside_spec
def analyze(self, program: VyperProgram, function: VyperFunction):
with self._spec_scope():
self.visit_nodes(function.postconditions, program, function)
self.visit_nodes(function.preconditions, program, function)
self.visit_nodes(function.checks, program, function)
self.generic_visit(function.node, program, function)
def visit_Assert(self, node: ast.Assert, program: VyperProgram, function: VyperFunction):
if isinstance(node.msg, ast.Name) and node.msg.id == names.UNREACHABLE:
function.analysis.uses_unreachable = True
self.generic_visit(node, program, function)
def visit_FunctionCall(self, node: ast.FunctionCall, program: VyperProgram, function: VyperFunction):
if node.name == names.ISSUED:
function.analysis.uses_issued = True
self.generic_visit(node, program, function)
def visit_For(self, node: ast.For, program: VyperProgram, function: VyperFunction):
with self._loop_scope():
self.generic_visit(node, program, function)
function.analysis.loop_used_names[node.target.id] = list(self.used_names)
with self._spec_scope():
self.visit_nodes(function.loop_invariants.get(node, []), program, function)
def visit_Name(self, node: ast.Name, program: VyperProgram, function: VyperFunction):
if self.inside_loop:
with switch(node.id) as case:
if case(names.MSG)\
or case(names.BLOCK)\
or case(names.CHAIN)\
or case(names.TX):
pass
else:
self.used_names.add(node.id)
self.generic_visit(node, program, function)
def visit_Raise(self, node: ast.Raise, program: VyperProgram, function: VyperFunction):
if isinstance(node.msg, ast.Name) and node.msg.id == names.UNREACHABLE:
function.analysis.uses_unreachable = True
self.generic_visit(node, program, function) | 2vyper | /2vyper-0.3.0.tar.gz/2vyper-0.3.0/src/twovyper/analysis/analyzer.py | analyzer.py |
from twovyper.utils import Subscriptable
def sign(a: int) -> int:
if a > 0:
return 1
elif a < 0:
return -1
else:
return 0
def div(a: int, b: int) -> int:
"""
Truncating division of two integers.
"""
return sign(a) * sign(b) * (abs(a) // abs(b))
def mod(a: int, b: int) -> int:
"""
Truncating modulo of two integers.
"""
return sign(a) * (abs(a) % abs(b))
class Decimal(object, metaclass=Subscriptable):
def __init__(self):
assert False
_cache = {}
def _subscript(number_of_digits: int):
# This is the function that gets called when using dictionary lookup syntax.
# For example, Decimal[10] returns a class of decimals with 10 digits. The class
# is cached so that Decimal[10] always returns the same class, which means that
# type(Decimal[10](1)) == type(Decimal[10](2)).
cached_class = Decimal._cache.get(number_of_digits)
if cached_class:
return cached_class
class _Decimal(Decimal):
def __init__(self, value: int = None, scaled_value: int = None):
assert (value is None) != (scaled_value is None)
self.number_of_digits = number_of_digits
self.scaling_factor = 10 ** number_of_digits
self.scaled_value = scaled_value if value is None else value * self.scaling_factor
def __eq__(self, other):
if isinstance(other, _Decimal):
return self.scaled_value == other.scaled_value
else:
return False
def __hash__(self):
return hash(self.scaled_value)
def __str__(self):
dv = div(self.scaled_value, self.scaling_factor)
md = mod(self.scaled_value, self.scaling_factor)
return f'{dv}.{str(md).zfill(self.number_of_digits)}'
Decimal._cache[number_of_digits] = _Decimal
return _Decimal | 2vyper | /2vyper-0.3.0.tar.gz/2vyper-0.3.0/src/twovyper/ast/arithmetic.py | arithmetic.py |
import os
from typing import Optional, List, Dict
from twovyper.ast import ast_nodes as ast, names
from twovyper.ast.visitors import NodeVisitor
from twovyper.exceptions import InvalidProgramException
class VyperType:
def __init__(self, id: str):
self.id = id
def __str__(self) -> str:
return self.id
def __eq__(self, other) -> bool:
if isinstance(other, VyperType):
return self.id == other.id
return NotImplemented
def __hash__(self) -> int:
return hash(self.id)
class FunctionType(VyperType):
def __init__(self, arg_types: List[VyperType], return_type: Optional[VyperType]):
self.arg_types = arg_types
self.return_type = return_type
arg_type_names = [str(arg) for arg in arg_types]
id = f'({", ".join(arg_type_names)}) -> {return_type}'
super().__init__(id)
class MapType(VyperType):
def __init__(self, key_type: VyperType, value_type: VyperType):
self.key_type = key_type
self.value_type = value_type
id = f'{names.MAP}({key_type}, {value_type})'
super().__init__(id)
class ArrayType(VyperType):
def __init__(self, element_type: VyperType, size: int, is_strict: bool = True):
self.element_type = element_type
self.size = size
self.is_strict = is_strict
id = f'{element_type}[{"" if is_strict else "<="}{size}]'
super().__init__(id)
class TupleType(VyperType):
def __init__(self, element_types: List[VyperType]):
self.element_types = element_types
id = f'({element_types})'
super().__init__(id)
class StructType(VyperType):
def __init__(self, name: str, member_types: Dict[str, VyperType]):
id = f'{self.kind} {name}'
super().__init__(id)
self.name = name
self.member_types = member_types
self.member_indices = {k: i for i, k in enumerate(member_types)}
@property
def kind(self) -> str:
return 'struct'
def add_member(self, name: str, type: VyperType):
self.member_types[name] = type
self.member_indices[name] = len(self.member_indices)
class AnyStructType(VyperType):
def __init__(self):
super().__init__('$AnyStruct')
class AnyAddressType(StructType):
def __init__(self, name: str, member_types: Dict[str, VyperType]):
assert names.ADDRESS_BALANCE not in member_types
assert names.ADDRESS_CODESIZE not in member_types
assert names.ADDRESS_IS_CONTRACT not in member_types
member_types[names.ADDRESS_BALANCE] = VYPER_WEI_VALUE
member_types[names.ADDRESS_CODESIZE] = VYPER_INT128
member_types[names.ADDRESS_IS_CONTRACT] = VYPER_BOOL
super().__init__(name, member_types)
class AddressType(AnyAddressType):
def __init__(self):
super().__init__(names.ADDRESS, {})
class SelfType(AnyAddressType):
def __init__(self, member_types: Dict[str, VyperType]):
super().__init__(names.SELF, member_types)
class ResourceType(StructType):
@property
def kind(self) -> str:
return 'resource'
class UnknownResourceType(ResourceType):
def __init__(self):
super().__init__('$unknown', {})
class DerivedResourceType(ResourceType):
def __init__(self, name: str, member_types: Dict[str, VyperType], underlying_resource: ResourceType):
super().__init__(name, member_types)
self.underlying_resource = underlying_resource
class ContractType(VyperType):
def __init__(self,
name: str,
function_types: Dict[str, FunctionType],
function_modifiers: Dict[str, str]):
id = f'contract {name}'
super().__init__(id)
self.name = name
self.function_types = function_types
self.function_modifiers = function_modifiers
class InterfaceType(VyperType):
def __init__(self, name: str):
id = f'interface {name}'
super().__init__(id)
self.name = name
class StringType(ArrayType):
def __init__(self, size: int):
super().__init__(VYPER_BYTE, size, False)
self.id = f'{names.STRING}[{size}]'
class PrimitiveType(VyperType):
def __init__(self, name: str):
super().__init__(name)
self.name = name
class BoundedType(PrimitiveType):
def __init__(self, name: str, lower: int, upper: int):
super().__init__(name)
self.lower = lower
self.upper = upper
class DecimalType(BoundedType):
def __init__(self, name: str, digits: int, lower: int, upper: int):
self.number_of_digits = digits
self.scaling_factor = 10 ** digits
lower *= self.scaling_factor
upper *= self.scaling_factor
super().__init__(name, lower, upper)
class EventType(VyperType):
def __init__(self, arg_types: List[VyperType]):
arg_type_names = [str(arg) for arg in arg_types]
id = f'event({", ".join(arg_type_names)})'
super().__init__(id)
self.arg_types = arg_types
VYPER_BOOL = PrimitiveType(names.BOOL)
VYPER_INT128 = BoundedType(names.INT128, -2 ** 127, 2 ** 127 - 1)
VYPER_UINT256 = BoundedType(names.UINT256, 0, 2 ** 256 - 1)
VYPER_DECIMAL = DecimalType(names.DECIMAL, 10, -2 ** 127, 2 ** 127 - 1)
VYPER_WEI_VALUE = VYPER_UINT256
VYPER_ADDRESS = BoundedType(names.ADDRESS, 0, 2 ** 160 - 1)
VYPER_BYTE = PrimitiveType(names.BYTE)
VYPER_BYTES32 = ArrayType(VYPER_BYTE, 32, True)
NON_NEGATIVE_INT = PrimitiveType(names.NON_NEGATIVE_INTEGER)
TYPES = {
VYPER_BOOL.name: VYPER_BOOL,
VYPER_WEI_VALUE.name: VYPER_WEI_VALUE,
VYPER_INT128.name: VYPER_INT128,
VYPER_UINT256.name: VYPER_UINT256,
VYPER_DECIMAL.name: VYPER_DECIMAL,
names.WEI_VALUE: VYPER_WEI_VALUE,
VYPER_ADDRESS.name: VYPER_ADDRESS,
VYPER_BYTE.name: VYPER_BYTE,
names.BYTES32: VYPER_BYTES32,
names.STRING: VYPER_BYTE,
names.TIMESTAMP: VYPER_UINT256,
names.TIMEDELTA: VYPER_UINT256
}
MSG_TYPE = StructType(names.MSG, {
names.MSG_SENDER: VYPER_ADDRESS,
names.MSG_VALUE: VYPER_WEI_VALUE,
names.MSG_GAS: VYPER_UINT256
})
BLOCK_TYPE = StructType(names.BLOCK, {
names.BLOCK_COINBASE: VYPER_ADDRESS,
names.BLOCK_DIFFICULTY: VYPER_UINT256,
names.BLOCK_NUMBER: VYPER_UINT256,
names.BLOCK_PREVHASH: VYPER_BYTES32,
names.BLOCK_TIMESTAMP: VYPER_UINT256
})
CHAIN_TYPE = StructType(names.CHAIN, {
names.CHAIN_ID: VYPER_UINT256
})
TX_TYPE = StructType(names.TX, {
names.TX_ORIGIN: VYPER_ADDRESS
})
def is_numeric(type: VyperType) -> bool:
return type in [VYPER_INT128, VYPER_UINT256, VYPER_DECIMAL, NON_NEGATIVE_INT]
def is_bounded(type: VyperType) -> bool:
return type in [VYPER_INT128, VYPER_UINT256, VYPER_DECIMAL, VYPER_ADDRESS]
def is_integer(type: VyperType) -> bool:
return type in [VYPER_INT128, VYPER_UINT256, NON_NEGATIVE_INT]
def is_unsigned(type: VyperType) -> bool:
return type in [VYPER_UINT256, VYPER_ADDRESS, NON_NEGATIVE_INT]
def has_strict_array_size(element_type: VyperType) -> bool:
return element_type != VYPER_BYTE
def is_bytes_array(type: VyperType):
return isinstance(type, ArrayType) and type.element_type == VYPER_BYTE
def matches(t: VyperType, m: VyperType):
"""
Determines whether a type t matches a required type m in the
specifications.
Usually the types have to be the same, except non-strict arrays
which may be shorter than the expected length, and contract types
which can be used as addresses. Also, all integer types are treated
as mathematical integers.
"""
if isinstance(t, MapType) and isinstance(m, MapType) and t.key_type == m.key_type:
return matches(t.value_type, m.value_type)
elif (isinstance(t, ArrayType) and (isinstance(m, ArrayType) and not m.is_strict)
and t.element_type == m.element_type):
return t.size <= m.size
elif is_integer(t) and is_integer(m):
return True
elif isinstance(t, ContractType) and m == VYPER_ADDRESS:
return True
elif isinstance(t, InterfaceType) and m == VYPER_ADDRESS:
return True
else:
return t == m
class TypeBuilder(NodeVisitor):
def __init__(self, type_map: Dict[str, VyperType], is_stub: bool = False):
self.type_map = type_map
self.is_stub = is_stub
def build(self, node) -> VyperType:
if self.is_stub:
return VyperType('$unknown')
return self.visit(node)
@property
def method_name(self):
return '_visit'
def generic_visit(self, node):
raise InvalidProgramException(node, 'invalid.type')
def _visit_Name(self, node: ast.Name) -> VyperType:
type = self.type_map.get(node.id) or TYPES.get(node.id)
if type is None:
raise InvalidProgramException(node, 'invalid.type')
return type
def _visit_StructDef(self, node: ast.StructDef) -> VyperType:
members = {n.target.id: self.visit(n.annotation) for n in node.body}
return StructType(node.name, members)
def _visit_EventDef(self, node: ast.EventDef) -> VyperType:
arg_types = [self.visit(n.annotation) for n in node.body]
return EventType(arg_types)
def _visit_FunctionStub(self, node: ast.FunctionStub) -> VyperType:
from twovyper.ast.nodes import Resource
name, is_derived = Resource.get_name_and_derived_flag(node)
members = {n.name: self.visit(n.annotation) for n in node.args}
contract_name = os.path.split(os.path.abspath(node.file))[1].split('.')[0]
resource_name = f'{contract_name}${name}'
if is_derived:
return DerivedResourceType(resource_name, members, UnknownResourceType())
return ResourceType(resource_name, members)
def _visit_ContractDef(self, node: ast.ContractDef) -> VyperType:
functions = {}
modifiers = {}
for f in node.body:
name = f.name
arg_types = [self.visit(arg.annotation) for arg in f.args]
return_type = None if f.returns is None else self.visit(f.returns)
functions[name] = FunctionType(arg_types, return_type)
modifiers[name] = f.body[0].value.id
return ContractType(node.name, functions, modifiers)
def _visit_FunctionCall(self, node: ast.FunctionCall) -> VyperType:
# We allow
# - public, indexed: not important for verification
# - map: map type
# - event: event type
# Not allowed is
# - constant: should already be replaced
# Anything else is treated as a unit
if node.name == names.PUBLICFIELD or node.name == names.INDEXED:
return self.visit(node.args[0])
elif node.name == names.MAP:
key_type = self.visit(node.args[0])
value_type = self.visit(node.args[1])
return MapType(key_type, value_type)
elif node.name == names.EVENT:
dict_literal = node.args[0]
arg_types = [self.visit(arg) for arg in dict_literal.values]
return EventType(arg_types)
else:
type = self.type_map.get(node.name) or TYPES.get(node.name)
if type is None:
raise InvalidProgramException(node, 'invalid.type')
return type
def _visit_Subscript(self, node: ast.Subscript) -> VyperType:
element_type = self.visit(node.value)
# Array size has to be an int or a constant
# (which has already been replaced by an int)
size = node.index.n
return ArrayType(element_type, size, has_strict_array_size(element_type))
def _visit_Tuple(self, node: ast.Tuple) -> VyperType:
element_types = [self.visit(n) for n in node.elements]
return TupleType(element_types) | 2vyper | /2vyper-0.3.0.tar.gz/2vyper-0.3.0/src/twovyper/ast/types.py | types.py |
from twovyper.vyper import select_version
# Constants for names in the original AST
# Decorators
PUBLIC = select_version({'^0.2.0': 'external', '>=0.1.0-beta.16 <0.1.0': 'public'})
PRIVATE = select_version({'^0.2.0': 'internal', '>=0.1.0-beta.16 <0.1.0': 'private'})
PAYABLE = 'payable'
CONSTANT = select_version({'^0.2.0': 'view', '>=0.1.0-beta.16 <0.1.0': 'constant'})
NONREENTRANT = 'nonreentrant'
PURE = 'pure'
INTERPRETED_DECORATOR = 'interpreted'
# Modifiers
assert CONSTANT == select_version({'^0.2.0': 'view', '>=0.1.0-beta.16 <0.1.0': 'constant'})
assert PURE == 'pure'
MODIFYING = select_version({'^0.2.0': 'payable', '>=0.1.0-beta.16 <0.1.0': 'modifying'})
NONPAYABLE = select_version({'^0.2.0': 'nonpayable'}, default="")
PUBLICFIELD = 'public'
# Types
BOOL = 'bool'
INT128 = 'int128'
UINT256 = 'uint256'
DECIMAL = 'decimal'
WEI_VALUE = 'wei_value'
TIMESTAMP = 'timestamp'
TIMEDELTA = 'timedelta'
ADDRESS = 'address'
BYTE = select_version({'^0.2.0': 'Bytes', '>=0.1.0-beta.16 <0.1.0': 'bytes'})
BYTES32 = 'bytes32'
STRING = select_version({'^0.2.0': 'String', '>=0.1.0-beta.16 <0.1.0': 'string'})
MAP = select_version({'^0.2.0': 'HashMap', '>=0.1.0-beta.16 <0.1.0': 'map'})
EVENT = 'event'
NON_NEGATIVE_INTEGER = "non_negative_integer"
# Functions
INIT = '__init__'
# Variables
ADDRESS_BALANCE = 'balance'
ADDRESS_CODESIZE = 'codesize'
ADDRESS_IS_CONTRACT = 'is_contract'
SELF = 'self'
MSG = 'msg'
MSG_SENDER = 'sender'
MSG_VALUE = 'value'
MSG_GAS = 'gas'
BLOCK = 'block'
BLOCK_COINBASE = 'coinbase'
BLOCK_DIFFICULTY = 'difficulty'
BLOCK_NUMBER = 'number'
BLOCK_PREVHASH = 'prevhash'
BLOCK_TIMESTAMP = 'timestamp'
CHAIN = 'chain'
CHAIN_ID = 'id'
TX = 'tx'
TX_ORIGIN = 'origin'
LOG = 'log'
LEMMA = 'lemma'
ENV_VARIABLES = [MSG, BLOCK, CHAIN, TX]
# Constants
EMPTY_BYTES32 = 'EMPTY_BYTES32'
ZERO_ADDRESS = 'ZERO_ADDRESS'
ZERO_WEI = 'ZERO_WEI'
MIN_INT128 = 'MIN_INT128'
MAX_INT128 = 'MAX_INT128'
MAX_UINT256 = 'MAX_UINT256'
MIN_DECIMAL = 'MIN_DECIMAL'
MAX_DECIMAL = 'MAX_DECIMAL'
CONSTANT_VALUES = {
EMPTY_BYTES32: 'b"' + '\\x00' * 32 + '"',
ZERO_ADDRESS: '0',
ZERO_WEI: '0',
MIN_INT128: f'-{2 ** 127}',
MAX_INT128: f'{2 ** 127 - 1}',
MAX_UINT256: f'{2 ** 256 - 1}',
MIN_DECIMAL: f'-{2 ** 127}.0',
MAX_DECIMAL: f'{2 ** 127 - 1}.0'
}
# Special
UNITS = 'units'
IMPLEMENTS = 'implements'
INDEXED = 'indexed'
UNREACHABLE = 'UNREACHABLE'
# Ether units
ETHER_UNITS = {
'wei': 1,
('femtoether', 'kwei', 'babbage'): 10 ** 3,
('picoether', 'mwei', 'lovelace'): 10 ** 6,
('nanoether', 'gwei', 'shannon'): 10 ** 9,
('microether', 'szabo'): 10 ** 12,
('milliether', 'finney'): 10 ** 15,
'ether': 10 ** 18,
('kether', 'grand'): 10 ** 21
}
# Built-in functions
MIN = 'min'
MAX = 'max'
ADDMOD = 'uint256_addmod'
MULMOD = 'uint256_mulmod'
SQRT = 'sqrt'
FLOOR = 'floor'
CEIL = 'ceil'
SHIFT = 'shift'
BITWISE_NOT = 'bitwise_not'
BITWISE_AND = 'bitwise_and'
BITWISE_OR = 'bitwise_or'
BITWISE_XOR = 'bitwise_xor'
AS_WEI_VALUE = 'as_wei_value'
AS_UNITLESS_NUMBER = 'as_unitless_number'
CONVERT = 'convert'
EXTRACT32 = 'extract32'
EXTRACT32_TYPE = select_version({'^0.2.0': 'output_type', '>=0.1.0-beta.16 <0.1.0': 'type'})
RANGE = 'range'
LEN = 'len'
CONCAT = 'concat'
KECCAK256 = 'keccak256'
SHA256 = 'sha256'
ECRECOVER = 'ecrecover'
ECADD = 'ecadd'
ECMUL = 'ecmul'
BLOCKHASH = 'blockhash'
METHOD_ID = 'method_id'
METHOD_ID_OUTPUT_TYPE = select_version({'^0.2.0': 'output_type'}, default="")
EMPTY = select_version({'^0.2.0': 'empty'}, default="")
ASSERT_MODIFIABLE = select_version({'>=0.1.0-beta.16 <0.1.0': 'assert_modifiable'}, default="")
CLEAR = 'clear'
SELFDESTRUCT = 'selfdestruct'
SEND = 'send'
RAW_CALL = 'raw_call'
RAW_CALL_OUTSIZE = select_version({'^0.2.0': 'max_outsize', '>=0.1.0-beta.16 <0.1.0': 'outsize'})
RAW_CALL_VALUE = 'value'
RAW_CALL_GAS = 'gas'
RAW_CALL_DELEGATE_CALL = 'delegate_call'
RAW_CALL_IS_STATIC_CALL = select_version({'^0.2.0': 'is_static_call'}, default="")
RAW_LOG = 'raw_log'
CREATE_FORWARDER_TO = 'create_forwarder_to'
CREATE_FORWARDER_TO_VALUE = 'value'
# Verification
INVARIANT = 'invariant'
INTER_CONTRACT_INVARIANTS = 'inter_contract_invariant'
GENERAL_POSTCONDITION = 'always_ensures'
GENERAL_CHECK = 'always_check'
POSTCONDITION = 'ensures'
PRECONDITION = 'requires'
CHECK = 'check'
CALLER_PRIVATE = 'caller_private'
PERFORMS = 'performs'
CONFIG = 'config'
CONFIG_ALLOCATION = 'allocation'
CONFIG_NO_GAS = 'no_gas'
CONFIG_NO_OVERFLOWS = 'no_overflows'
CONFIG_NO_PERFORMS = 'no_performs'
CONFIG_NO_DERIVED_WEI = 'no_derived_wei_resource'
CONFIG_TRUST_CASTS = 'trust_casts'
CONFIG_OPTIONS = [CONFIG_ALLOCATION, CONFIG_NO_GAS, CONFIG_NO_OVERFLOWS, CONFIG_NO_PERFORMS, CONFIG_NO_DERIVED_WEI,
CONFIG_TRUST_CASTS]
INTERFACE = 'internal_interface'
IMPLIES = 'implies'
FORALL = 'forall'
SUM = 'sum'
TUPLE = 'tuple'
RESULT = 'result'
RESULT_DEFAULT = 'default'
STORAGE = 'storage'
OLD = 'old'
PUBLIC_OLD = 'public_old'
ISSUED = 'issued'
SENT = 'sent'
RECEIVED = 'received'
ACCESSIBLE = 'accessible'
INDEPENDENT = 'independent'
REORDER_INDEPENDENT = 'reorder_independent'
assert EVENT == 'event' # EVENT = 'event'
assert SELFDESTRUCT == 'selfdestruct' # SELFDESTRUCT = 'selfdestruct'
assert IMPLEMENTS == 'implements' # IMPLEMENTS = 'implements'
LOCKED = 'locked'
REVERT = 'revert'
PREVIOUS = 'previous'
LOOP_ARRAY = 'loop_array'
LOOP_ITERATION = 'loop_iteration'
INTERPRETED = 'interpreted'
CONDITIONAL = 'conditional'
OVERFLOW = 'overflow'
OUT_OF_GAS = 'out_of_gas'
FAILED = 'failed'
CALLER = 'caller'
SUCCESS = 'success'
SUCCESS_IF_NOT = 'if_not'
SUCCESS_OVERFLOW = 'overflow'
SUCCESS_OUT_OF_GAS = 'out_of_gas'
SUCCESS_SENDER_FAILED = 'sender_failed'
SUCCESS_CONDITIONS = [SUCCESS_OVERFLOW, SUCCESS_OUT_OF_GAS, SUCCESS_SENDER_FAILED]
WEI = 'wei'
UNDERLYING_WEI = 'Wei'
ALLOCATED = 'allocated'
OFFERED = 'offered'
NO_OFFERS = 'no_offers'
TRUST_NO_ONE = 'trust_no_one'
TRUSTED = 'trusted'
TRUSTED_BY = 'by'
TRUSTED_WHERE = 'where'
REALLOCATE = 'reallocate'
REALLOCATE_TO = 'to'
REALLOCATE_ACTOR = 'actor'
RESOURCE_PAYOUT = 'payout'
RESOURCE_PAYOUT_ACTOR = 'actor'
RESOURCE_PAYABLE = 'payable'
RESOURCE_PAYABLE_ACTOR = 'actor'
FOREACH = 'foreach'
OFFER = 'offer'
OFFER_TO = 'to'
OFFER_ACTOR = 'actor'
OFFER_TIMES = 'times'
ALLOW_TO_DECOMPOSE = 'allow_to_decompose'
ALLOWED_TO_DECOMPOSE = 'allowed_to_decompose'
REVOKE = 'revoke'
REVOKE_TO = 'to'
REVOKE_ACTOR = 'actor'
EXCHANGE = 'exchange'
EXCHANGE_TIMES = 'times'
CREATE = 'create'
CREATE_TO = 'to'
CREATE_ACTOR = 'actor'
DESTROY = 'destroy'
DESTROY_ACTOR = 'actor'
TRUST = 'trust'
TRUST_ACTOR = 'actor'
ALLOCATE_UNTRACKED = 'allocate_untracked_wei'
CREATOR = 'creator'
RESOURCE_PREFIX = "r_"
DERIVED_RESOURCE_PREFIX = "d_"
GHOST_STATEMENTS = [REALLOCATE, FOREACH, OFFER, REVOKE, EXCHANGE, CREATE, DESTROY, TRUST, ALLOCATE_UNTRACKED,
ALLOW_TO_DECOMPOSE, RESOURCE_PAYABLE, RESOURCE_PAYOUT]
QUANTIFIED_GHOST_STATEMENTS = [OFFER, REVOKE, CREATE, DESTROY, TRUST]
SPECIAL_RESOURCES = [WEI, CREATOR]
ALLOCATION_SPECIFICATION_FUNCTIONS = [ALLOCATED, OFFERED, NO_OFFERS, TRUSTED, TRUST_NO_ONE, ALLOWED_TO_DECOMPOSE]
ALLOCATION_FUNCTIONS = [*ALLOCATION_SPECIFICATION_FUNCTIONS, *GHOST_STATEMENTS]
NOT_ALLOWED_BUT_IN_LOOP_INVARIANTS = [PREVIOUS, LOOP_ARRAY, LOOP_ITERATION]
NOT_ALLOWED_IN_SPEC = [ASSERT_MODIFIABLE, CLEAR, SEND, RAW_CALL, RAW_LOG, CREATE_FORWARDER_TO]
NOT_ALLOWED_IN_INVARIANT = [*NOT_ALLOWED_IN_SPEC, CALLER, OVERFLOW, OUT_OF_GAS, FAILED, ISSUED, BLOCKHASH,
INDEPENDENT, REORDER_INDEPENDENT, EVENT, PUBLIC_OLD, INTERPRETED, CONDITIONAL,
*NOT_ALLOWED_BUT_IN_LOOP_INVARIANTS, *GHOST_STATEMENTS]
NOT_ALLOWED_IN_LOOP_INVARIANT = [*NOT_ALLOWED_IN_SPEC, CALLER, ACCESSIBLE, OVERFLOW, OUT_OF_GAS, FAILED,
ACCESSIBLE, INDEPENDENT, REORDER_INDEPENDENT, INTERPRETED, CONDITIONAL,
*GHOST_STATEMENTS]
NOT_ALLOWED_IN_CHECK = [*NOT_ALLOWED_IN_SPEC, CALLER, INDEPENDENT, ACCESSIBLE, PUBLIC_OLD,
INTERPRETED, CONDITIONAL, *NOT_ALLOWED_BUT_IN_LOOP_INVARIANTS, *GHOST_STATEMENTS]
NOT_ALLOWED_IN_POSTCONDITION = [*NOT_ALLOWED_IN_SPEC, CALLER, ACCESSIBLE, INTERPRETED, CONDITIONAL,
*NOT_ALLOWED_BUT_IN_LOOP_INVARIANTS, *GHOST_STATEMENTS]
NOT_ALLOWED_IN_PRECONDITION = [*NOT_ALLOWED_IN_SPEC, CALLER, ACCESSIBLE, SUCCESS, REVERT, OVERFLOW, OUT_OF_GAS, FAILED,
RESULT, ACCESSIBLE, OLD, INDEPENDENT, REORDER_INDEPENDENT,
INTERPRETED, CONDITIONAL, *NOT_ALLOWED_BUT_IN_LOOP_INVARIANTS, *GHOST_STATEMENTS]
NOT_ALLOWED_IN_TRANSITIVE_POSTCONDITION = [*NOT_ALLOWED_IN_SPEC, CALLER, OVERFLOW, OUT_OF_GAS, FAILED,
INDEPENDENT, REORDER_INDEPENDENT, EVENT, ACCESSIBLE, PUBLIC_OLD,
INTERPRETED, CONDITIONAL, *NOT_ALLOWED_BUT_IN_LOOP_INVARIANTS,
*GHOST_STATEMENTS]
NOT_ALLOWED_IN_CALLER_PRIVATE = [*NOT_ALLOWED_IN_SPEC, IMPLIES, FORALL, SUM, RESULT, STORAGE, OLD, PUBLIC_OLD, ISSUED,
SENT, RECEIVED, ACCESSIBLE, INDEPENDENT, REORDER_INDEPENDENT, EVENT, SELFDESTRUCT,
IMPLEMENTS, LOCKED, REVERT, OVERFLOW, OUT_OF_GAS, FAILED, SUCCESS,
BLOCKHASH, *ALLOCATION_FUNCTIONS, INTERPRETED, *NOT_ALLOWED_BUT_IN_LOOP_INVARIANTS]
NOT_ALLOWED_IN_GHOST_CODE = [*NOT_ALLOWED_IN_SPEC, CALLER, OVERFLOW, OUT_OF_GAS, FAILED, INDEPENDENT,
REORDER_INDEPENDENT, ACCESSIBLE, PUBLIC_OLD, SELFDESTRUCT,
CONDITIONAL, *NOT_ALLOWED_BUT_IN_LOOP_INVARIANTS]
NOT_ALLOWED_IN_GHOST_FUNCTION = [*NOT_ALLOWED_IN_SPEC, CALLER, OVERFLOW, OUT_OF_GAS, FAILED,
STORAGE, OLD, PUBLIC_OLD, ISSUED, BLOCKHASH, SENT, RECEIVED, ACCESSIBLE,
INDEPENDENT, REORDER_INDEPENDENT, SELFDESTRUCT, INTERPRETED, CONDITIONAL,
*NOT_ALLOWED_BUT_IN_LOOP_INVARIANTS, *GHOST_STATEMENTS]
NOT_ALLOWED_IN_GHOST_STATEMENT = [*NOT_ALLOWED_IN_SPEC, CALLER, SUCCESS, REVERT, OVERFLOW, OUT_OF_GAS, FAILED, RESULT,
ACCESSIBLE, INDEPENDENT, REORDER_INDEPENDENT, PUBLIC_OLD, SELFDESTRUCT,
INTERPRETED, CONDITIONAL, *NOT_ALLOWED_BUT_IN_LOOP_INVARIANTS]
NOT_ALLOWED_IN_LEMMAS = [*NOT_ALLOWED_IN_SPEC, RESULT, STORAGE, OLD, PUBLIC_OLD, ISSUED, SENT, RECEIVED, ACCESSIBLE,
INDEPENDENT, REORDER_INDEPENDENT, EVENT, SELFDESTRUCT, IMPLEMENTS, LOCKED, REVERT,
INTERPRETED, CONDITIONAL, *NOT_ALLOWED_BUT_IN_LOOP_INVARIANTS, OVERFLOW, OUT_OF_GAS, FAILED,
CALLER, SUCCESS, *ALLOCATION_FUNCTIONS]
# Heuristics
WITHDRAW = 'withdraw' | 2vyper | /2vyper-0.3.0.tar.gz/2vyper-0.3.0/src/twovyper/ast/names.py | names.py |
from twovyper.ast import names
from twovyper.ast import types
from twovyper.ast.types import ContractType, FunctionType, ArrayType
VYPER_INTERFACES = ['vyper', 'interfaces']
ERC20 = 'ERC20'
ERC721 = 'ERC721'
ERC20_FUNCTIONS = {
'totalSupply': FunctionType([], types.VYPER_UINT256),
'balanceOf': FunctionType([types.VYPER_ADDRESS], types.VYPER_UINT256),
'allowance': FunctionType([types.VYPER_ADDRESS, types.VYPER_ADDRESS], types.VYPER_UINT256),
'transfer': FunctionType([types.VYPER_ADDRESS, types.VYPER_UINT256], types.VYPER_BOOL),
'transferFrom': FunctionType([types.VYPER_ADDRESS, types.VYPER_ADDRESS, types.VYPER_UINT256], types.VYPER_BOOL),
'approve': FunctionType([types.VYPER_ADDRESS, types.VYPER_UINT256], types.VYPER_BOOL)
}
ERC20_MODIFIERS = {
'totalSupply': names.CONSTANT,
'balanceOf': names.CONSTANT,
'allowance': names.MODIFYING,
'transfer': names.MODIFYING,
'transferFrom': names.MODIFYING,
'approve': names.MODIFYING
}
_BYTES1024 = ArrayType(types.VYPER_BYTE, 1024, False)
ERC721_FUNCTIONS = {
'supportsInterfaces': FunctionType([types.VYPER_BYTES32], types.VYPER_BOOL),
'balanceOf': FunctionType([types.VYPER_ADDRESS], types.VYPER_UINT256),
'ownerOf': FunctionType([types.VYPER_UINT256], types.VYPER_ADDRESS),
'getApproved': FunctionType([types.VYPER_UINT256], types.VYPER_ADDRESS),
'isApprovedForAll': FunctionType([types.VYPER_ADDRESS, types.VYPER_ADDRESS], types.VYPER_BOOL),
'transferFrom': FunctionType([types.VYPER_ADDRESS, types.VYPER_ADDRESS, types.VYPER_UINT256], types.VYPER_BOOL),
'safeTransferFrom': FunctionType([types.VYPER_ADDRESS, types.VYPER_ADDRESS, types.VYPER_UINT256, _BYTES1024], None),
'approve': FunctionType([types.VYPER_ADDRESS, types.VYPER_UINT256], None),
'setApprovalForAll': FunctionType([types.VYPER_ADDRESS, types.VYPER_BOOL], None)
}
ERC721_MODIFIERS = {
'supportsInterfaces': names.CONSTANT,
'balanceOf': names.CONSTANT,
'ownerOf': names.CONSTANT,
'getApproved': names.CONSTANT,
'isApprovedForAll': names.CONSTANT,
'transferFrom': names.MODIFYING,
'safeTransferFrom': names.MODIFYING,
'approve': names.MODIFYING,
'setApprovalForAll': names.MODIFYING
}
ERC20_TYPE = ContractType(ERC20, ERC20_FUNCTIONS, ERC20_MODIFIERS)
ERC721_TYPE = ContractType(ERC721, ERC721_FUNCTIONS, ERC721_MODIFIERS) | 2vyper | /2vyper-0.3.0.tar.gz/2vyper-0.3.0/src/twovyper/ast/interfaces.py | interfaces.py |
import os
from collections import defaultdict
from itertools import chain
from typing import Dict, Iterable, List, Optional, Set, Tuple, TYPE_CHECKING
from twovyper.ast import ast_nodes as ast, names
from twovyper.ast.types import (
VyperType, FunctionType, StructType, ResourceType, ContractType, EventType, InterfaceType, DerivedResourceType
)
if TYPE_CHECKING:
from twovyper.analysis.analyzer import FunctionAnalysis, ProgramAnalysis
class Config:
def __init__(self, options: List[str]):
self.options = options
def has_option(self, option: str) -> bool:
return option in self.options
class VyperVar:
def __init__(self, name: str, type: VyperType, node):
self.name = name
self.type = type
self.node = node
class VyperFunction:
def __init__(self,
name: str,
index: int,
args: Dict[str, VyperVar],
defaults: Dict[str, Optional[ast.Expr]],
type: FunctionType,
postconditions: List[ast.Expr],
preconditions: List[ast.Expr],
checks: List[ast.Expr],
loop_invariants: Dict[ast.For, List[ast.Expr]],
performs: List[ast.FunctionCall],
decorators: List[ast.Decorator],
node: Optional[ast.FunctionDef]):
self.name = name
self.index = index
self.args = args
self.defaults = defaults
self.type = type
self.postconditions = postconditions
self.preconditions = preconditions
self.checks = checks
self.loop_invariants = loop_invariants
self.performs = performs
self.decorators = decorators
self.node = node
# Gets set in the analyzer
self.analysis: Optional[FunctionAnalysis] = None
@property
def _decorator_names(self) -> Iterable[str]:
for dec in self.decorators:
yield dec.name
def is_public(self) -> bool:
return names.PUBLIC in self._decorator_names
def is_private(self) -> bool:
return names.PRIVATE in self._decorator_names
def is_payable(self) -> bool:
return names.PAYABLE in self._decorator_names
def is_constant(self) -> bool:
return names.CONSTANT in self._decorator_names
def is_pure(self) -> bool:
return names.PURE in self._decorator_names
def is_interpreted(self) -> bool:
return names.INTERPRETED_DECORATOR in self._decorator_names
def nonreentrant_keys(self) -> Iterable[str]:
for dec in self.decorators:
if dec.name == names.NONREENTRANT:
yield dec.args[0].s
class GhostFunction:
def __init__(self,
name: str,
args: Dict[str, VyperVar],
type: FunctionType,
node: ast.FunctionDef,
file: str):
self.name = name
self.args = args
self.type = type
self.node = node
self.file = file
@property
def interface(self):
return os.path.split(self.file)[1].split('.')[0] if self.file else ''
class VyperStruct:
def __init__(self,
name: str,
type: StructType,
node: Optional[ast.Node]):
self.name = name
self.type = type
self.node = node
class Resource(VyperStruct):
def __init__(self,
rtype: ResourceType,
node: Optional[ast.Node],
file: Optional[str],
underlying_resource_node: Optional[ast.Expr] = None):
super().__init__(rtype.name, rtype, node)
self.file = file
self.analysed = False
self._own_address = None
self.underlying_resource = underlying_resource_node
self.underlying_address = None
self.derived_resources = []
@property
def interface(self):
return os.path.split(self.file)[1].split('.')[0] if self.file else ''
@property
def underlying_resource_name(self):
return self.type.underlying_resource.name if isinstance(self.type, DerivedResourceType) else None
@property
def own_address(self):
if self.analysed:
raise AssertionError("The own address attribute is only available during the analysing phase.")
return self._own_address
@own_address.setter
def own_address(self, expr: ast.Expr):
self._own_address = expr
def is_derived_resource(self):
return self.underlying_resource_name is not None
@staticmethod
def get_name_and_derived_flag(node) -> Tuple[str, bool]:
if node.name.startswith(names.DERIVED_RESOURCE_PREFIX):
return node.name[len(names.DERIVED_RESOURCE_PREFIX):], True
assert node.name.startswith(names.RESOURCE_PREFIX)
return node.name[len(names.RESOURCE_PREFIX):], False
class VyperContract:
def __init__(self, name: str, type: ContractType, node: Optional[ast.ContractDef]):
self.name = name
self.type = type
self.node = node
class VyperEvent:
def __init__(self, name: str, type: EventType):
self.name = name
self.type = type
class VyperProgram:
def __init__(self,
node: ast.Module,
file: str,
config: Config,
fields: VyperStruct,
functions: Dict[str, VyperFunction],
interfaces: Dict[str, 'VyperInterface'],
structs: Dict[str, VyperStruct],
contracts: Dict[str, VyperContract],
events: Dict[str, VyperEvent],
resources: Dict[str, Resource],
local_state_invariants: List[ast.Expr],
inter_contract_invariants: List[ast.Expr],
general_postconditions: List[ast.Expr],
transitive_postconditions: List[ast.Expr],
general_checks: List[ast.Expr],
lemmas: Dict[str, VyperFunction],
implements: List[InterfaceType],
real_implements: List[InterfaceType],
ghost_function_implementations: Dict[str, GhostFunction]):
self.node = node
self.file = file
self.config = config
self.fields = fields
self.functions = functions
self.interfaces = interfaces
self.structs = structs
self.contracts = contracts
self.events = events
self.imported_resources: Dict[str, List[Resource]] = defaultdict(list)
for key, value in self._resources():
self.imported_resources[key].append(value)
self.own_resources = resources
self.declared_resources = dict((name, resource) for name, resource in self.own_resources.items()
if name != names.WEI and name != names.UNDERLYING_WEI)
self.resources: Dict[str, List[Resource]] = defaultdict(list, self.imported_resources)
for key, value in resources.items():
self.resources[key].append(value)
self.local_state_invariants = local_state_invariants
self.inter_contract_invariants = inter_contract_invariants
self.general_postconditions = general_postconditions
self.transitive_postconditions = transitive_postconditions
self.general_checks = general_checks
self.lemmas = lemmas
self.implements = implements
self.real_implements = real_implements
self.ghost_functions: Dict[str, List[GhostFunction]] = defaultdict(list)
for key, value in self._ghost_functions():
self.ghost_functions[key].append(value)
self.ghost_function_implementations = ghost_function_implementations
self.type = fields.type
# Is set in the analyzer
self.analysis: Optional[ProgramAnalysis] = None
def is_interface(self) -> bool:
return False
def nonreentrant_keys(self) -> Set[str]:
s = set()
for func in self.functions.values():
for key in func.nonreentrant_keys():
s.add(key)
return s
def _ghost_functions(self) -> Iterable[Tuple[str, GhostFunction]]:
for interface in self.interfaces.values():
for name, func in interface.own_ghost_functions.items():
yield name, func
def _resources(self) -> Iterable[Tuple[str, Resource]]:
for interface in self.interfaces.values():
for name, resource in interface.declared_resources.items():
yield name, resource
@property
def invariants(self):
return chain(self.local_state_invariants, self.inter_contract_invariants)
class VyperInterface(VyperProgram):
def __init__(self,
node: ast.Module,
file: str,
name: Optional[str],
config: Config,
functions: Dict[str, VyperFunction],
interfaces: Dict[str, 'VyperInterface'],
resources: Dict[str, Resource],
local_state_invariants: List[ast.Expr],
inter_contract_invariants: List[ast.Expr],
general_postconditions: List[ast.Expr],
transitive_postconditions: List[ast.Expr],
general_checks: List[ast.Expr],
caller_private: List[ast.Expr],
ghost_functions: Dict[str, GhostFunction],
type: InterfaceType,
is_stub=False):
struct_name = f'{name}$self'
empty_struct_type = StructType(struct_name, {})
empty_struct = VyperStruct(struct_name, empty_struct_type, None)
super().__init__(node,
file,
config,
empty_struct,
functions,
interfaces,
{}, {}, {},
resources,
local_state_invariants,
inter_contract_invariants,
general_postconditions,
transitive_postconditions,
general_checks,
{}, [], [], {})
self.name = name
self.imported_ghost_functions: Dict[str, List[GhostFunction]] = defaultdict(list)
for key, value in self._ghost_functions():
self.imported_ghost_functions[key].append(value)
self.own_ghost_functions = ghost_functions
self.ghost_functions: Dict[str, List[GhostFunction]] = defaultdict(list, self.imported_ghost_functions)
for key, value in ghost_functions.items():
self.ghost_functions[key].append(value)
self.type = type
self.caller_private = caller_private
self.is_stub = is_stub
def is_interface(self) -> bool:
return True | 2vyper | /2vyper-0.3.0.tar.gz/2vyper-0.3.0/src/twovyper/ast/nodes.py | nodes.py |
from typing import Iterable, List, Optional, Tuple, Union
from twovyper.ast import ast_nodes as ast
def children(node: ast.Node) -> Iterable[Tuple[str, Union[Optional[ast.Node], List[ast.Node]]]]:
for child in node.children:
yield child, getattr(node, child)
# TODO: improve
def descendants(node: ast.Node) -> Iterable[ast.Node]:
for _, child in children(node):
if child is None:
continue
elif isinstance(child, ast.Node):
yield child
for child_child in descendants(child):
yield child_child
elif isinstance(child, List):
for child_child in child:
yield child_child
for child_child_child in descendants(child_child):
yield child_child_child
class NodeVisitor:
@property
def method_name(self) -> str:
return 'visit'
def visit(self, node, *args):
method = f'{self.method_name}_{node.__class__.__name__}'
visitor = getattr(self, method, self.generic_visit)
return visitor(node, *args)
def visit_nodes(self, nodes: Iterable[ast.Node], *args):
for node in nodes:
self.visit(node, *args)
return None
def generic_visit(self, node, *args):
for _, value in children(node):
if value is None:
continue
elif isinstance(value, ast.Node):
self.visit(value, *args)
elif isinstance(value, list):
for item in value:
self.visit(item, *args)
else:
assert False
return None
class NodeTransformer(NodeVisitor):
def generic_visit(self, node, *args):
for field, old_value in children(node):
if old_value is None:
continue
elif isinstance(old_value, ast.Node):
new_node = self.visit(old_value, *args)
if new_node is None:
delattr(node, field)
else:
setattr(node, field, new_node)
elif isinstance(old_value, list):
new_values = []
for value in old_value:
new_value = self.visit(value)
if new_value is None:
continue
else:
new_values.append(new_value)
old_value[:] = new_values
else:
assert False
return node | 2vyper | /2vyper-0.3.0.tar.gz/2vyper-0.3.0/src/twovyper/ast/visitors.py | visitors.py |
from enum import Enum
from typing import List as ListT, Optional as OptionalT
class Node:
_children: ListT[str] = []
def __init__(self):
self.file = None
self.lineno = None
self.col_offset = None
self.end_lineno = None
self.end_col_offset = None
self.is_ghost_code = False
@property
def children(self):
return self._children
class AllowedInGhostCode:
pass
class Stmt(Node):
pass
class Expr(Node, AllowedInGhostCode):
def __init__(self):
super().__init__()
self.type = None
class Operator(AllowedInGhostCode):
pass
class BoolOperator(Operator, Enum):
AND = 'and'
OR = 'or'
IMPLIES = 'implies'
class BoolOp(Expr):
_children = ['left', 'right']
def __init__(self, left: Expr, op: BoolOperator, right: Expr):
super().__init__()
self.left = left
self.op = op
self.right = right
class Not(Expr):
_children = ['operand']
def __init__(self, operand: Expr):
super().__init__()
self.operand = operand
class ArithmeticOperator(Operator, Enum):
ADD = '+'
SUB = '-'
MUL = '*'
DIV = '/'
MOD = '%'
POW = '**'
class ArithmeticOp(Expr):
_children = ['left', 'right']
def __init__(self, left: Expr, op: ArithmeticOperator, right: Expr):
super().__init__()
self.left = left
self.op = op
self.right = right
class UnaryArithmeticOperator(Operator, Enum):
ADD = '+'
SUB = '-'
class UnaryArithmeticOp(Expr):
_children = ['operand']
def __init__(self, op: UnaryArithmeticOperator, operand: Expr):
super().__init__()
self.op = op
self.operand = operand
class ComparisonOperator(Operator, Enum):
LT = '<'
LTE = '<='
GTE = '>='
GT = '>'
class Comparison(Expr):
_children = ['left', 'right']
def __init__(self, left: Expr, op: ComparisonOperator, right: Expr):
super().__init__()
self.left = left
self.op = op
self.right = right
class ContainmentOperator(Operator, Enum):
IN = 'in'
NOT_IN = 'not in'
class Containment(Expr):
_children = ['value', 'list']
def __init__(self, value: Expr, op: ContainmentOperator, list_expr: Expr):
super().__init__()
self.value = value
self.op = op
self.list = list_expr
class EqualityOperator(Operator, Enum):
EQ = '=='
NEQ = '!='
class Equality(Expr):
_children = ['left', 'right']
def __init__(self, left: Expr, op: EqualityOperator, right: Expr):
super().__init__()
self.left = left
self.op = op
self.right = right
class IfExpr(Expr):
_children = ['test', 'body', 'orelse']
def __init__(self, test: Expr, body: Expr, orelse: Expr):
super().__init__()
self.test = test
self.body = body
self.orelse = orelse
class Set(Expr):
_children = ['elements']
def __init__(self, elements: ListT[Expr]):
super().__init__()
self.elements = elements
class Keyword(Node, AllowedInGhostCode):
_children = ['value']
def __init__(self, name: str, value: Expr):
super().__init__()
self.name = name
self.value = value
class FunctionCall(Expr):
_children = ['args', 'keywords', 'resource']
def __init__(self, name: str, args: ListT[Expr], keywords: ListT[Keyword], resource: OptionalT[Expr] = None):
super().__init__()
self.name = name
self.args = args
self.keywords = keywords
self.resource = resource
self.underlying_resource = None
class ReceiverCall(Expr):
_children = ['receiver', 'args', 'keywords']
def __init__(self, name: str, receiver: Expr, args: ListT[Expr], keywords: ListT[Keyword]):
super().__init__()
self.name = name
self.receiver = receiver
self.args = args
self.keywords = keywords
class Num(Expr):
def __init__(self, n):
super().__init__()
self.n = n
class Str(Expr):
def __init__(self, s: str):
super().__init__()
self.s = s
class Bytes(Expr):
def __init__(self, s: bytes):
super().__init__()
self.s = s
class Bool(Expr):
def __init__(self, value):
super().__init__()
self.value = value
class Ellipsis(Expr):
pass
class Exchange(Expr):
_children = ['left', 'right']
def __init__(self, left: Expr, right: Expr):
super().__init__()
self.left = left
self.right = right
class Attribute(Expr):
_children = ['value']
def __init__(self, value: Expr, attr: str):
super().__init__()
self.value = value
self.attr = attr
class Subscript(Expr):
_children = ['value', 'index']
def __init__(self, value: Expr, index: Expr):
super().__init__()
self.value = value
self.index = index
class Name(Expr):
def __init__(self, id_str: str):
super().__init__()
self.id = id_str
class Dict(Expr):
_children = ['keys', 'values']
def __init__(self, keys: ListT[Name], values: ListT[Expr]):
super().__init__()
self.keys = keys
self.values = values
class List(Expr):
_children = ['elements']
def __init__(self, elements: ListT[Expr]):
super().__init__()
self.elements = elements
class Tuple(Expr):
_children = ['elements']
def __init__(self, elements: ListT[Expr]):
super().__init__()
self.elements = elements
class Module(Node):
_children = ['stmts']
def __init__(self, stmts: ListT[Stmt]):
super().__init__()
self.stmts = stmts
class StructDef(Node):
_children = ['body']
def __init__(self, name: str, body: ListT[Stmt]):
super().__init__()
self.name = name
self.body = body
class ContractDef(Node):
_children = ['body']
def __init__(self, name: str, body: ListT[Stmt]):
super().__init__()
self.name = name
self.body = body
class EventDef(Node):
"""
Struct like event declaration
"""
_children = ['body']
def __init__(self, name: str, body: ListT[Stmt]):
super().__init__()
self.name = name
self.body = body
class Arg(Node, AllowedInGhostCode):
_children = ['annotation', 'default']
def __init__(self, name: str, annotation: Expr, default: OptionalT[Expr]):
super().__init__()
self.name = name
self.annotation = annotation
self.default = default
class Decorator(Node, AllowedInGhostCode):
_children = ['args']
def __init__(self, name: str, args: ListT[Expr]):
super().__init__()
self.name = name
self.args = args
class FunctionDef(Stmt, AllowedInGhostCode):
_children = ['args', 'body', 'decorators', 'returns']
def __init__(self, name: str, args: ListT[Arg], body: ListT[Stmt],
decorators: ListT[Decorator], returns: OptionalT[Expr]):
super().__init__()
self.name = name
self.args = args
self.body = body
self.decorators = decorators
self.returns = returns
self.is_lemma = False
class FunctionStub(Stmt, AllowedInGhostCode):
_children = ['args']
def __init__(self, name: str, args: ListT[Arg], returns: OptionalT[Expr]):
super().__init__()
self.name = name
self.args = args
self.returns = returns
class Return(Stmt):
_children = ['value']
def __init__(self, value: OptionalT[Expr]):
super().__init__()
self.value = value
class Assign(Stmt):
_children = ['target', 'value']
def __init__(self, target: Expr, value: Expr):
super().__init__()
self.target = target
self.value = value
class AugAssign(Stmt):
_children = ['target', 'value']
def __init__(self, target: Expr, op: ArithmeticOperator, value: Expr):
super().__init__()
self.target = target
self.op = op
self.value = value
class AnnAssign(Stmt):
_children = ['target', 'annotation', 'value']
def __init__(self, target: Name, annotation: Expr, value: Expr):
super().__init__()
self.target = target
self.annotation = annotation
self.value = value
class For(Stmt):
_children = ['target', 'iter', 'body']
def __init__(self, target: Name, iter_expr: Expr, body: ListT[Stmt]):
super().__init__()
self.target = target
self.iter = iter_expr
self.body = body
class If(Stmt, AllowedInGhostCode):
_children = ['test', 'body', 'orelse']
def __init__(self, test: Expr, body: ListT[Stmt], orelse: ListT[Stmt]):
super().__init__()
self.test = test
self.body = body
self.orelse = orelse
class Log(Stmt, AllowedInGhostCode):
_children = ['body']
def __init__(self, body: Expr):
super().__init__()
self.body = body
class Ghost(Stmt, AllowedInGhostCode):
_children = ['body']
def __init__(self, body: ListT[Stmt]):
super().__init__()
self.body = body
class Raise(Stmt, AllowedInGhostCode):
_children = ['msg']
def __init__(self, msg: Expr):
super().__init__()
self.msg = msg
class Assert(Stmt, AllowedInGhostCode):
_children = ['test', 'msg']
def __init__(self, test: Expr, msg: OptionalT[Expr]):
super().__init__()
self.test = test
self.msg = msg
self.is_lemma = False
class Alias(Node):
def __init__(self, name: str, asname: OptionalT[str]):
super().__init__()
self.name = name
self.asname = asname
class Import(Stmt):
_children = ['names']
def __init__(self, names: ListT[Alias]):
super().__init__()
self.names = names
class ImportFrom(Stmt):
_children = ['names']
def __init__(self, module: OptionalT[str], names: ListT[Alias], level: int):
super().__init__()
self.module = module
self.names = names
self.level = level
class ExprStmt(Stmt, AllowedInGhostCode):
_children = ['value']
def __init__(self, value: Expr):
super().__init__()
self.value = value
class Pass(Stmt, AllowedInGhostCode):
pass
class Break(Stmt):
pass
class Continue(Stmt):
pass
def compare_nodes(first_node: Node, second_node: Node) -> bool:
"""
Similar as vyper/ast/nodes.py:compare_nodes
"""
if not isinstance(first_node, type(second_node)):
return False
for field_name in (i for i in first_node.children):
left_value = getattr(first_node, field_name, None)
right_value = getattr(second_node, field_name, None)
# compare types instead of isinstance() in case one node class inherits the other
if type(left_value) is not type(right_value):
return False
if isinstance(left_value, list):
if next((i for i in zip(left_value, right_value) if not compare_nodes(*i)), None):
return False
elif isinstance(left_value, Node):
if not compare_nodes(left_value, right_value):
return False
elif left_value != right_value:
return False
return True | 2vyper | /2vyper-0.3.0.tar.gz/2vyper-0.3.0/src/twovyper/ast/ast_nodes.py | ast_nodes.py |
Subsets and Splits