Raiff1982 commited on
Commit
281e98f
·
verified ·
1 Parent(s): 4e9bed2

Add 49 files

Browse files
astype_copy.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9564b309cbf3441ff0a6e4468fddaca46230fab34f15c77d87025a455bdf59d9
3
+ size 716
codette.metadata.v9.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c645d26f746c7f66ec0be9ddac5f158008f8626b85b32aaf4ed8b530753f41aa
3
+ size 1765
codette.projects.v9.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:733fef2d250dc4691c54275301cc429928a9f17dc0b44051b1efee6538ddba6c
3
+ size 74518
codette.strings.v9.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bc83c69937b5eba9e6770ee23148ade429e801f4363c3853ca78aa32efa7926e
3
+ size 174680
codettev3.zip ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e19bf1008b054e1998d082e21ac08325de6b3a85ce563ccce891ce468c14d673
3
+ size 637284
deploy_space_action.yaml ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ name: Run Python script
2
+
3
+ on:
4
+ push:
5
+ branches:
6
+ - $branch
7
+
8
+ jobs:
9
+ build:
10
+ runs-on: ubuntu-latest
11
+
12
+ steps:
13
+ - name: Checkout
14
+ uses: actions/checkout@v2
15
+
16
+ - name: Set up Python
17
+ uses: actions/setup-python@v2
18
+ with:
19
+ python-version: '3.9'
20
+
21
+ - name: Install Gradio
22
+ run: python -m pip install gradio
23
+
24
+ - name: Log in to Hugging Face
25
+ run: python -c 'import huggingface_hub; huggingface_hub.login(token="${{ secrets.hf_token }}")'
26
+
27
+ - name: Deploy to Spaces
28
+ run: gradio deploy
deprecated.yaml ADDED
@@ -0,0 +1,134 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Deprecated function signatures. These are exposed in Python, but not included
2
+ # in the error message suggestions.
3
+
4
+ - name: add(Tensor self, Scalar alpha, Tensor other) -> Tensor
5
+ aten: add(self, other, alpha)
6
+
7
+ - name: add_(Tensor(a!) self, Scalar alpha, Tensor other) -> Tensor(a!)
8
+ aten: add_(self, other, alpha)
9
+
10
+ - name: add(Tensor self, Scalar alpha, Tensor other, *, Tensor(a!) out) -> Tensor(a!)
11
+ aten: add_out(out, self, other, alpha)
12
+
13
+ - name: addbmm(Scalar beta, Tensor self, Scalar alpha, Tensor batch1, Tensor batch2) -> Tensor
14
+ aten: addbmm(self, batch1, batch2, beta, alpha)
15
+
16
+ - name: addbmm_(Scalar beta, Tensor(a!) self, Scalar alpha, Tensor batch1, Tensor batch2) -> Tensor(a!)
17
+ aten: addbmm_(self, batch1, batch2, beta, alpha)
18
+
19
+ - name: addbmm(Scalar beta, Tensor self, Scalar alpha, Tensor batch1, Tensor batch2, *, Tensor(a!) out) -> Tensor(a!)
20
+ aten: addbmm_out(out, self, batch1, batch2, beta, alpha)
21
+
22
+ - name: addbmm(Scalar beta, Tensor self, Tensor batch1, Tensor batch2) -> Tensor
23
+ aten: addbmm(self, batch1, batch2, beta, 1)
24
+
25
+ - name: addbmm_(Scalar beta, Tensor(a!) self, Tensor batch1, Tensor batch2) -> Tensor(a!)
26
+ aten: addbmm_(self, batch1, batch2, beta, 1)
27
+
28
+ - name: addbmm(Scalar beta, Tensor self, Tensor batch1, Tensor batch2, *, Tensor(a!) out) -> Tensor(a!)
29
+ aten: addbmm_out(out, self, batch1, batch2, beta, 1)
30
+
31
+ - name: addcdiv(Tensor self, Scalar value, Tensor tensor1, Tensor tensor2) -> Tensor
32
+ aten: addcdiv(self, tensor1, tensor2, value)
33
+
34
+ - name: addcdiv_(Tensor(a!) self, Scalar value, Tensor tensor1, Tensor tensor2) -> Tensor(a!)
35
+ aten: addcdiv_(self, tensor1, tensor2, value)
36
+
37
+ - name: addcdiv(Tensor self, Scalar value, Tensor tensor1, Tensor tensor2, *, Tensor(a!) out) -> Tensor(a!)
38
+ aten: addcdiv_out(out, self, tensor1, tensor2, value)
39
+
40
+ - name: addcmul(Tensor self, Scalar value, Tensor tensor1, Tensor tensor2) -> Tensor
41
+ aten: addcmul(self, tensor1, tensor2, value)
42
+
43
+ - name: addcmul_(Tensor(a!) self, Scalar value, Tensor tensor1, Tensor tensor2) -> Tensor(a!)
44
+ aten: addcmul_(self, tensor1, tensor2, value)
45
+
46
+ - name: addcmul(Tensor self, Scalar value, Tensor tensor1, Tensor tensor2, *, Tensor(a!) out) -> Tensor(a!)
47
+ aten: addcmul_out(out, self, tensor1, tensor2, value)
48
+
49
+ - name: addmm(Scalar beta, Tensor self, Scalar alpha, Tensor mat1, Tensor mat2) -> Tensor
50
+ aten: addmm(self, mat1, mat2, beta, alpha)
51
+
52
+ - name: addmm_(Scalar beta, Tensor(a!) self, Scalar alpha, Tensor mat1, Tensor mat2) -> Tensor(a!)
53
+ aten: addmm_(self, mat1, mat2, beta, alpha)
54
+
55
+ - name: addmm(Scalar beta, Tensor self, Scalar alpha, Tensor mat1, Tensor mat2, *, Tensor(a!) out) -> Tensor(a!)
56
+ aten: addmm_out(out, self, mat1, mat2, beta, alpha)
57
+
58
+ - name: addmm(Scalar beta, Tensor self, Tensor mat1, Tensor mat2) -> Tensor
59
+ aten: addmm(self, mat1, mat2, beta, 1)
60
+
61
+ - name: addmm_(Scalar beta, Tensor(a!) self, Tensor mat1, Tensor mat2) -> Tensor(a!)
62
+ aten: addmm_(self, mat1, mat2, beta, 1)
63
+
64
+ - name: addmm(Scalar beta, Tensor self, Tensor mat1, Tensor mat2, *, Tensor(a!) out) -> Tensor(a!)
65
+ aten: addmm_out(out, self, mat1, mat2, beta, 1)
66
+
67
+ - name: sspaddmm(Scalar beta, Tensor self, Scalar alpha, Tensor mat1, Tensor mat2) -> Tensor
68
+ aten: sspaddmm(self, mat1, mat2, beta, alpha)
69
+
70
+ - name: sspaddmm(Scalar beta, Tensor self, Tensor mat1, Tensor mat2) -> Tensor
71
+ aten: sspaddmm(self, mat1, mat2, beta, 1)
72
+
73
+ - name: addmv(Scalar beta, Tensor self, Scalar alpha, Tensor mat, Tensor vec) -> Tensor
74
+ aten: addmv(self, mat, vec, beta, alpha)
75
+
76
+ - name: addmv_(Scalar beta, Tensor(a!) self, Scalar alpha, Tensor mat, Tensor vec) -> Tensor(a!)
77
+ aten: addmv_(self, mat, vec, beta, alpha)
78
+
79
+ - name: addmv(Scalar beta, Tensor self, Scalar alpha, Tensor mat, Tensor vec, *, Tensor(a!) out) -> Tensor(a!)
80
+ aten: addmv_out(out, self, mat, vec, beta, alpha)
81
+
82
+ - name: addmv(Scalar beta, Tensor self, Tensor mat, Tensor vec) -> Tensor
83
+ aten: addmv(self, mat, vec, beta, 1)
84
+
85
+ - name: addmv_(Scalar beta, Tensor(a!) self, Tensor mat, Tensor vec) -> Tensor(a!)
86
+ aten: addmv_(self, mat, vec, beta, 1)
87
+
88
+ - name: addmv(Scalar beta, Tensor self, Tensor mat, Tensor vec, *, Tensor(a!) out) -> Tensor(a!)
89
+ aten: addmv_out(out, self, mat, vec, beta, 1)
90
+
91
+ - name: addr(Scalar beta, Tensor self, Scalar alpha, Tensor vec1, Tensor vec2) -> Tensor
92
+ aten: addr(self, vec1, vec2, beta, alpha)
93
+
94
+ - name: addr_(Scalar beta, Tensor(a!) self, Scalar alpha, Tensor vec1, Tensor vec2) -> Tensor(a!)
95
+ aten: addr_(self, vec1, vec2, beta, alpha)
96
+
97
+ - name: addr(Scalar beta, Tensor self, Scalar alpha, Tensor vec1, Tensor vec2, *, Tensor(a!) out) -> Tensor(a!)
98
+ aten: addr_out(out, self, vec1, vec2, beta, alpha)
99
+
100
+ - name: addr(Scalar beta, Tensor self, Tensor vec1, Tensor vec2) -> Tensor
101
+ aten: addr(self, vec1, vec2, beta, 1)
102
+
103
+ - name: addr_(Scalar beta, Tensor(a!) self, Tensor vec1, Tensor vec2) -> Tensor(a!)
104
+ aten: addr_(self, vec1, vec2, beta, 1)
105
+
106
+ - name: addr(Scalar beta, Tensor self, Tensor vec1, Tensor vec2, *, Tensor(a!) out) -> Tensor(a!)
107
+ aten: addr_out(out, self, vec1, vec2, beta, 1)
108
+
109
+ - name: baddbmm(Scalar beta, Tensor self, Scalar alpha, Tensor batch1, Tensor batch2) -> Tensor
110
+ aten: baddbmm(self, batch1, batch2, beta, alpha)
111
+
112
+ - name: baddbmm_(Scalar beta, Tensor(a!) self, Scalar alpha, Tensor batch1, Tensor batch2) -> Tensor(a!)
113
+ aten: baddbmm_(self, batch1, batch2, beta, alpha)
114
+
115
+ - name: baddbmm(Scalar beta, Tensor self, Scalar alpha, Tensor batch1, Tensor batch2, *, Tensor(a!) out) -> Tensor(a!)
116
+ aten: baddbmm_out(out, self, batch1, batch2, beta, alpha)
117
+
118
+ - name: baddbmm(Scalar beta, Tensor self, Tensor batch1, Tensor batch2) -> Tensor
119
+ aten: baddbmm(self, batch1, batch2, beta, 1)
120
+
121
+ - name: baddbmm_(Scalar beta, Tensor(a!) self, Tensor batch1, Tensor batch2) -> Tensor(a!)
122
+ aten: baddbmm_(self, batch1, batch2, beta, 1)
123
+
124
+ - name: baddbmm(Scalar beta, Tensor self, Tensor batch1, Tensor batch2, *, Tensor(a!) out) -> Tensor(a!)
125
+ aten: baddbmm_out(out, self, batch1, batch2, beta, 1)
126
+
127
+ - name: sub(Tensor self, Scalar alpha, Tensor other) -> Tensor
128
+ aten: sub(self, other, alpha)
129
+
130
+ - name: sub_(Tensor(a!) self, Scalar alpha, Tensor other) -> Tensor(a!)
131
+ aten: sub_(self, other, alpha)
132
+
133
+ - name: sub(Tensor self, Scalar alpha, Tensor other, *, Tensor(a!) out) -> Tensor(a!)
134
+ aten: sub_out(out, self, other, alpha)
derivatives.yaml ADDED
The diff for this file is too large to render. See raw diff
 
images/My ChatGPT image (1).png ADDED

Git LFS Details

  • SHA256: c14de63a032561b95ef20186b58c3bfdc7f1bbf6adf6af9a972f3e38f58a5ffd
  • Pointer size: 132 Bytes
  • Size of remote file: 2.09 MB
images/My ChatGPT image (2).png ADDED

Git LFS Details

  • SHA256: 21c9ed4e037eb2a048d07e11ee80488ed6c1e28457256c70fc0f3a74827d3b9b
  • Pointer size: 132 Bytes
  • Size of remote file: 1.92 MB
images/My ChatGPT image.png ADDED

Git LFS Details

  • SHA256: 6187909a70afda522d305f5175d5fd57226923a139ce7dacb83b5f7d50dd4edd
  • Pointer size: 132 Bytes
  • Size of remote file: 2.86 MB
images/outline.png ADDED

Git LFS Details

  • SHA256: 2a888afff5b4669cc3e335e847f1ee67dbcb0cf44f9477b3e15e71b3368eb6dc
  • Pointer size: 128 Bytes
  • Size of remote file: 492 Bytes
joblib_0.10.0_pickle_py27_np17.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:89c4508e3dfbe01f801e4e739f1aded13f685941e89281c8050f0ca8aa3c97e5
3
+ size 986
joblib_0.10.0_pickle_py33_np18.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e064c2eecfdc58d552844467da7bd56eca596098322bfd266a7e1312abdd5735
3
+ size 1068
joblib_0.10.0_pickle_py34_np19.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1cbe456f5b91f5a3cb8e386838f276c30335432a351426686187761d5c34168b
3
+ size 1068
joblib_0.10.0_pickle_py35_np19.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:97b9ef2e896104321d3c5ce73b3de504788c38f04f08c8b56d7a29d6d1520a96
3
+ size 1068
joblib_0.11.0_pickle_py36_np111.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5e6b0e171782d5fd5a61d1844dc946eb27c5f6b2e8075d436b23808433142ebc
3
+ size 1068
joblib_0.9.2_pickle_py27_np16.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9da8a3764db121e29d21ade67c9c3426598e76d88deae44cd7238983af8cef73
3
+ size 670
joblib_0.9.2_pickle_py27_np17.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2f29d7f1d2ceca07f10df172c0e826ef08163a14b12c6ef3fa80ec53a5fcdc3c
3
+ size 670
joblib_0.9.2_pickle_py33_np18.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c3d4cbc690d3ce9e5323a714ea546f32c01ab1710285c420184f6cdf4b26fc25
3
+ size 691
joblib_0.9.2_pickle_py34_np19.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8a538100e6ae94b16f2ab0f7d92d4d7e7a622be2dfcc0f6b0b73b623bc513ae2
3
+ size 691
joblib_0.9.2_pickle_py35_np19.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:59f0d522a29c333ce1d60480b2121fcc1a08a5d2dd650b86efdc987f991fa4ea
3
+ size 691
language-model.lm.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4ed8f52ed0413045f0e9c65a4709953896b6ff9a99ebc10f7ef926f662c5e506
3
+ size 29208442
log_config.yml ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ version: 1
2
+ disable_existing_loggers: False
3
+ formatters:
4
+ default:
5
+ "()": uvicorn.logging.DefaultFormatter
6
+ format: '%(levelprefix)s [%(asctime)s] %(message)s'
7
+ use_colors: null
8
+ datefmt: '%d-%m-%Y %H:%M:%S'
9
+ access:
10
+ "()": uvicorn.logging.AccessFormatter
11
+ format: '%(levelprefix)s [%(asctime)s] %(client_addr)s - "%(request_line)s" %(status_code)s'
12
+ datefmt: '%d-%m-%Y %H:%M:%S'
13
+ handlers:
14
+ default:
15
+ formatter: default
16
+ class: logging.StreamHandler
17
+ stream: ext://sys.stderr
18
+ access:
19
+ formatter: access
20
+ class: logging.StreamHandler
21
+ stream: ext://sys.stdout
22
+ console:
23
+ class: logging.StreamHandler
24
+ stream: ext://sys.stdout
25
+ formatter: default
26
+ file:
27
+ class : logging.handlers.RotatingFileHandler
28
+ filename: chroma.log
29
+ formatter: default
30
+ loggers:
31
+ root:
32
+ level: WARN
33
+ handlers: [console, file]
34
+ chromadb:
35
+ level: DEBUG
36
+ uvicorn:
37
+ level: INFO
logreg_iris.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8224784c98d73412d9fd99abcd57a38568bd590980d0fbe5916464531c52e8fc
3
+ size 670
mul_1.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:71f431c4e9321ec6fbeb158d02ed240459a7dcc98673fa79a4f439ce42efaf10
3
+ size 130
my-test-package.zip ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:01845c437f4655e3cf9cc4fc4e49cfd607431f22675e1b611129a90239f34822
3
+ size 1809
native_functions.yaml ADDED
The diff for this file is too large to render. See raw diff
 
nev3.zip ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:309e7f47526f9de6eda266c9304d91a5aaa5da099501dfd5fcb86a58438f3248
3
+ size 8896313
port.yaml ADDED
@@ -0,0 +1,48 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ - package: markdown-it/markdown-it
2
+ version: 13.0.1
3
+ commit: e843acc9edad115cbf8cf85e676443f01658be08
4
+ date: May 3, 2022
5
+ notes:
6
+ - Rename variables that use python built-in names, e.g.
7
+ - `max` -> `maximum`
8
+ - `len` -> `length`
9
+ - `str` -> `string`
10
+ - |
11
+ Convert JS `for` loops to `while` loops
12
+ this is generally the main difference between the codes,
13
+ because in python you can't do e.g. `for {i=1;i<x;i++} {}`
14
+ - |
15
+ `env` is a common Python dictionary, and so does not have attribute access to keys,
16
+ as with JavaScript dictionaries.
17
+ `options` have attribute access only to core markdownit configuration options
18
+ - |
19
+ `Token.attrs` is a dictionary, instead of a list of lists.
20
+ Upstream the list format is only used to guarantee order: https://github.com/markdown-it/markdown-it/issues/142,
21
+ but in Python 3.7+ order of dictionaries is guaranteed.
22
+ One should anyhow use the `attrGet`, `attrSet`, `attrPush` and `attrJoin` methods
23
+ to manipulate `Token.attrs`, which have an identical signature to those upstream.
24
+ - Use python version of `charCodeAt`
25
+ - |
26
+ Use `str` units instead of `int`s to represent Unicode codepoints.
27
+ This provides a significant performance boost
28
+ - |
29
+ In markdown_it/rules_block/reference.py,
30
+ record line range in state.env["references"] and add state.env["duplicate_refs"]
31
+ This is to allow renderers to report on issues regarding references
32
+ - |
33
+ The `MarkdownIt.__init__` signature is slightly different for updating options,
34
+ since you must always specify the config first, e.g.
35
+ use `MarkdownIt("commonmark", {"html": False})` instead of `MarkdownIt({"html": False})`
36
+ - The default configuration preset for `MarkdownIt` is "commonmark" not "default"
37
+ - Allow custom renderer to be passed to `MarkdownIt`
38
+ - |
39
+ change render method signatures
40
+ `func(tokens, idx, options, env, slf)` to
41
+ `func(self, tokens, idx, options, env)`
42
+ - |
43
+ Extensions add render methods by format
44
+ `MarkdownIt.add_render_rule(name, function, fmt="html")`,
45
+ rather than `MarkdownIt.renderer.rules[name] = function`
46
+ and renderers should declare a class property `__output__ = "html"`.
47
+ This allows for extensibility to more than just HTML renderers
48
+ - inline tokens in tables are assigned a map (this is helpful for propagation to children)
schema.yaml ADDED
@@ -0,0 +1,462 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # @generated by update_schema.py
2
+ # checksum<<9e3a8f9389cd89b981fa654e4bf6ebc544f1696f9cac84818d60424143426f30>>
3
+ Argument:
4
+ kind: union
5
+ fields:
6
+ as_none:
7
+ type: bool
8
+ as_tensor:
9
+ type: TensorArgument
10
+ as_tensors:
11
+ type: List[TensorArgument]
12
+ as_int:
13
+ type: int
14
+ as_ints:
15
+ type: List[int]
16
+ as_float:
17
+ type: float
18
+ as_floats:
19
+ type: List[float]
20
+ as_string:
21
+ type: str
22
+ as_strings:
23
+ type: List[str]
24
+ as_sym_int:
25
+ type: SymIntArgument
26
+ as_sym_ints:
27
+ type: List[SymIntArgument]
28
+ as_scalar_type:
29
+ type: ScalarType
30
+ as_memory_format:
31
+ type: MemoryFormat
32
+ as_layout:
33
+ type: Layout
34
+ as_device:
35
+ type: Device
36
+ as_bool:
37
+ type: bool
38
+ as_bools:
39
+ type: List[bool]
40
+ as_sym_bool:
41
+ type: SymBoolArgument
42
+ as_sym_bools:
43
+ type: List[SymBoolArgument]
44
+ as_graph:
45
+ type: GraphArgument
46
+ as_optional_tensors:
47
+ type: List[OptionalTensorArgument]
48
+ as_custom_obj:
49
+ type: CustomObjArgument
50
+ as_operator:
51
+ type: str
52
+ as_sym_float:
53
+ type: SymFloatArgument
54
+ as_sym_floats:
55
+ type: List[SymFloatArgument]
56
+ BufferMutationSpec:
57
+ kind: struct
58
+ fields:
59
+ arg:
60
+ type: TensorArgument
61
+ buffer_name:
62
+ type: str
63
+ ConstantValue:
64
+ kind: union
65
+ fields:
66
+ as_none:
67
+ type: bool
68
+ as_int:
69
+ type: int
70
+ as_float:
71
+ type: float
72
+ as_string:
73
+ type: str
74
+ as_bool:
75
+ type: bool
76
+ CustomObjArgument:
77
+ kind: struct
78
+ fields:
79
+ name:
80
+ type: str
81
+ class_fqn:
82
+ type: str
83
+ Device:
84
+ kind: struct
85
+ fields:
86
+ type:
87
+ type: str
88
+ index:
89
+ type: Optional[int]
90
+ default: None
91
+ ExportedProgram:
92
+ kind: struct
93
+ fields:
94
+ graph_module:
95
+ type: GraphModule
96
+ opset_version:
97
+ type: Dict[str, int]
98
+ range_constraints:
99
+ type: Dict[str, RangeConstraint]
100
+ schema_version:
101
+ type: SchemaVersion
102
+ verifiers:
103
+ type: List[str]
104
+ default: '[]'
105
+ torch_version:
106
+ type: str
107
+ default: <=2.4
108
+ GradientToParameterSpec:
109
+ kind: struct
110
+ fields:
111
+ arg:
112
+ type: TensorArgument
113
+ parameter_name:
114
+ type: str
115
+ GradientToUserInputSpec:
116
+ kind: struct
117
+ fields:
118
+ arg:
119
+ type: TensorArgument
120
+ user_input_name:
121
+ type: str
122
+ Graph:
123
+ kind: struct
124
+ fields:
125
+ inputs:
126
+ type: List[Argument]
127
+ outputs:
128
+ type: List[Argument]
129
+ nodes:
130
+ type: List[Node]
131
+ tensor_values:
132
+ type: Dict[str, TensorMeta]
133
+ sym_int_values:
134
+ type: Dict[str, SymInt]
135
+ sym_bool_values:
136
+ type: Dict[str, SymBool]
137
+ is_single_tensor_return:
138
+ type: bool
139
+ default: 'False'
140
+ custom_obj_values:
141
+ type: Dict[str, CustomObjArgument]
142
+ default: '{}'
143
+ sym_float_values:
144
+ type: Dict[str, SymFloat]
145
+ default: '{}'
146
+ GraphArgument:
147
+ kind: struct
148
+ fields:
149
+ name:
150
+ type: str
151
+ graph:
152
+ type: Graph
153
+ GraphModule:
154
+ kind: struct
155
+ fields:
156
+ graph:
157
+ type: Graph
158
+ signature:
159
+ type: GraphSignature
160
+ module_call_graph:
161
+ type: List[ModuleCallEntry]
162
+ metadata:
163
+ type: Dict[str, str]
164
+ default: '{}'
165
+ GraphSignature:
166
+ kind: struct
167
+ fields:
168
+ input_specs:
169
+ type: List[InputSpec]
170
+ output_specs:
171
+ type: List[OutputSpec]
172
+ InputSpec:
173
+ kind: union
174
+ fields:
175
+ user_input:
176
+ type: UserInputSpec
177
+ parameter:
178
+ type: InputToParameterSpec
179
+ buffer:
180
+ type: InputToBufferSpec
181
+ tensor_constant:
182
+ type: InputToTensorConstantSpec
183
+ custom_obj:
184
+ type: InputToCustomObjSpec
185
+ token:
186
+ type: InputTokenSpec
187
+ constant_input:
188
+ type: InputToConstantInputSpec
189
+ InputToBufferSpec:
190
+ kind: struct
191
+ fields:
192
+ arg:
193
+ type: TensorArgument
194
+ buffer_name:
195
+ type: str
196
+ persistent:
197
+ type: bool
198
+ InputToConstantInputSpec:
199
+ kind: struct
200
+ fields:
201
+ name:
202
+ type: str
203
+ value:
204
+ type: ConstantValue
205
+ InputToCustomObjSpec:
206
+ kind: struct
207
+ fields:
208
+ arg:
209
+ type: CustomObjArgument
210
+ custom_obj_name:
211
+ type: str
212
+ InputToParameterSpec:
213
+ kind: struct
214
+ fields:
215
+ arg:
216
+ type: TensorArgument
217
+ parameter_name:
218
+ type: str
219
+ InputToTensorConstantSpec:
220
+ kind: struct
221
+ fields:
222
+ arg:
223
+ type: TensorArgument
224
+ tensor_constant_name:
225
+ type: str
226
+ InputTokenSpec:
227
+ kind: struct
228
+ fields:
229
+ arg:
230
+ type: TokenArgument
231
+ Layout:
232
+ kind: enum
233
+ fields:
234
+ Unknown: 0
235
+ SparseCoo: 1
236
+ SparseCsr: 2
237
+ SparseCsc: 3
238
+ SparseBsr: 4
239
+ SparseBsc: 5
240
+ _mkldnn: 6
241
+ Strided: 7
242
+ LossOutputSpec:
243
+ kind: struct
244
+ fields:
245
+ arg:
246
+ type: TensorArgument
247
+ MemoryFormat:
248
+ kind: enum
249
+ fields:
250
+ Unknown: 0
251
+ ContiguousFormat: 1
252
+ ChannelsLast: 2
253
+ ChannelsLast3d: 3
254
+ PreserveFormat: 4
255
+ ModuleCallEntry:
256
+ kind: struct
257
+ fields:
258
+ fqn:
259
+ type: str
260
+ signature:
261
+ type: Optional[ModuleCallSignature]
262
+ default: None
263
+ ModuleCallSignature:
264
+ kind: struct
265
+ fields:
266
+ inputs:
267
+ type: List[Argument]
268
+ outputs:
269
+ type: List[Argument]
270
+ in_spec:
271
+ type: str
272
+ out_spec:
273
+ type: str
274
+ forward_arg_names:
275
+ type: Optional[List[str]]
276
+ default: None
277
+ NamedArgument:
278
+ kind: struct
279
+ fields:
280
+ name:
281
+ type: str
282
+ arg:
283
+ type: Argument
284
+ Node:
285
+ kind: struct
286
+ fields:
287
+ target:
288
+ type: str
289
+ inputs:
290
+ type: List[NamedArgument]
291
+ outputs:
292
+ type: List[Argument]
293
+ metadata:
294
+ type: Dict[str, str]
295
+ OptionalTensorArgument:
296
+ kind: union
297
+ fields:
298
+ as_tensor:
299
+ type: TensorArgument
300
+ as_none:
301
+ type: bool
302
+ OutputSpec:
303
+ kind: union
304
+ fields:
305
+ user_output:
306
+ type: UserOutputSpec
307
+ loss_output:
308
+ type: LossOutputSpec
309
+ buffer_mutation:
310
+ type: BufferMutationSpec
311
+ gradient_to_parameter:
312
+ type: GradientToParameterSpec
313
+ gradient_to_user_input:
314
+ type: GradientToUserInputSpec
315
+ user_input_mutation:
316
+ type: UserInputMutationSpec
317
+ token:
318
+ type: OutputTokenSpec
319
+ OutputTokenSpec:
320
+ kind: struct
321
+ fields:
322
+ arg:
323
+ type: TokenArgument
324
+ RangeConstraint:
325
+ kind: struct
326
+ fields:
327
+ min_val:
328
+ type: Optional[int]
329
+ max_val:
330
+ type: Optional[int]
331
+ ScalarType:
332
+ kind: enum
333
+ fields:
334
+ UNKNOWN: 0
335
+ BYTE: 1
336
+ CHAR: 2
337
+ SHORT: 3
338
+ INT: 4
339
+ LONG: 5
340
+ HALF: 6
341
+ FLOAT: 7
342
+ DOUBLE: 8
343
+ COMPLEXHALF: 9
344
+ COMPLEXFLOAT: 10
345
+ COMPLEXDOUBLE: 11
346
+ BOOL: 12
347
+ BFLOAT16: 13
348
+ UINT16: 28
349
+ SchemaVersion:
350
+ kind: struct
351
+ fields:
352
+ major:
353
+ type: int
354
+ minor:
355
+ type: int
356
+ SymBool:
357
+ kind: union
358
+ fields:
359
+ as_expr:
360
+ type: SymExpr
361
+ as_bool:
362
+ type: bool
363
+ SymBoolArgument:
364
+ kind: union
365
+ fields:
366
+ as_name:
367
+ type: str
368
+ as_bool:
369
+ type: bool
370
+ SymExpr:
371
+ kind: struct
372
+ fields:
373
+ expr_str:
374
+ type: str
375
+ hint:
376
+ type: Optional[SymExprHint]
377
+ default: None
378
+ SymExprHint:
379
+ kind: union
380
+ fields:
381
+ as_int:
382
+ type: int
383
+ as_bool:
384
+ type: bool
385
+ as_float:
386
+ type: float
387
+ SymFloat:
388
+ kind: union
389
+ fields:
390
+ as_expr:
391
+ type: SymExpr
392
+ as_float:
393
+ type: float
394
+ SymFloatArgument:
395
+ kind: union
396
+ fields:
397
+ as_name:
398
+ type: str
399
+ as_float:
400
+ type: float
401
+ SymInt:
402
+ kind: union
403
+ fields:
404
+ as_expr:
405
+ type: SymExpr
406
+ as_int:
407
+ type: int
408
+ SymIntArgument:
409
+ kind: union
410
+ fields:
411
+ as_name:
412
+ type: str
413
+ as_int:
414
+ type: int
415
+ TensorArgument:
416
+ kind: struct
417
+ fields:
418
+ name:
419
+ type: str
420
+ TensorMeta:
421
+ kind: struct
422
+ fields:
423
+ dtype:
424
+ type: ScalarType
425
+ sizes:
426
+ type: List[SymInt]
427
+ requires_grad:
428
+ type: bool
429
+ device:
430
+ type: Device
431
+ strides:
432
+ type: List[SymInt]
433
+ storage_offset:
434
+ type: SymInt
435
+ layout:
436
+ type: Layout
437
+ TokenArgument:
438
+ kind: struct
439
+ fields:
440
+ name:
441
+ type: str
442
+ UserInputMutationSpec:
443
+ kind: struct
444
+ fields:
445
+ arg:
446
+ type: TensorArgument
447
+ user_input_name:
448
+ type: str
449
+ UserInputSpec:
450
+ kind: struct
451
+ fields:
452
+ arg:
453
+ type: Argument
454
+ UserOutputSpec:
455
+ kind: struct
456
+ fields:
457
+ arg:
458
+ type: Argument
459
+ SCHEMA_VERSION:
460
+ - 8
461
+ - 2
462
+ TREESPEC_VERSION: 1
sigmoid.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5340aba67a7e3475162ad794378af55f1718f55f9a5d74b4af60ecc7f7a624b6
3
+ size 103
tags.yaml ADDED
@@ -0,0 +1,80 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # This yaml file contains all the possible tags that can be defined in `tags` in `native_functions.yaml`
2
+
3
+ - tag: inplace_view
4
+ desc: |
5
+ This tag indicates if an operator *only* modifies the tensor metadata
6
+ - tag: pt2_compliant_tag
7
+ desc: |
8
+ This tag indicates if the operator is guaranteed to
9
+ work with the PT2 compilation APIs (torch.compile,
10
+ torch.export, etc). If you add this tag to an
11
+ operator, please use
12
+ `torch.testing._internal.optest.opcheck` to test that
13
+ the operator has been registered correctly and
14
+ works with torch.compile
15
+ - tag: view_copy
16
+ desc: |
17
+ This tag indicates operators that are *_copy* variants
18
+ of view/aliasing operators. If an operator has a view_copy tag,
19
+ then it should have the name {op}_copy, where {op} is a view operator.
20
+ - tag: dynamic_output_shape
21
+ desc: |
22
+ This tag indicates if an operator's output's shape depends on input Tensor
23
+ data.
24
+ - tag: data_dependent_output
25
+ desc: |
26
+ Operator has a non-Tensor output whose value is dependent on the data
27
+ of Tensor inputs. Among other things, this implies that this operator
28
+ cannot be run with meta tensor (since data is not available), nor
29
+ can it be symbolically traced.
30
+ - tag: generated
31
+ desc: |
32
+ This tag indicates that the operator doesn't have an explicit entry in
33
+ native_functions.yaml, and instead was generated automatically by the codegen.
34
+ - tag: nondeterministic_seeded
35
+ desc: |
36
+ This tag indicates if an operator is nondeterministically seeded
37
+ (i.e., is random) such that the operator intentionally produces
38
+ different results when run twice on the same inputs, but this randomness
39
+ is controlled by a Generator which, if reseeded would give you the
40
+ same result.
41
+ - tag: nondeterministic_bitwise
42
+ desc: |
43
+ This tag indicates if an operator doesn't guarantee bitwise equivalence
44
+ across different runs of an operator with identical inputs.
45
+ - tag: needs_fixed_stride_order
46
+ desc: |
47
+ This tag indicates that the operator should be passed Tensors following
48
+ the same stride permutation as observed in eager when compiled in inductor.
49
+ Only one of {needs_fixed_stride_order, flexible_layout} can apply; if
50
+ multiple are assigned then we assume the most restrictive one.
51
+ - tag: flexible_layout
52
+ desc: |
53
+ This tag indicates that the custom operator can accept inputs with varying
54
+ strides/storage_offset and that when compiled, Inductor is allowed to change
55
+ the strides/storage_offset of inputs to the custom operator.
56
+ Only one of {needs_fixed_stride_order, flexible_layout} can apply; if
57
+ multiple are assigned then we assume the most restrictive one.
58
+
59
+ # NOTE [Core ATen Ops]
60
+ - tag: core
61
+ desc: |
62
+ Core aten ops is a subset of aten ops that remains after aten-to-aten decomposition and
63
+ functionalization pass. Core aten ops are fully functional and adhere to single static
64
+ assignment (SSA): this implies there will be no `inplace` or `_out` variants in this opset.
65
+ This opset is designed to serve as the functional IR to interface with compiler backends.
66
+ In contrast to primTorch, core aten opset doesn't decompose ops into explicit
67
+ type promotion and broadcasting ops.
68
+ Core aten ops is also effectively the opset produced by torchdynamo.export(aten_graph=True),
69
+ and thus can be used as an opset for export purpose.
70
+ - tag: pointwise
71
+ desc: |
72
+ Pointwise operators are operators where each element of the output is computed only by accessing
73
+ the corresponding element of all the broadcasted inputs. The output shape will be the broadcasted
74
+ shape of the inputs.
75
+ - tag: maybe_aliasing_or_mutating
76
+ desc: |
77
+ For some ops, we can't statically determine whether the op is functional or not. Note that this is only
78
+ relevant to CIA ops that decompose before functionalization/autograd. It is useful to
79
+ know this information for export as we would want to decompose these ops as they are unsafe to be
80
+ preserved.
teamsapp.local.yml ADDED
@@ -0,0 +1,79 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # yaml-language-server: $schema=https://aka.ms/teams-toolkit/v1.7/yaml.schema.json
2
+ # Visit https://aka.ms/teamsfx-v5.0-guide for details on this file
3
+ # Visit https://aka.ms/teamsfx-actions for details on actions
4
+ version: v1.7
5
+
6
+ provision:
7
+ # Creates a Teams app
8
+ - uses: teamsApp/create
9
+ with:
10
+ # Teams app name
11
+ name: codette${{APP_NAME_SUFFIX}}
12
+ # Write the information of created resources into environment file for
13
+ # the specified environment variable(s).
14
+ writeToEnvironmentFile:
15
+ teamsAppId: TEAMS_APP_ID
16
+
17
+ # Create or reuse an existing Microsoft Entra application for bot.
18
+ - uses: aadApp/create
19
+ with:
20
+ # The Microsoft Entra application's display name
21
+ name: codette${{APP_NAME_SUFFIX}}
22
+ generateClientSecret: true
23
+ signInAudience: AzureADMultipleOrgs
24
+ writeToEnvironmentFile:
25
+ # The Microsoft Entra application's client id created for bot.
26
+ clientId: BOT_ID
27
+ # The Microsoft Entra application's client secret created for bot.
28
+ clientSecret: SECRET_BOT_PASSWORD
29
+ # The Microsoft Entra application's object id created for bot.
30
+ objectId: BOT_OBJECT_ID
31
+
32
+ # Generate runtime appsettings to JSON file
33
+ - uses: file/createOrUpdateJsonFile
34
+ with:
35
+ target: ../codette/appsettings.Development.json
36
+ content:
37
+ BOT_TYPE: 'MultiTenant'
38
+ BOT_ID: ${{BOT_ID}}
39
+ BOT_PASSWORD: ${{SECRET_BOT_PASSWORD}}
40
+ Azure:
41
+ OpenAIApiKey: ${{SECRET_AZURE_OPENAI_API_KEY}}
42
+ OpenAIEndpoint: ${{AZURE_OPENAI_ENDPOINT}}
43
+ OpenAIDeploymentName: ${{AZURE_OPENAI_DEPLOYMENT_NAME}}
44
+
45
+ # Create or update the bot registration on dev.botframework.com
46
+ - uses: botFramework/create
47
+ with:
48
+ botId: ${{BOT_ID}}
49
+ name: codette
50
+ messagingEndpoint: ${{BOT_ENDPOINT}}/api/messages
51
+ description: ""
52
+ channels:
53
+ - name: msteams
54
+ # Validate using manifest schema
55
+ - uses: teamsApp/validateManifest
56
+ with:
57
+ # Path to manifest template
58
+ manifestPath: ./appPackage/manifest.json
59
+ # Build Teams app package with latest env value
60
+ - uses: teamsApp/zipAppPackage
61
+ with:
62
+ # Path to manifest template
63
+ manifestPath: ./appPackage/manifest.json
64
+ outputZipPath: ./appPackage/build/appPackage.${{TEAMSFX_ENV}}.zip
65
+ outputFolder: ./appPackage/build
66
+ # Validate app package using validation rules
67
+ - uses: teamsApp/validateAppPackage
68
+ with:
69
+ # Relative path to this file. This is the path for built zip file.
70
+ appPackagePath: ./appPackage/build/appPackage.${{TEAMSFX_ENV}}.zip
71
+
72
+ # Apply the Teams app manifest to an existing Teams app in
73
+ # Teams Developer Portal.
74
+ # Will use the app id in manifest file to determine which Teams app to update.
75
+ - uses: teamsApp/update
76
+ with:
77
+ # Relative path to this file. This is the path for built zip file.
78
+ appPackagePath: ./appPackage/build/appPackage.${{TEAMSFX_ENV}}.zip
79
+
teamsapp.yml ADDED
@@ -0,0 +1,87 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # yaml-language-server: $schema=https://aka.ms/teams-toolkit/v1.7/yaml.schema.json
2
+ # Visit https://aka.ms/teamsfx-v5.0-guide for details on this file
3
+ # Visit https://aka.ms/teamsfx-actions for details on actions
4
+ version: v1.7
5
+
6
+ environmentFolderPath: ./env
7
+
8
+ # Triggered when 'teamsapp provision' is executed
9
+ provision:
10
+ # Creates a Teams app
11
+ - uses: teamsApp/create
12
+ with:
13
+ # Teams app name
14
+ name: codette${{APP_NAME_SUFFIX}}
15
+ # Write the information of created resources into environment file for
16
+ # the specified environment variable(s).
17
+ writeToEnvironmentFile:
18
+ teamsAppId: TEAMS_APP_ID
19
+
20
+ - uses: arm/deploy # Deploy given ARM templates parallelly.
21
+ with:
22
+ # AZURE_SUBSCRIPTION_ID is a built-in environment variable,
23
+ # if its value is empty, TeamsFx will prompt you to select a subscription.
24
+ # Referencing other environment variables with empty values
25
+ # will skip the subscription selection prompt.
26
+ subscriptionId: ${{AZURE_SUBSCRIPTION_ID}}
27
+ # AZURE_RESOURCE_GROUP_NAME is a built-in environment variable,
28
+ # if its value is empty, TeamsFx will prompt you to select or create one
29
+ # resource group.
30
+ # Referencing other environment variables with empty values
31
+ # will skip the resource group selection prompt.
32
+ resourceGroupName: ${{AZURE_RESOURCE_GROUP_NAME}}
33
+ templates:
34
+ - path: ./infra/azure.bicep # Relative path to this file
35
+ # Relative path to this yaml file.
36
+ # Placeholders will be replaced with corresponding environment
37
+ # variable before ARM deployment.
38
+ parameters: ./infra/azure.parameters.json
39
+ # Required when deploying ARM template
40
+ deploymentName: Create-resources-for-bot
41
+ # Teams Toolkit will download this bicep CLI version from github for you,
42
+ # will use bicep CLI in PATH if you remove this config.
43
+ bicepCliVersion: v0.9.1
44
+ # Validate using manifest schema
45
+ - uses: teamsApp/validateManifest
46
+ with:
47
+ # Path to manifest template
48
+ manifestPath: ./appPackage/manifest.json
49
+ # Build Teams app package with latest env value
50
+ - uses: teamsApp/zipAppPackage
51
+ with:
52
+ # Path to manifest template
53
+ manifestPath: ./appPackage/manifest.json
54
+ outputZipPath: ./appPackage/build/appPackage.${{TEAMSFX_ENV}}.zip
55
+ outputFolder: ./appPackage/build
56
+ # Validate app package using validation rules
57
+ - uses: teamsApp/validateAppPackage
58
+ with:
59
+ # Relative path to this file. This is the path for built zip file.
60
+ appPackagePath: ./appPackage/build/appPackage.${{TEAMSFX_ENV}}.zip
61
+ # Apply the Teams app manifest to an existing Teams app in
62
+ # Teams Developer Portal.
63
+ # Will use the app id in manifest file to determine which Teams app to update.
64
+ - uses: teamsApp/update
65
+ with:
66
+ # Relative path to this file. This is the path for built zip file.
67
+ appPackagePath: ./appPackage/build/appPackage.${{TEAMSFX_ENV}}.zip
68
+
69
+ # Triggered when 'teamsapp deploy' is executed
70
+ deploy:
71
+ - uses: cli/runDotnetCommand
72
+ with:
73
+ args: publish --configuration Release codette.csproj
74
+ workingDirectory: ../codette
75
+ # Deploy your application to Azure App Service using the zip deploy feature.
76
+ # For additional details, refer to https://aka.ms/zip-deploy-to-app-services.
77
+ - uses: azureAppService/zipDeploy
78
+ with:
79
+ # Deploy base folder
80
+ artifactFolder: bin/Release/net8.0/publish
81
+ # The resource id of the cloud resource to be deployed to.
82
+ # This key will be generated by arm/deploy action automatically.
83
+ # You can replace it with your existing Azure Resource id
84
+ # or add it to your environment variable file.
85
+ resourceId: ${{BOT_AZURE_APP_SERVICE_RESOURCE_ID}}
86
+ workingDirectory: ../codette
87
+ projectId: aed51bd9-1779-40ba-b1ea-6d41c7319608