commit
stringlengths
40
40
old_file
stringlengths
4
237
new_file
stringlengths
4
237
old_contents
stringlengths
1
4.24k
new_contents
stringlengths
5
4.84k
subject
stringlengths
15
778
message
stringlengths
16
6.86k
lang
stringlengths
1
30
license
stringclasses
13 values
repos
stringlengths
5
116k
config
stringlengths
1
30
content
stringlengths
105
8.72k
51575888408e4bf08e67bcefc463547d31874b23
test/CodeGen/X86/2009-06-07-ExpandMMXBitcast.ll
test/CodeGen/X86/2009-06-07-ExpandMMXBitcast.ll
; RUN: llc < %s -mtriple=x86_64-linux -mattr=+mmx | grep movd | count 2 define i64 @a(i32 %a, i32 %b) nounwind readnone { entry: %0 = insertelement <2 x i32> undef, i32 %a, i32 0 ; <<2 x i32>> [#uses=1] %1 = insertelement <2 x i32> %0, i32 %b, i32 1 ; <<2 x i32>> [#uses=1] %conv = bitcast <2 x i32> %1 to i64 ; <i64> [#uses=1] ret i64 %conv }
; RUN: llc < %s -mtriple=x86_64-linux -mcpu=corei7 -mattr=+mmx | grep movd | count 2 define i64 @a(i32 %a, i32 %b) nounwind readnone { entry: %0 = insertelement <2 x i32> undef, i32 %a, i32 0 ; <<2 x i32>> [#uses=1] %1 = insertelement <2 x i32> %0, i32 %b, i32 1 ; <<2 x i32>> [#uses=1] %conv = bitcast <2 x i32> %1 to i64 ; <i64> [#uses=1] ret i64 %conv }
Stabilize the test by specifying an exact cpu target
Stabilize the test by specifying an exact cpu target git-svn-id: 0ff597fd157e6f4fc38580e8d64ab130330d2411@143307 91177308-0d34-0410-b5e6-96231b3b80d8
LLVM
bsd-2-clause
dslab-epfl/asap,apple/swift-llvm,GPUOpen-Drivers/llvm,GPUOpen-Drivers/llvm,apple/swift-llvm,llvm-mirror/llvm,chubbymaggie/asap,llvm-mirror/llvm,chubbymaggie/asap,GPUOpen-Drivers/llvm,dslab-epfl/asap,llvm-mirror/llvm,llvm-mirror/llvm,llvm-mirror/llvm,apple/swift-llvm,dslab-epfl/asap,apple/swift-llvm,llvm-mirror/llvm,apple/swift-llvm,GPUOpen-Drivers/llvm,chubbymaggie/asap,apple/swift-llvm,dslab-epfl/asap,llvm-mirror/llvm,dslab-epfl/asap,GPUOpen-Drivers/llvm,GPUOpen-Drivers/llvm,chubbymaggie/asap,GPUOpen-Drivers/llvm,chubbymaggie/asap,dslab-epfl/asap,dslab-epfl/asap,llvm-mirror/llvm,GPUOpen-Drivers/llvm,chubbymaggie/asap,llvm-mirror/llvm,apple/swift-llvm,apple/swift-llvm
llvm
## Code Before: ; RUN: llc < %s -mtriple=x86_64-linux -mattr=+mmx | grep movd | count 2 define i64 @a(i32 %a, i32 %b) nounwind readnone { entry: %0 = insertelement <2 x i32> undef, i32 %a, i32 0 ; <<2 x i32>> [#uses=1] %1 = insertelement <2 x i32> %0, i32 %b, i32 1 ; <<2 x i32>> [#uses=1] %conv = bitcast <2 x i32> %1 to i64 ; <i64> [#uses=1] ret i64 %conv } ## Instruction: Stabilize the test by specifying an exact cpu target git-svn-id: 0ff597fd157e6f4fc38580e8d64ab130330d2411@143307 91177308-0d34-0410-b5e6-96231b3b80d8 ## Code After: ; RUN: llc < %s -mtriple=x86_64-linux -mcpu=corei7 -mattr=+mmx | grep movd | count 2 define i64 @a(i32 %a, i32 %b) nounwind readnone { entry: %0 = insertelement <2 x i32> undef, i32 %a, i32 0 ; <<2 x i32>> [#uses=1] %1 = insertelement <2 x i32> %0, i32 %b, i32 1 ; <<2 x i32>> [#uses=1] %conv = bitcast <2 x i32> %1 to i64 ; <i64> [#uses=1] ret i64 %conv }
3460e5b8ef4042ba05560b6f9a0343e8c4f4f918
bootstrap/docker.json
bootstrap/docker.json
{ "data-root": "/mnt/stateful_partition/docker", "experimental": true, "deprecated-key-path": "/var/lib/docker/key.json", "hosts": [ "tcp://127.0.0.1:2376", "unix:///var/run/docker.sock" ], "storage-driver": "devicemapper", "storage-opts": [ "dm.override_udev_sync_check=true" ] }
{ "bip": "10.250.0.0/16", "data-root": "/mnt/stateful_partition/docker", "experimental": true, "deprecated-key-path": "/var/lib/docker/key.json", "default-address-pools":[ { "base":"10.251.0.0/16", "size":24 }, { "base":"10.252.0.0/16", "size":24 } ], "hosts": [ "tcp://127.0.0.1:2376", "unix:///var/run/docker.sock" ], "storage-driver": "devicemapper", "storage-opts": [ "dm.override_udev_sync_check=true" ] }
Add custom addresses for Docker containers
Add custom addresses for Docker containers
JSON
mit
cartolari/dotfiles,cartolari/dotfiles,cartolari/dotfiles,cartolari/dotfiles
json
## Code Before: { "data-root": "/mnt/stateful_partition/docker", "experimental": true, "deprecated-key-path": "/var/lib/docker/key.json", "hosts": [ "tcp://127.0.0.1:2376", "unix:///var/run/docker.sock" ], "storage-driver": "devicemapper", "storage-opts": [ "dm.override_udev_sync_check=true" ] } ## Instruction: Add custom addresses for Docker containers ## Code After: { "bip": "10.250.0.0/16", "data-root": "/mnt/stateful_partition/docker", "experimental": true, "deprecated-key-path": "/var/lib/docker/key.json", "default-address-pools":[ { "base":"10.251.0.0/16", "size":24 }, { "base":"10.252.0.0/16", "size":24 } ], "hosts": [ "tcp://127.0.0.1:2376", "unix:///var/run/docker.sock" ], "storage-driver": "devicemapper", "storage-opts": [ "dm.override_udev_sync_check=true" ] }
3bf2779805eeb2fff9fdda563e9188814c0859cf
cookbooks/dns/templates/default/creds.json.erb
cookbooks/dns/templates/default/creds.json.erb
{ "gandi_v5": { "apikey": "<%= @passwords["gandi"] %>", "sharing_id": "7028b616-ba65-11e7-8343-00163ec31f40" }, "cloudflare": { "accountid": "049c95aba02c95fc1e78a9d255282e0f", "accountname": "OpenStreetMap", "apitoken": "<%= @passwords["cloudflare"] %>" } }
{ "gandi_v5": { "TYPE": "GANDI_V5", "apikey": "<%= @passwords["gandi"] %>", "sharing_id": "7028b616-ba65-11e7-8343-00163ec31f40" }, "cloudflare": { "TYPE": "CLOUDFLAREAPI", "accountid": "049c95aba02c95fc1e78a9d255282e0f", "accountname": "OpenStreetMap", "apitoken": "<%= @passwords["cloudflare"] %>" } }
Update dnscontrol auth to v3.16+ syntax
Update dnscontrol auth to v3.16+ syntax https://stackexchange.github.io/dnscontrol/creds-json#missing
HTML+ERB
apache-2.0
tomhughes/openstreetmap-chef,tomhughes/openstreetmap-chef,tomhughes/openstreetmap-chef,tomhughes/openstreetmap-chef,openstreetmap/chef,tomhughes/openstreetmap-chef,openstreetmap/chef,tomhughes/openstreetmap-chef,openstreetmap/chef,openstreetmap/chef,openstreetmap/chef,tomhughes/openstreetmap-chef,openstreetmap/chef,tomhughes/openstreetmap-chef,openstreetmap/chef,openstreetmap/chef
html+erb
## Code Before: { "gandi_v5": { "apikey": "<%= @passwords["gandi"] %>", "sharing_id": "7028b616-ba65-11e7-8343-00163ec31f40" }, "cloudflare": { "accountid": "049c95aba02c95fc1e78a9d255282e0f", "accountname": "OpenStreetMap", "apitoken": "<%= @passwords["cloudflare"] %>" } } ## Instruction: Update dnscontrol auth to v3.16+ syntax https://stackexchange.github.io/dnscontrol/creds-json#missing ## Code After: { "gandi_v5": { "TYPE": "GANDI_V5", "apikey": "<%= @passwords["gandi"] %>", "sharing_id": "7028b616-ba65-11e7-8343-00163ec31f40" }, "cloudflare": { "TYPE": "CLOUDFLAREAPI", "accountid": "049c95aba02c95fc1e78a9d255282e0f", "accountname": "OpenStreetMap", "apitoken": "<%= @passwords["cloudflare"] %>" } }
8e275a22aa419297eee0acc5ef65fc0cce4aa3cb
clients/web/test/spec/swaggerSpec.js
clients/web/test/spec/swaggerSpec.js
function jasmineTests() { var jasmineEnv = jasmine.getEnv(); var consoleReporter = new jasmine.ConsoleReporter(); window.jasmine_phantom_reporter = consoleReporter; jasmineEnv.addReporter(consoleReporter); function waitAndExecute() { if (!jasmineEnv.currentRunner().suites_.length) { window.setTimeout(waitAndExecute, 10); return; } jasmineEnv.execute(); } waitAndExecute(); describe('Test the swagger pages', function () { it('Test swagger', function () { waitsFor(function () { return $('li#resource_system.resource').length > 0; }, 'swagger docs to appear'); runs(function () { $('li#resource_system.resource .heading h2 a').click(); }); waitsFor(function () { return $('#system_getVersion:visible').length > 0; }, 'end points to be visible'); runs(function () { $('#system_getVersion h3 a').click(); }); waitsFor(function () { return $('#system_getVersion .sandbox_header input.submit[name="commit"]:visible').length > 0; }, 'version try out button to be visible'); runs(function () { $('#system_getVersion .sandbox_header input.submit[name="commit"]').click(); }); waitsFor(function () { return $('#system_getVersion .response_body.json').text().indexOf('apiVersion') >= 0; }, 'version information was returned'); }); }); } $(function () { $.getScript('/static/built/testing-no-cover.min.js', jasmineTests); });
function jasmineTests() { var jasmineEnv = jasmine.getEnv(); var consoleReporter = new jasmine.ConsoleReporter(); window.jasmine_phantom_reporter = consoleReporter; jasmineEnv.addReporter(consoleReporter); function waitAndExecute() { if (!jasmineEnv.currentRunner().suites_.length) { window.setTimeout(waitAndExecute, 10); return; } jasmineEnv.execute(); } waitAndExecute(); describe('Test the swagger pages', function () { it('Test swagger', function () { waitsFor(function () { return $('li#resource_system.resource').length > 0; }, 'swagger docs to appear'); runs(function () { $('li#resource_system.resource .heading h2 a').click(); }); waitsFor(function () { return $('#system_system_getVersion:visible').length > 0; }, 'end points to be visible'); runs(function () { $('#system_system_getVersion h3 a').click(); }); waitsFor(function () { return $('#system_system_getVersion .sandbox_header input.submit:visible').length > 0; }, 'version try out button to be visible'); runs(function () { $('#system_system_getVersion .sandbox_header input.submit').click(); }); waitsFor(function () { return $('#system_system_getVersion .response_body.json').text().indexOf('apiVersion') >= 0; }, 'version information was returned'); }); }); } $(function () { $.getScript('/static/built/testing-no-cover.min.js', jasmineTests); });
Update web client swagger test for swagger-ui 2.1.4
Update web client swagger test for swagger-ui 2.1.4
JavaScript
apache-2.0
adsorensen/girder,kotfic/girder,data-exp-lab/girder,jbeezley/girder,Kitware/girder,jbeezley/girder,adsorensen/girder,RafaelPalomar/girder,kotfic/girder,girder/girder,sutartmelson/girder,Xarthisius/girder,sutartmelson/girder,sutartmelson/girder,manthey/girder,data-exp-lab/girder,RafaelPalomar/girder,data-exp-lab/girder,manthey/girder,data-exp-lab/girder,RafaelPalomar/girder,Xarthisius/girder,adsorensen/girder,kotfic/girder,data-exp-lab/girder,adsorensen/girder,RafaelPalomar/girder,RafaelPalomar/girder,girder/girder,sutartmelson/girder,jbeezley/girder,adsorensen/girder,manthey/girder,Xarthisius/girder,kotfic/girder,Kitware/girder,girder/girder,Kitware/girder,Kitware/girder,sutartmelson/girder,Xarthisius/girder,girder/girder,manthey/girder,jbeezley/girder,kotfic/girder,Xarthisius/girder
javascript
## Code Before: function jasmineTests() { var jasmineEnv = jasmine.getEnv(); var consoleReporter = new jasmine.ConsoleReporter(); window.jasmine_phantom_reporter = consoleReporter; jasmineEnv.addReporter(consoleReporter); function waitAndExecute() { if (!jasmineEnv.currentRunner().suites_.length) { window.setTimeout(waitAndExecute, 10); return; } jasmineEnv.execute(); } waitAndExecute(); describe('Test the swagger pages', function () { it('Test swagger', function () { waitsFor(function () { return $('li#resource_system.resource').length > 0; }, 'swagger docs to appear'); runs(function () { $('li#resource_system.resource .heading h2 a').click(); }); waitsFor(function () { return $('#system_getVersion:visible').length > 0; }, 'end points to be visible'); runs(function () { $('#system_getVersion h3 a').click(); }); waitsFor(function () { return $('#system_getVersion .sandbox_header input.submit[name="commit"]:visible').length > 0; }, 'version try out button to be visible'); runs(function () { $('#system_getVersion .sandbox_header input.submit[name="commit"]').click(); }); waitsFor(function () { return $('#system_getVersion .response_body.json').text().indexOf('apiVersion') >= 0; }, 'version information was returned'); }); }); } $(function () { $.getScript('/static/built/testing-no-cover.min.js', jasmineTests); }); ## Instruction: Update web client swagger test for swagger-ui 2.1.4 ## Code After: function jasmineTests() { var jasmineEnv = jasmine.getEnv(); var consoleReporter = new jasmine.ConsoleReporter(); window.jasmine_phantom_reporter = consoleReporter; jasmineEnv.addReporter(consoleReporter); function waitAndExecute() { if (!jasmineEnv.currentRunner().suites_.length) { window.setTimeout(waitAndExecute, 10); return; } jasmineEnv.execute(); } waitAndExecute(); describe('Test the swagger pages', function () { it('Test swagger', function () { waitsFor(function () { return $('li#resource_system.resource').length > 0; }, 'swagger docs to appear'); runs(function () { $('li#resource_system.resource .heading h2 a').click(); }); waitsFor(function () { return $('#system_system_getVersion:visible').length > 0; }, 'end points to be visible'); runs(function () { $('#system_system_getVersion h3 a').click(); }); waitsFor(function () { return $('#system_system_getVersion .sandbox_header input.submit:visible').length > 0; }, 'version try out button to be visible'); runs(function () { $('#system_system_getVersion .sandbox_header input.submit').click(); }); waitsFor(function () { return $('#system_system_getVersion .response_body.json').text().indexOf('apiVersion') >= 0; }, 'version information was returned'); }); }); } $(function () { $.getScript('/static/built/testing-no-cover.min.js', jasmineTests); });
17a18f1f8765bda183de2d52834faaa310ca086a
java/build.xml
java/build.xml
<project name="LibOpenConnect" default="dist" basedir="."> <property name="build" location="build"/> <property name="dist" location="dist"/> <property name="src" location="src"/> <property name="jar_lib" location="${dist}/openconnect-wrapper.jar"/> <property name="jar_app" location="${dist}/example.jar"/> <target name="init"> <mkdir dir="${build}"/> </target> <target name="compile" depends="init"> <javac srcdir="${src}" destdir="${build}" includeantruntime="false"/> </target> <target name="dist" depends="compile"> <jar jarfile="${jar_lib}" basedir="${build}" includes="org/infradead/libopenconnect/*" /> <jar jarfile="${jar_app}" basedir="${build}"> <manifest> <attribute name="Main-Class" value="com.example.LibTest" /> </manifest> </jar> </target> <target name="clean"> <delete dir="${build}"/> <delete dir="${dist}"/> </target> </project>
<project name="LibOpenConnect" default="dist" basedir="."> <property name="build" location="build"/> <property name="dist" location="dist"/> <property name="src" location="src"/> <property name="jar_lib" location="${dist}/openconnect-wrapper.jar"/> <property name="jar_app" location="${dist}/example.jar"/> <target name="init"> <mkdir dir="${build}"/> </target> <target name="compile" depends="init"> <javac srcdir="${src}" destdir="${build}" source="1.5" target="1.5" includeantruntime="false"/> </target> <target name="dist" depends="compile"> <jar jarfile="${jar_lib}" basedir="${build}" includes="org/infradead/libopenconnect/*" /> <jar jarfile="${jar_app}" basedir="${build}"> <manifest> <attribute name="Main-Class" value="com.example.LibTest" /> </manifest> </jar> </target> <target name="clean"> <delete dir="${build}"/> <delete dir="${dist}"/> </target> </project>
Build for Java 1.5 to avoid compatibility issues
Build for Java 1.5 to avoid compatibility issues cf. https://code.google.com/p/android/issues/detail?id=22970 Signed-off-by: David Woodhouse <[email protected]>
XML
lgpl-2.1
cernekee/openconnect,matlockx/openconnect,nmav/openconnect-mine,lins05/openconnect,lins05/openconnect,mtmiller/openconnect,brandt/openconnect,atyndall/openconnect,mveplus/openconnect-client,don-johnny/openconnect,don-johnny/openconnect,mtmiller/openconnect,mtmiller/openconnect,liiir1985/openconnect,cernekee/openconnect,brandt/openconnect,lins05/openconnect,atyndall/openconnect,lins05/openconnect,mveplus/openconnect-client,matlockx/openconnect,cernekee/openconnect,liiir1985/openconnect,liiir1985/openconnect,liiir1985/openconnect,nmav/openconnect-mine,nmav/openconnect-mine,brandt/openconnect,don-johnny/openconnect,mveplus/openconnect-client,mtmiller/openconnect,matlockx/openconnect,mveplus/openconnect-client,atyndall/openconnect,brandt/openconnect,atyndall/openconnect,nmav/openconnect-mine,don-johnny/openconnect,cernekee/openconnect,matlockx/openconnect
xml
## Code Before: <project name="LibOpenConnect" default="dist" basedir="."> <property name="build" location="build"/> <property name="dist" location="dist"/> <property name="src" location="src"/> <property name="jar_lib" location="${dist}/openconnect-wrapper.jar"/> <property name="jar_app" location="${dist}/example.jar"/> <target name="init"> <mkdir dir="${build}"/> </target> <target name="compile" depends="init"> <javac srcdir="${src}" destdir="${build}" includeantruntime="false"/> </target> <target name="dist" depends="compile"> <jar jarfile="${jar_lib}" basedir="${build}" includes="org/infradead/libopenconnect/*" /> <jar jarfile="${jar_app}" basedir="${build}"> <manifest> <attribute name="Main-Class" value="com.example.LibTest" /> </manifest> </jar> </target> <target name="clean"> <delete dir="${build}"/> <delete dir="${dist}"/> </target> </project> ## Instruction: Build for Java 1.5 to avoid compatibility issues cf. https://code.google.com/p/android/issues/detail?id=22970 Signed-off-by: David Woodhouse <[email protected]> ## Code After: <project name="LibOpenConnect" default="dist" basedir="."> <property name="build" location="build"/> <property name="dist" location="dist"/> <property name="src" location="src"/> <property name="jar_lib" location="${dist}/openconnect-wrapper.jar"/> <property name="jar_app" location="${dist}/example.jar"/> <target name="init"> <mkdir dir="${build}"/> </target> <target name="compile" depends="init"> <javac srcdir="${src}" destdir="${build}" source="1.5" target="1.5" includeantruntime="false"/> </target> <target name="dist" depends="compile"> <jar jarfile="${jar_lib}" basedir="${build}" includes="org/infradead/libopenconnect/*" /> <jar jarfile="${jar_app}" basedir="${build}"> <manifest> <attribute name="Main-Class" value="com.example.LibTest" /> </manifest> </jar> </target> <target name="clean"> <delete dir="${build}"/> <delete dir="${dist}"/> </target> </project>
e358e2b6fc2b6923b37a88f1b3fddb5a2f96fe37
.travis.yml
.travis.yml
language: C addons: apt: packages: - libgmp-dev - libmpfr-dev cache: apt: true #directories: # - compiler/ #install: bash scripts/build_cross_compiler.sh script: - make run-tests
sudo: false language: c addons: apt: packages: # Libraries needed to build cross compiler - libgmp-dev - libmpfr-dev cache: apt: true directories: - compiler/ # Compiles the cross compiler into the projects root before_install: bash scripts/build_cross_compiler.sh script: - make - make run-tests
Build the CC, build the kernel, and run libk tests
Build the CC, build the kernel, and run libk tests
YAML
mit
iankronquist/kernel-of-truth,Herbstein/kernel-of-truth,Herbstein/kernel-of-truth,awensaunders/kernel-of-truth,iankronquist/kernel-of-truth,awensaunders/kernel-of-truth,iankronquist/kernel-of-truth,awensaunders/kernel-of-truth,Herbstein/kernel-of-truth,iankronquist/kernel-of-truth,iankronquist/kernel-of-truth
yaml
## Code Before: language: C addons: apt: packages: - libgmp-dev - libmpfr-dev cache: apt: true #directories: # - compiler/ #install: bash scripts/build_cross_compiler.sh script: - make run-tests ## Instruction: Build the CC, build the kernel, and run libk tests ## Code After: sudo: false language: c addons: apt: packages: # Libraries needed to build cross compiler - libgmp-dev - libmpfr-dev cache: apt: true directories: - compiler/ # Compiles the cross compiler into the projects root before_install: bash scripts/build_cross_compiler.sh script: - make - make run-tests
efb3b418eba8316b42ac6ee92249ba13598cbc9e
app.sh
app.sh
cat << EOS AkkeyLab The elapsed time does not matter. Because speed is important. EOS # # Install web apps. # echo " ----- Install web apps ------" brew install --cask 1password brew tap AdoptOpenJDK/openjdk brew install --cask adoptopenjdk brew install --cask android-studio brew install --cask cyberduck brew install --cask docker brew install --cask google-chrome brew install --cask iterm2 brew install --cask sourcetree brew install --cask vlc brew install --cask google-japanese-ime brew install --cask swimat brew install --cask spotify brew install --cask imageoptim brew install --cask soundflower brew install --cask visual-studio-code brew install --cask adobe-creative-cloud brew install --cask gyazo brew install --cask zoomus brew install --cask tandem brew install --cask discord brew install --cask notion brew install --cask figma while true; do read -p 'Add "need license" apps? [Y/n]' Answer case $Answer in '' | [Yy]* ) brew cask install microsoft-office brew cask install intellij-idea brew cask install clip-studio-paint break; ;; [Nn]* ) echo "Skip install" break; ;; * ) echo Please answer YES or NO. esac done; echo " ------------ END ------------"
cat << EOS AkkeyLab The elapsed time does not matter. Because speed is important. EOS # # Install web apps. # echo " ----- Install web apps ------" brew install --cask 1password brew tap AdoptOpenJDK/openjdk brew install --cask adoptopenjdk brew install --cask android-studio brew install --cask cyberduck brew install --cask docker brew install --cask google-chrome brew install --cask iterm2 brew install --cask vlc brew install --cask google-japanese-ime brew install --cask spotify brew install --cask imageoptim brew install --cask visual-studio-code brew install --cask adobe-creative-cloud brew install --cask gyazo brew install --cask zoomus brew install --cask discord brew install --cask notion brew install --cask figma brew install --cask tableplus brew install --cask postman while true; do read -p 'Add "need license" apps? [Y/n]' Answer case $Answer in '' | [Yy]* ) brew cask install microsoft-office brew cask install intellij-idea brew cask install clip-studio-paint break; ;; [Nn]* ) echo "Skip install" break; ;; * ) echo Please answer YES or NO. esac done; echo " ------------ END ------------"
Update brew cask install commands
:bulb: Update brew cask install commands
Shell
mit
AkkeyLab/mac-auto-setup
shell
## Code Before: cat << EOS AkkeyLab The elapsed time does not matter. Because speed is important. EOS # # Install web apps. # echo " ----- Install web apps ------" brew install --cask 1password brew tap AdoptOpenJDK/openjdk brew install --cask adoptopenjdk brew install --cask android-studio brew install --cask cyberduck brew install --cask docker brew install --cask google-chrome brew install --cask iterm2 brew install --cask sourcetree brew install --cask vlc brew install --cask google-japanese-ime brew install --cask swimat brew install --cask spotify brew install --cask imageoptim brew install --cask soundflower brew install --cask visual-studio-code brew install --cask adobe-creative-cloud brew install --cask gyazo brew install --cask zoomus brew install --cask tandem brew install --cask discord brew install --cask notion brew install --cask figma while true; do read -p 'Add "need license" apps? [Y/n]' Answer case $Answer in '' | [Yy]* ) brew cask install microsoft-office brew cask install intellij-idea brew cask install clip-studio-paint break; ;; [Nn]* ) echo "Skip install" break; ;; * ) echo Please answer YES or NO. esac done; echo " ------------ END ------------" ## Instruction: :bulb: Update brew cask install commands ## Code After: cat << EOS AkkeyLab The elapsed time does not matter. Because speed is important. EOS # # Install web apps. # echo " ----- Install web apps ------" brew install --cask 1password brew tap AdoptOpenJDK/openjdk brew install --cask adoptopenjdk brew install --cask android-studio brew install --cask cyberduck brew install --cask docker brew install --cask google-chrome brew install --cask iterm2 brew install --cask vlc brew install --cask google-japanese-ime brew install --cask spotify brew install --cask imageoptim brew install --cask visual-studio-code brew install --cask adobe-creative-cloud brew install --cask gyazo brew install --cask zoomus brew install --cask discord brew install --cask notion brew install --cask figma brew install --cask tableplus brew install --cask postman while true; do read -p 'Add "need license" apps? [Y/n]' Answer case $Answer in '' | [Yy]* ) brew cask install microsoft-office brew cask install intellij-idea brew cask install clip-studio-paint break; ;; [Nn]* ) echo "Skip install" break; ;; * ) echo Please answer YES or NO. esac done; echo " ------------ END ------------"
7a403519539146fc50c0897576f3453f11960032
cli/README.md
cli/README.md
The GitHub Actions for [Google Cloud Platform](https://cloud.google.com/) and wraps the [gcloud SDK](https://cloud.google.com/sdk/) to enable common Google Cloud commands. This is a thin wrapper around the `gcloud` utility. ## Usage An example workflow to list clusters on Google Cloud Platform: ``` workflow "Run gcloud command" { on = "push" resolves = "GCP Authenticate" } action "GCP Authenticate" { uses = "actions/gcloud/auth@master" secrets = ["GCLOUD_AUTH"] } action "GCP List Clusters" { needs = ["GCP Authenticate"] uses = "actions/gcloud/cli@master" args = "container clusters list" } ``` For more information about the authentication step, [see `auth`](/auth). ## License The Dockerfile and associated scripts and documentation in this project are released under the [MIT License](LICENSE). Container images built with this project include third party materials. See [THIRD_PARTY_NOTICE.md](THIRD_PARTY_NOTICE.md) for details.
The GitHub Actions for [Google Cloud Platform](https://cloud.google.com/) and wraps the [gcloud SDK](https://cloud.google.com/sdk/) to enable common Google Cloud commands. This is a thin wrapper around the `gcloud` utility. ## Usage An example workflow to list clusters on Google Cloud Platform: ``` workflow "Run gcloud command" { on = "push" resolves = "GCP List Clusters" } action "GCP Authenticate" { uses = "actions/gcloud/auth@master" secrets = ["GCLOUD_AUTH"] } action "GCP List Clusters" { needs = ["GCP Authenticate"] uses = "actions/gcloud/cli@master" args = "container clusters list" } ``` For more information about the authentication step, [see `auth`](/auth). ## License The Dockerfile and associated scripts and documentation in this project are released under the [MIT License](LICENSE). Container images built with this project include third party materials. See [THIRD_PARTY_NOTICE.md](THIRD_PARTY_NOTICE.md) for details.
Update text to GCP List Clusters
Update text to GCP List Clusters Addresses PR feedback, thanks @parkr!
Markdown
apache-2.0
google-github-actions/setup-gcloud,google-github-actions/setup-gcloud,google-github-actions/setup-gcloud
markdown
## Code Before: The GitHub Actions for [Google Cloud Platform](https://cloud.google.com/) and wraps the [gcloud SDK](https://cloud.google.com/sdk/) to enable common Google Cloud commands. This is a thin wrapper around the `gcloud` utility. ## Usage An example workflow to list clusters on Google Cloud Platform: ``` workflow "Run gcloud command" { on = "push" resolves = "GCP Authenticate" } action "GCP Authenticate" { uses = "actions/gcloud/auth@master" secrets = ["GCLOUD_AUTH"] } action "GCP List Clusters" { needs = ["GCP Authenticate"] uses = "actions/gcloud/cli@master" args = "container clusters list" } ``` For more information about the authentication step, [see `auth`](/auth). ## License The Dockerfile and associated scripts and documentation in this project are released under the [MIT License](LICENSE). Container images built with this project include third party materials. See [THIRD_PARTY_NOTICE.md](THIRD_PARTY_NOTICE.md) for details. ## Instruction: Update text to GCP List Clusters Addresses PR feedback, thanks @parkr! ## Code After: The GitHub Actions for [Google Cloud Platform](https://cloud.google.com/) and wraps the [gcloud SDK](https://cloud.google.com/sdk/) to enable common Google Cloud commands. This is a thin wrapper around the `gcloud` utility. ## Usage An example workflow to list clusters on Google Cloud Platform: ``` workflow "Run gcloud command" { on = "push" resolves = "GCP List Clusters" } action "GCP Authenticate" { uses = "actions/gcloud/auth@master" secrets = ["GCLOUD_AUTH"] } action "GCP List Clusters" { needs = ["GCP Authenticate"] uses = "actions/gcloud/cli@master" args = "container clusters list" } ``` For more information about the authentication step, [see `auth`](/auth). ## License The Dockerfile and associated scripts and documentation in this project are released under the [MIT License](LICENSE). Container images built with this project include third party materials. See [THIRD_PARTY_NOTICE.md](THIRD_PARTY_NOTICE.md) for details.
d84d5f831424806c82033f660205cbfb2c1e7341
src/app/header/header.component.html
src/app/header/header.component.html
<nav class="navbar fixed-top navbar-dark bg-inverse"> <button class="navbar-toggler navbar-toggler-right" type="button" data-toggle="collapse" data-target="#navbarNav" aria-controls="navbarNav" aria-expanded="false" aria-label="Toggle navigation"> <span class="navbar-toggler-icon"></span> </button> <a class="navbar-brand" [routerLink]="['/home']">Catalogo Ricette</a> <div class="collapse navbar-collapse" id="navbarNav"> <ul class="navbar-nav"> <li class="nav-item active"> <a class="nav-link" [routerLink]="['/home']">Home <span class="sr-only">(current)</span></a> </li> <li class="nav-item"> <a class="nav-link disabled">Ingredienti</a> </li> <li class="nav-item"> <a class="nav-link" [routerLink]="['/dishTypes']">Categorie</a> </li> <li class="nav-item"> <a class="nav-link disabled">Libri</a> </li> </ul> </div> </nav>
<nav class="navbar fixed-top navbar-toggleable-md navbar-inverse bg-inverse"> <button class="navbar-toggler navbar-toggler-right" type="button" data-toggle="collapse" data-target="#navbarSupportedContent" aria-controls="navbarSupportedContent" aria-expanded="false" aria-label="Toggle navigation"> <span class="navbar-toggler-icon"></span> </button> <a class="navbar-brand" [routerLink]="['/home']">Catalogo Ricette</a> <div class="collapse navbar-collapse" id="navbarSupportedContent"> <ul class="navbar-nav"> <li class="nav-item"> <a class="nav-link" [routerLink]="['/home']">Home <span class="sr-only">(current)</span></a> </li> <li class="nav-item"> <a class="nav-link disabled">Ingredienti</a> </li> <li class="nav-item"> <a class="nav-link" [routerLink]="['/dishTypes']">Categorie</a> </li> <li class="nav-item"> <a class="nav-link disabled">Libri</a> </li> </ul> </div> </nav>
Fix navigation bar to work with latest bootstrap
Fix navigation bar to work with latest bootstrap
HTML
mit
riccardobellini/recipecatalog-frontend-angular,riccardobellini/recipecatalog-frontend-angular,riccardobellini/recipecatalog-frontend-angular
html
## Code Before: <nav class="navbar fixed-top navbar-dark bg-inverse"> <button class="navbar-toggler navbar-toggler-right" type="button" data-toggle="collapse" data-target="#navbarNav" aria-controls="navbarNav" aria-expanded="false" aria-label="Toggle navigation"> <span class="navbar-toggler-icon"></span> </button> <a class="navbar-brand" [routerLink]="['/home']">Catalogo Ricette</a> <div class="collapse navbar-collapse" id="navbarNav"> <ul class="navbar-nav"> <li class="nav-item active"> <a class="nav-link" [routerLink]="['/home']">Home <span class="sr-only">(current)</span></a> </li> <li class="nav-item"> <a class="nav-link disabled">Ingredienti</a> </li> <li class="nav-item"> <a class="nav-link" [routerLink]="['/dishTypes']">Categorie</a> </li> <li class="nav-item"> <a class="nav-link disabled">Libri</a> </li> </ul> </div> </nav> ## Instruction: Fix navigation bar to work with latest bootstrap ## Code After: <nav class="navbar fixed-top navbar-toggleable-md navbar-inverse bg-inverse"> <button class="navbar-toggler navbar-toggler-right" type="button" data-toggle="collapse" data-target="#navbarSupportedContent" aria-controls="navbarSupportedContent" aria-expanded="false" aria-label="Toggle navigation"> <span class="navbar-toggler-icon"></span> </button> <a class="navbar-brand" [routerLink]="['/home']">Catalogo Ricette</a> <div class="collapse navbar-collapse" id="navbarSupportedContent"> <ul class="navbar-nav"> <li class="nav-item"> <a class="nav-link" [routerLink]="['/home']">Home <span class="sr-only">(current)</span></a> </li> <li class="nav-item"> <a class="nav-link disabled">Ingredienti</a> </li> <li class="nav-item"> <a class="nav-link" [routerLink]="['/dishTypes']">Categorie</a> </li> <li class="nav-item"> <a class="nav-link disabled">Libri</a> </li> </ul> </div> </nav>
2ad9290db1bdc9a155a6c327f92088209f13d141
lib/pdoc.rb
lib/pdoc.rb
DIR = File.dirname(__FILE__) OUTPUT_DIR = File.expand_path(File.join(DIR, "..", "output")) TEMPLATES_DIR = File.expand_path(File.join(DIR, "..", "templates")) PARSER_DIR = File.expand_path(File.join(DIR, "pdoc", "parser")) VENDOR_DIR = File.expand_path(File.join(DIR, "..", "vendor")) [DIR, VENDOR_DIR, PARSER_DIR, OUTPUT_DIR, TEMPLATES_DIR].each do |c| $:.unshift File.expand_path(c) end require 'erb' require 'fileutils' require File.expand_path(File.join(DIR, "pdoc", "runner")) require File.expand_path(File.join(DIR, "pdoc", "generators")) require File.expand_path(File.join(DIR, "pdoc", "parser")) module PDoc def self.copy_templates(template_type, destination) dir = File.expand_path(destination) raise "File already exists: #{destination}" if File.exist?(dir) FileUtils.cp_r("#{TEMPLATES_DIR}/#{template_type}", dir) end end
DIR = File.expand_path(File.dirname(__FILE__)) OUTPUT_DIR = File.join(DIR, '..', 'output') TEMPLATES_DIR = File.join(DIR, '..', 'templates') VENDOR_DIR = File.join(DIR, '..', 'vendor') PARSER_DIR = File.join(DIR, 'pdoc', 'parser') [DIR, VENDOR_DIR, PARSER_DIR, OUTPUT_DIR, TEMPLATES_DIR].each do |c| $:.unshift(c) end require 'rubygems' require 'erb' require 'fileutils' require 'pdoc/runner' require 'pdoc/generators' require 'pdoc/parser' module PDoc def self.copy_templates(template_type, destination) dir = File.expand_path(destination) raise "File already exists: #{destination}" if File.exist?(dir) FileUtils.cp_r("#{TEMPLATES_DIR}/#{template_type}", dir) end end
Clean up internal requires and external dependencies.
Clean up internal requires and external dependencies.
Ruby
mit
tobie/pdoc,tobie/pdoc
ruby
## Code Before: DIR = File.dirname(__FILE__) OUTPUT_DIR = File.expand_path(File.join(DIR, "..", "output")) TEMPLATES_DIR = File.expand_path(File.join(DIR, "..", "templates")) PARSER_DIR = File.expand_path(File.join(DIR, "pdoc", "parser")) VENDOR_DIR = File.expand_path(File.join(DIR, "..", "vendor")) [DIR, VENDOR_DIR, PARSER_DIR, OUTPUT_DIR, TEMPLATES_DIR].each do |c| $:.unshift File.expand_path(c) end require 'erb' require 'fileutils' require File.expand_path(File.join(DIR, "pdoc", "runner")) require File.expand_path(File.join(DIR, "pdoc", "generators")) require File.expand_path(File.join(DIR, "pdoc", "parser")) module PDoc def self.copy_templates(template_type, destination) dir = File.expand_path(destination) raise "File already exists: #{destination}" if File.exist?(dir) FileUtils.cp_r("#{TEMPLATES_DIR}/#{template_type}", dir) end end ## Instruction: Clean up internal requires and external dependencies. ## Code After: DIR = File.expand_path(File.dirname(__FILE__)) OUTPUT_DIR = File.join(DIR, '..', 'output') TEMPLATES_DIR = File.join(DIR, '..', 'templates') VENDOR_DIR = File.join(DIR, '..', 'vendor') PARSER_DIR = File.join(DIR, 'pdoc', 'parser') [DIR, VENDOR_DIR, PARSER_DIR, OUTPUT_DIR, TEMPLATES_DIR].each do |c| $:.unshift(c) end require 'rubygems' require 'erb' require 'fileutils' require 'pdoc/runner' require 'pdoc/generators' require 'pdoc/parser' module PDoc def self.copy_templates(template_type, destination) dir = File.expand_path(destination) raise "File already exists: #{destination}" if File.exist?(dir) FileUtils.cp_r("#{TEMPLATES_DIR}/#{template_type}", dir) end end
44e5a3a2338907b975a7042372739bf88de81c6e
gcloud/aliases.zsh
gcloud/aliases.zsh
alias gcrli='gcloud container images list --repository=asia.gcr.io/greensync-container-registry' gcrlt() { gcloud container images list-tags asia.gcr.io/greensync-container-registry/"$1" }
alias gcrli='gcloud container images list --repository=asia.gcr.io/greensync-container-registry' gcrlt() { local full_image_name if [[ "$1" =~ ^asia.gcr.io/greensync-container-registry/ ]]; then full_image_name="$1" else full_image_name="asia.gcr.io/greensync-container-registry/$1" fi gcloud container images list-tags "$full_image_name" --sort-by=TIMESTAMP } gcrl() { if [[ -n "$1" ]]; then gcrlt "$1" else gcrli fi }
Make gcrlt support either image name or full image name; Add gcrl
Make gcrlt support either image name or full image name; Add gcrl
Shell
mit
ZimbiX/dotfiles,ZimbiX/dotfiles
shell
## Code Before: alias gcrli='gcloud container images list --repository=asia.gcr.io/greensync-container-registry' gcrlt() { gcloud container images list-tags asia.gcr.io/greensync-container-registry/"$1" } ## Instruction: Make gcrlt support either image name or full image name; Add gcrl ## Code After: alias gcrli='gcloud container images list --repository=asia.gcr.io/greensync-container-registry' gcrlt() { local full_image_name if [[ "$1" =~ ^asia.gcr.io/greensync-container-registry/ ]]; then full_image_name="$1" else full_image_name="asia.gcr.io/greensync-container-registry/$1" fi gcloud container images list-tags "$full_image_name" --sort-by=TIMESTAMP } gcrl() { if [[ -n "$1" ]]; then gcrlt "$1" else gcrli fi }
95329797ba6e76fa2d0aecc348805ac06b6e3e0f
scripts/after.sh
scripts/after.sh
echo "Running after.sh" #!/bin/sh
echo "Running after.sh" #!/bin/sh apt-get -y install phantomjs screen mysql -u root -pvagrant -e "GRANT ALL PRIVILEGES ON homestead.* TO 'homestead'@'%' IDENTIFIED BY 'secret'" echo "Set Travis privs" (crontab -l 2>/dev/null; echo "@reboot screen -S server -d -m phantomjs --webdriver=4444") | crontab - echo "Added scree phantomjs to crontab" screen -S server -d -m phantomjs --webdriver=4444 echo "Started phantomjs is screen" perl -pi -e 's/upload_max_filesize = 2M/upload_max_filesize = 50M/g' /etc/php5/fpm/php.ini perl -pi -e 's/post_max_size = 8M/post_max_size = 51M/g' /etc/php5/fpm/php.ini service nginx restart service php5-fpm reload /usr/local/bin/composer self-update echo "" >> /home/vagrant/.bashrc echo "PATH=$PATH:/home/vagrant/memvents.dev/vendor/bin" >> /home/vagrant/.bashrc echo "export PATH" >> /home/vagrant/.bashrc echo Done!
Add codeception testing stuff to vagrant
Add codeception testing stuff to vagrant
Shell
mit
svpernova09/memvents,svpernova09/memvents,svpernova09/memvents,svpernova09/memvents
shell
## Code Before: echo "Running after.sh" #!/bin/sh ## Instruction: Add codeception testing stuff to vagrant ## Code After: echo "Running after.sh" #!/bin/sh apt-get -y install phantomjs screen mysql -u root -pvagrant -e "GRANT ALL PRIVILEGES ON homestead.* TO 'homestead'@'%' IDENTIFIED BY 'secret'" echo "Set Travis privs" (crontab -l 2>/dev/null; echo "@reboot screen -S server -d -m phantomjs --webdriver=4444") | crontab - echo "Added scree phantomjs to crontab" screen -S server -d -m phantomjs --webdriver=4444 echo "Started phantomjs is screen" perl -pi -e 's/upload_max_filesize = 2M/upload_max_filesize = 50M/g' /etc/php5/fpm/php.ini perl -pi -e 's/post_max_size = 8M/post_max_size = 51M/g' /etc/php5/fpm/php.ini service nginx restart service php5-fpm reload /usr/local/bin/composer self-update echo "" >> /home/vagrant/.bashrc echo "PATH=$PATH:/home/vagrant/memvents.dev/vendor/bin" >> /home/vagrant/.bashrc echo "export PATH" >> /home/vagrant/.bashrc echo Done!
58d56bd6b7608349fb491e9f281e10c3aea838ba
lib/mforwd/deliver/slack.rb
lib/mforwd/deliver/slack.rb
class MForwd::Deliver::Slack < MForwd::Deliver::HttpPost private # @param message [MForwd::Message] def create_body message { text: message.body }.to_json end end
class MForwd::Deliver::Slack < MForwd::Deliver::HttpPost @@color_of_kind = { 'good' => /^(green$|ok$|good$|recover)/i, 'warning' => /^(yellow$|warn)/i, 'danger' => /^(red$|fatal$|crit|error$|danger$)/i, } private # @param message [MForwd::Message] def create_body message param = { attachments: [{ fallback: "#{message.title}\n#{message.body}", fields: [{ title: message.title, value: message.body, }], }], } @@color_of_kind.find { |col, regexp| regexp.match message.kind }.tap do |col, regexp| param[:attachments][0][:color] = col end param.to_json end end
Set rich format for Slack POST
Set rich format for Slack POST
Ruby
mit
key-amb/poloxy,key-amb/poloxy
ruby
## Code Before: class MForwd::Deliver::Slack < MForwd::Deliver::HttpPost private # @param message [MForwd::Message] def create_body message { text: message.body }.to_json end end ## Instruction: Set rich format for Slack POST ## Code After: class MForwd::Deliver::Slack < MForwd::Deliver::HttpPost @@color_of_kind = { 'good' => /^(green$|ok$|good$|recover)/i, 'warning' => /^(yellow$|warn)/i, 'danger' => /^(red$|fatal$|crit|error$|danger$)/i, } private # @param message [MForwd::Message] def create_body message param = { attachments: [{ fallback: "#{message.title}\n#{message.body}", fields: [{ title: message.title, value: message.body, }], }], } @@color_of_kind.find { |col, regexp| regexp.match message.kind }.tap do |col, regexp| param[:attachments][0][:color] = col end param.to_json end end
6f2b4ccdfa297605fd3d7be63039a8fcbda714d1
routes/titles.rb
routes/titles.rb
module Sinatra module TreeStats module Routing module Titles def self.registered(app) app.get '/titles/?' do @titles = TitleHelper::TITLES @titles = @titles.select { |i,t| i > 0 }.sort_by { |i,t| t } not_found("No titles found.") if @titles.length < 1 haml :titles end app.get '/title/:title' do |title_name| titles = TitleHelper::TITLES title = titles.select { |i,title| title == title_name } not_found("Title #{title} not found.") if title.length == 0 title_id = title.first[0] @title_name = title.first[1] @characters = Character.where(:titles.in => [title_id]).limit(100).project() haml :title end end end end end end
module Sinatra module TreeStats module Routing module Titles def self.registered(app) app.get '/titles/?' do @titles = TitleHelper::TITLES @titles = @titles.select { |i,t| i > 0 }.sort_by { |i,t| t } not_found("No titles found.") if @titles.length < 1 haml :titles end app.get '/title/:title' do |title_name| titles = TitleHelper::TITLES title = titles.select { |i,title| title == title_name } not_found("Title #{title} not found.") if title.length == 0 title_id = title.first[0] @title_name = title.first[1] @characters = Character.where(:titles.in => [title_id]).sort(:s => 1, :n => 1).limit(100) haml :title end end end end end end
Sort retrieved chars with title by server then player name
Sort retrieved chars with title by server then player name
Ruby
mit
amoeba/treestats.net,amoeba/treestats.net,amoeba/treestats.net
ruby
## Code Before: module Sinatra module TreeStats module Routing module Titles def self.registered(app) app.get '/titles/?' do @titles = TitleHelper::TITLES @titles = @titles.select { |i,t| i > 0 }.sort_by { |i,t| t } not_found("No titles found.") if @titles.length < 1 haml :titles end app.get '/title/:title' do |title_name| titles = TitleHelper::TITLES title = titles.select { |i,title| title == title_name } not_found("Title #{title} not found.") if title.length == 0 title_id = title.first[0] @title_name = title.first[1] @characters = Character.where(:titles.in => [title_id]).limit(100).project() haml :title end end end end end end ## Instruction: Sort retrieved chars with title by server then player name ## Code After: module Sinatra module TreeStats module Routing module Titles def self.registered(app) app.get '/titles/?' do @titles = TitleHelper::TITLES @titles = @titles.select { |i,t| i > 0 }.sort_by { |i,t| t } not_found("No titles found.") if @titles.length < 1 haml :titles end app.get '/title/:title' do |title_name| titles = TitleHelper::TITLES title = titles.select { |i,title| title == title_name } not_found("Title #{title} not found.") if title.length == 0 title_id = title.first[0] @title_name = title.first[1] @characters = Character.where(:titles.in => [title_id]).sort(:s => 1, :n => 1).limit(100) haml :title end end end end end end
d8eb5c32f51407b905c99ab43fcf65d441cd9f84
app/controllers/financials_controller.rb
app/controllers/financials_controller.rb
class FinancialsController < ProductController before_action :find_product! def index unless params[:product_id] == 'coderwall' redirect_to product_path(@product) end end end
class FinancialsController < ProductController before_action :find_product! def index unless @product.slug == 'coderwall' redirect_to product_path(@product) end end end
Check slug instead of path param
Check slug instead of path param
Ruby
agpl-3.0
lachlanjc/meta,assemblymade/meta,lachlanjc/meta,lachlanjc/meta,assemblymade/meta,lachlanjc/meta,assemblymade/meta,assemblymade/meta
ruby
## Code Before: class FinancialsController < ProductController before_action :find_product! def index unless params[:product_id] == 'coderwall' redirect_to product_path(@product) end end end ## Instruction: Check slug instead of path param ## Code After: class FinancialsController < ProductController before_action :find_product! def index unless @product.slug == 'coderwall' redirect_to product_path(@product) end end end
49fb91671dc758d0d692c6d8ac90bbb63678ad75
mc/Voyage-Model-Tests.package/VOArangoRepositoryTest.class/instance/setUp.st
mc/Voyage-Model-Tests.package/VOArangoRepositoryTest.class/instance/setUp.st
running setUp repository := VOArangoRepository uri: 'http://localhost:8529'
running setUp repository := VOArangoRepository uri: 'http://localhost:8529'. PQQueryBuilder generatorClass: PQArangoQueryGenerator.
Set up generator class for arango
Set up generator class for arango
Smalltalk
mit
pharo-nosql/voyage,tinchodias/voyage,tinchodias/voyage,noha/voyage,noha/voyage
smalltalk
## Code Before: running setUp repository := VOArangoRepository uri: 'http://localhost:8529' ## Instruction: Set up generator class for arango ## Code After: running setUp repository := VOArangoRepository uri: 'http://localhost:8529'. PQQueryBuilder generatorClass: PQArangoQueryGenerator.
f822b890d91d20ccb5a7da2935a1d4981e524c0f
app/views/home/show.html.erb
app/views/home/show.html.erb
<%= render 'shared/header' %> <main class="mainPageContent"> <div class="l-container"> <h1><%= t('.heading_html') %></h1> <ul class="specialList"><%= t('.summary_list_html') %></ul> <div id="introductionText"><%= t('.introduction_text_html') %></div> </div> </main> <%= render 'shared/footer_site_links', locals: alternative_locales %> <%= render 'shared/footer_social_links' %>
<%= render 'shared/header' %> <main class="mainPageContent"> <div class="l-container"> <h1><%= t('.heading_html') %></h1> <ul class="list--iconBullets"><%= t('.summary_list_html') %></ul> <div id="introductionText"><%= t('.introduction_text_html') %></div> </div> </main> <%= render 'shared/footer_site_links', locals: alternative_locales %> <%= render 'shared/footer_social_links' %>
Replace specialList with a more semantic name.
Replace specialList with a more semantic name.
HTML+ERB
mit
moneyadviceservice/frontend,moneyadviceservice/frontend,moneyadviceservice/frontend,moneyadviceservice/frontend
html+erb
## Code Before: <%= render 'shared/header' %> <main class="mainPageContent"> <div class="l-container"> <h1><%= t('.heading_html') %></h1> <ul class="specialList"><%= t('.summary_list_html') %></ul> <div id="introductionText"><%= t('.introduction_text_html') %></div> </div> </main> <%= render 'shared/footer_site_links', locals: alternative_locales %> <%= render 'shared/footer_social_links' %> ## Instruction: Replace specialList with a more semantic name. ## Code After: <%= render 'shared/header' %> <main class="mainPageContent"> <div class="l-container"> <h1><%= t('.heading_html') %></h1> <ul class="list--iconBullets"><%= t('.summary_list_html') %></ul> <div id="introductionText"><%= t('.introduction_text_html') %></div> </div> </main> <%= render 'shared/footer_site_links', locals: alternative_locales %> <%= render 'shared/footer_social_links' %>
faa739924994acc3d61459db9b75ffa81cd87bcb
examples/reset_password.html
examples/reset_password.html
<!DOCTYPE html> <head> <meta name="viewport" content="width=device-width, initial-scale=1"> </head> {% if error %} <p>{{ error }}</p> {% endif %} <form method="POST" action="/reset-password"> <label for="password">New Password</label> <input type="password" name="password"><br> <label for="confirm">Confirm Password</label> <input type="password" name="confirm"><br> <input type="hidden" name="code" value="{{ code }}"> <input type="hidden" name="user_id" value="{{ user_id }}"> <input type="hidden" name="expire_at" value="{{ expire_at }}"> <input type="submit" value="Submit"> </form>
<!DOCTYPE html> <head> <meta name="viewport" content="width=device-width, initial-scale=1"> </head> {% if error %} <p>{{ error }}</p> {% endif %} <form method="POST" action="/auth/fogot_password/reset_password"> <label for="password">New Password</label> <input type="password" name="password"><br> <label for="confirm">Confirm Password</label> <input type="password" name="confirm"><br> <input type="hidden" name="code" value="{{ code }}"> <input type="hidden" name="user_id" value="{{ user_id }}"> <input type="hidden" name="expire_at" value="{{ expire_at }}"> <input type="submit" value="Submit"> </form>
Update template reset password url
Update template reset password url
HTML
apache-2.0
SkygearIO/skygear-server,SkygearIO/skygear-server,SkygearIO/skygear-server,SkygearIO/skygear-server
html
## Code Before: <!DOCTYPE html> <head> <meta name="viewport" content="width=device-width, initial-scale=1"> </head> {% if error %} <p>{{ error }}</p> {% endif %} <form method="POST" action="/reset-password"> <label for="password">New Password</label> <input type="password" name="password"><br> <label for="confirm">Confirm Password</label> <input type="password" name="confirm"><br> <input type="hidden" name="code" value="{{ code }}"> <input type="hidden" name="user_id" value="{{ user_id }}"> <input type="hidden" name="expire_at" value="{{ expire_at }}"> <input type="submit" value="Submit"> </form> ## Instruction: Update template reset password url ## Code After: <!DOCTYPE html> <head> <meta name="viewport" content="width=device-width, initial-scale=1"> </head> {% if error %} <p>{{ error }}</p> {% endif %} <form method="POST" action="/auth/fogot_password/reset_password"> <label for="password">New Password</label> <input type="password" name="password"><br> <label for="confirm">Confirm Password</label> <input type="password" name="confirm"><br> <input type="hidden" name="code" value="{{ code }}"> <input type="hidden" name="user_id" value="{{ user_id }}"> <input type="hidden" name="expire_at" value="{{ expire_at }}"> <input type="submit" value="Submit"> </form>
e53ecef685569dfad2c62cd38c53190a9b2012d0
metpy/gridding/__init__.py
metpy/gridding/__init__.py
from .gridding_functions import * # noqa: F403 from .points import * # noqa: F403 from .triangles import * # noqa: F403 from .polygons import * # noqa: F403 from .interpolation import * # noqa: F403 __all__ = gridding_functions.__all__[:] # pylint: disable=undefined-variable __all__.extend(interpolation.__all__) # pylint: disable=undefined-variable
from .gridding_functions import * # noqa: F403 from .interpolation import * # noqa: F403 __all__ = gridding_functions.__all__[:] # pylint: disable=undefined-variable __all__.extend(interpolation.__all__) # pylint: disable=undefined-variable
Remove some imported modules from gridding
MNT: Remove some imported modules from gridding We never intended these to be part of the supported API. The fact that these were available from metpy.gridding was revealed by the new autosummary docs.
Python
bsd-3-clause
ShawnMurd/MetPy,dopplershift/MetPy,ahaberlie/MetPy,jrleeman/MetPy,ahaberlie/MetPy,dopplershift/MetPy,Unidata/MetPy,jrleeman/MetPy,Unidata/MetPy
python
## Code Before: from .gridding_functions import * # noqa: F403 from .points import * # noqa: F403 from .triangles import * # noqa: F403 from .polygons import * # noqa: F403 from .interpolation import * # noqa: F403 __all__ = gridding_functions.__all__[:] # pylint: disable=undefined-variable __all__.extend(interpolation.__all__) # pylint: disable=undefined-variable ## Instruction: MNT: Remove some imported modules from gridding We never intended these to be part of the supported API. The fact that these were available from metpy.gridding was revealed by the new autosummary docs. ## Code After: from .gridding_functions import * # noqa: F403 from .interpolation import * # noqa: F403 __all__ = gridding_functions.__all__[:] # pylint: disable=undefined-variable __all__.extend(interpolation.__all__) # pylint: disable=undefined-variable
6bdc335ebb8ed009cbc926e681ee6be8f475f323
reviewboard/reviews/features.py
reviewboard/reviews/features.py
"""Feature definitions for reviews.""" from __future__ import unicode_literals from django.utils.translation import ugettext_lazy as _ from djblets.features import Feature, FeatureLevel class GeneralCommentsFeature(Feature): """A feature for general comments. General comments allow comments to be created directly on a review request without accompanying file attachment or diff. These can be used to raise issues with the review request itself, such as its summary or description, or general implementation issues. """ feature_id = 'reviews.general_comments' name = _('General Comments') level = FeatureLevel.STABLE summary = _('Allow comments on review requests without an associated file ' 'attachment or diff.') class StatusUpdatesFeature(Feature): """A feature for status updates. A status update is a way for third-party tools to provide feedback on a review request. In the past, this was done just as a normal review. Status updates allow those tools (via some integration like Review Bot) to mark their state (such as pending, success, failure, or error) and then associate that with a review. """ feature_id = 'reviews.status_updates' name = _('Status Updates') summary = _('A way for external tools to do checks on a review request ' 'and report the results of those checks.') general_comments_feature = GeneralCommentsFeature() status_updates_feature = StatusUpdatesFeature()
"""Feature definitions for reviews.""" from __future__ import unicode_literals from django.utils.translation import ugettext_lazy as _ from djblets.features import Feature, FeatureLevel class GeneralCommentsFeature(Feature): """A feature for general comments. General comments allow comments to be created directly on a review request without accompanying file attachment or diff. These can be used to raise issues with the review request itself, such as its summary or description, or general implementation issues. """ feature_id = 'reviews.general_comments' name = _('General Comments') level = FeatureLevel.STABLE summary = _('Allow comments on review requests without an associated file ' 'attachment or diff.') class StatusUpdatesFeature(Feature): """A feature for status updates. A status update is a way for third-party tools to provide feedback on a review request. In the past, this was done just as a normal review. Status updates allow those tools (via some integration like Review Bot) to mark their state (such as pending, success, failure, or error) and then associate that with a review. """ feature_id = 'reviews.status_updates' name = _('Status Updates') level = FeatureLevel.STABLE summary = _('A way for external tools to do checks on a review request ' 'and report the results of those checks.') general_comments_feature = GeneralCommentsFeature() status_updates_feature = StatusUpdatesFeature()
Mark status updates as stable.
Mark status updates as stable. This allows people to actually use the feature without turning it on in their `settings_local.py`. Reviewed at https://reviews.reviewboard.org/r/8569/
Python
mit
sgallagher/reviewboard,brennie/reviewboard,chipx86/reviewboard,chipx86/reviewboard,chipx86/reviewboard,davidt/reviewboard,reviewboard/reviewboard,davidt/reviewboard,reviewboard/reviewboard,chipx86/reviewboard,brennie/reviewboard,sgallagher/reviewboard,davidt/reviewboard,sgallagher/reviewboard,davidt/reviewboard,brennie/reviewboard,sgallagher/reviewboard,reviewboard/reviewboard,brennie/reviewboard,reviewboard/reviewboard
python
## Code Before: """Feature definitions for reviews.""" from __future__ import unicode_literals from django.utils.translation import ugettext_lazy as _ from djblets.features import Feature, FeatureLevel class GeneralCommentsFeature(Feature): """A feature for general comments. General comments allow comments to be created directly on a review request without accompanying file attachment or diff. These can be used to raise issues with the review request itself, such as its summary or description, or general implementation issues. """ feature_id = 'reviews.general_comments' name = _('General Comments') level = FeatureLevel.STABLE summary = _('Allow comments on review requests without an associated file ' 'attachment or diff.') class StatusUpdatesFeature(Feature): """A feature for status updates. A status update is a way for third-party tools to provide feedback on a review request. In the past, this was done just as a normal review. Status updates allow those tools (via some integration like Review Bot) to mark their state (such as pending, success, failure, or error) and then associate that with a review. """ feature_id = 'reviews.status_updates' name = _('Status Updates') summary = _('A way for external tools to do checks on a review request ' 'and report the results of those checks.') general_comments_feature = GeneralCommentsFeature() status_updates_feature = StatusUpdatesFeature() ## Instruction: Mark status updates as stable. This allows people to actually use the feature without turning it on in their `settings_local.py`. Reviewed at https://reviews.reviewboard.org/r/8569/ ## Code After: """Feature definitions for reviews.""" from __future__ import unicode_literals from django.utils.translation import ugettext_lazy as _ from djblets.features import Feature, FeatureLevel class GeneralCommentsFeature(Feature): """A feature for general comments. General comments allow comments to be created directly on a review request without accompanying file attachment or diff. These can be used to raise issues with the review request itself, such as its summary or description, or general implementation issues. """ feature_id = 'reviews.general_comments' name = _('General Comments') level = FeatureLevel.STABLE summary = _('Allow comments on review requests without an associated file ' 'attachment or diff.') class StatusUpdatesFeature(Feature): """A feature for status updates. A status update is a way for third-party tools to provide feedback on a review request. In the past, this was done just as a normal review. Status updates allow those tools (via some integration like Review Bot) to mark their state (such as pending, success, failure, or error) and then associate that with a review. """ feature_id = 'reviews.status_updates' name = _('Status Updates') level = FeatureLevel.STABLE summary = _('A way for external tools to do checks on a review request ' 'and report the results of those checks.') general_comments_feature = GeneralCommentsFeature() status_updates_feature = StatusUpdatesFeature()
0a3495e269ff97b7f8e83d6f23c2bab81c88f99f
travis-ci.sh
travis-ci.sh
set -ex set | sort curl -L -o release-1.8.0.zip \ https://github.com/google/googletest/archive/release-1.8.0.zip unzip release-1.8.0.zip cd googletest-release-1.8.0 cmake -DCMAKE_PREFIX_PATH=.. . make make install cd .. mkdir build cd build cmake .. make make test
set -ex set | sort curl -L -o release-1.8.0.zip \ https://github.com/google/googletest/archive/release-1.8.0.zip unzip release-1.8.0.zip cd googletest-release-1.8.0 cmake -DCMAKE_INSTALL_PREFIX=.. . make make install cd .. mkdir build cd build cmake .. make make test
Fix CMake install paths in travis
Fix CMake install paths in travis
Shell
isc
DemiMarie/SlipRock,DemiMarie/SlipRock,DemiMarie/SlipRock,DemiMarie/SlipRock,DemiMarie/SlipRock
shell
## Code Before: set -ex set | sort curl -L -o release-1.8.0.zip \ https://github.com/google/googletest/archive/release-1.8.0.zip unzip release-1.8.0.zip cd googletest-release-1.8.0 cmake -DCMAKE_PREFIX_PATH=.. . make make install cd .. mkdir build cd build cmake .. make make test ## Instruction: Fix CMake install paths in travis ## Code After: set -ex set | sort curl -L -o release-1.8.0.zip \ https://github.com/google/googletest/archive/release-1.8.0.zip unzip release-1.8.0.zip cd googletest-release-1.8.0 cmake -DCMAKE_INSTALL_PREFIX=.. . make make install cd .. mkdir build cd build cmake .. make make test
2d8cfb3e35d0728314cd262c26157e835c2f8369
quickstart/src/main/resources/archetype-catalog.xml
quickstart/src/main/resources/archetype-catalog.xml
<?xml version="1.0" encoding="UTF-8"?> <archetype-catalog> <archetypes> <archetype> <groupId>org.apache.tapestry</groupId> <artifactId>quickstart</artifactId> <version>@project.version@</version> <repository>http://repository.apache.org/snapshots</repository> <description>Tapestry @project.version@ Quickstart Project</description> </archetype> </archetypes> </archetype-catalog>
<?xml version="1.0" encoding="UTF-8"?> <archetype-catalog> <archetypes> <archetype> <groupId>org.apache.tapestry</groupId> <artifactId>quickstart</artifactId> <version>@project.version@</version> <repository>http://repository.apache.org/snapshots</repository> <description>Tapestry 5 Quickstart Project</description> </archetype> </archetypes> </archetype-catalog>
Change the description for the quickstart archetype to just "Tapestry 5 Quickstart Project"
TAP5-1265: Change the description for the quickstart archetype to just "Tapestry 5 Quickstart Project" git-svn-id: d9b8539636d91aff9cd33ed5cd52a0cf73394897@992517 13f79535-47bb-0310-9956-ffa450edef68
XML
apache-2.0
agileowl/tapestry-5,apache/tapestry-5,apache/tapestry-5,agileowl/tapestry-5,apache/tapestry-5,apache/tapestry-5,apache/tapestry-5,agileowl/tapestry-5
xml
## Code Before: <?xml version="1.0" encoding="UTF-8"?> <archetype-catalog> <archetypes> <archetype> <groupId>org.apache.tapestry</groupId> <artifactId>quickstart</artifactId> <version>@project.version@</version> <repository>http://repository.apache.org/snapshots</repository> <description>Tapestry @project.version@ Quickstart Project</description> </archetype> </archetypes> </archetype-catalog> ## Instruction: TAP5-1265: Change the description for the quickstart archetype to just "Tapestry 5 Quickstart Project" git-svn-id: d9b8539636d91aff9cd33ed5cd52a0cf73394897@992517 13f79535-47bb-0310-9956-ffa450edef68 ## Code After: <?xml version="1.0" encoding="UTF-8"?> <archetype-catalog> <archetypes> <archetype> <groupId>org.apache.tapestry</groupId> <artifactId>quickstart</artifactId> <version>@project.version@</version> <repository>http://repository.apache.org/snapshots</repository> <description>Tapestry 5 Quickstart Project</description> </archetype> </archetypes> </archetype-catalog>
040c2f0f52a606daf9092175363193cc4aa2cee2
lib/multitenancy/rack/filter.rb
lib/multitenancy/rack/filter.rb
module Multitenancy class Filter def initialize(app) @app = app end def call(env) tenant = Tenant.new tenant_id, sub_tenant_id Multitenancy.with_tenant tenant do @app.call env end end private def tenant_id # rack converts X_FOO to HTTP_X_FOO id = env["HTTP_#{Multitenancy.tenant_header}"] if Multitenancy.tenant_header_regexp.blank? id else match = Multitenancy.tenant_header_regexp.match(id) match ? match[0] : id end end def subtenant_id id = env["HTTP_#{Multitenancy.sub_tenant_header}"] if Multitenancy.sub_tenant_header_regexp.blank? id else match = Multitenancy.sub_tenant_header_regexp.match(id) match ? match[0] : id end end end end
module Multitenancy class Filter def initialize(app) @app = app end def call(env) tenant = Tenant.new tenant_id(env), sub_tenant_id(env) Multitenancy.with_tenant tenant do @app.call env end end private def tenant_id(env) # rack converts X_FOO to HTTP_X_FOO id = env["HTTP_#{Multitenancy.tenant_header}"] if Multitenancy.tenant_header_regexp.blank? id else match = Multitenancy.tenant_header_regexp.match(id) match ? match[0] : id end end def sub_tenant_id(env) id = env["HTTP_#{Multitenancy.sub_tenant_header}"] if Multitenancy.sub_tenant_header_regexp.blank? id else match = Multitenancy.sub_tenant_header_regexp.match(id) match ? match[0] : id end end end end
Fix syntax errors in regexp support.
Fix syntax errors in regexp support.
Ruby
mit
85x14/multitenancy
ruby
## Code Before: module Multitenancy class Filter def initialize(app) @app = app end def call(env) tenant = Tenant.new tenant_id, sub_tenant_id Multitenancy.with_tenant tenant do @app.call env end end private def tenant_id # rack converts X_FOO to HTTP_X_FOO id = env["HTTP_#{Multitenancy.tenant_header}"] if Multitenancy.tenant_header_regexp.blank? id else match = Multitenancy.tenant_header_regexp.match(id) match ? match[0] : id end end def subtenant_id id = env["HTTP_#{Multitenancy.sub_tenant_header}"] if Multitenancy.sub_tenant_header_regexp.blank? id else match = Multitenancy.sub_tenant_header_regexp.match(id) match ? match[0] : id end end end end ## Instruction: Fix syntax errors in regexp support. ## Code After: module Multitenancy class Filter def initialize(app) @app = app end def call(env) tenant = Tenant.new tenant_id(env), sub_tenant_id(env) Multitenancy.with_tenant tenant do @app.call env end end private def tenant_id(env) # rack converts X_FOO to HTTP_X_FOO id = env["HTTP_#{Multitenancy.tenant_header}"] if Multitenancy.tenant_header_regexp.blank? id else match = Multitenancy.tenant_header_regexp.match(id) match ? match[0] : id end end def sub_tenant_id(env) id = env["HTTP_#{Multitenancy.sub_tenant_header}"] if Multitenancy.sub_tenant_header_regexp.blank? id else match = Multitenancy.sub_tenant_header_regexp.match(id) match ? match[0] : id end end end end
75b5fdd2b6627081bd1eb095169ed5f8ac6f3090
about_me/index.html
about_me/index.html
<!doctype html> <html lang="en"> <head> <meta charset="utf-8"> <title>CS290: Homework</title> </head> <body> <main> <img alt="Andy Chase" src="portrait.jpeg" /><br /> <a href="mailto:[email protected]">[email protected]</a> </main> </body> </html>
<!doctype html> <html lang="en"> <head> <meta charset="utf-8"> <title>CS290: Homework</title> </head> <body> <main> <img alt="Andy Chase" src="portrait.jpeg"/><br/> <a href="mailto:[email protected]">[email protected]</a> <div class="schedule"> <h2>My schedule</h2> <table border="1"> <thead> <tr> <th></th> <th>Mon</th> <th>Tue</th> <th>Wed</th> <th>Thurs</th> <th>Fri</th> </tr> </thead> <tbody> <tr> <th scope="row">10a</th> <td></td> <td>CS 419</td> <td></td> <td>CS 419 </td> <td></td> </tr> <tr> <th>11a</th> <td></td> <td></td> <td></td> <td></td> <td></td> </tr> <tr> <th>12p</th> <td></td> <td>CS 461</td> <td></td> <td>CS 461</td> <td></td> </tr> <tr> <th>1p</th> <td></td> <td></td> <td></td> <td></td> <td></td> </tr> <tr> <th>2p</th> <td>CS 290</td> <td></td> <td>CS 290</td> <td></td> <td>CS 290</td> </tr> </tbody> </table> </div> </main> </body> </html>
Add schedule to "about me"
Add schedule to "about me"
HTML
mit
andychase/classwork,andychase/classwork,andychase/classwork,andychase/classwork,andychase/classwork,andychase/classwork,andychase/classwork,andychase/classwork
html
## Code Before: <!doctype html> <html lang="en"> <head> <meta charset="utf-8"> <title>CS290: Homework</title> </head> <body> <main> <img alt="Andy Chase" src="portrait.jpeg" /><br /> <a href="mailto:[email protected]">[email protected]</a> </main> </body> </html> ## Instruction: Add schedule to "about me" ## Code After: <!doctype html> <html lang="en"> <head> <meta charset="utf-8"> <title>CS290: Homework</title> </head> <body> <main> <img alt="Andy Chase" src="portrait.jpeg"/><br/> <a href="mailto:[email protected]">[email protected]</a> <div class="schedule"> <h2>My schedule</h2> <table border="1"> <thead> <tr> <th></th> <th>Mon</th> <th>Tue</th> <th>Wed</th> <th>Thurs</th> <th>Fri</th> </tr> </thead> <tbody> <tr> <th scope="row">10a</th> <td></td> <td>CS 419</td> <td></td> <td>CS 419 </td> <td></td> </tr> <tr> <th>11a</th> <td></td> <td></td> <td></td> <td></td> <td></td> </tr> <tr> <th>12p</th> <td></td> <td>CS 461</td> <td></td> <td>CS 461</td> <td></td> </tr> <tr> <th>1p</th> <td></td> <td></td> <td></td> <td></td> <td></td> </tr> <tr> <th>2p</th> <td>CS 290</td> <td></td> <td>CS 290</td> <td></td> <td>CS 290</td> </tr> </tbody> </table> </div> </main> </body> </html>
a887308a086c3cfa32e0802b3f3c950499347399
js/locales/bootstrap-datepicker.es.js
js/locales/bootstrap-datepicker.es.js
/** * Spanish translation for bootstrap-datepicker * Bruno Bonamin <[email protected]> */ ;(function($){ $.fn.datepicker.dates['es'] = { days: ["Domingo", "Lunes", "Martes", "Miércoles", "Jueves", "Viernes", "Sábado", "Domingo"], daysShort: ["Dom", "Lun", "Mar", "Mié", "Jue", "Vie", "Sáb", "Dom"], daysMin: ["Do", "Lu", "Ma", "Mi", "Ju", "Vi", "Sa", "Do"], months: ["Enero", "Febrero", "Marzo", "Abril", "Mayo", "Junio", "Julio", "Agosto", "Septiembre", "Octubre", "Noviembre", "Diciembre"], monthsShort: ["Ene", "Feb", "Mar", "Abr", "May", "Jun", "Jul", "Ago", "Sep", "Oct", "Nov", "Dic"], today: "Hoy" }; }(jQuery));
/** * Spanish translation for bootstrap-datepicker * Bruno Bonamin <[email protected]> */ ;(function($){ $.fn.datepicker.dates['es'] = { days: ["Domingo", "Lunes", "Martes", "Miércoles", "Jueves", "Viernes", "Sábado", "Domingo"], daysShort: ["Dom", "Lun", "Mar", "Mié", "Jue", "Vie", "Sáb", "Dom"], daysMin: ["Do", "Lu", "Ma", "Mi", "Ju", "Vi", "Sa", "Do"], months: ["Enero", "Febrero", "Marzo", "Abril", "Mayo", "Junio", "Julio", "Agosto", "Septiembre", "Octubre", "Noviembre", "Diciembre"], monthsShort: ["Ene", "Feb", "Mar", "Abr", "May", "Jun", "Jul", "Ago", "Sep", "Oct", "Nov", "Dic"], today: "Hoy", clear: "Limpiar", weekStart: 1, format: "dd.mm.yyyy" }; }(jQuery));
Add clear, week start and format.
Add clear, week start and format.
JavaScript
apache-2.0
NFC-DITO/bootstrap-datepicker,nerionavea/bootstrap-datepicker,kiwiupover/bootstrap-datepicker,aldano/bootstrap-datepicker,inway/bootstrap-datepicker,janusnic/bootstrap-datepicker,Tekzenit/bootstrap-datepicker,josegomezr/bootstrap-datepicker,defrian8/bootstrap-datepicker,vegardok/bootstrap-datepicker,otnavle/bootstrap-datepicker,xutongtong/bootstrap-datepicker,jhalak/bootstrap-datepicker,inway/bootstrap-datepicker,Invulner/bootstrap-datepicker,gn0st1k4m/bootstrap-datepicker,acrobat/bootstrap-datepicker,kiwiupover/bootstrap-datepicker,CherryDT/bootstrap-datepicker,wearespindle/datepicker-js,hebbet/bootstrap-datepicker,Azaret/bootstrap-datepicker,SimpleUpdates/bootstrap-datepicker,christophercaburog/bootstrap-datepicker,kevintvh/bootstrap-datepicker,inukshuk/bootstrap-datepicker,mreiden/bootstrap-datepicker,Invulner/bootstrap-datepicker,yangyichen/bootstrap-datepicker,rocLv/bootstrap-datepicker,champierre/bootstrap-datepicker,dckesler/bootstrap-datepicker,vsn4ik/bootstrap-datepicker,steffendietz/bootstrap-datepicker,menatoric59/bootstrap-datepicker,osama9/bootstrap-datepicker,saurabh27125/bootstrap-datepicker,jhalak/bootstrap-datepicker,dswitzer/bootstrap-datepicker,bitzesty/bootstrap-datepicker,stenmuchow/bootstrap-datepicker,rocLv/bootstrap-datepicker,wetet2/bootstrap-datepicker,saurabh27125/bootstrap-datepicker,i-am-kenny/bootstrap-datepicker,huang-x-h/bootstrap-datepicker,eternicode/bootstrap-datepicker,inukshuk/bootstrap-datepicker,MartijnDwars/bootstrap-datepicker,pacozaa/bootstrap-datepicker,parkeugene/bootstrap-datepicker,thetimbanks/bootstrap-datepicker,gn0st1k4m/bootstrap-datepicker,joubertredrat/bootstrap-datepicker,zhaoyan158567/bootstrap-datepicker,josegomezr/bootstrap-datepicker,vegardok/bootstrap-datepicker,dckesler/bootstrap-datepicker,MartijnDwars/bootstrap-datepicker,martinRjs/bootstrap-datepicker,Mteuahasan/bootstrap-datepicker,Mteuahasan/bootstrap-datepicker,inspur-tobacco/bootstrap-datepicker,acrobat/bootstrap-datepicker,wulinmengzhu/bootstrap-datepicker,thetimbanks/bootstrap-datepicker,ibcooley/bootstrap-datepicker,darluc/bootstrap-datepicker,vgrish/bootstrap-datepicker,dswitzer/bootstrap-datepicker,HuBandiT/bootstrap-datepicker,riyan8250/bootstrap-datepicker,keiwerkgvr/bootstrap-datepicker,NFC-DITO/bootstrap-datepicker,daniyel/bootstrap-datepicker,cherylyan/bootstrap-datepicker,nilbus/bootstrap-datepicker,zhaoyan158567/bootstrap-datepicker,daniyel/bootstrap-datepicker,mreiden/bootstrap-datepicker,Doublemedia/bootstrap-datepicker,ibcooley/bootstrap-datepicker,cbryer/bootstrap-datepicker,stenmuchow/bootstrap-datepicker,christophercaburog/bootstrap-datepicker,ashleymcdonald/bootstrap-datepicker,chengky/bootstrap-datepicker,SimpleUpdates/bootstrap-datepicker,HuBandiT/bootstrap-datepicker,josegomezr/bootstrap-datepicker,defrian8/bootstrap-datepicker,bitzesty/bootstrap-datepicker,aldano/bootstrap-datepicker,nilbus/bootstrap-datepicker,CherryDT/bootstrap-datepicker,rstone770/bootstrap-datepicker,cbryer/bootstrap-datepicker,Sprinkle7/bootstrap-datepicker,chengky/bootstrap-datepicker,janusnic/bootstrap-datepicker,i-am-kenny/bootstrap-datepicker,uxsolutions/bootstrap-datepicker,Azaret/bootstrap-datepicker,Sprinkle7/bootstrap-datepicker,CheapSk8/bootstrap-datepicker,ashleymcdonald/bootstrap-datepicker,vgrish/bootstrap-datepicker,menatoric59/bootstrap-datepicker,joubertredrat/bootstrap-datepicker,otnavle/bootstrap-datepicker,jesperronn/bootstrap-datepicker,WeiLend/bootstrap-datepicker,huang-x-h/bootstrap-datepicker,wetet2/bootstrap-datepicker,pacozaa/bootstrap-datepicker,champierre/bootstrap-datepicker,wulinmengzhu/bootstrap-datepicker,Tekzenit/bootstrap-datepicker,osama9/bootstrap-datepicker,yangyichen/bootstrap-datepicker,tcrossland/bootstrap-datepicker,Doublemedia/bootstrap-datepicker,HNygard/bootstrap-datepicker,uxsolutions/bootstrap-datepicker,riyan8250/bootstrap-datepicker,parkeugene/bootstrap-datepicker,Youraiseme/bootstrap-datepicker,keiwerkgvr/bootstrap-datepicker,WeiLend/bootstrap-datepicker,martinRjs/bootstrap-datepicker,steffendietz/bootstrap-datepicker,kevintvh/bootstrap-datepicker,HNygard/bootstrap-datepicker,rstone770/bootstrap-datepicker,xutongtong/bootstrap-datepicker,Youraiseme/bootstrap-datepicker,cherylyan/bootstrap-datepicker,darluc/bootstrap-datepicker,eternicode/bootstrap-datepicker,vsn4ik/bootstrap-datepicker,jesperronn/bootstrap-datepicker,nerionavea/bootstrap-datepicker,hemp/bootstrap-datepicker,CheapSk8/bootstrap-datepicker,hemp/bootstrap-datepicker,inspur-tobacco/bootstrap-datepicker,tcrossland/bootstrap-datepicker,hebbet/bootstrap-datepicker
javascript
## Code Before: /** * Spanish translation for bootstrap-datepicker * Bruno Bonamin <[email protected]> */ ;(function($){ $.fn.datepicker.dates['es'] = { days: ["Domingo", "Lunes", "Martes", "Miércoles", "Jueves", "Viernes", "Sábado", "Domingo"], daysShort: ["Dom", "Lun", "Mar", "Mié", "Jue", "Vie", "Sáb", "Dom"], daysMin: ["Do", "Lu", "Ma", "Mi", "Ju", "Vi", "Sa", "Do"], months: ["Enero", "Febrero", "Marzo", "Abril", "Mayo", "Junio", "Julio", "Agosto", "Septiembre", "Octubre", "Noviembre", "Diciembre"], monthsShort: ["Ene", "Feb", "Mar", "Abr", "May", "Jun", "Jul", "Ago", "Sep", "Oct", "Nov", "Dic"], today: "Hoy" }; }(jQuery)); ## Instruction: Add clear, week start and format. ## Code After: /** * Spanish translation for bootstrap-datepicker * Bruno Bonamin <[email protected]> */ ;(function($){ $.fn.datepicker.dates['es'] = { days: ["Domingo", "Lunes", "Martes", "Miércoles", "Jueves", "Viernes", "Sábado", "Domingo"], daysShort: ["Dom", "Lun", "Mar", "Mié", "Jue", "Vie", "Sáb", "Dom"], daysMin: ["Do", "Lu", "Ma", "Mi", "Ju", "Vi", "Sa", "Do"], months: ["Enero", "Febrero", "Marzo", "Abril", "Mayo", "Junio", "Julio", "Agosto", "Septiembre", "Octubre", "Noviembre", "Diciembre"], monthsShort: ["Ene", "Feb", "Mar", "Abr", "May", "Jun", "Jul", "Ago", "Sep", "Oct", "Nov", "Dic"], today: "Hoy", clear: "Limpiar", weekStart: 1, format: "dd.mm.yyyy" }; }(jQuery));
ef156b7e8fed483f33be8243a627363984aee96c
blueprints/in-repo-addon/files/lib/__name__/index.js
blueprints/in-repo-addon/files/lib/__name__/index.js
/*jshint node:true*/ module.exports = { name: '<%= dasherizedModuleName %>', isDevelopingAddon: function() { return true; } };
/*jshint node:true*/ 'use strict'; module.exports = { name: '<%= dasherizedModuleName %>', isDevelopingAddon: function() { return true; } };
Make in-repo-addon blueprint 'use strict'.
Make in-repo-addon blueprint 'use strict'.
JavaScript
mit
kellyselden/ember-cli,Turbo87/ember-cli,jrjohnson/ember-cli,mike-north/ember-cli,HeroicEric/ember-cli,Turbo87/ember-cli,ef4/ember-cli,calderas/ember-cli,seawatts/ember-cli,sivakumar-kailasam/ember-cli,kategengler/ember-cli,kriswill/ember-cli,asakusuma/ember-cli,thoov/ember-cli,akatov/ember-cli,gfvcastro/ember-cli,patocallaghan/ember-cli,buschtoens/ember-cli,ef4/ember-cli,pzuraq/ember-cli,jgwhite/ember-cli,sivakumar-kailasam/ember-cli,twokul/ember-cli,kellyselden/ember-cli,kellyselden/ember-cli,HeroicEric/ember-cli,mike-north/ember-cli,Turbo87/ember-cli,calderas/ember-cli,rtablada/ember-cli,seawatts/ember-cli,patocallaghan/ember-cli,patocallaghan/ember-cli,kellyselden/ember-cli,cibernox/ember-cli,Turbo87/ember-cli,jgwhite/ember-cli,akatov/ember-cli,pzuraq/ember-cli,fpauser/ember-cli,romulomachado/ember-cli,patocallaghan/ember-cli,trentmwillis/ember-cli,calderas/ember-cli,romulomachado/ember-cli,rtablada/ember-cli,twokul/ember-cli,gfvcastro/ember-cli,asakusuma/ember-cli,mike-north/ember-cli,cibernox/ember-cli,romulomachado/ember-cli,mike-north/ember-cli,thoov/ember-cli,kriswill/ember-cli,ef4/ember-cli,fpauser/ember-cli,twokul/ember-cli,fpauser/ember-cli,trentmwillis/ember-cli,trentmwillis/ember-cli,raycohen/ember-cli,ember-cli/ember-cli,trentmwillis/ember-cli,elwayman02/ember-cli,calderas/ember-cli,kanongil/ember-cli,ember-cli/ember-cli,ember-cli/ember-cli,jrjohnson/ember-cli,kanongil/ember-cli,seawatts/ember-cli,balinterdi/ember-cli,balinterdi/ember-cli,buschtoens/ember-cli,pzuraq/ember-cli,kriswill/ember-cli,HeroicEric/ember-cli,rtablada/ember-cli,raycohen/ember-cli,fpauser/ember-cli,seawatts/ember-cli,twokul/ember-cli,pzuraq/ember-cli,kriswill/ember-cli,jgwhite/ember-cli,gfvcastro/ember-cli,thoov/ember-cli,kategengler/ember-cli,jgwhite/ember-cli,akatov/ember-cli,kanongil/ember-cli,thoov/ember-cli,ef4/ember-cli,romulomachado/ember-cli,gfvcastro/ember-cli,sivakumar-kailasam/ember-cli,HeroicEric/ember-cli,elwayman02/ember-cli,cibernox/ember-cli,kanongil/ember-cli,rtablada/ember-cli,akatov/ember-cli,cibernox/ember-cli,sivakumar-kailasam/ember-cli
javascript
## Code Before: /*jshint node:true*/ module.exports = { name: '<%= dasherizedModuleName %>', isDevelopingAddon: function() { return true; } }; ## Instruction: Make in-repo-addon blueprint 'use strict'. ## Code After: /*jshint node:true*/ 'use strict'; module.exports = { name: '<%= dasherizedModuleName %>', isDevelopingAddon: function() { return true; } };
9f79c567d0f46f02f77ea516180801b25fcafd6a
lib/nmap/run_stat.rb
lib/nmap/run_stat.rb
module Nmap # # Represents the runstats of a scan. # class RunStat < Struct.new(:end_time, :elapsed, :summary, :exit_status) # # Converts the stats to a String. # # @return [String] # The String form of the scan. # def to_s "#{self.end_time} #{self.elapsed} #{self.exit_status}" end end end
module Nmap # # Represents the runstats of a scan. # # @since 0.7.0 # class RunStat < Struct.new(:end_time, :elapsed, :summary, :exit_status) # # Converts the stats to a String. # # @return [String] # The String form of the scan. # def to_s "#{self.end_time} #{self.elapsed} #{self.exit_status}" end end end
Mark RunStat as being added in 0.7.0.
Mark RunStat as being added in 0.7.0.
Ruby
mit
squid108/ruby-nmap,sophsec/ruby-nmap
ruby
## Code Before: module Nmap # # Represents the runstats of a scan. # class RunStat < Struct.new(:end_time, :elapsed, :summary, :exit_status) # # Converts the stats to a String. # # @return [String] # The String form of the scan. # def to_s "#{self.end_time} #{self.elapsed} #{self.exit_status}" end end end ## Instruction: Mark RunStat as being added in 0.7.0. ## Code After: module Nmap # # Represents the runstats of a scan. # # @since 0.7.0 # class RunStat < Struct.new(:end_time, :elapsed, :summary, :exit_status) # # Converts the stats to a String. # # @return [String] # The String form of the scan. # def to_s "#{self.end_time} #{self.elapsed} #{self.exit_status}" end end end
c1330b3e685ca1ab4bc41ab16bc47631afab887a
lib/mpd-child-process.js
lib/mpd-child-process.js
var Q = require('Q'), childProcess = require('child_process'); function processPath(exec) { var exec = exec || Q.denodeify(childProcess.exec); var deferred = Q.defer(); function handleDone(stdout, stderr) { var location = stdout[0].trim().replace(/\n$/, ''); deferred.resolve(location); } function handleError(error) { deferred.reject(error); } exec('which mpd').done(handleDone, handleError); return deferred.promise; } function create(configFilepath, spawn) { var spawn = spawn || childProcess.spawn; var dfd = Q.defer(); this.processPath().then(function (path) { var mpdProcess = spawn(path, [configFilepath, '--no-daemon']); mpdProcess.stdout.on('data', function (data) { dfd.resolve(); }); mpdProcess.on('error', function (error) { dfd.reject(); }); }, function() { dfd.reject(); }); return dfd.promise; } module.exports = { processPath: processPath, create: create };
var Q = require('Q'), childProcess = require('child_process'); function processPath(exec) { var exec = exec || Q.denodeify(childProcess.exec); var deferred = Q.defer(); function handleDone(stdout, stderr) { var location = stdout[0].trim().replace(/\n$/, ''); deferred.resolve(location); } function handleError(error) { deferred.reject(error); } exec('which mpd').done(handleDone, handleError); return deferred.promise; } function create(configFilepath, spawn) { var spawn = spawn || childProcess.spawn; var dfd = Q.defer(); this.processPath().then(function (path) { var mpdProcess = spawn(path, [configFilepath, '--no-daemon']); mpdProcess.stdout.on('data', function (data) { dfd.resolve(); }); mpdProcess.on('error', function (error) { dfd.reject(); }); }, function() { dfd.reject('Could not find MPD binary'); }); return dfd.promise; } module.exports = { processPath: processPath, create: create };
Add failure message when MPD path is not found
Add failure message when MPD path is not found
JavaScript
apache-2.0
radiodan/radiodan.js,radiodan/radiodan.js
javascript
## Code Before: var Q = require('Q'), childProcess = require('child_process'); function processPath(exec) { var exec = exec || Q.denodeify(childProcess.exec); var deferred = Q.defer(); function handleDone(stdout, stderr) { var location = stdout[0].trim().replace(/\n$/, ''); deferred.resolve(location); } function handleError(error) { deferred.reject(error); } exec('which mpd').done(handleDone, handleError); return deferred.promise; } function create(configFilepath, spawn) { var spawn = spawn || childProcess.spawn; var dfd = Q.defer(); this.processPath().then(function (path) { var mpdProcess = spawn(path, [configFilepath, '--no-daemon']); mpdProcess.stdout.on('data', function (data) { dfd.resolve(); }); mpdProcess.on('error', function (error) { dfd.reject(); }); }, function() { dfd.reject(); }); return dfd.promise; } module.exports = { processPath: processPath, create: create }; ## Instruction: Add failure message when MPD path is not found ## Code After: var Q = require('Q'), childProcess = require('child_process'); function processPath(exec) { var exec = exec || Q.denodeify(childProcess.exec); var deferred = Q.defer(); function handleDone(stdout, stderr) { var location = stdout[0].trim().replace(/\n$/, ''); deferred.resolve(location); } function handleError(error) { deferred.reject(error); } exec('which mpd').done(handleDone, handleError); return deferred.promise; } function create(configFilepath, spawn) { var spawn = spawn || childProcess.spawn; var dfd = Q.defer(); this.processPath().then(function (path) { var mpdProcess = spawn(path, [configFilepath, '--no-daemon']); mpdProcess.stdout.on('data', function (data) { dfd.resolve(); }); mpdProcess.on('error', function (error) { dfd.reject(); }); }, function() { dfd.reject('Could not find MPD binary'); }); return dfd.promise; } module.exports = { processPath: processPath, create: create };
bf514f23138cf1fbf4971341ad4d9aff915884cc
wiki/README.md
wiki/README.md
You may have to create a docker volume for MySQL first: docker volume create mysql-data To start the services, just run this command inside the `wiki` directory: docker-compose up
You may have to create a docker volume for MySQL first: docker volume create mysql-data To start the services, just run this command inside the `wiki` directory: docker-compose up To view the wiki page locally on mac: In another terminal window, inside the `wiki` directory, connect to the running docker-machine and then list its IP address: docker-machine env eval $(docker-machine env) Locate the IP address in the DOCKER_HOST variable. Enter this IP address in your browser, followed by port 8080. Eg: 192.168.99.100:8080
Add instructions to view wiki on mac
Add instructions to view wiki on mac
Markdown
mit
oc4d/main,oc4d/main,oc4d/main
markdown
## Code Before: You may have to create a docker volume for MySQL first: docker volume create mysql-data To start the services, just run this command inside the `wiki` directory: docker-compose up ## Instruction: Add instructions to view wiki on mac ## Code After: You may have to create a docker volume for MySQL first: docker volume create mysql-data To start the services, just run this command inside the `wiki` directory: docker-compose up To view the wiki page locally on mac: In another terminal window, inside the `wiki` directory, connect to the running docker-machine and then list its IP address: docker-machine env eval $(docker-machine env) Locate the IP address in the DOCKER_HOST variable. Enter this IP address in your browser, followed by port 8080. Eg: 192.168.99.100:8080
c15a83118c98a5746136167768975ad97c733205
app/views/characters/show.haml
app/views/characters/show.haml
- content_for :breadcrumbs do = link_to "Characters", characters_path &raquo; %b= @character.name - if @character.icon = icon_tag @character.icon %br %b= @character.screenname %br = @character.name - if @character.template = surround '(', ')' do = link_to @character.template.name, template_path(@character.template) - if @character.pb %br Played by: = @character.pb - if logged_in? && current_user.id == @character.user_id %br = link_to 'Edit character', edit_character_path(@character) %br %br - if @character.recent_posts.present? = render partial: 'posts/list', locals: {posts: @character.recent_posts, title: 'Recent Threads' }
- content_for :breadcrumbs do = link_to "Characters", characters_path &raquo; %b= @character.name - if @character.icon = icon_tag @character.icon %br %b= @character.screenname %br = @character.name - if @character.template = surround '(', ')' do = link_to @character.template.name, template_path(@character.template) - if @character.template_name %br Nickname: #{@character.template_name} - if @character.pb %br Played by: = @character.pb - if logged_in? && current_user.id == @character.user_id %br = link_to 'Edit character', edit_character_path(@character) %br %br - if @character.recent_posts.present? = render partial: 'posts/list', locals: {posts: @character.recent_posts, title: 'Recent Threads' }
Add nickname to character page
Add nickname to character page
Haml
mit
Marri/glowfic,Marri/glowfic,Marri/glowfic,Marri/glowfic
haml
## Code Before: - content_for :breadcrumbs do = link_to "Characters", characters_path &raquo; %b= @character.name - if @character.icon = icon_tag @character.icon %br %b= @character.screenname %br = @character.name - if @character.template = surround '(', ')' do = link_to @character.template.name, template_path(@character.template) - if @character.pb %br Played by: = @character.pb - if logged_in? && current_user.id == @character.user_id %br = link_to 'Edit character', edit_character_path(@character) %br %br - if @character.recent_posts.present? = render partial: 'posts/list', locals: {posts: @character.recent_posts, title: 'Recent Threads' } ## Instruction: Add nickname to character page ## Code After: - content_for :breadcrumbs do = link_to "Characters", characters_path &raquo; %b= @character.name - if @character.icon = icon_tag @character.icon %br %b= @character.screenname %br = @character.name - if @character.template = surround '(', ')' do = link_to @character.template.name, template_path(@character.template) - if @character.template_name %br Nickname: #{@character.template_name} - if @character.pb %br Played by: = @character.pb - if logged_in? && current_user.id == @character.user_id %br = link_to 'Edit character', edit_character_path(@character) %br %br - if @character.recent_posts.present? = render partial: 'posts/list', locals: {posts: @character.recent_posts, title: 'Recent Threads' }
4b91cb368de4eaa86f5ca1c7bf03784252c68077
spec/support/generator_spec_support.rb
spec/support/generator_spec_support.rb
module GeneratorSpecSupport def assert_expectation_file(path, expectation_filepath = nil) expectation_filepath ||= File.expand_path("../../generators/expectations/#{path}", __FILE__) expectation_filepath.sub!(/_spec\.rb\z/, '_spek.rb') assert_file(path, File.read(expectation_filepath)) end end RSpec.configure do |config| config.include GeneratorSpecSupport, type: :generator end
module GeneratorSpecSupport def assert_expectation_file(path, expectation_filepath = nil) expectation_filepath ||= File.expand_path("../../../examples/#{path}", __FILE__) expectation_filepath.sub!(/_spec\.rb\z/, '_spek.rb') assert_file(path, File.read(expectation_filepath)) end end RSpec.configure do |config| config.include GeneratorSpecSupport, type: :generator end
Use examples instead of spec/generators/expectations
Use examples instead of spec/generators/expectations
Ruby
mit
akm/model_base_generators,akm/model_base_generators,akm/model_base_generators
ruby
## Code Before: module GeneratorSpecSupport def assert_expectation_file(path, expectation_filepath = nil) expectation_filepath ||= File.expand_path("../../generators/expectations/#{path}", __FILE__) expectation_filepath.sub!(/_spec\.rb\z/, '_spek.rb') assert_file(path, File.read(expectation_filepath)) end end RSpec.configure do |config| config.include GeneratorSpecSupport, type: :generator end ## Instruction: Use examples instead of spec/generators/expectations ## Code After: module GeneratorSpecSupport def assert_expectation_file(path, expectation_filepath = nil) expectation_filepath ||= File.expand_path("../../../examples/#{path}", __FILE__) expectation_filepath.sub!(/_spec\.rb\z/, '_spek.rb') assert_file(path, File.read(expectation_filepath)) end end RSpec.configure do |config| config.include GeneratorSpecSupport, type: :generator end
aaf80ab5b3700b32b20e0b3d069462c50d1a0aca
scgi.gemspec
scgi.gemspec
spec = Gem::Specification.new do |s| s.name = "scgi" s.version = "0.9.0" s.author = "Jeremy Evans" s.email = "[email protected]" s.platform = Gem::Platform::RUBY s.summary = "Simple support for using SCGI in ruby apps, such as Rails" s.files = %w'LICENSE README lib/scgi.rb' s.extra_rdoc_files = ['README'] s.require_paths = ["lib"] s.has_rdoc = true s.rdoc_options = %w'--inline-source --line-numbers' s.rubyforge_project = 'scgi' end
spec = Gem::Specification.new do |s| s.name = "scgi" s.version = "0.9.0" s.author = "Jeremy Evans" s.email = "[email protected]" s.platform = Gem::Platform::RUBY s.summary = "Simple support for using SCGI in ruby apps, such as Rails" s.files = %w'LICENSE README lib/scgi.rb' s.require_paths = ["lib"] s.has_rdoc = true s.rdoc_options = %w'--inline-source --line-numbers README lib' s.rubyforge_project = 'scgi' end
Make README the default RDoc page
Make README the default RDoc page
Ruby
mit
jeremyevans/ruby-scgi
ruby
## Code Before: spec = Gem::Specification.new do |s| s.name = "scgi" s.version = "0.9.0" s.author = "Jeremy Evans" s.email = "[email protected]" s.platform = Gem::Platform::RUBY s.summary = "Simple support for using SCGI in ruby apps, such as Rails" s.files = %w'LICENSE README lib/scgi.rb' s.extra_rdoc_files = ['README'] s.require_paths = ["lib"] s.has_rdoc = true s.rdoc_options = %w'--inline-source --line-numbers' s.rubyforge_project = 'scgi' end ## Instruction: Make README the default RDoc page ## Code After: spec = Gem::Specification.new do |s| s.name = "scgi" s.version = "0.9.0" s.author = "Jeremy Evans" s.email = "[email protected]" s.platform = Gem::Platform::RUBY s.summary = "Simple support for using SCGI in ruby apps, such as Rails" s.files = %w'LICENSE README lib/scgi.rb' s.require_paths = ["lib"] s.has_rdoc = true s.rdoc_options = %w'--inline-source --line-numbers README lib' s.rubyforge_project = 'scgi' end
461281865ccd7521f25b2d4e893eb55c036f56dc
static/js/crowdsource.interceptor.js
static/js/crowdsource.interceptor.js
/** * Date: 1/9/15 * Author: Shirish Goyal */ (function () { 'use strict'; angular .module('crowdsource.interceptor', []) .factory('AuthHttpResponseInterceptor', AuthHttpResponseInterceptor); AuthHttpResponseInterceptor.$inject = ['$location', '$log', '$injector', '$q']; function AuthHttpResponseInterceptor($location, $log, $injector, $q) { return { responseError: function (rejection) { if (rejection.status === 403) { if (rejection.hasOwnProperty('data') && rejection.data.hasOwnProperty('detail') && (rejection.data.detail.indexOf("Authentication credentials were not provided") != -1) ) { var $http = $injector.get('$http'); var Authentication = $injector.get('Authentication'); if (!Authentication.isAuthenticated()) { $location.path('/login'); }else{ deferred.resolve(data); } } } return $q.reject(rejection); } } } })();
/** * Date: 1/9/15 * Author: Shirish Goyal */ (function () { 'use strict'; angular .module('crowdsource.interceptor', []) .factory('AuthHttpResponseInterceptor', AuthHttpResponseInterceptor); AuthHttpResponseInterceptor.$inject = ['$location', '$log', '$injector', '$q']; function AuthHttpResponseInterceptor($location, $log, $injector, $q) { return { responseError: function (rejection) { if (rejection.status === 403) { if (rejection.hasOwnProperty('data') && rejection.data.hasOwnProperty('detail') && (rejection.data.detail.indexOf("Authentication credentials were not provided") != -1) ) { var $http = $injector.get('$http'); var Authentication = $injector.get('Authentication'); if (!Authentication.isAuthenticated()) { $location.path('/login'); } } } return $q.reject(rejection); } } } })();
Remove deferred as not needed
Remove deferred as not needed
JavaScript
mit
crowdresearch/crowdsource-platform,crowdresearch/crowdsource-platform,crowdresearch/daemo,rakshit-agrawal/crowdsource-platform,Milstein/crowdsource-platform,DESHRAJ/crowdsource-platform,Macbull/crowdsource-platform,crowdresearch/daemo,xasos/crowdsource-platform,crowdresearch/crowdsource-platform,shirishgoyal/crowdsource-platform,xasos/crowdsource-platform,xasos/crowdsource-platform,rakshit-agrawal/crowdsource-platform,aginzberg/crowdsource-platform,DESHRAJ/crowdsource-platform,aginzberg/crowdsource-platform,crowdresearch/daemo,Macbull/crowdsource-platform,shirishgoyal/crowdsource-platform,crowdresearch/crowdsource-platform,aginzberg/crowdsource-platform,Milstein/crowdsource-platform,tanujs22/crowdsource-platform,crowdresearch/daemo,DESHRAJ/crowdsource-platform,aginzberg/crowdsource-platform,shirishgoyal/crowdsource-platform,Macbull/crowdsource-platform,DESHRAJ/crowdsource-platform,rakshit-agrawal/crowdsource-platform,tanujs22/crowdsource-platform,rakshit-agrawal/crowdsource-platform,tanujs22/crowdsource-platform,Milstein/crowdsource-platform,Macbull/crowdsource-platform,xasos/crowdsource-platform,Milstein/crowdsource-platform,shirishgoyal/crowdsource-platform,Milstein/crowdsource-platform,tanujs22/crowdsource-platform
javascript
## Code Before: /** * Date: 1/9/15 * Author: Shirish Goyal */ (function () { 'use strict'; angular .module('crowdsource.interceptor', []) .factory('AuthHttpResponseInterceptor', AuthHttpResponseInterceptor); AuthHttpResponseInterceptor.$inject = ['$location', '$log', '$injector', '$q']; function AuthHttpResponseInterceptor($location, $log, $injector, $q) { return { responseError: function (rejection) { if (rejection.status === 403) { if (rejection.hasOwnProperty('data') && rejection.data.hasOwnProperty('detail') && (rejection.data.detail.indexOf("Authentication credentials were not provided") != -1) ) { var $http = $injector.get('$http'); var Authentication = $injector.get('Authentication'); if (!Authentication.isAuthenticated()) { $location.path('/login'); }else{ deferred.resolve(data); } } } return $q.reject(rejection); } } } })(); ## Instruction: Remove deferred as not needed ## Code After: /** * Date: 1/9/15 * Author: Shirish Goyal */ (function () { 'use strict'; angular .module('crowdsource.interceptor', []) .factory('AuthHttpResponseInterceptor', AuthHttpResponseInterceptor); AuthHttpResponseInterceptor.$inject = ['$location', '$log', '$injector', '$q']; function AuthHttpResponseInterceptor($location, $log, $injector, $q) { return { responseError: function (rejection) { if (rejection.status === 403) { if (rejection.hasOwnProperty('data') && rejection.data.hasOwnProperty('detail') && (rejection.data.detail.indexOf("Authentication credentials were not provided") != -1) ) { var $http = $injector.get('$http'); var Authentication = $injector.get('Authentication'); if (!Authentication.isAuthenticated()) { $location.path('/login'); } } } return $q.reject(rejection); } } } })();
e07d7bba31ffe9cca422417891fee20bf07f462b
lib/lokka/models/tag.rb
lib/lokka/models/tag.rb
class Tag < ActiveRecord::Base has_many :taggings, dependent: :destroy has_many :entries, through: :taggings, source: :tag, class_name: 'Entry' validates :name, presence: true, uniqueness: true scope :any, ->(list) { where(list.map {|name| sanitize_sql(['name = ?', name]) }.join(' OR ')) } def link "/tags/#{name}" end end def Tag(name) Tag.where(name: name).first end
class Tag < ActiveRecord::Base has_many :taggings, dependent: :destroy has_many :entries, through: :taggings, source: :taggable, source_type: 'Entry' validates :name, presence: true, uniqueness: true scope :any, ->(list) { where(list.map {|name| sanitize_sql(['name = ?', name]) }.join(' OR ')) } def link "/tags/#{name}" end end def Tag(name) Tag.where(name: name).first end
Fix Tag to Entry association
Fix Tag to Entry association
Ruby
mit
morygonzalez/portalshit.net,morygonzalez/portalshit.net,morygonzalez/portalshit.net,morygonzalez/portalshit.net
ruby
## Code Before: class Tag < ActiveRecord::Base has_many :taggings, dependent: :destroy has_many :entries, through: :taggings, source: :tag, class_name: 'Entry' validates :name, presence: true, uniqueness: true scope :any, ->(list) { where(list.map {|name| sanitize_sql(['name = ?', name]) }.join(' OR ')) } def link "/tags/#{name}" end end def Tag(name) Tag.where(name: name).first end ## Instruction: Fix Tag to Entry association ## Code After: class Tag < ActiveRecord::Base has_many :taggings, dependent: :destroy has_many :entries, through: :taggings, source: :taggable, source_type: 'Entry' validates :name, presence: true, uniqueness: true scope :any, ->(list) { where(list.map {|name| sanitize_sql(['name = ?', name]) }.join(' OR ')) } def link "/tags/#{name}" end end def Tag(name) Tag.where(name: name).first end
24105a13247aba8509c6a0f6ee74a039bf7f77ae
README.md
README.md
![travis-badge](https://api.travis-ci.org/arthurmco/ghostwriter.svg?branch=master) Ghostwriter is a software for those that want a platform for article/blog editing and management, but don't want a predefined way to display it, with limited choice for themes, complicated APIs and a difficult configuration This software gives *you* the control of how the posts are shown, offering a simple REST API for creating and showing them, along a simple interface for managing them. We don't offer a way to show them, we let you do it at your way. **THIS SOFTWARE IS UNDER DEVELOPMENT** This software is licensed under the MIT License ## Installing - Clone this repository - Create a virtualenv and enter it - Run `pip install -e . ` - Run `FLASK_APP=ghostwriter flask initdb`. This will create the database and the needed tables ## Running To test, simply run flask with the `FLASK_APP` environment variable as `ghostwriter`. You may want to run ghostwriter in a WSGI environment. Although isn't supported yet, I'd love some testing. The '/' route contains a test message. In the future, that message will be removable The default login is 'admin' with password 'admin'
[https://travis-ci.org/arthurmco/ghostwriter](![Travis Status Badge](https://api.travis-ci.org/arthurmco/ghostwriter.svg?branch=master)) Ghostwriter is a software for those that want a platform for article/blog editing and management, but don't want a predefined way to display it, with limited choice for themes, complicated APIs and a difficult configuration This software gives *you* the control of how the posts are shown, offering a simple REST API for creating and showing them, along with a simple interface for managing them. We don't offer a way to show them, we let you do it at your way. **THIS SOFTWARE IS UNDER DEVELOPMENT** This software is licensed under the MIT License ## Installing - Clone this repository - Create a virtualenv and enter it - Run `pip install -e . ` - Run `FLASK_APP=ghostwriter flask initdb`. This will create the database and the needed tables ## Running To test, simply run flask with the `FLASK_APP` environment variable as `ghostwriter`. You may want to run ghostwriter in a WSGI environment. Although isn't supported yet, I'd love some testing. The '/' route contains a test message. In the future, that message will be removable The default login is 'admin' with password 'admin'. You can access the admin interface through the '/admin' entry point
Add link to Travis page in readme
Add link to Travis page in readme
Markdown
mit
arthurmco/ghostwriter,arthurmco/ghostwriter
markdown
## Code Before: ![travis-badge](https://api.travis-ci.org/arthurmco/ghostwriter.svg?branch=master) Ghostwriter is a software for those that want a platform for article/blog editing and management, but don't want a predefined way to display it, with limited choice for themes, complicated APIs and a difficult configuration This software gives *you* the control of how the posts are shown, offering a simple REST API for creating and showing them, along a simple interface for managing them. We don't offer a way to show them, we let you do it at your way. **THIS SOFTWARE IS UNDER DEVELOPMENT** This software is licensed under the MIT License ## Installing - Clone this repository - Create a virtualenv and enter it - Run `pip install -e . ` - Run `FLASK_APP=ghostwriter flask initdb`. This will create the database and the needed tables ## Running To test, simply run flask with the `FLASK_APP` environment variable as `ghostwriter`. You may want to run ghostwriter in a WSGI environment. Although isn't supported yet, I'd love some testing. The '/' route contains a test message. In the future, that message will be removable The default login is 'admin' with password 'admin' ## Instruction: Add link to Travis page in readme ## Code After: [https://travis-ci.org/arthurmco/ghostwriter](![Travis Status Badge](https://api.travis-ci.org/arthurmco/ghostwriter.svg?branch=master)) Ghostwriter is a software for those that want a platform for article/blog editing and management, but don't want a predefined way to display it, with limited choice for themes, complicated APIs and a difficult configuration This software gives *you* the control of how the posts are shown, offering a simple REST API for creating and showing them, along with a simple interface for managing them. We don't offer a way to show them, we let you do it at your way. **THIS SOFTWARE IS UNDER DEVELOPMENT** This software is licensed under the MIT License ## Installing - Clone this repository - Create a virtualenv and enter it - Run `pip install -e . ` - Run `FLASK_APP=ghostwriter flask initdb`. This will create the database and the needed tables ## Running To test, simply run flask with the `FLASK_APP` environment variable as `ghostwriter`. You may want to run ghostwriter in a WSGI environment. Although isn't supported yet, I'd love some testing. The '/' route contains a test message. In the future, that message will be removable The default login is 'admin' with password 'admin'. You can access the admin interface through the '/admin' entry point
98c9ddd43d7503044ce135680b9a2966d64d8096
source/_splash.erb
source/_splash.erb
<div class="row"> <div class="col-6 col-lg-6"> <img src="images/foodcritic.png" /> </div> <div class="col-6 col-lg-6"> <h1>About Foodcritic</h1> <p> Foodcritic is a helpful lint tool you can use to check your Chef cookbooks for common problems. </p> <p> It comes with <b><%= rule_count %></b> built-in rules that identify problems ranging from simple style inconsistencies to difficult to diagnose issues that will hurt in production. </p> <p> Various nice people in the Chef community have also written extra rules for foodcritic that you can install and run. Or write your own! </p> <p> It's easy to get started: </p> <p> <code>$ gem install foodcritic</code><br /> <code>$ foodcritic my_cookbook_dir</code> </p> </div> </div>
<div class="row"> <div class="col-6 col-lg-6"> <img src="images/foodcritic.png" /> </div> <div class="col-6 col-lg-6"> <h1>About Foodcritic</h1> <p> Foodcritic is a helpful lint tool you can use to check your Chef cookbooks for common problems. </p> <p> It comes with <b><%= rule_count %></b> built-in rules that identify problems ranging from simple style inconsistencies to difficult to diagnose issues that will hurt in production. </p> <p> Various nice people in the Chef community have also written extra rules for foodcritic that you can install and run. Or write your own! </p> <p> It's easy to get started with Foodcritic in Chef-DK already </p> <p> <code>$ foodcritic my_cookbook_dir</code> </p> </div> </div>
Update for Foodcritic in ChefDK
Update for Foodcritic in ChefDK Signed-off-by: Tim Smith <[email protected]>
HTML+ERB
mit
acrmp/foodcritic-site,Foodcritic/foodcritic-site,acrmp/foodcritic-site,Foodcritic/foodcritic-site
html+erb
## Code Before: <div class="row"> <div class="col-6 col-lg-6"> <img src="images/foodcritic.png" /> </div> <div class="col-6 col-lg-6"> <h1>About Foodcritic</h1> <p> Foodcritic is a helpful lint tool you can use to check your Chef cookbooks for common problems. </p> <p> It comes with <b><%= rule_count %></b> built-in rules that identify problems ranging from simple style inconsistencies to difficult to diagnose issues that will hurt in production. </p> <p> Various nice people in the Chef community have also written extra rules for foodcritic that you can install and run. Or write your own! </p> <p> It's easy to get started: </p> <p> <code>$ gem install foodcritic</code><br /> <code>$ foodcritic my_cookbook_dir</code> </p> </div> </div> ## Instruction: Update for Foodcritic in ChefDK Signed-off-by: Tim Smith <[email protected]> ## Code After: <div class="row"> <div class="col-6 col-lg-6"> <img src="images/foodcritic.png" /> </div> <div class="col-6 col-lg-6"> <h1>About Foodcritic</h1> <p> Foodcritic is a helpful lint tool you can use to check your Chef cookbooks for common problems. </p> <p> It comes with <b><%= rule_count %></b> built-in rules that identify problems ranging from simple style inconsistencies to difficult to diagnose issues that will hurt in production. </p> <p> Various nice people in the Chef community have also written extra rules for foodcritic that you can install and run. Or write your own! </p> <p> It's easy to get started with Foodcritic in Chef-DK already </p> <p> <code>$ foodcritic my_cookbook_dir</code> </p> </div> </div>
1e1c164fb1eae8d873c4aba244d89cb8fe42dbc1
integration-tests/spec/spec_helper_integ.rb
integration-tests/spec/spec_helper_integ.rb
require 'ostruct' require 'capybara/dsl' require 'jmx4r' require 'stilts-stomp-client' require 'driver_helper' def mbean(name) url = JMX::JDKHelper.find_local_url(/org.jboss.as.standalone/i) puts "default jmx url: " + url url.gsub!(/\d+\.\d+\.\d+\.\d+/, "127.0.0.1") puts "using jmx url: " + url JMX::MBean.establish_connection :url => url #:command => /org.jboss.as.standalone/i JMX::MBean.find_by_name(name) end def get_msc_service_state(service_name) @msc ||= mbean('jboss.msc:type=container,name=jboss-as') @msc.getServiceStatus(service_name).get('stateName') end def set_msc_service_mode(service_name, mode) @msc ||= mbean('jboss.msc:type=container,name=jboss-as') @msc.setServiceMode(service_name, mode) end def verify_msc_service_state(service_name, state, options={}) options[:timeout] ||= 30 # default to wait 30 seconds service_state = nil elapsed_seconds = 0 until service_state == state || elapsed_seconds > options[:timeout] do service_state = get_msc_service_state(service_name) sleep(0.5) end service_state.should == state end
require 'ostruct' require 'capybara/dsl' require 'jmx4r' require 'stilts-stomp-client' require 'driver_helper' def mbean(name) JMX::MBean.establish_connection :command => /org.jboss.as.standalone/i JMX::MBean.find_by_name(name) end def get_msc_service_state(service_name) @msc ||= mbean('jboss.msc:type=container,name=jboss-as') @msc.getServiceStatus(service_name).get('stateName') end def set_msc_service_mode(service_name, mode) @msc ||= mbean('jboss.msc:type=container,name=jboss-as') @msc.setServiceMode(service_name, mode) end def verify_msc_service_state(service_name, state, options={}) options[:timeout] ||= 30 # default to wait 30 seconds service_state = nil elapsed_seconds = 0 until service_state == state || elapsed_seconds > options[:timeout] do service_state = get_msc_service_state(service_name) sleep(0.5) end service_state.should == state end
Revert "A total f'ing shot in the dark."
Revert "A total f'ing shot in the dark." This reverts commit c04c63fd9d0160bb5b22dc475e64eb7e64a196ef.
Ruby
apache-2.0
mje113/torquebox,torquebox/torquebox,vaskoz/torquebox,torquebox/torquebox,vaskoz/torquebox,ksw2599/torquebox,torquebox/torquebox-release,torquebox/torquebox-release,mje113/torquebox,mje113/torquebox,ksw2599/torquebox,torquebox/torquebox-release,torquebox/torquebox,samwgoldman/torquebox,vaskoz/torquebox,torquebox/torquebox,samwgoldman/torquebox,mje113/torquebox,samwgoldman/torquebox,ksw2599/torquebox,vaskoz/torquebox,samwgoldman/torquebox,torquebox/torquebox-release,ksw2599/torquebox
ruby
## Code Before: require 'ostruct' require 'capybara/dsl' require 'jmx4r' require 'stilts-stomp-client' require 'driver_helper' def mbean(name) url = JMX::JDKHelper.find_local_url(/org.jboss.as.standalone/i) puts "default jmx url: " + url url.gsub!(/\d+\.\d+\.\d+\.\d+/, "127.0.0.1") puts "using jmx url: " + url JMX::MBean.establish_connection :url => url #:command => /org.jboss.as.standalone/i JMX::MBean.find_by_name(name) end def get_msc_service_state(service_name) @msc ||= mbean('jboss.msc:type=container,name=jboss-as') @msc.getServiceStatus(service_name).get('stateName') end def set_msc_service_mode(service_name, mode) @msc ||= mbean('jboss.msc:type=container,name=jboss-as') @msc.setServiceMode(service_name, mode) end def verify_msc_service_state(service_name, state, options={}) options[:timeout] ||= 30 # default to wait 30 seconds service_state = nil elapsed_seconds = 0 until service_state == state || elapsed_seconds > options[:timeout] do service_state = get_msc_service_state(service_name) sleep(0.5) end service_state.should == state end ## Instruction: Revert "A total f'ing shot in the dark." This reverts commit c04c63fd9d0160bb5b22dc475e64eb7e64a196ef. ## Code After: require 'ostruct' require 'capybara/dsl' require 'jmx4r' require 'stilts-stomp-client' require 'driver_helper' def mbean(name) JMX::MBean.establish_connection :command => /org.jboss.as.standalone/i JMX::MBean.find_by_name(name) end def get_msc_service_state(service_name) @msc ||= mbean('jboss.msc:type=container,name=jboss-as') @msc.getServiceStatus(service_name).get('stateName') end def set_msc_service_mode(service_name, mode) @msc ||= mbean('jboss.msc:type=container,name=jboss-as') @msc.setServiceMode(service_name, mode) end def verify_msc_service_state(service_name, state, options={}) options[:timeout] ||= 30 # default to wait 30 seconds service_state = nil elapsed_seconds = 0 until service_state == state || elapsed_seconds > options[:timeout] do service_state = get_msc_service_state(service_name) sleep(0.5) end service_state.should == state end
2e397e534b4ad6732e57b9d7638df7e494339d85
source/stylesheets/site.scss
source/stylesheets/site.scss
@charset "utf-8"; @import "vendor/normalize"; @import "vendor/animate"; // http://bourbon.io @import "bourbon/bourbon"; @import "base/base"; @import "neat/neat"; // Partials @import "partials/layout"; @import "partials/nav-transparent"; @import "partials/nav"; @import "partials/alpha"; @import "partials/earthscape"; @import "partials/lessons"; @import "partials/lessonx"; @import "partials/faith-cometh"; @import "partials/talking-points"; @import "partials/call-to-action"; @import "partials/video"; @import "partials/training"; @import "partials/downloads"; @import "partials/contact"; @import "partials/purchase"; @import "partials/bulk-purchase"; @import "partials/thanks"; @import "partials/callout-box"; @import "partials/breadcrumbs"; @import "partials/copyright-footer"; @import "partials/universal-footer"; @import "partials/404"; //////////////////////// SANDBOX //////////////////////// //// Try out new code here, then extract to partial. //// /////////////////////////////////////////////////////////
@charset "utf-8"; @import "vendor/normalize"; @import "vendor/animate"; // http://bourbon.io @import "bourbon/bourbon"; @import "base/grid-settings"; @import "neat/neat"; @import "base/base"; // Partials @import "partials/layout"; @import "partials/nav-transparent"; @import "partials/nav"; @import "partials/alpha"; @import "partials/earthscape"; @import "partials/lessons"; @import "partials/lessonx"; @import "partials/faith-cometh"; @import "partials/talking-points"; @import "partials/call-to-action"; @import "partials/video"; @import "partials/training"; @import "partials/downloads"; @import "partials/contact"; @import "partials/purchase"; @import "partials/bulk-purchase"; @import "partials/thanks"; @import "partials/callout-box"; @import "partials/breadcrumbs"; @import "partials/copyright-footer"; @import "partials/universal-footer"; @import "partials/404"; //////////////////////// SANDBOX //////////////////////// //// Try out new code here, then extract to partial. //// /////////////////////////////////////////////////////////
Update sass imports order for new bourbon/neat/bitters requirements.
Update sass imports order for new bourbon/neat/bitters requirements.
SCSS
mit
euroteamoutreach/getbiblefirst.com,joshukraine/getbiblefirst.com,joshukraine/getbiblefirst.com,euroteamoutreach/getbiblefirst.com,joshukraine/getbiblefirst.com,joshukraine/getbiblefirst.com,euroteamoutreach/getbiblefirst.com
scss
## Code Before: @charset "utf-8"; @import "vendor/normalize"; @import "vendor/animate"; // http://bourbon.io @import "bourbon/bourbon"; @import "base/base"; @import "neat/neat"; // Partials @import "partials/layout"; @import "partials/nav-transparent"; @import "partials/nav"; @import "partials/alpha"; @import "partials/earthscape"; @import "partials/lessons"; @import "partials/lessonx"; @import "partials/faith-cometh"; @import "partials/talking-points"; @import "partials/call-to-action"; @import "partials/video"; @import "partials/training"; @import "partials/downloads"; @import "partials/contact"; @import "partials/purchase"; @import "partials/bulk-purchase"; @import "partials/thanks"; @import "partials/callout-box"; @import "partials/breadcrumbs"; @import "partials/copyright-footer"; @import "partials/universal-footer"; @import "partials/404"; //////////////////////// SANDBOX //////////////////////// //// Try out new code here, then extract to partial. //// ///////////////////////////////////////////////////////// ## Instruction: Update sass imports order for new bourbon/neat/bitters requirements. ## Code After: @charset "utf-8"; @import "vendor/normalize"; @import "vendor/animate"; // http://bourbon.io @import "bourbon/bourbon"; @import "base/grid-settings"; @import "neat/neat"; @import "base/base"; // Partials @import "partials/layout"; @import "partials/nav-transparent"; @import "partials/nav"; @import "partials/alpha"; @import "partials/earthscape"; @import "partials/lessons"; @import "partials/lessonx"; @import "partials/faith-cometh"; @import "partials/talking-points"; @import "partials/call-to-action"; @import "partials/video"; @import "partials/training"; @import "partials/downloads"; @import "partials/contact"; @import "partials/purchase"; @import "partials/bulk-purchase"; @import "partials/thanks"; @import "partials/callout-box"; @import "partials/breadcrumbs"; @import "partials/copyright-footer"; @import "partials/universal-footer"; @import "partials/404"; //////////////////////// SANDBOX //////////////////////// //// Try out new code here, then extract to partial. //// /////////////////////////////////////////////////////////
befb57176f20692ce514ebce37282592e6bfe882
hammerspoon/window-expand.lua
hammerspoon/window-expand.lua
local function pressFn(mods, key) if key == nil then key = mods mods = {} end return function() hs.eventtap.keyStroke(mods, key, 1000) end end local function remap(mods, key, pressFn) hs.hotkey.bind(mods, key, pressFn, nil, pressFn) end remap({"cmd", "alt", "ctrl", "shift"}, "delete", pressFn({"cmd", "shift"}, "Return")) hs.hotkey.bind({"cmd", "alt", "ctrl", "shift"}, "Return", function() local win = hs.window.focusedWindow() local f = win:frame() local max = win:screen():frame() f.x = max.x f.y = max.y f.w = max.w f.h = max.h win:setFrame(f) end) hs.hotkey.bind({"cmd", "alt", "ctrl", "shift"}, "F", function() local win = hs.window.focusedWindow() win:toggleFullScreen() end)
local function pressFn(mods, key) if key == nil then key = mods mods = {} end return function() hs.eventtap.keyStroke(mods, key, 1000) end end local function remap(mods, key, pressFn) hs.hotkey.bind(mods, key, pressFn, nil, pressFn) end remap({"cmd", "alt", "ctrl", "shift"}, "delete", pressFn({"cmd", "shift"}, "Return")) hs.hotkey.bind({"cmd", "alt", "ctrl", "shift"}, "Return", function() local win = hs.window.focusedWindow() local f = win:frame() local max = win:screen():frame() f.x = max.x f.y = max.y f.w = max.w f.h = max.h win:setFrame(f) end) hs.hotkey.bind({"cmd", "alt", "ctrl", "shift"}, "F", function() local win = hs.window.focusedWindow() win:toggleFullScreen() end) -- cycle through all windows of the current application hs.hotkey.bind({"cmd", "alt", "ctrl", "shift"}, "space", function() local app = hs.application.frontmostApplication() local current = 1 local next = current for i, window in pairs(app:allWindows()) do if app:focusedWindow() == app:allWindows()[i] then current = i end if i > current then next = i end end app:allWindows()[next]:focus() end)
Add hyper+space for window cycling
Add hyper+space for window cycling
Lua
mit
bsingr/dotfiles,bsingr/dotfiles
lua
## Code Before: local function pressFn(mods, key) if key == nil then key = mods mods = {} end return function() hs.eventtap.keyStroke(mods, key, 1000) end end local function remap(mods, key, pressFn) hs.hotkey.bind(mods, key, pressFn, nil, pressFn) end remap({"cmd", "alt", "ctrl", "shift"}, "delete", pressFn({"cmd", "shift"}, "Return")) hs.hotkey.bind({"cmd", "alt", "ctrl", "shift"}, "Return", function() local win = hs.window.focusedWindow() local f = win:frame() local max = win:screen():frame() f.x = max.x f.y = max.y f.w = max.w f.h = max.h win:setFrame(f) end) hs.hotkey.bind({"cmd", "alt", "ctrl", "shift"}, "F", function() local win = hs.window.focusedWindow() win:toggleFullScreen() end) ## Instruction: Add hyper+space for window cycling ## Code After: local function pressFn(mods, key) if key == nil then key = mods mods = {} end return function() hs.eventtap.keyStroke(mods, key, 1000) end end local function remap(mods, key, pressFn) hs.hotkey.bind(mods, key, pressFn, nil, pressFn) end remap({"cmd", "alt", "ctrl", "shift"}, "delete", pressFn({"cmd", "shift"}, "Return")) hs.hotkey.bind({"cmd", "alt", "ctrl", "shift"}, "Return", function() local win = hs.window.focusedWindow() local f = win:frame() local max = win:screen():frame() f.x = max.x f.y = max.y f.w = max.w f.h = max.h win:setFrame(f) end) hs.hotkey.bind({"cmd", "alt", "ctrl", "shift"}, "F", function() local win = hs.window.focusedWindow() win:toggleFullScreen() end) -- cycle through all windows of the current application hs.hotkey.bind({"cmd", "alt", "ctrl", "shift"}, "space", function() local app = hs.application.frontmostApplication() local current = 1 local next = current for i, window in pairs(app:allWindows()) do if app:focusedWindow() == app:allWindows()[i] then current = i end if i > current then next = i end end app:allWindows()[next]:focus() end)
8af7e9a2e63a155289c2840dd0c8950c95f9153b
tests/spec/storm/storm_spec.rb
tests/spec/storm/storm_spec.rb
require 'faraday' require 'json' describe 'Apache Storm Mesos release' do context 'Storm UI is running' do let(:storm_ui_url) { 'http://10.244.1.22:8080/' } it 'has an accessible home page' do status_code = Faraday.head(storm_ui_url).status expect(status_code).to eq(200) end end context 'Storm Nimbus is running' do let(:mesos_master_state) { 'http://10.244.1.6:5050/master/state.json' } it 'is registered with Mesos master' do raw_state = Faraday.get mesos_master_state state_json = JSON.parse(raw_state.body) storm = state_json.fetch('frameworks').find do |framework| framework.fetch('name') == 'Storm!!!' end expect(storm).not_to be_nil expect(storm['name']).to eql('Storm!!!') end xit 'accepts a tuple' do end end end
require 'spec_helper' describe 'Apache Storm Mesos release' do context 'Storm UI is running' do let(:storm_ui_url) { 'http://10.244.1.22:8080/' } it 'has an accessible home page' do status_code = Faraday.head(storm_ui_url).status expect(status_code).to eq(200) end end context 'Storm Nimbus is running' do it 'is registered with Mesos master' do storm = mesos_master_state_json.fetch('frameworks').find do |framework| framework.fetch('name') == 'Storm!!!' end expect(storm).not_to be_nil expect(storm['name']).to eql('Storm!!!') end xit 'accepts a tuple' do end end end
Move shared Mesos and Bosh logic to spec helper
Move shared Mesos and Bosh logic to spec helper
Ruby
apache-2.0
CloudCredo/mesos-boshrelease,CloudCredo/mesos-boshrelease
ruby
## Code Before: require 'faraday' require 'json' describe 'Apache Storm Mesos release' do context 'Storm UI is running' do let(:storm_ui_url) { 'http://10.244.1.22:8080/' } it 'has an accessible home page' do status_code = Faraday.head(storm_ui_url).status expect(status_code).to eq(200) end end context 'Storm Nimbus is running' do let(:mesos_master_state) { 'http://10.244.1.6:5050/master/state.json' } it 'is registered with Mesos master' do raw_state = Faraday.get mesos_master_state state_json = JSON.parse(raw_state.body) storm = state_json.fetch('frameworks').find do |framework| framework.fetch('name') == 'Storm!!!' end expect(storm).not_to be_nil expect(storm['name']).to eql('Storm!!!') end xit 'accepts a tuple' do end end end ## Instruction: Move shared Mesos and Bosh logic to spec helper ## Code After: require 'spec_helper' describe 'Apache Storm Mesos release' do context 'Storm UI is running' do let(:storm_ui_url) { 'http://10.244.1.22:8080/' } it 'has an accessible home page' do status_code = Faraday.head(storm_ui_url).status expect(status_code).to eq(200) end end context 'Storm Nimbus is running' do it 'is registered with Mesos master' do storm = mesos_master_state_json.fetch('frameworks').find do |framework| framework.fetch('name') == 'Storm!!!' end expect(storm).not_to be_nil expect(storm['name']).to eql('Storm!!!') end xit 'accepts a tuple' do end end end
fbd0ce006326e2e410084297e64afdfb9052d2a1
.travis.yml
.travis.yml
language: node_js node_js: - "6.9.2" addons: firefox: "latest" before_install: - ./tools/travis/init-env - ./tools/travis/install-geckodriver - ./tools/travis/start-xvfb - . ./tools/travis/env.sh - export DISPLAY=$ENV_DISPLAY - export PATH=$PATH:$ENV_LOCAL_BIN_DIR install: - npm install --global gulp-cli - npm run bootstrap script: - gulp unit-test-with-coverage - gulp dist - gulp start-server - gulp acceptance-test - gulp stop-server - gulp lint - gulp docs after_script: - gulp publish-coverage
language: node_js node_js: - "6.9.2" addons: firefox: "latest" before_install: - ./tools/travis/init-env - ./tools/travis/install-geckodriver - ./tools/travis/start-xvfb - . ./tools/travis/env.sh - export DISPLAY=$ENV_DISPLAY - export PATH=$PATH:$ENV_LOCAL_BIN_DIR install: - npm install --global gulp-cli phantomjs - npm run bootstrap script: - gulp unit-test-with-coverage - gulp dist - gulp start-server - gulp acceptance-test:server - ENV_WEBDRIVER_BROWSER_NAME=phantomjs gulp acceptance-test:client - ENV_WEBDRIVER_BROWSER_NAME=firefox gulp acceptance-test:client - gulp stop-server - gulp lint - gulp docs after_script: - gulp publish-coverage
Add PhantomJS to list of browsers to be tested on Travis
Add PhantomJS to list of browsers to be tested on Travis
YAML
mit
ssoloff/dice-server-js,ssoloff/dice-server-js,ssoloff/dice-server-js,ssoloff/dice-server-js
yaml
## Code Before: language: node_js node_js: - "6.9.2" addons: firefox: "latest" before_install: - ./tools/travis/init-env - ./tools/travis/install-geckodriver - ./tools/travis/start-xvfb - . ./tools/travis/env.sh - export DISPLAY=$ENV_DISPLAY - export PATH=$PATH:$ENV_LOCAL_BIN_DIR install: - npm install --global gulp-cli - npm run bootstrap script: - gulp unit-test-with-coverage - gulp dist - gulp start-server - gulp acceptance-test - gulp stop-server - gulp lint - gulp docs after_script: - gulp publish-coverage ## Instruction: Add PhantomJS to list of browsers to be tested on Travis ## Code After: language: node_js node_js: - "6.9.2" addons: firefox: "latest" before_install: - ./tools/travis/init-env - ./tools/travis/install-geckodriver - ./tools/travis/start-xvfb - . ./tools/travis/env.sh - export DISPLAY=$ENV_DISPLAY - export PATH=$PATH:$ENV_LOCAL_BIN_DIR install: - npm install --global gulp-cli phantomjs - npm run bootstrap script: - gulp unit-test-with-coverage - gulp dist - gulp start-server - gulp acceptance-test:server - ENV_WEBDRIVER_BROWSER_NAME=phantomjs gulp acceptance-test:client - ENV_WEBDRIVER_BROWSER_NAME=firefox gulp acceptance-test:client - gulp stop-server - gulp lint - gulp docs after_script: - gulp publish-coverage
edebe37458da391723e3206c63102cbb69606c5b
ideascube/conf/idb_irq_bardarash.py
ideascube/conf/idb_irq_bardarash.py
"""Bardarash in Kurdistan, Iraq""" from .idb_jor_azraq import * # pragma: no flakes from django.utils.translation import ugettext_lazy as _ USER_FORM_FIELDS = ( ('Ideasbox', ['serial', 'box_awareness']), (_('Personal informations'), ['short_name', 'full_name', 'latin_name', 'birth_year', 'gender', 'country_of_origin_occupation', 'school_level']), # noqa (_('Family'), ['marital_status', 'family_status', 'children_under_12', 'children_under_18', 'children_above_18']), # noqa (_('In the camp'), ['camp_entry_date', 'camp_activities', 'current_occupation', 'camp_address']), # noqa (_('Language skills'), ['ar_level', 'ku_level', 'sdb_level', 'en_level']), ) MONITORING_ENTRY_EXPORT_FIELDS = ['serial', 'birth_year', 'gender'] ENTRY_ACTIVITY_CHOICES = []
"""Bardarash in Kurdistan, Iraq""" from .idb_jor_azraq import * # pragma: no flakes from django.utils.translation import ugettext_lazy as _ USER_FORM_FIELDS = ( ('Ideasbox', ['serial', 'box_awareness']), (_('Personal informations'), ['short_name', 'full_name', 'latin_name', 'birth_year', 'gender', 'country_of_origin_occupation', 'school_level']), # noqa (_('Family'), ['marital_status', 'family_status', 'children_under_12', 'children_under_18', 'children_above_18']), # noqa (_('In the camp'), ['camp_entry_date', 'camp_activities', 'current_occupation', 'camp_address']), # noqa (_('Language skills'), ['ar_level', 'ku_level', 'sdb_level', 'en_level']), ) MONITORING_ENTRY_EXPORT_FIELDS = ['serial', 'birth_year', 'gender'] ENTRY_ACTIVITY_CHOICES = [] HOME_CARDS = STAFF_HOME_CARDS + [ { 'id': 'blog', }, { 'id': 'mediacenter', }, { 'id': 'library', } ]
Remove Kalite until Arabic language is available
Remove Kalite until Arabic language is available
Python
agpl-3.0
ideascube/ideascube,ideascube/ideascube,ideascube/ideascube,ideascube/ideascube
python
## Code Before: """Bardarash in Kurdistan, Iraq""" from .idb_jor_azraq import * # pragma: no flakes from django.utils.translation import ugettext_lazy as _ USER_FORM_FIELDS = ( ('Ideasbox', ['serial', 'box_awareness']), (_('Personal informations'), ['short_name', 'full_name', 'latin_name', 'birth_year', 'gender', 'country_of_origin_occupation', 'school_level']), # noqa (_('Family'), ['marital_status', 'family_status', 'children_under_12', 'children_under_18', 'children_above_18']), # noqa (_('In the camp'), ['camp_entry_date', 'camp_activities', 'current_occupation', 'camp_address']), # noqa (_('Language skills'), ['ar_level', 'ku_level', 'sdb_level', 'en_level']), ) MONITORING_ENTRY_EXPORT_FIELDS = ['serial', 'birth_year', 'gender'] ENTRY_ACTIVITY_CHOICES = [] ## Instruction: Remove Kalite until Arabic language is available ## Code After: """Bardarash in Kurdistan, Iraq""" from .idb_jor_azraq import * # pragma: no flakes from django.utils.translation import ugettext_lazy as _ USER_FORM_FIELDS = ( ('Ideasbox', ['serial', 'box_awareness']), (_('Personal informations'), ['short_name', 'full_name', 'latin_name', 'birth_year', 'gender', 'country_of_origin_occupation', 'school_level']), # noqa (_('Family'), ['marital_status', 'family_status', 'children_under_12', 'children_under_18', 'children_above_18']), # noqa (_('In the camp'), ['camp_entry_date', 'camp_activities', 'current_occupation', 'camp_address']), # noqa (_('Language skills'), ['ar_level', 'ku_level', 'sdb_level', 'en_level']), ) MONITORING_ENTRY_EXPORT_FIELDS = ['serial', 'birth_year', 'gender'] ENTRY_ACTIVITY_CHOICES = [] HOME_CARDS = STAFF_HOME_CARDS + [ { 'id': 'blog', }, { 'id': 'mediacenter', }, { 'id': 'library', } ]
9eafe8385c8134d60241ef653a9342884c04d3e3
tox.ini
tox.ini
[tox] envlist = py26, py27, py33, py34, py35 [testenv] setenv = PYTHONPATH = {toxinidir}:{toxinidir}/regressors commands = python setup.py test deps = pip==7.0.2 -r{toxinidir}/requirements_dev.txt [testenv:py26] ; Python 2.6 tests need to be run with nose to work with unittest2 commands = nosetests tests/test_regressors.py [testenv:py35] ; Output coverage report in last environment tested commands = coverage erase coverage run --source regressors setup.py test coverage combine coverage report [flake8] max-line-length = 80
[tox] envlist = py26, py27, py33, py34, py35 [testenv] setenv = PYTHONPATH = {toxinidir}:{toxinidir}/regressors commands = python setup.py test deps = pip==7.0.2 -r{toxinidir}/requirements_dev.txt [testenv:py26] ; Python 2.6 tests need to be run with nose to work with unittest2 commands = nosetests tests/test_regressors.py [testenv:py33] ; Need to discover tests with unittest to properly pick them up in Python 3 commands = coverage erase coverage run --source regressors -m unittest discover [testenv:py34] ; Need to discover tests with unittest to properly pick them up in Python 3 commands = coverage erase coverage run --source regressors -m unittest discover [testenv:py35] ; Output coverage report in last environment tested commands = coverage erase coverage run --source regressors -m unittest discover coverage combine coverage report [flake8] max-line-length = 80
Fix tests not getting picked up in Python 3
Fix tests not getting picked up in Python 3
INI
isc
nsh87/regressors
ini
## Code Before: [tox] envlist = py26, py27, py33, py34, py35 [testenv] setenv = PYTHONPATH = {toxinidir}:{toxinidir}/regressors commands = python setup.py test deps = pip==7.0.2 -r{toxinidir}/requirements_dev.txt [testenv:py26] ; Python 2.6 tests need to be run with nose to work with unittest2 commands = nosetests tests/test_regressors.py [testenv:py35] ; Output coverage report in last environment tested commands = coverage erase coverage run --source regressors setup.py test coverage combine coverage report [flake8] max-line-length = 80 ## Instruction: Fix tests not getting picked up in Python 3 ## Code After: [tox] envlist = py26, py27, py33, py34, py35 [testenv] setenv = PYTHONPATH = {toxinidir}:{toxinidir}/regressors commands = python setup.py test deps = pip==7.0.2 -r{toxinidir}/requirements_dev.txt [testenv:py26] ; Python 2.6 tests need to be run with nose to work with unittest2 commands = nosetests tests/test_regressors.py [testenv:py33] ; Need to discover tests with unittest to properly pick them up in Python 3 commands = coverage erase coverage run --source regressors -m unittest discover [testenv:py34] ; Need to discover tests with unittest to properly pick them up in Python 3 commands = coverage erase coverage run --source regressors -m unittest discover [testenv:py35] ; Output coverage report in last environment tested commands = coverage erase coverage run --source regressors -m unittest discover coverage combine coverage report [flake8] max-line-length = 80
2ed094d6902272f95dfd556c2244b13f9fbbc76b
build.sh
build.sh
cd $(dirname $0) file=$(perl -e ' use JSON::PP; local $/; $info = decode_json <>; print $info->{name}."_".$info->{version}' \ src/info.json) mkdir -p build rm -rf build/* git archive HEAD:src --prefix=$file/ -9 -o build/$file.zip
cd $(dirname $0) file=$(perl -e ' use JSON::PP; local $/; $info = decode_json <>; print $info->{name}."_".$info->{version}' \ src/info.json) mkdir -p build rm -rf build/* git archive HEAD:src --prefix=$file/ -9 -o build/$file.zip # Include LICENSE.md with release cd build mkdir -p $file cp ../LICENSE.md $file/LICENSE.md zip -q -9 $file.zip $file/LICENSE.md rm -rf $file
Add LICENSE.md to built .zip
Add LICENSE.md to built .zip
Shell
mit
jpas/factorio-expanded-tools
shell
## Code Before: cd $(dirname $0) file=$(perl -e ' use JSON::PP; local $/; $info = decode_json <>; print $info->{name}."_".$info->{version}' \ src/info.json) mkdir -p build rm -rf build/* git archive HEAD:src --prefix=$file/ -9 -o build/$file.zip ## Instruction: Add LICENSE.md to built .zip ## Code After: cd $(dirname $0) file=$(perl -e ' use JSON::PP; local $/; $info = decode_json <>; print $info->{name}."_".$info->{version}' \ src/info.json) mkdir -p build rm -rf build/* git archive HEAD:src --prefix=$file/ -9 -o build/$file.zip # Include LICENSE.md with release cd build mkdir -p $file cp ../LICENSE.md $file/LICENSE.md zip -q -9 $file.zip $file/LICENSE.md rm -rf $file
dcf0f8aaf8a45ea5be82b3de8b57e9f9580feab1
spec/support/backend.rb
spec/support/backend.rb
shared_examples "a Que backend" do it "should be able to drop and create the jobs table" do DB.table_exists?(:que_jobs).should be true Que.drop! DB.table_exists?(:que_jobs).should be false Que.create! DB.table_exists?(:que_jobs).should be true end it "should be able to clear the jobs table" do DB[:que_jobs].insert :type => "Que::Job" DB[:que_jobs].count.should be 1 Que.clear! DB[:que_jobs].count.should be 0 end end
shared_examples "a Que backend" do it "should be able to drop and create the jobs table" do DB.table_exists?(:que_jobs).should be true Que.drop! DB.table_exists?(:que_jobs).should be false Que.create! DB.table_exists?(:que_jobs).should be true end it "should be able to clear the jobs table" do DB[:que_jobs].insert :type => "Que::Job" DB[:que_jobs].count.should be 1 Que.clear! DB[:que_jobs].count.should be 0 end describe "when queueing jobs" do it "should be able to queue a job with arguments" do pending end it "should be able to queue a job with a specific time to run" do pending end it "should be able to queue a job with a specific priority" do pending end it "should be able to queue a job with queueing options in addition to argument options" do pending end it "should respect a default (but overridable) priority for the job class" do pending end it "should respect a default (but overridable) run_at for the job class" do pending end it "should raise an error if given arguments that can't convert to and from JSON unambiguously" do pending end end end
Add some pending specs on queueing jobs.
Add some pending specs on queueing jobs.
Ruby
mit
chanks/que,zloydadka/que,hardbap/que,gocardless/que,goddamnhippie/que,joevandyk/que,brandur/que,gocardless/que,godfat/que,heroku/que
ruby
## Code Before: shared_examples "a Que backend" do it "should be able to drop and create the jobs table" do DB.table_exists?(:que_jobs).should be true Que.drop! DB.table_exists?(:que_jobs).should be false Que.create! DB.table_exists?(:que_jobs).should be true end it "should be able to clear the jobs table" do DB[:que_jobs].insert :type => "Que::Job" DB[:que_jobs].count.should be 1 Que.clear! DB[:que_jobs].count.should be 0 end end ## Instruction: Add some pending specs on queueing jobs. ## Code After: shared_examples "a Que backend" do it "should be able to drop and create the jobs table" do DB.table_exists?(:que_jobs).should be true Que.drop! DB.table_exists?(:que_jobs).should be false Que.create! DB.table_exists?(:que_jobs).should be true end it "should be able to clear the jobs table" do DB[:que_jobs].insert :type => "Que::Job" DB[:que_jobs].count.should be 1 Que.clear! DB[:que_jobs].count.should be 0 end describe "when queueing jobs" do it "should be able to queue a job with arguments" do pending end it "should be able to queue a job with a specific time to run" do pending end it "should be able to queue a job with a specific priority" do pending end it "should be able to queue a job with queueing options in addition to argument options" do pending end it "should respect a default (but overridable) priority for the job class" do pending end it "should respect a default (but overridable) run_at for the job class" do pending end it "should raise an error if given arguments that can't convert to and from JSON unambiguously" do pending end end end
c8918e5ee06c3f7ec84124ebfccddb738dca2bfb
test/python_api/default-constructor/sb_communication.py
test/python_api/default-constructor/sb_communication.py
import sys import lldb def fuzz_obj(obj): broadcaster = obj.GetBroadcaster() # Do fuzz testing on the broadcaster obj, it should not crash lldb. import sb_broadcaster sb_broadcaster.fuzz_obj(broadcaster) obj.AdoptFileDesriptor(0, False) obj.AdoptFileDesriptor(1, False) obj.AdoptFileDesriptor(2, False) obj.Connect("file:/tmp/myfile") obj.Disconnect() obj.IsConnected() obj.GetCloseOnEOF() obj.SetCloseOnEOF(True) obj.SetCloseOnEOF(False) #obj.Write(None, sys.maxint, None) #obj.Read(None, sys.maxint, 0xffffffff, None) obj.ReadThreadStart() obj.ReadThreadStop() obj.ReadThreadIsRunning() obj.SetReadThreadBytesReceivedCallback(None, None)
import sys import lldb def fuzz_obj(obj): broadcaster = obj.GetBroadcaster() # Do fuzz testing on the broadcaster obj, it should not crash lldb. import sb_broadcaster sb_broadcaster.fuzz_obj(broadcaster) obj.AdoptFileDesriptor(0, False) obj.AdoptFileDesriptor(1, False) obj.AdoptFileDesriptor(2, False) obj.Connect("file:/tmp/myfile") obj.Connect(None) obj.Disconnect() obj.IsConnected() obj.GetCloseOnEOF() obj.SetCloseOnEOF(True) obj.SetCloseOnEOF(False) #obj.Write(None, sys.maxint, None) #obj.Read(None, sys.maxint, 0xffffffff, None) obj.ReadThreadStart() obj.ReadThreadStop() obj.ReadThreadIsRunning() obj.SetReadThreadBytesReceivedCallback(None, None)
Add a fuzz call for SBCommunication: obj.connect(None).
Add a fuzz call for SBCommunication: obj.connect(None). git-svn-id: 4c4cc70b1ef44ba2b7963015e681894188cea27e@146912 91177308-0d34-0410-b5e6-96231b3b80d8
Python
apache-2.0
llvm-mirror/lldb,apple/swift-lldb,llvm-mirror/lldb,apple/swift-lldb,llvm-mirror/lldb,llvm-mirror/lldb,apple/swift-lldb,apple/swift-lldb,apple/swift-lldb,llvm-mirror/lldb,apple/swift-lldb
python
## Code Before: import sys import lldb def fuzz_obj(obj): broadcaster = obj.GetBroadcaster() # Do fuzz testing on the broadcaster obj, it should not crash lldb. import sb_broadcaster sb_broadcaster.fuzz_obj(broadcaster) obj.AdoptFileDesriptor(0, False) obj.AdoptFileDesriptor(1, False) obj.AdoptFileDesriptor(2, False) obj.Connect("file:/tmp/myfile") obj.Disconnect() obj.IsConnected() obj.GetCloseOnEOF() obj.SetCloseOnEOF(True) obj.SetCloseOnEOF(False) #obj.Write(None, sys.maxint, None) #obj.Read(None, sys.maxint, 0xffffffff, None) obj.ReadThreadStart() obj.ReadThreadStop() obj.ReadThreadIsRunning() obj.SetReadThreadBytesReceivedCallback(None, None) ## Instruction: Add a fuzz call for SBCommunication: obj.connect(None). git-svn-id: 4c4cc70b1ef44ba2b7963015e681894188cea27e@146912 91177308-0d34-0410-b5e6-96231b3b80d8 ## Code After: import sys import lldb def fuzz_obj(obj): broadcaster = obj.GetBroadcaster() # Do fuzz testing on the broadcaster obj, it should not crash lldb. import sb_broadcaster sb_broadcaster.fuzz_obj(broadcaster) obj.AdoptFileDesriptor(0, False) obj.AdoptFileDesriptor(1, False) obj.AdoptFileDesriptor(2, False) obj.Connect("file:/tmp/myfile") obj.Connect(None) obj.Disconnect() obj.IsConnected() obj.GetCloseOnEOF() obj.SetCloseOnEOF(True) obj.SetCloseOnEOF(False) #obj.Write(None, sys.maxint, None) #obj.Read(None, sys.maxint, 0xffffffff, None) obj.ReadThreadStart() obj.ReadThreadStop() obj.ReadThreadIsRunning() obj.SetReadThreadBytesReceivedCallback(None, None)
ddef37f7e6fc75b314fdf64ec74929f3e305390b
developer/js/tests/test-compile-trie.ts
developer/js/tests/test-compile-trie.ts
import LexicalModelCompiler from '../'; import {assert} from 'chai'; import 'mocha'; import {makePathToFixture, compileModelSourceCode} from './helpers'; describe('LexicalModelCompiler', function () { describe('#generateLexicalModelCode', function () { const MODEL_ID = 'example.qaa.trivial'; const PATH = makePathToFixture(MODEL_ID); it('should compile a trivial word list', function () { let compiler = new LexicalModelCompiler; let code = compiler.generateLexicalModelCode(MODEL_ID, { format: 'trie-1.0', sources: ['wordlist.tsv'] }, PATH) as string; let result = compileModelSourceCode(code); assert.isFalse(result.hasSyntaxError); assert.isNotNull(result.exportedModel); assert.equal(result.modelConstructorName, 'TrieModel'); // Sanity check: the word list has three total unweighted words, with a // total weight of 3! assert.match(code, /\btotalWeight\b["']?:\s*3\b/); }); }) });
import LexicalModelCompiler from '../'; import {assert} from 'chai'; import 'mocha'; import {makePathToFixture, compileModelSourceCode} from './helpers'; describe('LexicalModelCompiler', function () { describe('#generateLexicalModelCode', function () { it('should compile a trivial word list', function () { const MODEL_ID = 'example.qaa.trivial'; const PATH = makePathToFixture(MODEL_ID); let compiler = new LexicalModelCompiler; let code = compiler.generateLexicalModelCode(MODEL_ID, { format: 'trie-1.0', sources: ['wordlist.tsv'] }, PATH) as string; let result = compileModelSourceCode(code); assert.isFalse(result.hasSyntaxError); assert.isNotNull(result.exportedModel); assert.equal(result.modelConstructorName, 'TrieModel'); // Sanity check: the word list has three total unweighted words, with a // total weight of 3! assert.match(code, /\btotalWeight\b["']?:\s*3\b/); }); it('should compile a word list exported by Microsoft Excel', function () { const MODEL_ID = 'example.qaa.utf16le'; const PATH = makePathToFixture(MODEL_ID); let compiler = new LexicalModelCompiler; let code = compiler.generateLexicalModelCode(MODEL_ID, { format: 'trie-1.0', sources: ['wordlist.txt'] }, PATH) as string; let result = compileModelSourceCode(code); assert.isFalse(result.hasSyntaxError); assert.isNotNull(result.exportedModel); assert.equal(result.modelConstructorName, 'TrieModel'); // Sanity check: the word list has three total unweighted words, with a // total weight of 44,103! assert.match(code, /\btotalWeight\b["']?:\s*44103\b/); }); }) });
Test a UTF-16LE file compiles properly.
Test a UTF-16LE file compiles properly.
TypeScript
apache-2.0
tavultesoft/keymanweb,tavultesoft/keymanweb
typescript
## Code Before: import LexicalModelCompiler from '../'; import {assert} from 'chai'; import 'mocha'; import {makePathToFixture, compileModelSourceCode} from './helpers'; describe('LexicalModelCompiler', function () { describe('#generateLexicalModelCode', function () { const MODEL_ID = 'example.qaa.trivial'; const PATH = makePathToFixture(MODEL_ID); it('should compile a trivial word list', function () { let compiler = new LexicalModelCompiler; let code = compiler.generateLexicalModelCode(MODEL_ID, { format: 'trie-1.0', sources: ['wordlist.tsv'] }, PATH) as string; let result = compileModelSourceCode(code); assert.isFalse(result.hasSyntaxError); assert.isNotNull(result.exportedModel); assert.equal(result.modelConstructorName, 'TrieModel'); // Sanity check: the word list has three total unweighted words, with a // total weight of 3! assert.match(code, /\btotalWeight\b["']?:\s*3\b/); }); }) }); ## Instruction: Test a UTF-16LE file compiles properly. ## Code After: import LexicalModelCompiler from '../'; import {assert} from 'chai'; import 'mocha'; import {makePathToFixture, compileModelSourceCode} from './helpers'; describe('LexicalModelCompiler', function () { describe('#generateLexicalModelCode', function () { it('should compile a trivial word list', function () { const MODEL_ID = 'example.qaa.trivial'; const PATH = makePathToFixture(MODEL_ID); let compiler = new LexicalModelCompiler; let code = compiler.generateLexicalModelCode(MODEL_ID, { format: 'trie-1.0', sources: ['wordlist.tsv'] }, PATH) as string; let result = compileModelSourceCode(code); assert.isFalse(result.hasSyntaxError); assert.isNotNull(result.exportedModel); assert.equal(result.modelConstructorName, 'TrieModel'); // Sanity check: the word list has three total unweighted words, with a // total weight of 3! assert.match(code, /\btotalWeight\b["']?:\s*3\b/); }); it('should compile a word list exported by Microsoft Excel', function () { const MODEL_ID = 'example.qaa.utf16le'; const PATH = makePathToFixture(MODEL_ID); let compiler = new LexicalModelCompiler; let code = compiler.generateLexicalModelCode(MODEL_ID, { format: 'trie-1.0', sources: ['wordlist.txt'] }, PATH) as string; let result = compileModelSourceCode(code); assert.isFalse(result.hasSyntaxError); assert.isNotNull(result.exportedModel); assert.equal(result.modelConstructorName, 'TrieModel'); // Sanity check: the word list has three total unweighted words, with a // total weight of 44,103! assert.match(code, /\btotalWeight\b["']?:\s*44103\b/); }); }) });
416bdc29aa6cf1ab764623f5f226e96e4b0ae081
file-processing/ignore-script-comments/ignore-script-comments-commands.sh
file-processing/ignore-script-comments/ignore-script-comments-commands.sh
sed '/^[ ]*#/d' "${sourceFile}" sed '/^[ ]*#/d' < "${sourceFile}" grep -v '^[ ]*#' "${sourceFile}" grep -v '^[ ]*#' < "${sourceFile}" egrep -v '^[ ]*#' "${sourceFile}" egrep -v '^[ ]*#' < "${sourceFile}" awk '$0 !~ /^[ ]*#/ { print $0 }' "${sourceFile}" awk '$0 !~ /^[ ]*#/ { print $0 }' < "${sourceFile}" awk '{ if ( $0 !~ "^[ ]*#" ) { print $0 } }' "${sourceFile}" awk '{ if ( $0 !~ "^[ ]*#" ) { print $0 } }' < "${sourceFile}"
sed '/^[ ]*#/d' "${sourceFile}" sed '/^[ ]*#/d' < "${sourceFile}" grep -v '^[ ]*#' "${sourceFile}" grep -v '^[ ]*#' < "${sourceFile}" grep -v '^\\s*#' "${sourceFile}" grep -v '^\\s*#' < "${sourceFile}" egrep -v '^[ ]*#' "${sourceFile}" egrep -v '^[ ]*#' < "${sourceFile}" egrep -v '^\\s*#' "${sourceFile}" egrep -v '^\\s*#' < "${sourceFile}" awk '$0 !~ /^[ ]*#/ { print $0 }' "${sourceFile}" awk '$0 !~ /^[ ]*#/ { print $0 }' < "${sourceFile}" awk '{ if ( $0 !~ "^[ ]*#" ) { print $0 } }' "${sourceFile}" awk '{ if ( $0 !~ "^[ ]*#" ) { print $0 } }' < "${sourceFile}"
Add \s for space/tab variant in grep/egrep
Add \s for space/tab variant in grep/egrep
Shell
mit
Sylvain-Bugat/linux-commands-reference,Sylvain-Bugat/linux-commands-reference
shell
## Code Before: sed '/^[ ]*#/d' "${sourceFile}" sed '/^[ ]*#/d' < "${sourceFile}" grep -v '^[ ]*#' "${sourceFile}" grep -v '^[ ]*#' < "${sourceFile}" egrep -v '^[ ]*#' "${sourceFile}" egrep -v '^[ ]*#' < "${sourceFile}" awk '$0 !~ /^[ ]*#/ { print $0 }' "${sourceFile}" awk '$0 !~ /^[ ]*#/ { print $0 }' < "${sourceFile}" awk '{ if ( $0 !~ "^[ ]*#" ) { print $0 } }' "${sourceFile}" awk '{ if ( $0 !~ "^[ ]*#" ) { print $0 } }' < "${sourceFile}" ## Instruction: Add \s for space/tab variant in grep/egrep ## Code After: sed '/^[ ]*#/d' "${sourceFile}" sed '/^[ ]*#/d' < "${sourceFile}" grep -v '^[ ]*#' "${sourceFile}" grep -v '^[ ]*#' < "${sourceFile}" grep -v '^\\s*#' "${sourceFile}" grep -v '^\\s*#' < "${sourceFile}" egrep -v '^[ ]*#' "${sourceFile}" egrep -v '^[ ]*#' < "${sourceFile}" egrep -v '^\\s*#' "${sourceFile}" egrep -v '^\\s*#' < "${sourceFile}" awk '$0 !~ /^[ ]*#/ { print $0 }' "${sourceFile}" awk '$0 !~ /^[ ]*#/ { print $0 }' < "${sourceFile}" awk '{ if ( $0 !~ "^[ ]*#" ) { print $0 } }' "${sourceFile}" awk '{ if ( $0 !~ "^[ ]*#" ) { print $0 } }' < "${sourceFile}"
b0f7a0474f84b9ce7652c93d8e7821075f20b2df
config/deploy.rb
config/deploy.rb
require "bundler/capistrano" set :deploy_via, :remote_cache set :application, "devcon" set :repository, "git://github.com/devcon-ph/devcon.git" set :deploy_to, "/home/deploy/apps/devcon-release/" set :scm, :git set :branch, "release" default_run_options[:pty] = true default_run_options[:shell] = "/bin/bash --login" server "devcon.ph", :app, :web, :db, :primary => true set :user, "deploy" set :use_sudo, false depend :remote, :gem, "bundler" set :rails_env, "production" # for carrierwave set :shared_children, shared_children + %w{public/uploads} before "deploy:finalize_update", :copy_production_database_configuration, :replace_secret_token task :copy_production_database_configuration do run "cp #{shared_path}/config/database.yml #{release_path}/config/database.yml" end task :replace_secret_token do run "cp #{shared_path}/config/secret_token.rb #{release_path}/config/initializers/secret_token.rb" end after "deploy:update", "deploy:cleanup", "deploy:migrate" require "capistrano-unicorn" after 'deploy:restart', 'unicorn:reload'
require "bundler/capistrano" set :deploy_via, :remote_cache set :application, "devcon" set :repository, "git://github.com/devcon-ph/devcon.git" set :deploy_to, "/home/deploy/apps/devcon-release/" set :scm, :git set :branch, "release" default_run_options[:pty] = true default_run_options[:shell] = "/bin/bash --login" server "devcon.ph", :app, :web, :db, :primary => true set :user, "deploy" set :use_sudo, false depend :remote, :gem, "bundler" set :rails_env, "production" # for carrierwave set :shared_children, shared_children + %w{public/uploads tmp/sockets} before "deploy:finalize_update", :copy_production_database_configuration, :replace_secret_token task :copy_production_database_configuration do run "cp #{shared_path}/config/database.yml #{release_path}/config/database.yml" end task :replace_secret_token do run "cp #{shared_path}/config/secret_token.rb #{release_path}/config/initializers/secret_token.rb" end after "deploy:update", "deploy:cleanup", "deploy:migrate" require "capistrano-unicorn" after 'deploy:restart', 'unicorn:reload'
Fix missing shared sockets folder
Fix missing shared sockets folder [ci skip]
Ruby
mit
devcon-ph/devcon,devcon-ph/devcon,devcon-ph/devcon
ruby
## Code Before: require "bundler/capistrano" set :deploy_via, :remote_cache set :application, "devcon" set :repository, "git://github.com/devcon-ph/devcon.git" set :deploy_to, "/home/deploy/apps/devcon-release/" set :scm, :git set :branch, "release" default_run_options[:pty] = true default_run_options[:shell] = "/bin/bash --login" server "devcon.ph", :app, :web, :db, :primary => true set :user, "deploy" set :use_sudo, false depend :remote, :gem, "bundler" set :rails_env, "production" # for carrierwave set :shared_children, shared_children + %w{public/uploads} before "deploy:finalize_update", :copy_production_database_configuration, :replace_secret_token task :copy_production_database_configuration do run "cp #{shared_path}/config/database.yml #{release_path}/config/database.yml" end task :replace_secret_token do run "cp #{shared_path}/config/secret_token.rb #{release_path}/config/initializers/secret_token.rb" end after "deploy:update", "deploy:cleanup", "deploy:migrate" require "capistrano-unicorn" after 'deploy:restart', 'unicorn:reload' ## Instruction: Fix missing shared sockets folder [ci skip] ## Code After: require "bundler/capistrano" set :deploy_via, :remote_cache set :application, "devcon" set :repository, "git://github.com/devcon-ph/devcon.git" set :deploy_to, "/home/deploy/apps/devcon-release/" set :scm, :git set :branch, "release" default_run_options[:pty] = true default_run_options[:shell] = "/bin/bash --login" server "devcon.ph", :app, :web, :db, :primary => true set :user, "deploy" set :use_sudo, false depend :remote, :gem, "bundler" set :rails_env, "production" # for carrierwave set :shared_children, shared_children + %w{public/uploads tmp/sockets} before "deploy:finalize_update", :copy_production_database_configuration, :replace_secret_token task :copy_production_database_configuration do run "cp #{shared_path}/config/database.yml #{release_path}/config/database.yml" end task :replace_secret_token do run "cp #{shared_path}/config/secret_token.rb #{release_path}/config/initializers/secret_token.rb" end after "deploy:update", "deploy:cleanup", "deploy:migrate" require "capistrano-unicorn" after 'deploy:restart', 'unicorn:reload'
27375ed871235c8f03a6f641201eea8400451bac
.github/workflows/main.yml
.github/workflows/main.yml
name: main on: push: pull_request: jobs: build: runs-on: windows-latest strategy: fail-fast: false matrix: include: [ { msystem: MINGW64, arch: x86_64 }, { msystem: MINGW32, arch: i686 } ] defaults: run: shell: msys2 {0} steps: - uses: actions/checkout@v2 with: fetch-depth: 0 - uses: msys2/setup-msys2@v2 with: msystem: ${{ matrix.msystem }} install: git base-devel mingw-w64-${{ matrix.arch }}-toolchain update: true - name: CI-Build run: MINGW_INSTALLS=${{ matrix.msystem }} ./ci-build.sh - name: "Upload binaries" uses: actions/upload-artifact@v2 with: name: ${{ matrix.msystem }}-packages path: artifacts/*.pkg.tar.* - name: "Upload sources" uses: actions/upload-artifact@v2 with: name: sources path: artifacts/*.src.tar.*
name: main on: push: pull_request: jobs: build: runs-on: windows-latest strategy: fail-fast: false matrix: include: [ { msystem: MINGW64, arch: x86_64 }, { msystem: MINGW32, arch: i686 } ] steps: - uses: actions/checkout@v2 with: path: temp fetch-depth: 0 - uses: msys2/setup-msys2@v2 with: msystem: ${{ matrix.msystem }} install: git base-devel mingw-w64-${{ matrix.arch }}-toolchain update: true - name: Move Checkout run: | Copy-Item -Path ".\temp" -Destination "C:\_" -Recurse - name: CI-Build shell: msys2 {0} run: | cd /C/_ MINGW_INSTALLS=${{ matrix.msystem }} ./ci-build.sh - name: "Upload binaries" uses: actions/upload-artifact@v2 with: name: ${{ matrix.msystem }}-packages path: artifacts/*.pkg.tar.* - name: "Upload sources" uses: actions/upload-artifact@v2 with: name: sources path: artifacts/*.src.tar.*
Move the checkout to C:/_
CI: Move the checkout to C:/_ To (a) get more space and (b) reduce the path lengths
YAML
bsd-3-clause
Ede123/MINGW-packages,Optiligence/MINGW-packages,JPeterMugaas/MINGW-packages,JPeterMugaas/MINGW-packages,stahta01/MINGW-packages,oroppas/MINGW-packages,lhmouse/MINGW-packages,mati865/MINGW-packages,Chocobo1/MINGW-packages,stahta01/MINGW-packages,oroppas/MINGW-packages,fabiangreffrath/MINGW-packages,RickLeaf/MINGW-packages,RickLeaf/MINGW-packages,nazar554/MINGW-packages,Alexpux/MINGW-packages,peterbud/MINGW-packages,jtanx/MINGW-packages,stahta01/MINGW-packages,bjlemke/MINGW-packages,Ede123/MINGW-packages,JPeterMugaas/MINGW-packages,Alexpux/MINGW-packages,lhmouse/MINGW-packages,Ede123/MINGW-packages,oroppas/MINGW-packages,nazar554/MINGW-packages,Chocobo1/MINGW-packages,JPeterMugaas/MINGW-packages,oscarfv/MINGW-packages,fabiangreffrath/MINGW-packages,oscarfv/MINGW-packages,peterbud/MINGW-packages,InitialForce/MINGW-packages,Ede123/MINGW-packages,Alexpux/MINGW-packages,jtanx/MINGW-packages,stahta01/MINGW-packages,Optiligence/MINGW-packages,fabiangreffrath/MINGW-packages,bjlemke/MINGW-packages,stahta01/MINGW-packages,lhmouse/MINGW-packages,jtanx/MINGW-packages,JPeterMugaas/MINGW-packages,bjlemke/MINGW-packages,oscarfv/MINGW-packages,Alexpux/MINGW-packages,lhmouse/MINGW-packages,Optiligence/MINGW-packages,InitialForce/MINGW-packages,RickLeaf/MINGW-packages,mati865/MINGW-packages,oscarfv/MINGW-packages,JPeterMugaas/MINGW-packages,oscarfv/MINGW-packages,nazar554/MINGW-packages,oroppas/MINGW-packages,stahta01/MINGW-packages,oroppas/MINGW-packages,InitialForce/MINGW-packages,mati865/MINGW-packages,oscarfv/MINGW-packages,Optiligence/MINGW-packages,peterbud/MINGW-packages,RickLeaf/MINGW-packages,mati865/MINGW-packages,oroppas/MINGW-packages,fabiangreffrath/MINGW-packages,mati865/MINGW-packages,nazar554/MINGW-packages,Alexpux/MINGW-packages,bjlemke/MINGW-packages,lhmouse/MINGW-packages,fabiangreffrath/MINGW-packages,peterbud/MINGW-packages,fabiangreffrath/MINGW-packages,Chocobo1/MINGW-packages,lhmouse/MINGW-packages,jtanx/MINGW-packages,Alexpux/MINGW-packages
yaml
## Code Before: name: main on: push: pull_request: jobs: build: runs-on: windows-latest strategy: fail-fast: false matrix: include: [ { msystem: MINGW64, arch: x86_64 }, { msystem: MINGW32, arch: i686 } ] defaults: run: shell: msys2 {0} steps: - uses: actions/checkout@v2 with: fetch-depth: 0 - uses: msys2/setup-msys2@v2 with: msystem: ${{ matrix.msystem }} install: git base-devel mingw-w64-${{ matrix.arch }}-toolchain update: true - name: CI-Build run: MINGW_INSTALLS=${{ matrix.msystem }} ./ci-build.sh - name: "Upload binaries" uses: actions/upload-artifact@v2 with: name: ${{ matrix.msystem }}-packages path: artifacts/*.pkg.tar.* - name: "Upload sources" uses: actions/upload-artifact@v2 with: name: sources path: artifacts/*.src.tar.* ## Instruction: CI: Move the checkout to C:/_ To (a) get more space and (b) reduce the path lengths ## Code After: name: main on: push: pull_request: jobs: build: runs-on: windows-latest strategy: fail-fast: false matrix: include: [ { msystem: MINGW64, arch: x86_64 }, { msystem: MINGW32, arch: i686 } ] steps: - uses: actions/checkout@v2 with: path: temp fetch-depth: 0 - uses: msys2/setup-msys2@v2 with: msystem: ${{ matrix.msystem }} install: git base-devel mingw-w64-${{ matrix.arch }}-toolchain update: true - name: Move Checkout run: | Copy-Item -Path ".\temp" -Destination "C:\_" -Recurse - name: CI-Build shell: msys2 {0} run: | cd /C/_ MINGW_INSTALLS=${{ matrix.msystem }} ./ci-build.sh - name: "Upload binaries" uses: actions/upload-artifact@v2 with: name: ${{ matrix.msystem }}-packages path: artifacts/*.pkg.tar.* - name: "Upload sources" uses: actions/upload-artifact@v2 with: name: sources path: artifacts/*.src.tar.*
1bf81b51ad340d9b29c7d8f0593f2b911598fad3
install.sh
install.sh
function bootstrap { mkdir -p ~/.vim/{backups,swaps,undo} echo '" This is where your personal configuration goes. " Whatever you write in here will override the default. ' > ~/.vim/config.vim cd ~/.vim/ git clone https://github.com/umayr/vimrc.git runtime } function extended { cd ~/.vim/runtime echo 'set runtimepath+=~/.vim/runtime source ~/.vim/runtime/config/basic.vim source ~/.vim/runtime/config/filetypes.vim source ~/.vim/runtime/config/plugins_config.vim source ~/.vim/runtime/config/extended.vim try source ~/.vim/runtime/config.vim catch endtry' > ~/.vimrc echo "Extended Vim configurations has been installed." exit 0; } function basic { cd ~/.vim/runtime cat ~/.vim/runtime/config/basic.vim > ~/.vimrc echo "Basic Vim configurations has been installed." exit 0; } bootstrap; [[ $* == *--extended* ]] && extended; basic;
function bootstrap { mkdir -p ~/.vim/{backups,swaps,undo} echo '" This is where your personal configuration goes. " Whatever you write in here will override the default. ' > ~/.vim/config.vim cd ~/.vim/ git clone --recursive https://github.com/umayr/vimrc.git runtime } function extended { cd ~/.vim/runtime echo 'set runtimepath+=~/.vim/runtime source ~/.vim/runtime/config/basic.vim source ~/.vim/runtime/config/filetypes.vim source ~/.vim/runtime/config/plugins_config.vim source ~/.vim/runtime/config/extended.vim try source ~/.vim/runtime/config.vim catch endtry' > ~/.vimrc echo "Extended Vim configurations has been installed." exit 0; } function basic { cd ~/.vim/runtime cat ~/.vim/runtime/config/basic.vim > ~/.vimrc echo "Basic Vim configurations has been installed." exit 0; } bootstrap; [[ $* == *--extended* ]] && extended; basic;
Add --recursive flag to clone repo with sub-modules
Add --recursive flag to clone repo with sub-modules
Shell
mit
umayr/vimrc
shell
## Code Before: function bootstrap { mkdir -p ~/.vim/{backups,swaps,undo} echo '" This is where your personal configuration goes. " Whatever you write in here will override the default. ' > ~/.vim/config.vim cd ~/.vim/ git clone https://github.com/umayr/vimrc.git runtime } function extended { cd ~/.vim/runtime echo 'set runtimepath+=~/.vim/runtime source ~/.vim/runtime/config/basic.vim source ~/.vim/runtime/config/filetypes.vim source ~/.vim/runtime/config/plugins_config.vim source ~/.vim/runtime/config/extended.vim try source ~/.vim/runtime/config.vim catch endtry' > ~/.vimrc echo "Extended Vim configurations has been installed." exit 0; } function basic { cd ~/.vim/runtime cat ~/.vim/runtime/config/basic.vim > ~/.vimrc echo "Basic Vim configurations has been installed." exit 0; } bootstrap; [[ $* == *--extended* ]] && extended; basic; ## Instruction: Add --recursive flag to clone repo with sub-modules ## Code After: function bootstrap { mkdir -p ~/.vim/{backups,swaps,undo} echo '" This is where your personal configuration goes. " Whatever you write in here will override the default. ' > ~/.vim/config.vim cd ~/.vim/ git clone --recursive https://github.com/umayr/vimrc.git runtime } function extended { cd ~/.vim/runtime echo 'set runtimepath+=~/.vim/runtime source ~/.vim/runtime/config/basic.vim source ~/.vim/runtime/config/filetypes.vim source ~/.vim/runtime/config/plugins_config.vim source ~/.vim/runtime/config/extended.vim try source ~/.vim/runtime/config.vim catch endtry' > ~/.vimrc echo "Extended Vim configurations has been installed." exit 0; } function basic { cd ~/.vim/runtime cat ~/.vim/runtime/config/basic.vim > ~/.vimrc echo "Basic Vim configurations has been installed." exit 0; } bootstrap; [[ $* == *--extended* ]] && extended; basic;
b21854fe55dd70c284570648a8c8fa431551c178
src/test/java/by/triumgroup/recourse/supplier/entity/model/EntityIntegerPKSupplier.java
src/test/java/by/triumgroup/recourse/supplier/entity/model/EntityIntegerPKSupplier.java
package by.triumgroup.recourse.supplier.entity.model; import by.triumgroup.recourse.entity.model.BaseEntity; import org.springframework.data.util.Pair; public interface EntityIntegerPKSupplier<E extends BaseEntity<Integer>> extends EntitySupplier<E, Integer>{ @Override default Integer getAnyId(){ return 1; } @Override default Pair<Integer, Integer> getDifferentIds(){ return Pair.of(1, 2); } }
package by.triumgroup.recourse.supplier.entity.model; import by.triumgroup.recourse.entity.model.BaseEntity; import org.springframework.data.util.Pair; import java.util.Random; public interface EntityIntegerPKSupplier<E extends BaseEntity<Integer>> extends EntitySupplier<E, Integer>{ @Override default Integer getAnyId(){ return new Random().ints(1, Integer.MAX_VALUE).findFirst().getAsInt(); } @Override default Pair<Integer, Integer> getDifferentIds(){ Integer first = getAnyId(); Integer second = first + 1; return Pair.of(first, second); } }
Add random id generating for tests
Add random id generating for tests
Java
mit
TriumGroup/ReCourse,TriumGroup/ReCourse,TriumGroup/ReCourse
java
## Code Before: package by.triumgroup.recourse.supplier.entity.model; import by.triumgroup.recourse.entity.model.BaseEntity; import org.springframework.data.util.Pair; public interface EntityIntegerPKSupplier<E extends BaseEntity<Integer>> extends EntitySupplier<E, Integer>{ @Override default Integer getAnyId(){ return 1; } @Override default Pair<Integer, Integer> getDifferentIds(){ return Pair.of(1, 2); } } ## Instruction: Add random id generating for tests ## Code After: package by.triumgroup.recourse.supplier.entity.model; import by.triumgroup.recourse.entity.model.BaseEntity; import org.springframework.data.util.Pair; import java.util.Random; public interface EntityIntegerPKSupplier<E extends BaseEntity<Integer>> extends EntitySupplier<E, Integer>{ @Override default Integer getAnyId(){ return new Random().ints(1, Integer.MAX_VALUE).findFirst().getAsInt(); } @Override default Pair<Integer, Integer> getDifferentIds(){ Integer first = getAnyId(); Integer second = first + 1; return Pair.of(first, second); } }
df41be792be5b5b44f75d3848bd01809a05eefd2
README.md
README.md
tacopajen.se ===================
[tacopajen.se](http://tacopajen.se) =================== ##### What is this? [tacopajen.se](http://tacopajen.se) is a simple webpage by and for taco pie lovers to share the best taco pie recipe. ##### The code and the commit message look awfully! Yeah, I know.. and I'm sorry for that but [tacopajen.se](http://tacopajen.se) was made on airports, bus stops etc. and in extremely short time periods, meaning each development round lasted only for a few minutes. ##### Why? This project is one of the contestants in a SEO contest. The word "tacopaj" (taco pie) is one of the most googled word in Sweden so the goal is to get as good position as possible on the SERP with the word "tacopaj" on Google. ##### Did you win? Too early to say. The competition will end February 20. ##### Can I help? Absolute! Share the URL [tacopajen.se](http://tacopajen.se), visit the site, write comments. I will love you for this.
Update readme with some valid information
Update readme with some valid information Updated the readme file with some relevant information about the project
Markdown
mit
tjoskar/tacopajen.se,tjoskar/tacopajen.se
markdown
## Code Before: tacopajen.se =================== ## Instruction: Update readme with some valid information Updated the readme file with some relevant information about the project ## Code After: [tacopajen.se](http://tacopajen.se) =================== ##### What is this? [tacopajen.se](http://tacopajen.se) is a simple webpage by and for taco pie lovers to share the best taco pie recipe. ##### The code and the commit message look awfully! Yeah, I know.. and I'm sorry for that but [tacopajen.se](http://tacopajen.se) was made on airports, bus stops etc. and in extremely short time periods, meaning each development round lasted only for a few minutes. ##### Why? This project is one of the contestants in a SEO contest. The word "tacopaj" (taco pie) is one of the most googled word in Sweden so the goal is to get as good position as possible on the SERP with the word "tacopaj" on Google. ##### Did you win? Too early to say. The competition will end February 20. ##### Can I help? Absolute! Share the URL [tacopajen.se](http://tacopajen.se), visit the site, write comments. I will love you for this.
10a345c82df463291a60957e160bdea87f729ffd
README.md
README.md
* `bower install angular-chart-donut` * `npm install angular-chart-donut` ```js angular.module('mymodule', ['chart.donut']); ``` ```html <chart-donut data="modeldata" title="Percentage awesomeness" size="150"></chart-donut> ``` <img src="https://nervetattoo.github.io/angular-chart-donut/images/shot.png"> ### Element attribute options * `data` Bind to an integer ranging from 0 to 100 * `title` Use this title for the drop-pin * `color` Optionally set a color. This can also be achieved through CSS overrides * `symbol` Defaults to `%` but can be set to other values, like empty string * `size` Pixel size of chart, if not set the chart will attempt to size itself by its parent, but if parent does not have a size while the chart is rendered this will fail. ### Contributors * [Raymond Julin](http://twitter.com/nervetattoo) * [Christian Leon Christensen](http://twitter.com/chrleon) — Design
A simple donut chart for Angular built on d3 * `bower install angular-chart-donut` * `npm install angular-chart-donut` ```js angular.module('mymodule', ['chart.donut']); ``` ```html <chart-donut data="modeldata" title="Percentage awesomeness" size="150"></chart-donut> ``` <img src="https://nervetattoo.github.io/angular-chart-donut/images/shot.png"> ### Element attribute options * `data` Bind to an integer ranging from 0 to 100 * `title` Use this title for the drop-pin * `color` Optionally set a color. This can also be achieved through CSS overrides * `symbol` Defaults to `%` but can be set to other values, like empty string * `size` Pixel size of chart, if not set the chart will attempt to size itself by its parent, but if parent does not have a size while the chart is rendered this will fail. ### Contributors * [Raymond Julin](http://twitter.com/nervetattoo) * [Christian Leon Christensen](http://twitter.com/chrleon) — Design
Add simple description to readme
Add simple description to readme
Markdown
mit
nervetattoo/angular-chart-donut
markdown
## Code Before: * `bower install angular-chart-donut` * `npm install angular-chart-donut` ```js angular.module('mymodule', ['chart.donut']); ``` ```html <chart-donut data="modeldata" title="Percentage awesomeness" size="150"></chart-donut> ``` <img src="https://nervetattoo.github.io/angular-chart-donut/images/shot.png"> ### Element attribute options * `data` Bind to an integer ranging from 0 to 100 * `title` Use this title for the drop-pin * `color` Optionally set a color. This can also be achieved through CSS overrides * `symbol` Defaults to `%` but can be set to other values, like empty string * `size` Pixel size of chart, if not set the chart will attempt to size itself by its parent, but if parent does not have a size while the chart is rendered this will fail. ### Contributors * [Raymond Julin](http://twitter.com/nervetattoo) * [Christian Leon Christensen](http://twitter.com/chrleon) — Design ## Instruction: Add simple description to readme ## Code After: A simple donut chart for Angular built on d3 * `bower install angular-chart-donut` * `npm install angular-chart-donut` ```js angular.module('mymodule', ['chart.donut']); ``` ```html <chart-donut data="modeldata" title="Percentage awesomeness" size="150"></chart-donut> ``` <img src="https://nervetattoo.github.io/angular-chart-donut/images/shot.png"> ### Element attribute options * `data` Bind to an integer ranging from 0 to 100 * `title` Use this title for the drop-pin * `color` Optionally set a color. This can also be achieved through CSS overrides * `symbol` Defaults to `%` but can be set to other values, like empty string * `size` Pixel size of chart, if not set the chart will attempt to size itself by its parent, but if parent does not have a size while the chart is rendered this will fail. ### Contributors * [Raymond Julin](http://twitter.com/nervetattoo) * [Christian Leon Christensen](http://twitter.com/chrleon) — Design
89b7a6bd1f19405be5426d32ac55fdb718b8868c
test/powershell/Tee-Object.Tests.ps1
test/powershell/Tee-Object.Tests.ps1
Describe "Tee-Object" { Context "Validate Tee-Object is correctly forking output" { $testfile = Join-Path -Path (Join-Path -Path $PSScriptRoot -ChildPath assets) -ChildPath testfile.txt It "Should return the output to the screen and to the variable" { $teefile = $testfile echo teeobjecttest1 | Tee-Object -variable teeresults $teeresults | Should Be "teeobjecttest1" Remove-Item $teefile -ErrorAction SilentlyContinue } It "Should tee the output to a file" { $teefile = $testfile echo teeobjecttest3 | Tee-Object $teefile Get-Content $teefile | Should Be "teeobjecttest3" Remove-Item $teefile -ErrorAction SilentlyContinue } } }
Describe "Tee-Object" { Context "Validate Tee-Object is correctly forking output" { $testfile = Join-Path -Path (Join-Path -Path $PSScriptRoot -ChildPath assets) -ChildPath testfile.txt It "Should return the output to the screen and to the variable" { $teefile = $testfile echo teeobjecttest1 | Tee-Object -variable teeresults $teeresults | Should Be "teeobjecttest1" Remove-Item $teefile -ErrorAction SilentlyContinue } It "Should tee the output to a file" { $teefile = $testfile echo teeobjecttest3 | Tee-Object $teefile Get-Content $teefile | Should Be "teeobjecttest3" Remove-Item $teefile -ErrorAction SilentlyContinue } } } Describe "Tee-Object DRT Unit Tests" -Tags DRT { $tempDirectory = Join-Path $TestDrive -ChildPath "TeeObjectTestsTempDir" $tempFile = "TeeObjectTestsTempFile" New-Item $tempDirectory -ItemType Directory -Force It "Positive File Test" { $expected = "1", "2", "3" $filePath = Join-Path $tempDirectory -ChildPath $tempFile $results = $expected | Tee-Object -FilePath $filePath $results.Length | Should be 3 $results | Should Be $expected $content = Get-Content $filePath $content | Should Be $expected } It "Positive Var Test" { $expected = "1", "2", "3" $varName = "teeObjectTestVar" $results = $expected | Tee-Object -Variable $varName $results.Length | Should be 3 $results | Should Be $expected $results = Get-Variable -Name $varName -ValueOnly $results.Length | Should be 3 $results | Should Be $expected } }
Add Unit Test For Tee-Object
Add Unit Test For Tee-Object
PowerShell
mit
bingbing8/PowerShell,JamesWTruher/PowerShell-1,JamesWTruher/PowerShell-1,bmanikm/PowerShell,bmanikm/PowerShell,TravisEz13/PowerShell,daxian-dbw/PowerShell,bingbing8/PowerShell,kmosher/PowerShell,KarolKaczmarek/PowerShell,kmosher/PowerShell,KarolKaczmarek/PowerShell,PaulHigin/PowerShell,bingbing8/PowerShell,PaulHigin/PowerShell,bmanikm/PowerShell,PaulHigin/PowerShell,bingbing8/PowerShell,jsoref/PowerShell,JamesWTruher/PowerShell-1,jsoref/PowerShell,bingbing8/PowerShell,jsoref/PowerShell,PaulHigin/PowerShell,daxian-dbw/PowerShell,bmanikm/PowerShell,TravisEz13/PowerShell,bmanikm/PowerShell,daxian-dbw/PowerShell,jsoref/PowerShell,kmosher/PowerShell,daxian-dbw/PowerShell,JamesWTruher/PowerShell-1,kmosher/PowerShell,KarolKaczmarek/PowerShell,TravisEz13/PowerShell,kmosher/PowerShell,KarolKaczmarek/PowerShell,KarolKaczmarek/PowerShell,TravisEz13/PowerShell,jsoref/PowerShell
powershell
## Code Before: Describe "Tee-Object" { Context "Validate Tee-Object is correctly forking output" { $testfile = Join-Path -Path (Join-Path -Path $PSScriptRoot -ChildPath assets) -ChildPath testfile.txt It "Should return the output to the screen and to the variable" { $teefile = $testfile echo teeobjecttest1 | Tee-Object -variable teeresults $teeresults | Should Be "teeobjecttest1" Remove-Item $teefile -ErrorAction SilentlyContinue } It "Should tee the output to a file" { $teefile = $testfile echo teeobjecttest3 | Tee-Object $teefile Get-Content $teefile | Should Be "teeobjecttest3" Remove-Item $teefile -ErrorAction SilentlyContinue } } } ## Instruction: Add Unit Test For Tee-Object ## Code After: Describe "Tee-Object" { Context "Validate Tee-Object is correctly forking output" { $testfile = Join-Path -Path (Join-Path -Path $PSScriptRoot -ChildPath assets) -ChildPath testfile.txt It "Should return the output to the screen and to the variable" { $teefile = $testfile echo teeobjecttest1 | Tee-Object -variable teeresults $teeresults | Should Be "teeobjecttest1" Remove-Item $teefile -ErrorAction SilentlyContinue } It "Should tee the output to a file" { $teefile = $testfile echo teeobjecttest3 | Tee-Object $teefile Get-Content $teefile | Should Be "teeobjecttest3" Remove-Item $teefile -ErrorAction SilentlyContinue } } } Describe "Tee-Object DRT Unit Tests" -Tags DRT { $tempDirectory = Join-Path $TestDrive -ChildPath "TeeObjectTestsTempDir" $tempFile = "TeeObjectTestsTempFile" New-Item $tempDirectory -ItemType Directory -Force It "Positive File Test" { $expected = "1", "2", "3" $filePath = Join-Path $tempDirectory -ChildPath $tempFile $results = $expected | Tee-Object -FilePath $filePath $results.Length | Should be 3 $results | Should Be $expected $content = Get-Content $filePath $content | Should Be $expected } It "Positive Var Test" { $expected = "1", "2", "3" $varName = "teeObjectTestVar" $results = $expected | Tee-Object -Variable $varName $results.Length | Should be 3 $results | Should Be $expected $results = Get-Variable -Name $varName -ValueOnly $results.Length | Should be 3 $results | Should Be $expected } }
12ccfab25a885a6c69685bb48dc783ecc358c86b
demo/router-components/echo.tsx
demo/router-components/echo.tsx
import app, { Component } from '../../index-jsx'; var model = 'world'; const Hello = ({name}) => <div>Hello: {name}</div>; const view = (val) => { return <div> <Hello name={val}/> <input value={val} oninput={function() { app.run('render', this.value)}}/> </div> }; const update = { '#echo': (model, pushState) => pushState || model, 'render': (_, val) => { history.pushState(null, null, '#echo/' + val); return val; } } export default (element) => app.start(element, model, view, update);
import app from '../../index-jsx'; var model = 'world'; const Hello = ({name}) => <div>Hello: {name}</div>; const view = (val) => { return <div> <Hello name={val}/> <input value={val} oninput={function() { app.run('render', this.value)}}/> </div> }; const update = { '#echo': (model, pushState) => pushState || model, 'render': (_, val) => { history.pushState(null, null, '#echo/' + val); return val; } } export default (element) => app.start(element, model, view, update);
Fix demo to use app.start instead of new Component
Fix demo to use app.start instead of new Component
TypeScript
mit
yysun/apprun,yysun/apprun,yysun/apprun
typescript
## Code Before: import app, { Component } from '../../index-jsx'; var model = 'world'; const Hello = ({name}) => <div>Hello: {name}</div>; const view = (val) => { return <div> <Hello name={val}/> <input value={val} oninput={function() { app.run('render', this.value)}}/> </div> }; const update = { '#echo': (model, pushState) => pushState || model, 'render': (_, val) => { history.pushState(null, null, '#echo/' + val); return val; } } export default (element) => app.start(element, model, view, update); ## Instruction: Fix demo to use app.start instead of new Component ## Code After: import app from '../../index-jsx'; var model = 'world'; const Hello = ({name}) => <div>Hello: {name}</div>; const view = (val) => { return <div> <Hello name={val}/> <input value={val} oninput={function() { app.run('render', this.value)}}/> </div> }; const update = { '#echo': (model, pushState) => pushState || model, 'render': (_, val) => { history.pushState(null, null, '#echo/' + val); return val; } } export default (element) => app.start(element, model, view, update);
cdc4160c8aff31bdb7859285d021f5a122ed755d
docs/sample_code/MYINST/MYINST.h
docs/sample_code/MYINST/MYINST.h
class MYINST : public Instrument { public: MYINST(); virtual ~MYINST(); virtual int init(double *, int); virtual int configure(); virtual int run(); private: void doupdate(); int _nargs, _inchan, _branch; float _amp, _pan; float *_in; };
class MYINST : public Instrument { public: MYINST(); virtual ~MYINST(); virtual int init(double *, int); virtual int configure(); virtual int run(); private: void doupdate(); float *_in; int _nargs, _inchan, _branch; float _amp, _pan; };
Change order of declarations to suppress compiler warning.
Change order of declarations to suppress compiler warning.
C
apache-2.0
RTcmix/RTcmix,RTcmix/RTcmix,RTcmix/RTcmix,RTcmix/RTcmix,RTcmix/RTcmix,RTcmix/RTcmix
c
## Code Before: class MYINST : public Instrument { public: MYINST(); virtual ~MYINST(); virtual int init(double *, int); virtual int configure(); virtual int run(); private: void doupdate(); int _nargs, _inchan, _branch; float _amp, _pan; float *_in; }; ## Instruction: Change order of declarations to suppress compiler warning. ## Code After: class MYINST : public Instrument { public: MYINST(); virtual ~MYINST(); virtual int init(double *, int); virtual int configure(); virtual int run(); private: void doupdate(); float *_in; int _nargs, _inchan, _branch; float _amp, _pan; };
762d803b88da17766b8f0347f0dd6b09cab8160f
modules/analyse/src/main/TimeChart.scala
modules/analyse/src/main/TimeChart.scala
package lila.analyse import scala.concurrent.Future import chess.Color import play.api.libs.json.Json import lila.game.{ Game, Namer } final class TimeChart(game: Game, moves: List[String]) { private val pgnMoves = moves.toIndexedSeq def series = Json stringify { Json.obj( "white" -> points(true), "black" -> points(false) ) } private def points(white: Boolean) = indexedMoveTimes collect { case (m, ply) if (white ^ (ply % 2 == 1)) => val mt = if (m < 0.5) 0 else m val san = ~(pgnMoves lift ply) val turn = 1 + (ply - 1) / 2 val dots = if (ply % 2 == 1) "..." else "." Json.obj( "name" -> s"$turn$dots $san", "x" -> ply, "y" -> (if (white) mt else -mt) ) } def maxTime = moveTimes.foldLeft(0f) { case (x, y) => if (y > x) y else x } private val moveTimes = game.moveTimesInSeconds private val indexedMoveTimes = game.moveTimesInSeconds.zipWithIndex }
package lila.analyse import scala.concurrent.Future import chess.Color import play.api.libs.json.Json import lila.game.{ Game, Namer } final class TimeChart(game: Game, moves: List[String]) { private val pgnMoves = moves.toIndexedSeq def series = Json stringify { Json.obj( "white" -> points(true), "black" -> points(false) ) } private def points(white: Boolean) = indexedMoveTimes collect { case (m, ply) if (white ^ (ply % 2 == 1)) => val mt = if (m < 0.5) 0 else m val san = ~(pgnMoves lift ply) val turn = ((ply/2).floor + 1).toInt val dots = if (ply % 2 == 1) "..." else "." Json.obj( "name" -> s"$turn$dots $san", "x" -> ply, "y" -> (if (white) mt else -mt) ) } def maxTime = moveTimes.foldLeft(0f) { case (x, y) => if (y > x) y else x } private val moveTimes = game.moveTimesInSeconds private val indexedMoveTimes = game.moveTimesInSeconds.zipWithIndex }
Fix move numbering in timeChart
Fix move numbering in timeChart
Scala
agpl-3.0
r0k3/lila,pawank/lila,TangentialAlan/lila,r0k3/lila,pawank/lila,bjhaid/lila,Happy0/lila,TangentialAlan/lila,Unihedro/lila,pavelo65/lila,elioair/lila,clarkerubber/lila,Unihedro/lila,ccampo133/lila,clarkerubber/lila,arex1337/lila,pavelo65/lila,r0k3/lila,terokinnunen/lila,systemovich/lila,clarkerubber/lila,pawank/lila,systemovich/lila,luanlv/lila,pawank/lila,elioair/lila,arex1337/lila,elioair/lila,luanlv/lila,pavelo65/lila,pavelo65/lila,r0k3/lila,bjhaid/lila,bjhaid/lila,pawank/lila,Unihedro/lila,Unihedro/lila,elioair/lila,TangentialAlan/lila,arex1337/lila,danilovsergey/i-bur,Enigmahack/lila,samuel-soubeyran/lila,systemovich/lila,JimmyMow/lila,ccampo133/lila,clarkerubber/lila,Happy0/lila,abougouffa/lila,pawank/lila,Unihedro/lila,TangentialAlan/lila,JimmyMow/lila,arex1337/lila,Happy0/lila,abougouffa/lila,ccampo133/lila,elioair/lila,samuel-soubeyran/lila,JimmyMow/lila,danilovsergey/i-bur,luanlv/lila,terokinnunen/lila,clarkerubber/lila,samuel-soubeyran/lila,r0k3/lila,samuel-soubeyran/lila,bjhaid/lila,ccampo133/lila,r0k3/lila,terokinnunen/lila,JimmyMow/lila,danilovsergey/i-bur,abougouffa/lila,JimmyMow/lila,terokinnunen/lila,bjhaid/lila,terokinnunen/lila,systemovich/lila,pavelo65/lila,clarkerubber/lila,Enigmahack/lila,ccampo133/lila,Enigmahack/lila,clarkerubber/lila,arex1337/lila,luanlv/lila,TangentialAlan/lila,abougouffa/lila,Happy0/lila,pavelo65/lila,Enigmahack/lila,Enigmahack/lila,TangentialAlan/lila,samuel-soubeyran/lila,danilovsergey/i-bur,Happy0/lila,systemovich/lila,arex1337/lila,luanlv/lila,bjhaid/lila,Happy0/lila,Unihedro/lila,systemovich/lila,luanlv/lila,Unihedro/lila,arex1337/lila,abougouffa/lila,ccampo133/lila,pavelo65/lila,danilovsergey/i-bur,JimmyMow/lila,bjhaid/lila,Enigmahack/lila,JimmyMow/lila,Enigmahack/lila,danilovsergey/i-bur,samuel-soubeyran/lila,TangentialAlan/lila,r0k3/lila,pawank/lila,luanlv/lila,terokinnunen/lila,Happy0/lila,systemovich/lila,abougouffa/lila,ccampo133/lila,danilovsergey/i-bur,terokinnunen/lila,abougouffa/lila,elioair/lila,elioair/lila,samuel-soubeyran/lila
scala
## Code Before: package lila.analyse import scala.concurrent.Future import chess.Color import play.api.libs.json.Json import lila.game.{ Game, Namer } final class TimeChart(game: Game, moves: List[String]) { private val pgnMoves = moves.toIndexedSeq def series = Json stringify { Json.obj( "white" -> points(true), "black" -> points(false) ) } private def points(white: Boolean) = indexedMoveTimes collect { case (m, ply) if (white ^ (ply % 2 == 1)) => val mt = if (m < 0.5) 0 else m val san = ~(pgnMoves lift ply) val turn = 1 + (ply - 1) / 2 val dots = if (ply % 2 == 1) "..." else "." Json.obj( "name" -> s"$turn$dots $san", "x" -> ply, "y" -> (if (white) mt else -mt) ) } def maxTime = moveTimes.foldLeft(0f) { case (x, y) => if (y > x) y else x } private val moveTimes = game.moveTimesInSeconds private val indexedMoveTimes = game.moveTimesInSeconds.zipWithIndex } ## Instruction: Fix move numbering in timeChart ## Code After: package lila.analyse import scala.concurrent.Future import chess.Color import play.api.libs.json.Json import lila.game.{ Game, Namer } final class TimeChart(game: Game, moves: List[String]) { private val pgnMoves = moves.toIndexedSeq def series = Json stringify { Json.obj( "white" -> points(true), "black" -> points(false) ) } private def points(white: Boolean) = indexedMoveTimes collect { case (m, ply) if (white ^ (ply % 2 == 1)) => val mt = if (m < 0.5) 0 else m val san = ~(pgnMoves lift ply) val turn = ((ply/2).floor + 1).toInt val dots = if (ply % 2 == 1) "..." else "." Json.obj( "name" -> s"$turn$dots $san", "x" -> ply, "y" -> (if (white) mt else -mt) ) } def maxTime = moveTimes.foldLeft(0f) { case (x, y) => if (y > x) y else x } private val moveTimes = game.moveTimesInSeconds private val indexedMoveTimes = game.moveTimesInSeconds.zipWithIndex }
69b2da9f8eab86b0e4bad789edcea0332e6d692b
requirements-dev.txt
requirements-dev.txt
pytest>=2.7.0,<=3.0.7 mock>=1.0.1 tox>=1.9.2 behave>=1.2.4 pexpect==3.3 pre-commit>=1.16.0 coverage==4.3.4 codecov>=1.5.1 docutils>=0.13.1 autopep8==1.3.3 click==6.7 twine==1.11.0
pytest>=2.7.0,<=3.0.7 mock>=1.0.1 tox>=1.9.2 behave>=1.2.4 pexpect==3.3 pre-commit>=1.16.0 coverage==4.3.4 codecov>=1.5.1 docutils>=0.13.1 autopep8==1.3.3 click==6.7 twine==1.11.0 wheel==0.33.6
Add wheel to dev reqs (needed to release).
Add wheel to dev reqs (needed to release).
Text
bsd-3-clause
dbcli/pgcli,dbcli/pgcli
text
## Code Before: pytest>=2.7.0,<=3.0.7 mock>=1.0.1 tox>=1.9.2 behave>=1.2.4 pexpect==3.3 pre-commit>=1.16.0 coverage==4.3.4 codecov>=1.5.1 docutils>=0.13.1 autopep8==1.3.3 click==6.7 twine==1.11.0 ## Instruction: Add wheel to dev reqs (needed to release). ## Code After: pytest>=2.7.0,<=3.0.7 mock>=1.0.1 tox>=1.9.2 behave>=1.2.4 pexpect==3.3 pre-commit>=1.16.0 coverage==4.3.4 codecov>=1.5.1 docutils>=0.13.1 autopep8==1.3.3 click==6.7 twine==1.11.0 wheel==0.33.6
355286252d8fedb670da44dac191414d7a7e837e
about/index.html
about/index.html
--- layout: default title: About --- <article class="page container"> {% include intro.html %} <section class="post-list"> <div class="container"> <ul class="project"> <li> <a href="http://dx.doi.org/10.14279/depositonce-4310"><img width="150px" src="../assets/icons/thesis.png"/> <h3><span>PhD thesis</span></h3></a> </li> <li> <a href="http://twoears.eu"><img width="150px" src="../assets/icons/twoears.png"/> <h3><span>Two!Ears</span></h3></a> </li> <li> <a href="http://sfstoolbox.org"><img width="150px" src="../assets/icons/sfs.png"/> <h3><span>SFS Toolbox</span></h3></a> </li> <li> <a href="http://gnuplotting.org"><img width="150px" src="../assets/icons/gnuplotting.png"/> <h3><span>Gnuplotting</span></h3></a> </li> </ul> </div> </section> </article>
--- layout: default title: About permalink: /about/ --- <article class="page container"> {% include intro.html %} <section class="post-list"> <div class="container"> <ul class="project"> <li> <a href="http://dx.doi.org/10.14279/depositonce-4310"><img width="150px" src="../assets/icons/thesis.png"/> <h3><span>PhD thesis</span></h3></a> </li> <li> <a href="http://twoears.eu"><img width="150px" src="../assets/icons/twoears.png"/> <h3><span>Two!Ears</span></h3></a> </li> <li> <a href="http://sfstoolbox.org"><img width="150px" src="../assets/icons/sfs.png"/> <h3><span>SFS Toolbox</span></h3></a> </li> <li> <a href="http://gnuplotting.org"><img width="150px" src="../assets/icons/gnuplotting.png"/> <h3><span>Gnuplotting</span></h3></a> </li> </ul> </div> </section> </article>
Set permalink name for about
Set permalink name for about
HTML
mit
hagenw/hagenw.github.io,hagenw/hagenw.github.io
html
## Code Before: --- layout: default title: About --- <article class="page container"> {% include intro.html %} <section class="post-list"> <div class="container"> <ul class="project"> <li> <a href="http://dx.doi.org/10.14279/depositonce-4310"><img width="150px" src="../assets/icons/thesis.png"/> <h3><span>PhD thesis</span></h3></a> </li> <li> <a href="http://twoears.eu"><img width="150px" src="../assets/icons/twoears.png"/> <h3><span>Two!Ears</span></h3></a> </li> <li> <a href="http://sfstoolbox.org"><img width="150px" src="../assets/icons/sfs.png"/> <h3><span>SFS Toolbox</span></h3></a> </li> <li> <a href="http://gnuplotting.org"><img width="150px" src="../assets/icons/gnuplotting.png"/> <h3><span>Gnuplotting</span></h3></a> </li> </ul> </div> </section> </article> ## Instruction: Set permalink name for about ## Code After: --- layout: default title: About permalink: /about/ --- <article class="page container"> {% include intro.html %} <section class="post-list"> <div class="container"> <ul class="project"> <li> <a href="http://dx.doi.org/10.14279/depositonce-4310"><img width="150px" src="../assets/icons/thesis.png"/> <h3><span>PhD thesis</span></h3></a> </li> <li> <a href="http://twoears.eu"><img width="150px" src="../assets/icons/twoears.png"/> <h3><span>Two!Ears</span></h3></a> </li> <li> <a href="http://sfstoolbox.org"><img width="150px" src="../assets/icons/sfs.png"/> <h3><span>SFS Toolbox</span></h3></a> </li> <li> <a href="http://gnuplotting.org"><img width="150px" src="../assets/icons/gnuplotting.png"/> <h3><span>Gnuplotting</span></h3></a> </li> </ul> </div> </section> </article>
babfc82613508e619e5debf044e68eaeae03ba11
lib/node_modules/@stdlib/math/base/dist/degenerate/logcdf/lib/logcdf.js
lib/node_modules/@stdlib/math/base/dist/degenerate/logcdf/lib/logcdf.js
'use strict'; // MODULES // var isnan = require( '@stdlib/math/base/utils/is-nan' ); var NINF = require( '@stdlib/math/constants/float64-ninf' ); // MAIN // /** * Evaluates the logarithm of the cumulative distribution function (CDF) for a degenerate distribution with a point mass at `mu`. * * @param {number} x - input value * @param {number} mu - constant value of distribution * @returns {Probability} evaluated logarithm of cumulative distribution function * * @example * var y = logcdf( 2.0, 3.0 ); * // returns Number.NEGATIVE_INFINITY * @example * var y = logcdf( 4.0, 3.0 ); * // returns 0 * @example * var y = logcdf( 3.0, 3.0 ); * // returns 0 * @example * var y = logcdf( NaN, 0.0 ); * // returns NaN * @example * var y = logcdf( 0.0, NaN ); * // returns NaN */ function logcdf( x, mu ) { if ( isnan( x ) || isnan( mu ) ) { return NaN; } return x < mu ? NINF : 0.0; } // end FUNCTION logcdf() // EXPORTS // module.exports = logcdf;
'use strict'; // MODULES // var isnan = require( '@stdlib/math/base/utils/is-nan' ); var NINF = require( '@stdlib/math/constants/float64-ninf' ); // MAIN // /** * Evaluates the natural logarithm of the cumulative distribution function (logCDF) for a degenerate distribution with mean `mu`. * * @param {number} x - input value * @param {number} mu - constant value of distribution * @returns {number} natural logarithm of cumulative distribution function * * @example * var y = logcdf( 2.0, 3.0 ); * // returns Number.NEGATIVE_INFINITY * @example * var y = logcdf( 4.0, 3.0 ); * // returns 0.0 * @example * var y = logcdf( 3.0, 3.0 ); * // returns 0.0 * @example * var y = logcdf( NaN, 0.0 ); * // returns NaN * @example * var y = logcdf( 0.0, NaN ); * // returns NaN */ function logcdf( x, mu ) { if ( isnan( x ) || isnan( mu ) ) { return NaN; } return ( x < mu ) ? NINF : 0.0; } // end FUNCTION logcdf() // EXPORTS // module.exports = logcdf;
Fix return type and add decimal points
Fix return type and add decimal points
JavaScript
apache-2.0
stdlib-js/stdlib,stdlib-js/stdlib,stdlib-js/stdlib,stdlib-js/stdlib,stdlib-js/stdlib,stdlib-js/stdlib,stdlib-js/stdlib,stdlib-js/stdlib
javascript
## Code Before: 'use strict'; // MODULES // var isnan = require( '@stdlib/math/base/utils/is-nan' ); var NINF = require( '@stdlib/math/constants/float64-ninf' ); // MAIN // /** * Evaluates the logarithm of the cumulative distribution function (CDF) for a degenerate distribution with a point mass at `mu`. * * @param {number} x - input value * @param {number} mu - constant value of distribution * @returns {Probability} evaluated logarithm of cumulative distribution function * * @example * var y = logcdf( 2.0, 3.0 ); * // returns Number.NEGATIVE_INFINITY * @example * var y = logcdf( 4.0, 3.0 ); * // returns 0 * @example * var y = logcdf( 3.0, 3.0 ); * // returns 0 * @example * var y = logcdf( NaN, 0.0 ); * // returns NaN * @example * var y = logcdf( 0.0, NaN ); * // returns NaN */ function logcdf( x, mu ) { if ( isnan( x ) || isnan( mu ) ) { return NaN; } return x < mu ? NINF : 0.0; } // end FUNCTION logcdf() // EXPORTS // module.exports = logcdf; ## Instruction: Fix return type and add decimal points ## Code After: 'use strict'; // MODULES // var isnan = require( '@stdlib/math/base/utils/is-nan' ); var NINF = require( '@stdlib/math/constants/float64-ninf' ); // MAIN // /** * Evaluates the natural logarithm of the cumulative distribution function (logCDF) for a degenerate distribution with mean `mu`. * * @param {number} x - input value * @param {number} mu - constant value of distribution * @returns {number} natural logarithm of cumulative distribution function * * @example * var y = logcdf( 2.0, 3.0 ); * // returns Number.NEGATIVE_INFINITY * @example * var y = logcdf( 4.0, 3.0 ); * // returns 0.0 * @example * var y = logcdf( 3.0, 3.0 ); * // returns 0.0 * @example * var y = logcdf( NaN, 0.0 ); * // returns NaN * @example * var y = logcdf( 0.0, NaN ); * // returns NaN */ function logcdf( x, mu ) { if ( isnan( x ) || isnan( mu ) ) { return NaN; } return ( x < mu ) ? NINF : 0.0; } // end FUNCTION logcdf() // EXPORTS // module.exports = logcdf;
b9a8ad6ade66070f022f9b808df08cc9ff8d8ab2
05-ext/CMakeLists.txt
05-ext/CMakeLists.txt
cmake_minimum_required(VERSION 3.5) project(main) include(ExternalProject) ExternalProject_Add(fooProject SOURCE_DIR ${CMAKE_CURRENT_SOURCE_DIR}/foo1 CMAKE_ARGS -DCMAKE_INSTALL_PREFIX=${CMAKE_BINARY_DIR}/ext ) ExternalProject_Get_Property(fooProject install_dir) message("fooProject install_dir=${install_dir}") ExternalProject_Add(helloProject SOURCE_DIR ${CMAKE_CURRENT_SOURCE_DIR}/hello CMAKE_ARGS -DCMAKE_MODULE_PATH=${CMAKE_BINARY_DIR}/fooProject-prefix/src/fooProject-build )
cmake_minimum_required(VERSION 3.5) project(main) include(ExternalProject) ExternalProject_Add(fooProject SOURCE_DIR ${CMAKE_CURRENT_SOURCE_DIR}/foo1 CMAKE_ARGS -DCMAKE_INSTALL_PREFIX=${CMAKE_BINARY_DIR}/ext ) ExternalProject_Get_Property(fooProject install_dir) ExternalProject_Get_Property(fooProject source_dir) ExternalProject_Get_Property(fooProject binary_dir) message("fooProject install_dir=${install_dir}") message("fooProject source_dir=${source_dir}") message("fooProject binary_dir=${binary_dir}") ExternalProject_Add(helloProject SOURCE_DIR ${CMAKE_CURRENT_SOURCE_DIR}/hello CMAKE_ARGS -DCMAKE_PREFIX_PATH=${binary_dir} )
Fix the cmake include directory
Fix the cmake include directory Signed-off-by: Francis Giraldeau <[email protected]>
Text
mit
giraldeau/cmake-tutorial,giraldeau/cmake-tutorial
text
## Code Before: cmake_minimum_required(VERSION 3.5) project(main) include(ExternalProject) ExternalProject_Add(fooProject SOURCE_DIR ${CMAKE_CURRENT_SOURCE_DIR}/foo1 CMAKE_ARGS -DCMAKE_INSTALL_PREFIX=${CMAKE_BINARY_DIR}/ext ) ExternalProject_Get_Property(fooProject install_dir) message("fooProject install_dir=${install_dir}") ExternalProject_Add(helloProject SOURCE_DIR ${CMAKE_CURRENT_SOURCE_DIR}/hello CMAKE_ARGS -DCMAKE_MODULE_PATH=${CMAKE_BINARY_DIR}/fooProject-prefix/src/fooProject-build ) ## Instruction: Fix the cmake include directory Signed-off-by: Francis Giraldeau <[email protected]> ## Code After: cmake_minimum_required(VERSION 3.5) project(main) include(ExternalProject) ExternalProject_Add(fooProject SOURCE_DIR ${CMAKE_CURRENT_SOURCE_DIR}/foo1 CMAKE_ARGS -DCMAKE_INSTALL_PREFIX=${CMAKE_BINARY_DIR}/ext ) ExternalProject_Get_Property(fooProject install_dir) ExternalProject_Get_Property(fooProject source_dir) ExternalProject_Get_Property(fooProject binary_dir) message("fooProject install_dir=${install_dir}") message("fooProject source_dir=${source_dir}") message("fooProject binary_dir=${binary_dir}") ExternalProject_Add(helloProject SOURCE_DIR ${CMAKE_CURRENT_SOURCE_DIR}/hello CMAKE_ARGS -DCMAKE_PREFIX_PATH=${binary_dir} )
6149b251296bac9e26bd81763cbf8379a743e3b5
DEVELOPER.md
DEVELOPER.md
1. Google HTML CSS Style Guide (https://google.github.io/styleguide/htmlcssguide.xml) 2. ...to be continue...
1. Google HTML CSS Style Guide (https://google.github.io/styleguide/htmlcssguide.xml) 2. Coding Style Guide - Don't add the hard code style if it can be done by CSS. ...to be continue...
Update the coding style guide
Update the coding style guide
Markdown
mit
GoEngin/GoEngin,GoEngin/GoEngin,GoEngin/GoEngin
markdown
## Code Before: 1. Google HTML CSS Style Guide (https://google.github.io/styleguide/htmlcssguide.xml) 2. ...to be continue... ## Instruction: Update the coding style guide ## Code After: 1. Google HTML CSS Style Guide (https://google.github.io/styleguide/htmlcssguide.xml) 2. Coding Style Guide - Don't add the hard code style if it can be done by CSS. ...to be continue...
a08294c731f9bd190786911830cc636f00b2af7d
console/kubernetes/darzee-service.yml
console/kubernetes/darzee-service.yml
apiVersion: apps/v1 kind: Service metadata: labels: app: darzee name: darzee-svc spec: type: LoadBalancer ports: # NOTE: Exposting ports - port: 8080 targetPort: 8080 - port: 9090 targetPort: 9090 selector: app: darzee sessionAffinity: None
apiVersion: v1 kind: Service metadata: labels: app: darzee name: darzee-svc spec: type: LoadBalancer ports: # NOTE: Exposting ports - port: 8080 name: nodejs-server-port targetPort: 8080 - port: 9090 name: prometheus-port targetPort: 9090 selector: app: darzee sessionAffinity: None
Add name to ports. Fix API version.
Add name to ports. Fix API version.
YAML
mit
emc-mongoose/console,emc-mongoose/console,emc-mongoose/console
yaml
## Code Before: apiVersion: apps/v1 kind: Service metadata: labels: app: darzee name: darzee-svc spec: type: LoadBalancer ports: # NOTE: Exposting ports - port: 8080 targetPort: 8080 - port: 9090 targetPort: 9090 selector: app: darzee sessionAffinity: None ## Instruction: Add name to ports. Fix API version. ## Code After: apiVersion: v1 kind: Service metadata: labels: app: darzee name: darzee-svc spec: type: LoadBalancer ports: # NOTE: Exposting ports - port: 8080 name: nodejs-server-port targetPort: 8080 - port: 9090 name: prometheus-port targetPort: 9090 selector: app: darzee sessionAffinity: None
1bac0ee98dc7150e51dd699da52f741924c09abb
.dotfiles/env.sh
.dotfiles/env.sh
export PATH=$PATH:/mingw64/bin export PATH=/cmd:$PATH export TERM=cygwin export GOPATH=$HOME/go export PATH=$PATH:$GOPATH/bin export PATH="$PATH:$HOME/npm/bin" export PATH="$PATH:$HOME/npm" . ~/.dotfiles/env/docker.sh . ~/.dotfiles/env/git-ssh-forwarding.sh . ~/.dotfiles/env/git.sh
export PATH=$PATH:/mingw64/bin export PATH=/cmd:$PATH export TERM=cygwin export GOPATH=$HOME/go export PATH=$PATH:$GOPATH/bin export PATH="$PATH:$HOME/npm/bin" export PATH="$PATH:$HOME/npm" export ANDROID_HOME="$LOCALAPPDATA/Android/Sdk" export PATH="$PATH:$ANDROID_HOME/platform-tools" . ~/.dotfiles/env/docker.sh . ~/.dotfiles/env/git-ssh-forwarding.sh . ~/.dotfiles/env/git.sh
Add android tools support - for react-native tools
Add android tools support - for react-native tools
Shell
mit
micnigh/windows-dotfiles
shell
## Code Before: export PATH=$PATH:/mingw64/bin export PATH=/cmd:$PATH export TERM=cygwin export GOPATH=$HOME/go export PATH=$PATH:$GOPATH/bin export PATH="$PATH:$HOME/npm/bin" export PATH="$PATH:$HOME/npm" . ~/.dotfiles/env/docker.sh . ~/.dotfiles/env/git-ssh-forwarding.sh . ~/.dotfiles/env/git.sh ## Instruction: Add android tools support - for react-native tools ## Code After: export PATH=$PATH:/mingw64/bin export PATH=/cmd:$PATH export TERM=cygwin export GOPATH=$HOME/go export PATH=$PATH:$GOPATH/bin export PATH="$PATH:$HOME/npm/bin" export PATH="$PATH:$HOME/npm" export ANDROID_HOME="$LOCALAPPDATA/Android/Sdk" export PATH="$PATH:$ANDROID_HOME/platform-tools" . ~/.dotfiles/env/docker.sh . ~/.dotfiles/env/git-ssh-forwarding.sh . ~/.dotfiles/env/git.sh
c6f913f5c74c54ec839cf78ca6d42112246617af
docs/source/releases/v1.1.rst
docs/source/releases/v1.1.rst
=========================================== Oscar 1.1 release notes - UNDER DEVELOPMENT =========================================== :release: TBD .. warning:: Since v1.1 has not been released yet, these release notes are still a work-in-progress. Welcome to Oscar 1.1! Table of contents: .. contents:: :local: :depth: 1 .. _compatibility_of_1.1: Compatibility ------------- .. _new_in_1.1: What's new in Oscar 1.1? ------------------------ .. _minor_changes_in_1.1: Minor changes ~~~~~~~~~~~~~ .. _incompatible_changes_in_1.1: Backwards incompatible changes in 1.1 ------------------------------------- Misc ~~~~ Migrations ~~~~~~~~~~ .. _deprecated_features_in_1.1: Deprecated features ~~~~~~~~~~~~~~~~~~~ The following features have been deprecated in this release: Removal of deprecated features ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ These methods have been removed: These classes have been removed:
=========================================== Oscar 1.1 release notes - UNDER DEVELOPMENT =========================================== :release: TBD .. warning:: Since v1.1 has not been released yet, these release notes are still a work-in-progress. Welcome to Oscar 1.1! Table of contents: .. contents:: :local: :depth: 1 .. _compatibility_of_1.1: Compatibility ------------- .. _new_in_1.1: What's new in Oscar 1.1? ------------------------ .. _minor_changes_in_1.1: Minor changes ~~~~~~~~~~~~~ .. _incompatible_changes_in_1.1: Backwards incompatible changes in 1.1 ------------------------------------- Misc ~~~~ * The :class:`~oscar.apps.shipping.abstract_models.AbstractWeightBased` shipping method now allows zero-weight baskets to have a non-zero shipping cost (`#1565`_). This means that sites that rely on zero-weight baskets having no change will need to introduce a new weight band that covers this edge case. .. _`#1565`: https://github.com/django-oscar/django-oscar/pull/1565 Migrations ~~~~~~~~~~ .. _deprecated_features_in_1.1: Deprecated features ~~~~~~~~~~~~~~~~~~~ The following features have been deprecated in this release: Removal of deprecated features ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ These methods have been removed: These classes have been removed:
Update release note for change to shipping method
Update release note for change to shipping method
reStructuredText
bsd-3-clause
pasqualguerrero/django-oscar,WillisXChen/django-oscar,spartonia/django-oscar,adamend/django-oscar,thechampanurag/django-oscar,sasha0/django-oscar,john-parton/django-oscar,bnprk/django-oscar,nickpack/django-oscar,binarydud/django-oscar,MatthewWilkes/django-oscar,ademuk/django-oscar,sonofatailor/django-oscar,jlmadurga/django-oscar,saadatqadri/django-oscar,sonofatailor/django-oscar,eddiep1101/django-oscar,ka7eh/django-oscar,rocopartners/django-oscar,bschuon/django-oscar,nickpack/django-oscar,michaelkuty/django-oscar,rocopartners/django-oscar,pasqualguerrero/django-oscar,jinnykoo/christmas,dongguangming/django-oscar,michaelkuty/django-oscar,eddiep1101/django-oscar,saadatqadri/django-oscar,manevant/django-oscar,kapari/django-oscar,jmt4/django-oscar,sasha0/django-oscar,anentropic/django-oscar,thechampanurag/django-oscar,vovanbo/django-oscar,WadeYuChen/django-oscar,pasqualguerrero/django-oscar,spartonia/django-oscar,spartonia/django-oscar,bschuon/django-oscar,QLGu/django-oscar,mexeniz/django-oscar,monikasulik/django-oscar,jlmadurga/django-oscar,kapari/django-oscar,MatthewWilkes/django-oscar,ademuk/django-oscar,lijoantony/django-oscar,ka7eh/django-oscar,jlmadurga/django-oscar,pdonadeo/django-oscar,Jannes123/django-oscar,nfletton/django-oscar,rocopartners/django-oscar,amirrpp/django-oscar,thechampanurag/django-oscar,Jannes123/django-oscar,kapari/django-oscar,WillisXChen/django-oscar,okfish/django-oscar,itbabu/django-oscar,ademuk/django-oscar,josesanch/django-oscar,jinnykoo/wuyisj.com,jlmadurga/django-oscar,adamend/django-oscar,sasha0/django-oscar,jinnykoo/wuyisj.com,lijoantony/django-oscar,Jannes123/django-oscar,anentropic/django-oscar,WillisXChen/django-oscar,WillisXChen/django-oscar,rocopartners/django-oscar,spartonia/django-oscar,adamend/django-oscar,vovanbo/django-oscar,nickpack/django-oscar,dongguangming/django-oscar,pdonadeo/django-oscar,Bogh/django-oscar,pasqualguerrero/django-oscar,john-parton/django-oscar,Bogh/django-oscar,anentropic/django-oscar,taedori81/django-oscar,monikasulik/django-oscar,jinnykoo/wuyisj,itbabu/django-oscar,manevant/django-oscar,okfish/django-oscar,django-oscar/django-oscar,monikasulik/django-oscar,jmt4/django-oscar,solarissmoke/django-oscar,pdonadeo/django-oscar,jinnykoo/wuyisj.com,mexeniz/django-oscar,adamend/django-oscar,kapari/django-oscar,bschuon/django-oscar,okfish/django-oscar,taedori81/django-oscar,ka7eh/django-oscar,machtfit/django-oscar,josesanch/django-oscar,anentropic/django-oscar,sonofatailor/django-oscar,binarydud/django-oscar,WillisXChen/django-oscar,WadeYuChen/django-oscar,mexeniz/django-oscar,django-oscar/django-oscar,amirrpp/django-oscar,jmt4/django-oscar,QLGu/django-oscar,john-parton/django-oscar,django-oscar/django-oscar,michaelkuty/django-oscar,machtfit/django-oscar,nfletton/django-oscar,nfletton/django-oscar,josesanch/django-oscar,jinnykoo/wuyisj,eddiep1101/django-oscar,solarissmoke/django-oscar,binarydud/django-oscar,solarissmoke/django-oscar,faratro/django-oscar,WadeYuChen/django-oscar,jinnykoo/wuyisj,django-oscar/django-oscar,vovanbo/django-oscar,binarydud/django-oscar,MatthewWilkes/django-oscar,sonofatailor/django-oscar,solarissmoke/django-oscar,nickpack/django-oscar,thechampanurag/django-oscar,dongguangming/django-oscar,mexeniz/django-oscar,nfletton/django-oscar,jinnykoo/wuyisj.com,WillisXChen/django-oscar,QLGu/django-oscar,john-parton/django-oscar,pdonadeo/django-oscar,machtfit/django-oscar,monikasulik/django-oscar,manevant/django-oscar,lijoantony/django-oscar,bnprk/django-oscar,saadatqadri/django-oscar,ka7eh/django-oscar,jinnykoo/christmas,ademuk/django-oscar,itbabu/django-oscar,Jannes123/django-oscar,itbabu/django-oscar,amirrpp/django-oscar,MatthewWilkes/django-oscar,okfish/django-oscar,WadeYuChen/django-oscar,faratro/django-oscar,taedori81/django-oscar,jmt4/django-oscar,saadatqadri/django-oscar,Bogh/django-oscar,jinnykoo/wuyisj,vovanbo/django-oscar,faratro/django-oscar,dongguangming/django-oscar,faratro/django-oscar,taedori81/django-oscar,michaelkuty/django-oscar,sasha0/django-oscar,bnprk/django-oscar,eddiep1101/django-oscar,Bogh/django-oscar,amirrpp/django-oscar,manevant/django-oscar,lijoantony/django-oscar,bschuon/django-oscar,QLGu/django-oscar,jinnykoo/christmas,bnprk/django-oscar
restructuredtext
## Code Before: =========================================== Oscar 1.1 release notes - UNDER DEVELOPMENT =========================================== :release: TBD .. warning:: Since v1.1 has not been released yet, these release notes are still a work-in-progress. Welcome to Oscar 1.1! Table of contents: .. contents:: :local: :depth: 1 .. _compatibility_of_1.1: Compatibility ------------- .. _new_in_1.1: What's new in Oscar 1.1? ------------------------ .. _minor_changes_in_1.1: Minor changes ~~~~~~~~~~~~~ .. _incompatible_changes_in_1.1: Backwards incompatible changes in 1.1 ------------------------------------- Misc ~~~~ Migrations ~~~~~~~~~~ .. _deprecated_features_in_1.1: Deprecated features ~~~~~~~~~~~~~~~~~~~ The following features have been deprecated in this release: Removal of deprecated features ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ These methods have been removed: These classes have been removed: ## Instruction: Update release note for change to shipping method ## Code After: =========================================== Oscar 1.1 release notes - UNDER DEVELOPMENT =========================================== :release: TBD .. warning:: Since v1.1 has not been released yet, these release notes are still a work-in-progress. Welcome to Oscar 1.1! Table of contents: .. contents:: :local: :depth: 1 .. _compatibility_of_1.1: Compatibility ------------- .. _new_in_1.1: What's new in Oscar 1.1? ------------------------ .. _minor_changes_in_1.1: Minor changes ~~~~~~~~~~~~~ .. _incompatible_changes_in_1.1: Backwards incompatible changes in 1.1 ------------------------------------- Misc ~~~~ * The :class:`~oscar.apps.shipping.abstract_models.AbstractWeightBased` shipping method now allows zero-weight baskets to have a non-zero shipping cost (`#1565`_). This means that sites that rely on zero-weight baskets having no change will need to introduce a new weight band that covers this edge case. .. _`#1565`: https://github.com/django-oscar/django-oscar/pull/1565 Migrations ~~~~~~~~~~ .. _deprecated_features_in_1.1: Deprecated features ~~~~~~~~~~~~~~~~~~~ The following features have been deprecated in this release: Removal of deprecated features ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ These methods have been removed: These classes have been removed:
190a71eed1e158ad61b44f8f7c51b764f52264fe
src/js/plugins/bandcamp.js
src/js/plugins/bandcamp.js
"use strict"; var $ = require("jquery"); var Plugin = require("../modules/Plugin"); var Utils = require("../modules/Utilities"); var bandcamp = Object.create(Plugin); bandcamp.init("bandcamp", "Bandcamp"); bandcamp.scrape = function () { var discover, info, isTrack, pageTitle; discover = window.location.pathname.slice(1) === "discover"; info = { stopped: !$(".inline_player .playbutton").hasClass("playing") }; if (!info.stopped) { if (discover) { info.artist = $("#detail_body_container .itemsubtext a").text(); } else { info.artist = $("span[itemprop=byArtist]").text(); } info.title = $(".track_info .title").first().text(); info.duration = Utils.calculateDuration($(".inline_player .track_info .time_total").text()); info.elapsed = Utils.calculateDuration($(".inline_player .track_info .time_elapsed").text()); if (!info.title) { info.title = $(".trackTitle").first().text(); } } return info; }; module.exports = bandcamp;
"use strict"; var $ = require("jquery"); var Plugin = require("../modules/Plugin"); var Utils = require("../modules/Utilities"); var bandcamp = Object.create(Plugin); bandcamp.init("bandcamp", "Bandcamp"); bandcamp.scrape = function () { var discover, info, isTrack, isAlbum, pageTitle; discover = window.location.hash === "#discover"; isTrack = window.location.pathname.split("/")[1] === "track"; isAlbum = window.location.pathname.split("/")[1] === "album"; info = { stopped: !$(".inline_player .playbutton").hasClass("playing") }; if (!info.stopped) { if (discover) { info.artist = $("#detail_body_container .itemsubtext a").text(); } else { info.artist = $("span[itemprop=byArtist]").text(); } if (isTrack) { info.album = $("#name-section h3 span a span").text(); } else if (isAlbum) { info.album = $("#name-section > h2").text().trim(); } else if (discover) { info.album = $("div.itemtext a").filter(function (i, e) { return e.href.match(/album/)}).text().trim(); } info.title = $(".track_info .title").first().text(); info.duration = Utils.calculateDuration($(".inline_player .track_info .time_total").text()); info.elapsed = Utils.calculateDuration($(".inline_player .track_info .time_elapsed").text()); if (!info.title) { info.title = $(".trackTitle").first().text(); } } return info; }; module.exports = bandcamp;
Add album support, fix Discover detection
Add album support, fix Discover detection
JavaScript
mit
m9x/np2file,nsue/scroblr,nwalker/scroblr,Nicholi/scroblr,Amir-Saeed/scroblr,Nicholi/scroblr,nsue/scroblr,m9x/np2file,mziyut/scroblr,cgravolet/scroblr,Amir-Saeed/scroblr,cgravolet/scroblr,nwalker/scroblr,bdukes/scroblr,mziyut/scroblr,bdukes/scroblr
javascript
## Code Before: "use strict"; var $ = require("jquery"); var Plugin = require("../modules/Plugin"); var Utils = require("../modules/Utilities"); var bandcamp = Object.create(Plugin); bandcamp.init("bandcamp", "Bandcamp"); bandcamp.scrape = function () { var discover, info, isTrack, pageTitle; discover = window.location.pathname.slice(1) === "discover"; info = { stopped: !$(".inline_player .playbutton").hasClass("playing") }; if (!info.stopped) { if (discover) { info.artist = $("#detail_body_container .itemsubtext a").text(); } else { info.artist = $("span[itemprop=byArtist]").text(); } info.title = $(".track_info .title").first().text(); info.duration = Utils.calculateDuration($(".inline_player .track_info .time_total").text()); info.elapsed = Utils.calculateDuration($(".inline_player .track_info .time_elapsed").text()); if (!info.title) { info.title = $(".trackTitle").first().text(); } } return info; }; module.exports = bandcamp; ## Instruction: Add album support, fix Discover detection ## Code After: "use strict"; var $ = require("jquery"); var Plugin = require("../modules/Plugin"); var Utils = require("../modules/Utilities"); var bandcamp = Object.create(Plugin); bandcamp.init("bandcamp", "Bandcamp"); bandcamp.scrape = function () { var discover, info, isTrack, isAlbum, pageTitle; discover = window.location.hash === "#discover"; isTrack = window.location.pathname.split("/")[1] === "track"; isAlbum = window.location.pathname.split("/")[1] === "album"; info = { stopped: !$(".inline_player .playbutton").hasClass("playing") }; if (!info.stopped) { if (discover) { info.artist = $("#detail_body_container .itemsubtext a").text(); } else { info.artist = $("span[itemprop=byArtist]").text(); } if (isTrack) { info.album = $("#name-section h3 span a span").text(); } else if (isAlbum) { info.album = $("#name-section > h2").text().trim(); } else if (discover) { info.album = $("div.itemtext a").filter(function (i, e) { return e.href.match(/album/)}).text().trim(); } info.title = $(".track_info .title").first().text(); info.duration = Utils.calculateDuration($(".inline_player .track_info .time_total").text()); info.elapsed = Utils.calculateDuration($(".inline_player .track_info .time_elapsed").text()); if (!info.title) { info.title = $(".trackTitle").first().text(); } } return info; }; module.exports = bandcamp;
4870b09309293fb053f743068f603747177088eb
org.metaborg.spoofax.core/src/main/java/org/metaborg/spoofax/core/messages/ISourceRegion.java
org.metaborg.spoofax.core/src/main/java/org/metaborg/spoofax/core/messages/ISourceRegion.java
package org.metaborg.spoofax.core.messages; public interface ISourceRegion { public int startOffset(); public int startRow(); public int startColumn(); public int endOffset(); public int endRow(); public int endColumn(); }
package org.metaborg.spoofax.core.messages; /** * Interface for representing a finite region in source code text. A region has: * <ul> * <li>Offset - number of characters from the beginning of the source text, with interval [0,#chars).</li> * <li>Row - row or line in the source text, with interval [0,#rows).</li> * <li>Column - column in the source text, with interval [0,#columns@row).</li> * </ul> * Both the starting and ending locations for these numbers are available. */ public interface ISourceRegion { public int startOffset(); public int startRow(); public int startColumn(); public int endOffset(); public int endRow(); public int endColumn(); }
Clarify intervals of source region.
Clarify intervals of source region.
Java
apache-2.0
metaborg/spoofax-eclipse,metaborg/spoofax,metaborg/spoofax,metaborg/spoofax,metaborg/spoofax
java
## Code Before: package org.metaborg.spoofax.core.messages; public interface ISourceRegion { public int startOffset(); public int startRow(); public int startColumn(); public int endOffset(); public int endRow(); public int endColumn(); } ## Instruction: Clarify intervals of source region. ## Code After: package org.metaborg.spoofax.core.messages; /** * Interface for representing a finite region in source code text. A region has: * <ul> * <li>Offset - number of characters from the beginning of the source text, with interval [0,#chars).</li> * <li>Row - row or line in the source text, with interval [0,#rows).</li> * <li>Column - column in the source text, with interval [0,#columns@row).</li> * </ul> * Both the starting and ending locations for these numbers are available. */ public interface ISourceRegion { public int startOffset(); public int startRow(); public int startColumn(); public int endOffset(); public int endRow(); public int endColumn(); }
cdc915a97a99f28332e24c4b1cac9a4c87466233
README.md
README.md
Query the host networking when working in a Docker environment on Rackspace ``` $ docker run --net=host racknet/ip public 104.130.0.127 ``` ``` $ docker run --net=host racknet/ip service 10.184.13.252 ``` ## Usage ``` racknet public [ipv4|ipv6] racknet service [ipv4|ipv6] Examples: $ racknet public 104.130.0.127 $ racknet service ipv6 fe80::be76:4eff:fe20:b452 Examples when run with Docker: $ docker run --net=host racknet/ip public 104.130.0.127 ```
Query the host networking when working in a Docker environment on Rackspace ``` $ docker run --net=host --rm racknet/ip public 104.130.0.127 ``` ``` $ docker run --net=host --rm racknet/ip service 10.184.13.252 ``` ## Usage ``` racknet public [ipv4|ipv6] racknet service [ipv4|ipv6] Examples: $ racknet public 104.130.0.127 $ racknet service ipv6 fe80::be76:4eff:fe20:b452 Examples when run with Docker: $ docker run --net=host racknet/ip public 104.130.0.127 ```
Add --rm flag, courtesy decaturdev
Add --rm flag, courtesy decaturdev
Markdown
apache-2.0
rgbkrk/docker-rackspace-net
markdown
## Code Before: Query the host networking when working in a Docker environment on Rackspace ``` $ docker run --net=host racknet/ip public 104.130.0.127 ``` ``` $ docker run --net=host racknet/ip service 10.184.13.252 ``` ## Usage ``` racknet public [ipv4|ipv6] racknet service [ipv4|ipv6] Examples: $ racknet public 104.130.0.127 $ racknet service ipv6 fe80::be76:4eff:fe20:b452 Examples when run with Docker: $ docker run --net=host racknet/ip public 104.130.0.127 ``` ## Instruction: Add --rm flag, courtesy decaturdev ## Code After: Query the host networking when working in a Docker environment on Rackspace ``` $ docker run --net=host --rm racknet/ip public 104.130.0.127 ``` ``` $ docker run --net=host --rm racknet/ip service 10.184.13.252 ``` ## Usage ``` racknet public [ipv4|ipv6] racknet service [ipv4|ipv6] Examples: $ racknet public 104.130.0.127 $ racknet service ipv6 fe80::be76:4eff:fe20:b452 Examples when run with Docker: $ docker run --net=host racknet/ip public 104.130.0.127 ```
1e22a392f818a5b309342500b5a7f81fa063ff3b
cp-content-shell.sh
cp-content-shell.sh
rm -rf ~/content-shell rm -f ~/content-shell.tar.gz mkdir ~/content-shell cd ~/src/out/Default cp icudtl.dat ~/content-shell cp *.bin ~/content-shell cp *.pak ~/content-shell cp *.so ~/content-shell cp content_shell ~/content-shell cd ~ tar -zvcf content-shell.tar.gz content-shell
rm -rf ~/content-shell rm -f ~/content-shell.tar.gz mkdir ~/content-shell cd ~/src/out/Default cp icudtl.dat ~/content-shell cp *.bin ~/content-shell cp *.pak ~/content-shell cp *.so ~/content-shell cp content_shell ~/content-shell cp chrome ~/content-shell cp -r locales ~/content-shell cp -r resources ~/content-shell cp -r ui ~/content-shell cd ~ tar -zvcf content-shell.tar.gz content-shell
Add needed files for chrome binary
Add needed files for chrome binary
Shell
mit
bjwbell/ironframe,bjwbell/ironframe,bjwbell/ironframe,bjwbell/ironframe,bjwbell/ironframe
shell
## Code Before: rm -rf ~/content-shell rm -f ~/content-shell.tar.gz mkdir ~/content-shell cd ~/src/out/Default cp icudtl.dat ~/content-shell cp *.bin ~/content-shell cp *.pak ~/content-shell cp *.so ~/content-shell cp content_shell ~/content-shell cd ~ tar -zvcf content-shell.tar.gz content-shell ## Instruction: Add needed files for chrome binary ## Code After: rm -rf ~/content-shell rm -f ~/content-shell.tar.gz mkdir ~/content-shell cd ~/src/out/Default cp icudtl.dat ~/content-shell cp *.bin ~/content-shell cp *.pak ~/content-shell cp *.so ~/content-shell cp content_shell ~/content-shell cp chrome ~/content-shell cp -r locales ~/content-shell cp -r resources ~/content-shell cp -r ui ~/content-shell cd ~ tar -zvcf content-shell.tar.gz content-shell
8f1c475347ca637417ede6b573a81aa4e8145c0b
.travis.yml
.travis.yml
language: generic services: - docker cache: directories: - $HOME/.ccache env: global: - CCACHE_DIR=$HOME/.ccache matrix: - ROS_DISTRO="kinetic" ROS_REPO=ros UPSTREAM_WORKSPACE=file CATKIN_LINT=true - ROS_DISTRO="melodic" ROS_REPO=ros UPSTREAM_WORKSPACE=file CATKIN_LINT=true - ROS_DISTRO="kinetic" CLANG_FORMAT_CHECK=file CLANG_FORMAT_VERSION=6.0 install: - git clone --depth=1 https://github.com/ros-industrial/industrial_ci.git .ci_config script: - .ci_config/travis.sh
language: generic services: - docker cache: directories: - $HOME/.ccache env: global: - CCACHE_DIR=$HOME/.ccache matrix: - ROS_DISTRO="kinetic" ROS_REPO=ros UPSTREAM_WORKSPACE=file - ROS_DISTRO="melodic" ROS_REPO=ros UPSTREAM_WORKSPACE=file - ROS_DISTRO="kinetic" CLANG_FORMAT_CHECK=file CLANG_FORMAT_VERSION=6.0 install: - git clone --depth=1 https://github.com/ros-industrial/industrial_ci.git .ci_config script: - .ci_config/travis.sh
Revert "explicitly add catkin_lint to the ci pipeline"
Revert "explicitly add catkin_lint to the ci pipeline" This reverts commit 3108a45a6ea3fa81533e7e50f62d3b1bfdc8f864.
YAML
apache-2.0
UniversalRobots/Universal_Robots_Client_Library,UniversalRobots/Universal_Robots_Client_Library
yaml
## Code Before: language: generic services: - docker cache: directories: - $HOME/.ccache env: global: - CCACHE_DIR=$HOME/.ccache matrix: - ROS_DISTRO="kinetic" ROS_REPO=ros UPSTREAM_WORKSPACE=file CATKIN_LINT=true - ROS_DISTRO="melodic" ROS_REPO=ros UPSTREAM_WORKSPACE=file CATKIN_LINT=true - ROS_DISTRO="kinetic" CLANG_FORMAT_CHECK=file CLANG_FORMAT_VERSION=6.0 install: - git clone --depth=1 https://github.com/ros-industrial/industrial_ci.git .ci_config script: - .ci_config/travis.sh ## Instruction: Revert "explicitly add catkin_lint to the ci pipeline" This reverts commit 3108a45a6ea3fa81533e7e50f62d3b1bfdc8f864. ## Code After: language: generic services: - docker cache: directories: - $HOME/.ccache env: global: - CCACHE_DIR=$HOME/.ccache matrix: - ROS_DISTRO="kinetic" ROS_REPO=ros UPSTREAM_WORKSPACE=file - ROS_DISTRO="melodic" ROS_REPO=ros UPSTREAM_WORKSPACE=file - ROS_DISTRO="kinetic" CLANG_FORMAT_CHECK=file CLANG_FORMAT_VERSION=6.0 install: - git clone --depth=1 https://github.com/ros-industrial/industrial_ci.git .ci_config script: - .ci_config/travis.sh
b064a09395534c553a385fb81ca7548359747ba1
tools/Setup.hx
tools/Setup.hx
import com.haxepunk.utils.HaxelibInfo; class Setup { public static function setup() { installDependencies(); Sys.command('haxelib run $OPENFL setup'); } public static function update() { Sys.command("haxelib update HaxePunk"); installDependencies(); } public static function installDependencies() { for (field in Reflect.fields(HaxelibInfo.install)) { Sys.command("haxelib install " + field + " " + Reflect.field(HaxelibInfo.install, field)); } } private static inline var OPENFL = "lime"; }
import com.haxepunk.utils.HaxelibInfo; class Setup { public static function setup() { installDependencies(); Sys.command("haxelib", ["run", OPENFL, "setup"]); } public static function update() { Sys.command("haxelib" ["update", "HaxePunk"]); installDependencies(); } public static function installDependencies() { for (field in Reflect.fields(HaxelibInfo.install)) { Sys.command("haxelib" ["install", field, Reflect.field(HaxelibInfo.install, field)]); } } private static inline var OPENFL = "lime"; }
Fix command call in setup tool
Fix command call in setup tool
Haxe
mit
scriptorum/HaxePunk,wselken/HaxePunk,scriptorum/HaxePunk,HaxePunk/HaxePunk,MattTuttle/HaxePunk
haxe
## Code Before: import com.haxepunk.utils.HaxelibInfo; class Setup { public static function setup() { installDependencies(); Sys.command('haxelib run $OPENFL setup'); } public static function update() { Sys.command("haxelib update HaxePunk"); installDependencies(); } public static function installDependencies() { for (field in Reflect.fields(HaxelibInfo.install)) { Sys.command("haxelib install " + field + " " + Reflect.field(HaxelibInfo.install, field)); } } private static inline var OPENFL = "lime"; } ## Instruction: Fix command call in setup tool ## Code After: import com.haxepunk.utils.HaxelibInfo; class Setup { public static function setup() { installDependencies(); Sys.command("haxelib", ["run", OPENFL, "setup"]); } public static function update() { Sys.command("haxelib" ["update", "HaxePunk"]); installDependencies(); } public static function installDependencies() { for (field in Reflect.fields(HaxelibInfo.install)) { Sys.command("haxelib" ["install", field, Reflect.field(HaxelibInfo.install, field)]); } } private static inline var OPENFL = "lime"; }
8658ce434085f0262a2e3e72e7014ac8c459c180
HISTORY.md
HISTORY.md
3.2.0 / unreleased ================== * new: Ukrainian support 3.1.0 / 2015-06-25 ================== * new: Greek support 3.0.0 / 2015-05-29 ================== * new: `largest` option to get largest units * update: languages can change decimal point (which can be overridden) * update: fix some unintuitive `round` behavior * update: specify units in shorthand rather than longhand * update: use the Unlicense * update: `nob` is now `no` * update: use underscores instead of dashes for Chinese * remove: half units 2.8.0 / 2015-05-10 ================== * new: `getSupportedLanguages` 2.7.0 / 2015-05-08 ================== * new: changelog * new: Arabic support * new: Italian support * remove: (spoken) languages from package.json keywords
3.2.0 / 2015-07-15 ================== * new: Ukrainian support * fix: things would break in global strict mode 3.1.0 / 2015-06-25 ================== * new: Greek support 3.0.0 / 2015-05-29 ================== * new: `largest` option to get largest units * update: languages can change decimal point (which can be overridden) * update: fix some unintuitive `round` behavior * update: specify units in shorthand rather than longhand * update: use the Unlicense * update: `nob` is now `no` * update: use underscores instead of dashes for Chinese * remove: half units 2.8.0 / 2015-05-10 ================== * new: `getSupportedLanguages` 2.7.0 / 2015-05-08 ================== * new: changelog * new: Arabic support * new: Italian support * remove: (spoken) languages from package.json keywords
Update changelog in preparation for 3.2.0 release
Update changelog in preparation for 3.2.0 release
Markdown
unlicense
EvanHahn/HumanizeDuration.js,smnbbrv/HumanizeDuration.js,TioNoob/HumanizeDuration.js
markdown
## Code Before: 3.2.0 / unreleased ================== * new: Ukrainian support 3.1.0 / 2015-06-25 ================== * new: Greek support 3.0.0 / 2015-05-29 ================== * new: `largest` option to get largest units * update: languages can change decimal point (which can be overridden) * update: fix some unintuitive `round` behavior * update: specify units in shorthand rather than longhand * update: use the Unlicense * update: `nob` is now `no` * update: use underscores instead of dashes for Chinese * remove: half units 2.8.0 / 2015-05-10 ================== * new: `getSupportedLanguages` 2.7.0 / 2015-05-08 ================== * new: changelog * new: Arabic support * new: Italian support * remove: (spoken) languages from package.json keywords ## Instruction: Update changelog in preparation for 3.2.0 release ## Code After: 3.2.0 / 2015-07-15 ================== * new: Ukrainian support * fix: things would break in global strict mode 3.1.0 / 2015-06-25 ================== * new: Greek support 3.0.0 / 2015-05-29 ================== * new: `largest` option to get largest units * update: languages can change decimal point (which can be overridden) * update: fix some unintuitive `round` behavior * update: specify units in shorthand rather than longhand * update: use the Unlicense * update: `nob` is now `no` * update: use underscores instead of dashes for Chinese * remove: half units 2.8.0 / 2015-05-10 ================== * new: `getSupportedLanguages` 2.7.0 / 2015-05-08 ================== * new: changelog * new: Arabic support * new: Italian support * remove: (spoken) languages from package.json keywords
eee0f7795d1da18931b27b15abcf17287b02b3af
_data/messages.yaml
_data/messages.yaml
- content: Everything up and running status: green timestamp: 2015-04-09 12:00 +0200
- content: This is a test of rss feed hooks status: green timestamp: 2015-04-09 17:40 +0200 - content: Everything up and running status: green timestamp: 2015-04-09 12:00 +0200
Add rss feed test message
Add rss feed test message
YAML
mit
frigg/status.frigg.io
yaml
## Code Before: - content: Everything up and running status: green timestamp: 2015-04-09 12:00 +0200 ## Instruction: Add rss feed test message ## Code After: - content: This is a test of rss feed hooks status: green timestamp: 2015-04-09 17:40 +0200 - content: Everything up and running status: green timestamp: 2015-04-09 12:00 +0200
f319b5ed12e04827aeed9d72d225eb227e948328
README.md
README.md
DevOps/DX for Linux. Continuous integration snippets and common tasks. * [ ] Adding automatic build to Linux projects through Travis - [![Travis](https://img.shields.io/travis/yakshaveinc/linux.svg)](https://travis-ci.org/yakshaveinc/linux) * [ ] Adding build status indicators to your development dashboard --- * [ ] Adding static site pages through Netlify * [ ] design previews * [ ] Django (https://filmfest.netlify.com) * [ ] host documentation, JAMStack sites (https://jamstack.org/) * [ ] automatic deployment * [ ] PR previews
Linux DevOps snippets, recipes and services. #### Elements * [x] GitHub (https://github.com/yakshaveinc/linux) * [x] Travis CI (https://travis-ci.org/yakshaveinc/linux) #### Ops * [x] Adding automatic builds to Linux projects through Travis - [![Travis](https://img.shields.io/travis/yakshaveinc/linux.svg)](https://travis-ci.org/yakshaveinc/linux) * [ ] Adding build status indicators to your development dashboard * [ ] Automatic image uploads from Travis CI to DockerHub --- * [ ] Adding static site pages through Netlify * [ ] design previews * [ ] Django (https://filmfest.netlify.com) * [ ] host documentation, JAMStack sites (https://jamstack.org/) * [ ] automatic deployment * [ ] PR previews
Add DockerHub to the list of services
Add DockerHub to the list of services
Markdown
unlicense
yakshaveinc/linux
markdown
## Code Before: DevOps/DX for Linux. Continuous integration snippets and common tasks. * [ ] Adding automatic build to Linux projects through Travis - [![Travis](https://img.shields.io/travis/yakshaveinc/linux.svg)](https://travis-ci.org/yakshaveinc/linux) * [ ] Adding build status indicators to your development dashboard --- * [ ] Adding static site pages through Netlify * [ ] design previews * [ ] Django (https://filmfest.netlify.com) * [ ] host documentation, JAMStack sites (https://jamstack.org/) * [ ] automatic deployment * [ ] PR previews ## Instruction: Add DockerHub to the list of services ## Code After: Linux DevOps snippets, recipes and services. #### Elements * [x] GitHub (https://github.com/yakshaveinc/linux) * [x] Travis CI (https://travis-ci.org/yakshaveinc/linux) #### Ops * [x] Adding automatic builds to Linux projects through Travis - [![Travis](https://img.shields.io/travis/yakshaveinc/linux.svg)](https://travis-ci.org/yakshaveinc/linux) * [ ] Adding build status indicators to your development dashboard * [ ] Automatic image uploads from Travis CI to DockerHub --- * [ ] Adding static site pages through Netlify * [ ] design previews * [ ] Django (https://filmfest.netlify.com) * [ ] host documentation, JAMStack sites (https://jamstack.org/) * [ ] automatic deployment * [ ] PR previews
329ab24aed5a9848e98d6a504ad8b2997142c23b
lib/main.js
lib/main.js
var buffer = require("./buffer"); /** * @namespace Splat */ module.exports = { makeBuffer: buffer.makeBuffer, flipBufferHorizontally: buffer.flipBufferHorizontally, flipBufferVertically: buffer.flipBufferVertically, ads: require("./ads"), AStar: require("./astar"), BinaryHeap: require("./binary-heap"), Game: require("./game"), iap: require("./iap"), Input: require("./input"), leaderboards: require("./leaderboards"), math: require("./math"), openUrl: require("./openUrl"), NinePatch: require("./ninepatch"), Particles: require("./particles"), saveData: require("./save-data"), Scene: require("./scene"), components: { animation: require("./components/animation"), camera: require("./components/camera"), friction: require("./components/friction"), image: require("./components/image"), movement2d: require("./components/movement-2d"), playableArea: require("./components/playable-area"), playerController2d: require("./components/player-controller-2d"), position: require("./components/position"), size: require("./components/size"), timers: require("./components/timers"), velocity: require("./components/velocity") } };
var buffer = require("./buffer"); /** * @namespace Splat */ module.exports = { makeBuffer: buffer.makeBuffer, flipBufferHorizontally: buffer.flipBufferHorizontally, flipBufferVertically: buffer.flipBufferVertically, ads: require("./ads"), AStar: require("./astar"), BinaryHeap: require("./binary-heap"), Game: require("./game"), iap: require("./iap"), Input: require("./input"), leaderboards: require("./leaderboards"), math: require("./math"), openUrl: require("./openUrl"), NinePatch: require("./ninepatch"), Particles: require("./particles"), saveData: require("./save-data"), Scene: require("./scene") };
Remove big export of components
Remove big export of components
JavaScript
mit
SplatJS/splat-ecs
javascript
## Code Before: var buffer = require("./buffer"); /** * @namespace Splat */ module.exports = { makeBuffer: buffer.makeBuffer, flipBufferHorizontally: buffer.flipBufferHorizontally, flipBufferVertically: buffer.flipBufferVertically, ads: require("./ads"), AStar: require("./astar"), BinaryHeap: require("./binary-heap"), Game: require("./game"), iap: require("./iap"), Input: require("./input"), leaderboards: require("./leaderboards"), math: require("./math"), openUrl: require("./openUrl"), NinePatch: require("./ninepatch"), Particles: require("./particles"), saveData: require("./save-data"), Scene: require("./scene"), components: { animation: require("./components/animation"), camera: require("./components/camera"), friction: require("./components/friction"), image: require("./components/image"), movement2d: require("./components/movement-2d"), playableArea: require("./components/playable-area"), playerController2d: require("./components/player-controller-2d"), position: require("./components/position"), size: require("./components/size"), timers: require("./components/timers"), velocity: require("./components/velocity") } }; ## Instruction: Remove big export of components ## Code After: var buffer = require("./buffer"); /** * @namespace Splat */ module.exports = { makeBuffer: buffer.makeBuffer, flipBufferHorizontally: buffer.flipBufferHorizontally, flipBufferVertically: buffer.flipBufferVertically, ads: require("./ads"), AStar: require("./astar"), BinaryHeap: require("./binary-heap"), Game: require("./game"), iap: require("./iap"), Input: require("./input"), leaderboards: require("./leaderboards"), math: require("./math"), openUrl: require("./openUrl"), NinePatch: require("./ninepatch"), Particles: require("./particles"), saveData: require("./save-data"), Scene: require("./scene") };
d312446805ee548188bcdf423a854b5a780f8749
.travis.yml
.travis.yml
language: csharp dist: trusty dotnet: 1.0.1 mono: none env: global: - DOTNET_SKIP_FIRST_TIME_EXPERIENCE: 1 - DOTNET_CLI_TELEMETRY_OPTOUT: 1 script: - ./bin/fetch-configlet - ./bin/configlet lint . - ./build.sh
language: csharp dist: trusty dotnet: 1.0.1 mono: none env: global: - DOTNET_SKIP_FIRST_TIME_EXPERIENCE: 1 - DOTNET_CLI_TELEMETRY_OPTOUT: 1 script: - ./bin/fetch-configlet - ./bin/configlet lint . - ./build.sh cache: directories: - $HOME/.nuget/packages
Add NuGet package caching to Travis script
Add NuGet package caching to Travis script
YAML
mit
ErikSchierboom/xcsharp,GKotfis/csharp,robkeim/xcsharp,robkeim/xcsharp,GKotfis/csharp,exercism/xcsharp,exercism/xcsharp,ErikSchierboom/xcsharp
yaml
## Code Before: language: csharp dist: trusty dotnet: 1.0.1 mono: none env: global: - DOTNET_SKIP_FIRST_TIME_EXPERIENCE: 1 - DOTNET_CLI_TELEMETRY_OPTOUT: 1 script: - ./bin/fetch-configlet - ./bin/configlet lint . - ./build.sh ## Instruction: Add NuGet package caching to Travis script ## Code After: language: csharp dist: trusty dotnet: 1.0.1 mono: none env: global: - DOTNET_SKIP_FIRST_TIME_EXPERIENCE: 1 - DOTNET_CLI_TELEMETRY_OPTOUT: 1 script: - ./bin/fetch-configlet - ./bin/configlet lint . - ./build.sh cache: directories: - $HOME/.nuget/packages
7b57656633836b48a0565a2559a9c3709d93ce5b
composer.json
composer.json
{ "name": "tomkyle/database-servicelocator", "replace": {"tomkyle/databases": "self.version"}, "type": "library", "description": "A nifty Database ServiceLocator, Pimple-style", "keywords": ["dependency injection", "database", "service locator", "mysql", "aura", "sql"], "license": "MIT", "authors": [ { "name": "Carsten Witt", "email": "[email protected]" } ], "require": { "php": ">=5.4.0", "pimple/pimple": "dev-master" }, "suggest": { "aura/sql": "Aura.SQL – Provides adapters for SQL database access around PDO" }, "autoload": { "psr-4": { "tomkyle\\Databases\\": "src/" } } }
{ "name": "tomkyle/databases", "replace": {"tomkyle/databases": "self.version"}, "type": "library", "description": "A nifty Database ServiceLocator, Pimple-style", "keywords": ["dependency injection", "database", "service locator", "mysql", "aura", "sql"], "license": "MIT", "authors": [ { "name": "Carsten Witt", "email": "[email protected]" } ], "require": { "php": ">=5.4.0", "pimple/pimple": "dev-master" }, "suggest": { "aura/sql": "Aura.SQL – Provides adapters for SQL database access around PDO" }, "autoload": { "psr-4": { "tomkyle\\Databases\\": "src/" } } }
Update with 'new old' package name…
Update with 'new old' package name…
JSON
mit
tomkyle/Databases
json
## Code Before: { "name": "tomkyle/database-servicelocator", "replace": {"tomkyle/databases": "self.version"}, "type": "library", "description": "A nifty Database ServiceLocator, Pimple-style", "keywords": ["dependency injection", "database", "service locator", "mysql", "aura", "sql"], "license": "MIT", "authors": [ { "name": "Carsten Witt", "email": "[email protected]" } ], "require": { "php": ">=5.4.0", "pimple/pimple": "dev-master" }, "suggest": { "aura/sql": "Aura.SQL – Provides adapters for SQL database access around PDO" }, "autoload": { "psr-4": { "tomkyle\\Databases\\": "src/" } } } ## Instruction: Update with 'new old' package name… ## Code After: { "name": "tomkyle/databases", "replace": {"tomkyle/databases": "self.version"}, "type": "library", "description": "A nifty Database ServiceLocator, Pimple-style", "keywords": ["dependency injection", "database", "service locator", "mysql", "aura", "sql"], "license": "MIT", "authors": [ { "name": "Carsten Witt", "email": "[email protected]" } ], "require": { "php": ">=5.4.0", "pimple/pimple": "dev-master" }, "suggest": { "aura/sql": "Aura.SQL – Provides adapters for SQL database access around PDO" }, "autoload": { "psr-4": { "tomkyle\\Databases\\": "src/" } } }
24faff8b034a8d57a9e67249dbf182c2c1c8ca9c
.github/workflows/build_msw_xml_libs.yml
.github/workflows/build_msw_xml_libs.yml
name: XML MSW binaries on: workflow_dispatch: jobs: build-xml-libs: name: Build XML binaries runs-on: ubuntu-latest steps: - name: Checkout uses: actions/checkout@v2 - name: Install Compiler run: | sudo apt-get update -qq sudo apt-get install -qq --no-install-recommends g++-mingw-w64-i686 g++-mingw-w64-x86-64 - name: Build 32-bit Libraries run: | HOST=i686-w64-mingw32 ./scripts/ci/install.sh - name: Build 64-bit Libraries run: | HOST=x86_64-w64-mingw32 ./scripts/ci/install.sh - name: Upload uses: actions/upload-artifact@v3 with: name: xml-libs path: /usr/local
name: XML MSW binaries on: workflow_dispatch: jobs: build-xml-libs: name: Build XML binaries runs-on: ubuntu-latest steps: - name: Checkout uses: actions/checkout@v2 - name: Install Compiler run: | sudo apt-get update -qq sudo apt-get install -qq --no-install-recommends g++-mingw-w64-i686 g++-mingw-w64-x86-64 - name: Build 32-bit Libraries run: | HOST=i686-w64-mingw32 ./scripts/install_deps.sh - name: Build 64-bit Libraries run: | HOST=x86_64-w64-mingw32 ./scripts/install_deps.sh - name: Upload uses: actions/upload-artifact@v3 with: name: xml-libs path: /usr/local
Fix wrong script name used in the workflow file
Fix wrong script name used in the workflow file
YAML
bsd-3-clause
vslavik/xmlwrapp,vslavik/xmlwrapp,vslavik/xmlwrapp
yaml
## Code Before: name: XML MSW binaries on: workflow_dispatch: jobs: build-xml-libs: name: Build XML binaries runs-on: ubuntu-latest steps: - name: Checkout uses: actions/checkout@v2 - name: Install Compiler run: | sudo apt-get update -qq sudo apt-get install -qq --no-install-recommends g++-mingw-w64-i686 g++-mingw-w64-x86-64 - name: Build 32-bit Libraries run: | HOST=i686-w64-mingw32 ./scripts/ci/install.sh - name: Build 64-bit Libraries run: | HOST=x86_64-w64-mingw32 ./scripts/ci/install.sh - name: Upload uses: actions/upload-artifact@v3 with: name: xml-libs path: /usr/local ## Instruction: Fix wrong script name used in the workflow file ## Code After: name: XML MSW binaries on: workflow_dispatch: jobs: build-xml-libs: name: Build XML binaries runs-on: ubuntu-latest steps: - name: Checkout uses: actions/checkout@v2 - name: Install Compiler run: | sudo apt-get update -qq sudo apt-get install -qq --no-install-recommends g++-mingw-w64-i686 g++-mingw-w64-x86-64 - name: Build 32-bit Libraries run: | HOST=i686-w64-mingw32 ./scripts/install_deps.sh - name: Build 64-bit Libraries run: | HOST=x86_64-w64-mingw32 ./scripts/install_deps.sh - name: Upload uses: actions/upload-artifact@v3 with: name: xml-libs path: /usr/local
8d1929c2b56630889e64a8671ac635633a015632
README.rst
README.rst
Swoopi ====== .. image:: https://img.shields.io/travis/swoopi/swoopi/master.svg?maxAge=3600?style=flat-square :target: https://travis-ci.org/swoopi/swoopi .. image:: https://img.shields.io/coveralls/swoopi/swoopi/master.svg?maxAge=3600?style=flat-square :target: https://coveralls.io/github/swoopi/swoopi?branch=master Camera interactions for the Raspberry Pi.
Swoopi ====== |Travis|_ |Coveralls|_ Camera interactions for the Raspberry Pi. .. |Travis| image:: https://img.shields.io/travis/swoopi/swoopi/master.svg?maxAge=3600?style=flat-square .. _Travis: https://travis-ci.org/swoopi/swoopi .. |Coveralls| image:: https://img.shields.io/coveralls/swoopi/swoopi/master.svg?maxAge=3600?style=flat-square .. _Coveralls: https://coveralls.io/github/swoopi/swoopi?branch=master
Change shields to rst references
Change shields to rst references
reStructuredText
apache-2.0
swoopi/swoopi,YouLookFamiliar/swoop
restructuredtext
## Code Before: Swoopi ====== .. image:: https://img.shields.io/travis/swoopi/swoopi/master.svg?maxAge=3600?style=flat-square :target: https://travis-ci.org/swoopi/swoopi .. image:: https://img.shields.io/coveralls/swoopi/swoopi/master.svg?maxAge=3600?style=flat-square :target: https://coveralls.io/github/swoopi/swoopi?branch=master Camera interactions for the Raspberry Pi. ## Instruction: Change shields to rst references ## Code After: Swoopi ====== |Travis|_ |Coveralls|_ Camera interactions for the Raspberry Pi. .. |Travis| image:: https://img.shields.io/travis/swoopi/swoopi/master.svg?maxAge=3600?style=flat-square .. _Travis: https://travis-ci.org/swoopi/swoopi .. |Coveralls| image:: https://img.shields.io/coveralls/swoopi/swoopi/master.svg?maxAge=3600?style=flat-square .. _Coveralls: https://coveralls.io/github/swoopi/swoopi?branch=master
02b60cc05265429b90ac7289086661a7d8bd7814
pac_script.js
pac_script.js
var BOARD_HEIGHT = 288; var BOARD_WIDTH = 224; var VERT_TILES = BOARD_HEIGHT / 8; var HORIZ_TILES = BOARD_WIDTH / 8; gameBoard = new Array(VERT_TILES); for(var y = 0; y < VERT_TILES; y++) { gameBoard[y] = new Array(HORIZ_TILES); for(var x = 0; x < HORIZ_TILES; x++) { gameBoard[y][x] = 0; } } var canvas, context; var ready = function(fun) { if(document.readyState != "loading") { fun(); } else if(document.addEventListener) { document.addEventListener("DOMContentLoaded", fun); } else { document.attachEvent("onreadystatechange", function() { if(document.readyState != "loading") { fun(); } }); } } ready(function() { canvas = document.getElementById("board"); context = canvas.getContext("2d"); });
var BOARD_HEIGHT = 288; var BOARD_WIDTH = 224; var VERT_TILES = BOARD_HEIGHT / 8; var HORIZ_TILES = BOARD_WIDTH / 8; gameBoard = new Array(VERT_TILES); for(var y = 0; y < VERT_TILES; y++) { gameBoard[y] = new Array(HORIZ_TILES); for(var x = 0; x < HORIZ_TILES; x++) { gameBoard[y][x] = ""; } } var canvas, context; var drawObject = function(args) { args.context.clearRect(0, 0, BOARD_WIDTH, BOARD_HEIGHT); args.context.fillStyle = args.color; args.objectArr.forEach(function(row, rIndex) { row.forEach(function(col, cIndex) { if(col == 1) { args.context.fillRect(args.x + cIndex, args.y + rIndex, 1, 1); } }); }); } var ready = function(fun) { if(document.readyState != "loading") { fun(); } else if(document.addEventListener) { document.addEventListener("DOMContentLoaded", fun); } else { document.attachEvent("onreadystatechange", function() { if(document.readyState != "loading") { fun(); } }); } } ready(function() { canvas = document.getElementById("board"); context = canvas.getContext("2d"); drawObject({ x: 80, y: 80, objectArr: charset["A"], color: "#FFF", context: context }); });
Add drawObject function to draw matrix of pixels
Add drawObject function to draw matrix of pixels
JavaScript
mit
peternatewood/pac-man-replica,peternatewood/pac-man-replica
javascript
## Code Before: var BOARD_HEIGHT = 288; var BOARD_WIDTH = 224; var VERT_TILES = BOARD_HEIGHT / 8; var HORIZ_TILES = BOARD_WIDTH / 8; gameBoard = new Array(VERT_TILES); for(var y = 0; y < VERT_TILES; y++) { gameBoard[y] = new Array(HORIZ_TILES); for(var x = 0; x < HORIZ_TILES; x++) { gameBoard[y][x] = 0; } } var canvas, context; var ready = function(fun) { if(document.readyState != "loading") { fun(); } else if(document.addEventListener) { document.addEventListener("DOMContentLoaded", fun); } else { document.attachEvent("onreadystatechange", function() { if(document.readyState != "loading") { fun(); } }); } } ready(function() { canvas = document.getElementById("board"); context = canvas.getContext("2d"); }); ## Instruction: Add drawObject function to draw matrix of pixels ## Code After: var BOARD_HEIGHT = 288; var BOARD_WIDTH = 224; var VERT_TILES = BOARD_HEIGHT / 8; var HORIZ_TILES = BOARD_WIDTH / 8; gameBoard = new Array(VERT_TILES); for(var y = 0; y < VERT_TILES; y++) { gameBoard[y] = new Array(HORIZ_TILES); for(var x = 0; x < HORIZ_TILES; x++) { gameBoard[y][x] = ""; } } var canvas, context; var drawObject = function(args) { args.context.clearRect(0, 0, BOARD_WIDTH, BOARD_HEIGHT); args.context.fillStyle = args.color; args.objectArr.forEach(function(row, rIndex) { row.forEach(function(col, cIndex) { if(col == 1) { args.context.fillRect(args.x + cIndex, args.y + rIndex, 1, 1); } }); }); } var ready = function(fun) { if(document.readyState != "loading") { fun(); } else if(document.addEventListener) { document.addEventListener("DOMContentLoaded", fun); } else { document.attachEvent("onreadystatechange", function() { if(document.readyState != "loading") { fun(); } }); } } ready(function() { canvas = document.getElementById("board"); context = canvas.getContext("2d"); drawObject({ x: 80, y: 80, objectArr: charset["A"], color: "#FFF", context: context }); });
9504529dd4b9140be0026d0b30a0e88e5dea5e25
rtrss/config.py
rtrss/config.py
import os import logging import importlib # All configuration defaults are set in this module TRACKER_HOST = 'rutracker.org' # Timeone for the tracker times TZNAME = 'Europe/Moscow' LOGLEVEL = logging.INFO LOG_FORMAT_LOGENTRIES = '%(levelname)s %(name)s %(message)s' LOG_FORMAT_BRIEF = '%(asctime)s %(levelname)s %(name)s %(message)s' ADMIN_LOGIN = os.environ.get('ADMIN_LOGIN', 'admin') ADMIN_PASSWORD = os.environ.get('ADMIN_PASSWORD', 'admin') ADMIN_EMAIL = os.environ.get('ADMIN_EMAIL', 'admin@localhost') # path to save torrent files TORRENT_PATH_PATTERN = 'torrents/{}.torrent' APP_ENVIRONMENT = os.environ.get('RTRSS_ENVIRONMENT') if not APP_ENVIRONMENT: raise EnvironmentError('RTRSS_ENVIRONMENT must be set') _mod = importlib.import_module('rtrss.config_{}'.format(APP_ENVIRONMENT)) _envconf = {k: v for k, v in _mod.__dict__.items() if k == k.upper()} globals().update(_envconf)
import os import logging import importlib # All configuration defaults are set in this module TRACKER_HOST = 'rutracker.org' # Timeone for the tracker times TZNAME = 'Europe/Moscow' LOGLEVEL = logging.INFO LOG_FORMAT_LOGENTRIES = '%(levelname)s %(name)s %(message)s' LOG_FORMAT_BRIEF = '%(asctime)s %(levelname)s %(name)s %(message)s' ADMIN_LOGIN = os.environ.get('ADMIN_LOGIN', 'admin') ADMIN_PASSWORD = os.environ.get('ADMIN_PASSWORD', 'admin') ADMIN_EMAIL = os.environ.get('ADMIN_EMAIL', 'admin@localhost') # path to save torrent files TORRENT_PATH_PATTERN = 'torrents/{}.torrent' APP_ENVIRONMENT = os.environ.get('RTRSS_ENVIRONMENT') if not APP_ENVIRONMENT: raise EnvironmentError('RTRSS_ENVIRONMENT must be set') IP = '0.0.0.0' PORT = 8080 _mod = importlib.import_module('rtrss.config_{}'.format(APP_ENVIRONMENT)) _envconf = {k: v for k, v in _mod.__dict__.items() if k == k.upper()} globals().update(_envconf)
Add default IP and PORT
Add default IP and PORT
Python
apache-2.0
notapresent/rtrss,notapresent/rtrss,notapresent/rtrss,notapresent/rtrss
python
## Code Before: import os import logging import importlib # All configuration defaults are set in this module TRACKER_HOST = 'rutracker.org' # Timeone for the tracker times TZNAME = 'Europe/Moscow' LOGLEVEL = logging.INFO LOG_FORMAT_LOGENTRIES = '%(levelname)s %(name)s %(message)s' LOG_FORMAT_BRIEF = '%(asctime)s %(levelname)s %(name)s %(message)s' ADMIN_LOGIN = os.environ.get('ADMIN_LOGIN', 'admin') ADMIN_PASSWORD = os.environ.get('ADMIN_PASSWORD', 'admin') ADMIN_EMAIL = os.environ.get('ADMIN_EMAIL', 'admin@localhost') # path to save torrent files TORRENT_PATH_PATTERN = 'torrents/{}.torrent' APP_ENVIRONMENT = os.environ.get('RTRSS_ENVIRONMENT') if not APP_ENVIRONMENT: raise EnvironmentError('RTRSS_ENVIRONMENT must be set') _mod = importlib.import_module('rtrss.config_{}'.format(APP_ENVIRONMENT)) _envconf = {k: v for k, v in _mod.__dict__.items() if k == k.upper()} globals().update(_envconf) ## Instruction: Add default IP and PORT ## Code After: import os import logging import importlib # All configuration defaults are set in this module TRACKER_HOST = 'rutracker.org' # Timeone for the tracker times TZNAME = 'Europe/Moscow' LOGLEVEL = logging.INFO LOG_FORMAT_LOGENTRIES = '%(levelname)s %(name)s %(message)s' LOG_FORMAT_BRIEF = '%(asctime)s %(levelname)s %(name)s %(message)s' ADMIN_LOGIN = os.environ.get('ADMIN_LOGIN', 'admin') ADMIN_PASSWORD = os.environ.get('ADMIN_PASSWORD', 'admin') ADMIN_EMAIL = os.environ.get('ADMIN_EMAIL', 'admin@localhost') # path to save torrent files TORRENT_PATH_PATTERN = 'torrents/{}.torrent' APP_ENVIRONMENT = os.environ.get('RTRSS_ENVIRONMENT') if not APP_ENVIRONMENT: raise EnvironmentError('RTRSS_ENVIRONMENT must be set') IP = '0.0.0.0' PORT = 8080 _mod = importlib.import_module('rtrss.config_{}'.format(APP_ENVIRONMENT)) _envconf = {k: v for k, v in _mod.__dict__.items() if k == k.upper()} globals().update(_envconf)
c937f502cd1efcf01eb4b9fe899912c75dae08f7
.appveyor.yml
.appveyor.yml
version: 2.0.{build} # select image with MinGW-w64 available image: Visual Studio 2013 clone_folder: c:\projects\sif environment: global: PATH: C:\mingw-w64\i686-5.3.0-posix-dwarf-rt_v4-rev0\mingw32\bin;C:\msys64\usr\bin;$(PATH) CXX: g++ CC: gcc GCOV: gcov MAKE: make install: # show all env vars #- set # display available tools - '%CC% --version' - pwd build_script: - cmd /c build_all.bat artifacts: - path: obj32\*.exe name: built sif2
version: 2.0.{build} # select image with MinGW-w64 available image: Visual Studio 2013 environment: global: PATH: C:\mingw-w64\i686-5.3.0-posix-dwarf-rt_v4-rev0\mingw32\bin;C:\msys64\usr\bin;$(PATH) CXX: g++ CC: gcc GCOV: gcov MAKE: make install: # show PATH env var #- echo %PATH% - '%CC% --version' - pwd build_script: - cmd /c build_all.bat artifacts: - path: obj32\*.exe name: built sif2
Remove clone folder setting from AppVeyor config
Remove clone folder setting from AppVeyor config
YAML
mit
goriy/sif,goriy/sif
yaml
## Code Before: version: 2.0.{build} # select image with MinGW-w64 available image: Visual Studio 2013 clone_folder: c:\projects\sif environment: global: PATH: C:\mingw-w64\i686-5.3.0-posix-dwarf-rt_v4-rev0\mingw32\bin;C:\msys64\usr\bin;$(PATH) CXX: g++ CC: gcc GCOV: gcov MAKE: make install: # show all env vars #- set # display available tools - '%CC% --version' - pwd build_script: - cmd /c build_all.bat artifacts: - path: obj32\*.exe name: built sif2 ## Instruction: Remove clone folder setting from AppVeyor config ## Code After: version: 2.0.{build} # select image with MinGW-w64 available image: Visual Studio 2013 environment: global: PATH: C:\mingw-w64\i686-5.3.0-posix-dwarf-rt_v4-rev0\mingw32\bin;C:\msys64\usr\bin;$(PATH) CXX: g++ CC: gcc GCOV: gcov MAKE: make install: # show PATH env var #- echo %PATH% - '%CC% --version' - pwd build_script: - cmd /c build_all.bat artifacts: - path: obj32\*.exe name: built sif2
5d80a93196f1822f3688e0612fa6c73a7679235b
steer_bot_hardware_gazebo/config/joints.yaml
steer_bot_hardware_gazebo/config/joints.yaml
steer_bot_hardware_gazebo: rear_wheel : 'base_to_wheel' front_steer : 'base_to_steer' virtual_rear_wheels : ['base_to_right_rear_wheel', 'base_to_left_rear_wheel'] virtual_front_wheels : ['base_to_right_front_wheel', 'base_to_left_front_wheel'] virtual_front_steers : ['base_to_right_front_steer', 'base_to_left_front_steer'] # ackermann link mechanism enable_ackermann_link: 1 wheel_separation_w : 0.5 wheel_separation_h : 0.79
steer_bot_hardware_gazebo: rear_wheel : 'rear_wheel_joint' front_steer : 'front_steer_joint' virtual_rear_wheels : ['base_to_right_rear_wheel', 'base_to_left_rear_wheel'] virtual_front_wheels : ['base_to_right_front_wheel', 'base_to_left_front_wheel'] virtual_front_steers : ['base_to_right_front_steer', 'base_to_left_front_steer'] # ackermann link mechanism enable_ackermann_link: 1 wheel_separation_w : 0.5 wheel_separation_h : 0.79
Update joint setting of steer_bot for new steer_drive configulation.
Update joint setting of steer_bot for new steer_drive configulation.
YAML
mit
CIR-KIT/steer_drive_ros,CIR-KIT/steer_drive_ros
yaml
## Code Before: steer_bot_hardware_gazebo: rear_wheel : 'base_to_wheel' front_steer : 'base_to_steer' virtual_rear_wheels : ['base_to_right_rear_wheel', 'base_to_left_rear_wheel'] virtual_front_wheels : ['base_to_right_front_wheel', 'base_to_left_front_wheel'] virtual_front_steers : ['base_to_right_front_steer', 'base_to_left_front_steer'] # ackermann link mechanism enable_ackermann_link: 1 wheel_separation_w : 0.5 wheel_separation_h : 0.79 ## Instruction: Update joint setting of steer_bot for new steer_drive configulation. ## Code After: steer_bot_hardware_gazebo: rear_wheel : 'rear_wheel_joint' front_steer : 'front_steer_joint' virtual_rear_wheels : ['base_to_right_rear_wheel', 'base_to_left_rear_wheel'] virtual_front_wheels : ['base_to_right_front_wheel', 'base_to_left_front_wheel'] virtual_front_steers : ['base_to_right_front_steer', 'base_to_left_front_steer'] # ackermann link mechanism enable_ackermann_link: 1 wheel_separation_w : 0.5 wheel_separation_h : 0.79
3b879db6b81844cdd64c4f3fb34814daa18bb68d
_source/_sass/_notices.scss
_source/_sass/_notices.scss
// NOTICES // -------------------------------------------------- @mixin notice($notice-color) { position: relative; padding: 1.5em; font-family: $alt-font; @include font-size(14,39); color: $white; background-color: $notice-color; border-radius: $border-radius; } .notice { @include notice($primary-color); } .notice-inverse { @include notice($white); color: $text-color; } .notice-info { @include notice($info-color); } .notice-warning { @include notice($warning-color); } .notice-success { @include notice($success-color); } .notice-danger { @include notice($danger-color); }
// NOTICES // -------------------------------------------------- @mixin notice($notice-color) { position: relative; padding: 1.5em; font-family: $alt-font; @include font-size(14,39); color: $white; background-color: $notice-color; border-radius: $border-radius; a { color: $white; border-bottom: 1px dotted $white !important; } } .notice { @include notice($primary-color); } .notice-inverse { @include notice($white); color: $text-color; a { color: $text-color; border-bottom: 1px dotted $text-color !important; } } .notice-info { @include notice($info-color); } .notice-warning { @include notice($warning-color); } .notice-success { @include notice($success-color); } .notice-danger { @include notice($danger-color); }
Fix link colors on notices
Fix link colors on notices
SCSS
mit
BrianLusina/brianlusina.github.io,QBayLogic/qbaylogic.com,ucuptimposu/ucuptimposu.github.io,hootnot/skinny-bones-jekyll,dgrtwo/rgallery-default,tzandrki/tzandrki.github.io,Shinobi881/skinny-bones-jekyll,CmdrKerfy/old_ind-volcanoes,dgrtwo/broom-gallery,tobbio/tobbio.github.io,frassati-new-haven/frassati-new-haven.github.io,chengdubuddy/chengdubuddy.github.io,l1905/l1905.github.io,jpfreeley/jpfreeley.github.io,dgrtwo/rgallery-default,piapandelakis/piapandelakis.github.io,teear/teear.github.com,plinioseniore/skinnybonesjeckyll,philippeowagner/skinny-bones-jekyll,cooldepartment/skinny-bones-jekyll,swat-ds/zombies-reimagined,cwigington3/cwigington3.github.io,pricelessbrewing/pricelessbrewing.github.io,jasonpchang/jasonpchang.github.io,gladtosee/skinny-bones-jekyll,nikolspencer/skinny-bones-jekyll,juvham/skinny-bones-jekyll,pricelessbrewing/pricelessbrewing.github.io,IncaProject/IncaProject.github.io,joehalliwell/joehalliwell.github.io,heeryung/skinny-bones-jekyll,iambsk1/iambsk1.github.io,open-learning/open-learning.github.io,cooldepartment/skinny-bones-jekyll,mightymetris/mightymetris.github.io,Shinobi881/skinny-bones-jekyll,warrenryu/warrenryu.github.io,iambsk1/iambsk1.github.io,hbc/hbc.github.io,massich/massich.github.io,BrianLusina/brianlusina.github.io,sgng/sgng.github.io,nagoya-workshop/skinny-bones-jekyll,GuyPaddock/guypaddock.github.io,adeldaoud/adeldaoud.github.io,teelahti/kerkesix.fi,BetsyHamlinArt/BetsyHamlinArt.github.io,zomeelee/skinny-bones-jekyll,warrenryu/warrenryu.github.io,iambsk1/iambsk1.github.io,ckochx/vapeblog,swat-ds/zombies-reimagined,AswiniC/skinny-bones-jekyll,BetsyHamlinArt/BetsyHamlinArt.github.io,chroni-blog/chroni,teear/teear.github.com,hootnot/skinny-bones-jekyll,cyproassist/cyproassist.github.io,AZURECITADEL/azurecitadel.github.io,Bigmonstercai/skinny-bones-jekyll,deaconblade/deaconblade.github.io,CmdrKerfy/old_ind-volcanoes,hsaghir/hsaghir.github.io,piapandelakis/piapandelakis.github.io,cyproassist/cyproassist.github.io,deaconblade/deaconblade.github.io,hsaghir/hsaghir.github.io,SusannStruwe/cyproassist.github.io,AswiniC/skinny-bones-jekyll,chengdubuddy/chengdubuddy.github.io,damienkilgannon/damienkilgannon.github.io,isbead/skinny-bones-jekyll,heeryung/skinny-bones-jekyll,alumnosdcao/alumnosdcao.github.io,SusannStruwe/cyproassist.github.io,emptymalei/planets-jekyll,SusannStruwe/cyproassist.github.io,joehalliwell/joehalliwell.github.io,mightymetris/mightymetris.github.io,vynci/phosabyss-dev,zomeelee/skinny-bones-jekyll,cooldepartment/skinny-bones-jekyll,ucuptimposu/ucuptimposu.github.io,philippeowagner/skinny-bones-jekyll,zomeelee/skinny-bones-jekyll,adeldaoud/adeldaoud.github.io,opengers/opengers.github.io,QN7Gardening/QN7Gardening.github.io,chengdubuddy/chengdubuddy.github.io,carlosdelfino/carlosdelfino.github.com,l1905/l1905.github.io,damienkilgannon/damienkilgannon.github.io,sgng/sgng.github.io,alumnosdcao/alumnosdcao.github.io,AswiniC/skinny-bones-jekyll,QN7Gardening/QN7Gardening.github.io,eliteconceptsmn/main-web,zbintliff/zbintliff.github.io,tobbio/tobbio.github.io,GuyPaddock/guypaddock.github.io,imran031387/mimran.github.io,vynci/phosabyss-dev,adeldaoud/adeldaoud.github.io,BetsyHamlinArt/BetsyHamlinArt.github.io,luislhl/luislhl.github.io,juvham/skinny-bones-jekyll,iyamegane/iyamegane.github.io,heeryung/skinny-bones-jekyll,IncaProject/IncaProject.github.io,CmdrKerfy/old_ind-volcanoes,emptymalei/planets-jekyll,hootnot/skinny-bones-jekyll,nagoya-workshop/skinny-bones-jekyll,piapandelakis/piapandelakis.github.io,vynci/phosabyss-dev,tobbio/tobbio.github.io,mikaelkaron/open-learning.github.io,hbc/hbc.github.io,deaconblade/deaconblade.github.io,teear/teear.github.com,eliteconceptsmn/main-web,Bigmonstercai/skinny-bones-jekyll,laredaccion/laredaccion.github.io,QBayLogic/qbaylogic.com,philippeowagner/skinny-bones-jekyll,sgng/sgng.github.io,mikaelkaron/open-learning.github.io,chroni-blog/chroni,jpfreeley/jpfreeley.github.io,carlosdelfino/carlosdelfino.github.com,mikaelkaron/open-learning.github.io,mmistakes/skinny-bones-jekyll,QBayLogic/qbaylogic.com,massich/massich.github.io,hbc/hbc.github.io,mmistakes/skinny-bones-jekyll,pricelessbrewing/pricelessbrewing.github.io,baumanab/baumanab.github.io,thecraftingcorollary/thecraftingcorollary.github.io,juvham/skinny-bones-jekyll,thecraftingcorollary/thecraftingcorollary.github.io,dgrtwo/rgallery-default,carlesgutierrez/carlesgutierrez.github.io,dgrtwo/broom-gallery,cwigington3/cwigington3.github.io,dgrtwo/broom-gallery,baumanab/baumanab.github.io,AZURECITADEL/azurecitadel.github.io,thecraftingcorollary/thecraftingcorollary.github.io,GuyPaddock/guypaddock.github.io,l1905/l1905.github.io,eliteconceptsmn/main-web,nikolspencer/skinny-bones-jekyll,cyproassist/cyproassist.github.io,swat-ds/zombies-reimagined,AlexPutz/alexputz.github.io,mmistakes/skinny-bones-jekyll,baumanab/baumanab.github.io,luislhl/luislhl.github.io,jasonpchang/jasonpchang.github.io,nagoya-workshop/skinny-bones-jekyll,luislhl/luislhl.github.io,jasonpchang/jasonpchang.github.io,open-learning/open-learning.github.io,IncaProject/IncaProject.github.io,IncaProject/IncaProject.github.io,AZURECITADEL/azurecitadel.github.io,damienkilgannon/damienkilgannon.github.io,frassati-new-haven/frassati-new-haven.github.io,warrenryu/warrenryu.github.io,isbead/skinny-bones-jekyll,isbead/skinny-bones-jekyll,bruce-lcf/bruce-lcf.github.io,iyamegane/iyamegane.github.io,carlosdelfino/carlosdelfino.github.com,carlosdelfino/carlosdelfino.github.com,open-learning/open-learning.github.io,ckochx/vapeblog,praxeos/praxeos.github.io,plinioseniore/skinnybonesjeckyll,frassati-new-haven/frassati-new-haven.github.io,jpfreeley/jpfreeley.github.io,emptymalei/planets-jekyll,laredaccion/laredaccion.github.io,carlesgutierrez/carlesgutierrez.github.io,AlexPutz/alexputz.github.io,BrianLusina/brianlusina.github.io,nikolspencer/skinny-bones-jekyll,massich/massich.github.io,rasnes/rasnes.github.io,chroni-blog/chroni,zbintliff/zbintliff.github.io,rasnes/rasnes.github.io,praxeos/praxeos.github.io,imran031387/mimran.github.io,Bigmonstercai/skinny-bones-jekyll,teelahti/kerkesix.fi,QN7Gardening/QN7Gardening.github.io,alumnosdcao/alumnosdcao.github.io,mightymetris/mightymetris.github.io,gladtosee/skinny-bones-jekyll,Shinobi881/skinny-bones-jekyll,hsaghir/hsaghir.github.io,gladtosee/skinny-bones-jekyll,cwigington3/cwigington3.github.io,imran031387/mimran.github.io,carlesgutierrez/carlesgutierrez.github.io,laredaccion/laredaccion.github.io,joehalliwell/joehalliwell.github.io,hsaghir/hsaghir.github.io,ucuptimposu/ucuptimposu.github.io,emptymalei/planets-jekyll,iyamegane/iyamegane.github.io,AlexPutz/alexputz.github.io,rasnes/rasnes.github.io,BrianLusina/brianlusina.github.io,plinioseniore/skinnybonesjeckyll,zbintliff/zbintliff.github.io
scss
## Code Before: // NOTICES // -------------------------------------------------- @mixin notice($notice-color) { position: relative; padding: 1.5em; font-family: $alt-font; @include font-size(14,39); color: $white; background-color: $notice-color; border-radius: $border-radius; } .notice { @include notice($primary-color); } .notice-inverse { @include notice($white); color: $text-color; } .notice-info { @include notice($info-color); } .notice-warning { @include notice($warning-color); } .notice-success { @include notice($success-color); } .notice-danger { @include notice($danger-color); } ## Instruction: Fix link colors on notices ## Code After: // NOTICES // -------------------------------------------------- @mixin notice($notice-color) { position: relative; padding: 1.5em; font-family: $alt-font; @include font-size(14,39); color: $white; background-color: $notice-color; border-radius: $border-radius; a { color: $white; border-bottom: 1px dotted $white !important; } } .notice { @include notice($primary-color); } .notice-inverse { @include notice($white); color: $text-color; a { color: $text-color; border-bottom: 1px dotted $text-color !important; } } .notice-info { @include notice($info-color); } .notice-warning { @include notice($warning-color); } .notice-success { @include notice($success-color); } .notice-danger { @include notice($danger-color); }
d3fd194c82b2d8f701b52d67a95a54bb622eaa53
multeval.sh
multeval.sh
set -eo pipefail version=0.1 scriptDir=$(dirname $0) $scriptDir/get_deps.sh java -XX:+UseCompressedOops -Xmx2g -cp $scriptDir/lib/tercom-0.7.26.jar:$scriptDir/lib/meteor-1.2/meteor-1.2.jar:$scriptDir/multeval-${version}.jar multeval.MultEval $@
set -eo pipefail version=0.1 scriptDir=$(dirname $0) $scriptDir/get_deps.sh # Specify -Dfile.encoding so that the meteor paraphrase tables load properly java -Dfile.encoding=UTF8 -XX:+UseCompressedOops -Xmx2g \ -cp $scriptDir/lib/tercom-0.7.26.jar:$scriptDir/lib/meteor-1.2/meteor-1.2.jar:$scriptDir/multeval-${version}.jar \ multeval.MultEval $@
Use file.encoding to fix loading meteor paraphrase tables
Use file.encoding to fix loading meteor paraphrase tables
Shell
mpl-2.0
jhclark/multeval,jhclark/multeval,jhclark/multeval,jhclark/multeval
shell
## Code Before: set -eo pipefail version=0.1 scriptDir=$(dirname $0) $scriptDir/get_deps.sh java -XX:+UseCompressedOops -Xmx2g -cp $scriptDir/lib/tercom-0.7.26.jar:$scriptDir/lib/meteor-1.2/meteor-1.2.jar:$scriptDir/multeval-${version}.jar multeval.MultEval $@ ## Instruction: Use file.encoding to fix loading meteor paraphrase tables ## Code After: set -eo pipefail version=0.1 scriptDir=$(dirname $0) $scriptDir/get_deps.sh # Specify -Dfile.encoding so that the meteor paraphrase tables load properly java -Dfile.encoding=UTF8 -XX:+UseCompressedOops -Xmx2g \ -cp $scriptDir/lib/tercom-0.7.26.jar:$scriptDir/lib/meteor-1.2/meteor-1.2.jar:$scriptDir/multeval-${version}.jar \ multeval.MultEval $@
06a9c42fb2163156ff2147fccc88bb73db08e76b
exports.js
exports.js
// Matching the public exports in husl-colors/husl module['exports'] = {}; module['exports']["fromRGB"] = husl.Husl.rgbToHusl; module['exports']["fromHex"] = husl.Husl.hexToHusl; module['exports']["toRGB"] = husl.Husl.huslToRgb; module['exports']["toHex"] = husl.Husl.huslToHex; module['exports']['p'] = {}; module['exports']['p']["fromRGB"] = husl.Husl.rgbToHuslp; module['exports']['p']["fromHex"] = husl.Husl.hexToHuslp; module['exports']['p']["toRGB"] = husl.Husl.huslpToRgb; module['exports']['p']["toHex"] = husl.Husl.huslpToHex;
// Matching the public exports in husl-colors/husl function expandParams(f) { return function(c1, c2, c3) { return f([c1, c2, c3]) } } module['exports'] = {}; module['exports']["fromRGB"] = expandParams(husl.Husl.rgbToHusl); module['exports']["fromHex"] = husl.Husl.hexToHusl; module['exports']["toRGB"] = expandParams(husl.Husl.huslToRgb); module['exports']["toHex"] = expandParams(husl.Husl.huslToHex); module['exports']['p'] = {}; module['exports']['p']["fromRGB"] = expandParams(husl.Husl.rgbToHuslp); module['exports']['p']["fromHex"] = husl.Husl.hexToHuslp; module['exports']['p']["toRGB"] = expandParams(husl.Husl.huslpToRgb); module['exports']['p']["toHex"] = expandParams(husl.Husl.huslpToHex);
Add adaptor for multiple argument functions
Add adaptor for multiple argument functions
JavaScript
mit
hsluv/hsluv,hsluv/hsluv
javascript
## Code Before: // Matching the public exports in husl-colors/husl module['exports'] = {}; module['exports']["fromRGB"] = husl.Husl.rgbToHusl; module['exports']["fromHex"] = husl.Husl.hexToHusl; module['exports']["toRGB"] = husl.Husl.huslToRgb; module['exports']["toHex"] = husl.Husl.huslToHex; module['exports']['p'] = {}; module['exports']['p']["fromRGB"] = husl.Husl.rgbToHuslp; module['exports']['p']["fromHex"] = husl.Husl.hexToHuslp; module['exports']['p']["toRGB"] = husl.Husl.huslpToRgb; module['exports']['p']["toHex"] = husl.Husl.huslpToHex; ## Instruction: Add adaptor for multiple argument functions ## Code After: // Matching the public exports in husl-colors/husl function expandParams(f) { return function(c1, c2, c3) { return f([c1, c2, c3]) } } module['exports'] = {}; module['exports']["fromRGB"] = expandParams(husl.Husl.rgbToHusl); module['exports']["fromHex"] = husl.Husl.hexToHusl; module['exports']["toRGB"] = expandParams(husl.Husl.huslToRgb); module['exports']["toHex"] = expandParams(husl.Husl.huslToHex); module['exports']['p'] = {}; module['exports']['p']["fromRGB"] = expandParams(husl.Husl.rgbToHuslp); module['exports']['p']["fromHex"] = husl.Husl.hexToHuslp; module['exports']['p']["toRGB"] = expandParams(husl.Husl.huslpToRgb); module['exports']['p']["toHex"] = expandParams(husl.Husl.huslpToHex);
5acad5c62ec204c5a190bb3f5b2516153b4869b1
.travis.yml
.travis.yml
node_js: - "6" - "7" - "8" - "9" language: node_js script: "npm run-script test-travis" after_script: "npm install coveralls@2 && cat ./coverage/lcov.info | coveralls" before_script: - curl http://ffmpeg.gusari.org/static/64bit/ffmpeg.static.64bit.latest.tar.gz | sudo tar -C /usr/local/bin/ -xvz
node_js: - "6" - "7" - "8" - "9" language: node_js script: "npm run-script test-travis" after_script: "npm install coveralls@2 && cat ./coverage/lcov.info | coveralls" before_script: - curl https://johnvansickle.com/ffmpeg/builds/ffmpeg-git-64bit-static.tar.xz | sudo tar -C /usr/local/bin/ -xvJ
Fix FFMPEG download URL for tests
:green_heart: Fix FFMPEG download URL for tests
YAML
mit
caffco/get-video-duration,caffco/get-video-duration
yaml
## Code Before: node_js: - "6" - "7" - "8" - "9" language: node_js script: "npm run-script test-travis" after_script: "npm install coveralls@2 && cat ./coverage/lcov.info | coveralls" before_script: - curl http://ffmpeg.gusari.org/static/64bit/ffmpeg.static.64bit.latest.tar.gz | sudo tar -C /usr/local/bin/ -xvz ## Instruction: :green_heart: Fix FFMPEG download URL for tests ## Code After: node_js: - "6" - "7" - "8" - "9" language: node_js script: "npm run-script test-travis" after_script: "npm install coveralls@2 && cat ./coverage/lcov.info | coveralls" before_script: - curl https://johnvansickle.com/ffmpeg/builds/ffmpeg-git-64bit-static.tar.xz | sudo tar -C /usr/local/bin/ -xvJ
52ffc2b264cbacaee56017cd4a67df4511d60392
celery/managers.py
celery/managers.py
from django.db import models from celery.registry import tasks from datetime import datetime, timedelta __all__ = ["TaskManager", "PeriodicTaskManager"] class TaskManager(models.Manager): def get_task(self, task_id): task, created = self.get_or_create(task_id=task_id) return task def is_done(self, task_id): return self.get_task(task_id).is_done def get_all_expired(self): return self.filter(date_done__lt=datetime.now() - timedelta(days=5)) def delete_expired(self): self.get_all_expired().delete() def mark_as_done(self, task_id): task, created = self.get_or_create(task_id=task_id, defaults={ "is_done": True}) if not created: task.is_done = True task.save() class PeriodicTaskManager(models.Manager): def get_waiting_tasks(self): periodic_tasks = tasks.get_all_periodic() waiting = [] for task_name, task in periodic_tasks.items(): task_meta, created = self.get_or_create(name=task_name) # task_run.every must be a timedelta object. run_at = task_meta.last_run_at + task.run_every if datetime.now() > run_at: waiting.append(task_meta) return waiting
from django.db import models from celery.registry import tasks from datetime import datetime, timedelta __all__ = ["TaskManager", "PeriodicTaskManager"] class TaskManager(models.Manager): def get_task(self, task_id): task, created = self.get_or_create(task_id=task_id) return task def is_done(self, task_id): return self.get_task(task_id).is_done def get_all_expired(self): return self.filter(date_done__lt=datetime.now() - timedelta(days=5), is_done=True) def delete_expired(self): self.get_all_expired().delete() def mark_as_done(self, task_id): task, created = self.get_or_create(task_id=task_id, defaults={ "is_done": True}) if not created: task.is_done = True task.save() class PeriodicTaskManager(models.Manager): def get_waiting_tasks(self): periodic_tasks = tasks.get_all_periodic() waiting = [] for task_name, task in periodic_tasks.items(): task_meta, created = self.get_or_create(name=task_name) # task_run.every must be a timedelta object. run_at = task_meta.last_run_at + task.run_every if datetime.now() > run_at: waiting.append(task_meta) return waiting
Add is_done=True to get_all_expired filter.
Add is_done=True to get_all_expired filter.
Python
bsd-3-clause
WoLpH/celery,cbrepo/celery,cbrepo/celery,ask/celery,frac/celery,WoLpH/celery,mitsuhiko/celery,ask/celery,mitsuhiko/celery,frac/celery
python
## Code Before: from django.db import models from celery.registry import tasks from datetime import datetime, timedelta __all__ = ["TaskManager", "PeriodicTaskManager"] class TaskManager(models.Manager): def get_task(self, task_id): task, created = self.get_or_create(task_id=task_id) return task def is_done(self, task_id): return self.get_task(task_id).is_done def get_all_expired(self): return self.filter(date_done__lt=datetime.now() - timedelta(days=5)) def delete_expired(self): self.get_all_expired().delete() def mark_as_done(self, task_id): task, created = self.get_or_create(task_id=task_id, defaults={ "is_done": True}) if not created: task.is_done = True task.save() class PeriodicTaskManager(models.Manager): def get_waiting_tasks(self): periodic_tasks = tasks.get_all_periodic() waiting = [] for task_name, task in periodic_tasks.items(): task_meta, created = self.get_or_create(name=task_name) # task_run.every must be a timedelta object. run_at = task_meta.last_run_at + task.run_every if datetime.now() > run_at: waiting.append(task_meta) return waiting ## Instruction: Add is_done=True to get_all_expired filter. ## Code After: from django.db import models from celery.registry import tasks from datetime import datetime, timedelta __all__ = ["TaskManager", "PeriodicTaskManager"] class TaskManager(models.Manager): def get_task(self, task_id): task, created = self.get_or_create(task_id=task_id) return task def is_done(self, task_id): return self.get_task(task_id).is_done def get_all_expired(self): return self.filter(date_done__lt=datetime.now() - timedelta(days=5), is_done=True) def delete_expired(self): self.get_all_expired().delete() def mark_as_done(self, task_id): task, created = self.get_or_create(task_id=task_id, defaults={ "is_done": True}) if not created: task.is_done = True task.save() class PeriodicTaskManager(models.Manager): def get_waiting_tasks(self): periodic_tasks = tasks.get_all_periodic() waiting = [] for task_name, task in periodic_tasks.items(): task_meta, created = self.get_or_create(name=task_name) # task_run.every must be a timedelta object. run_at = task_meta.last_run_at + task.run_every if datetime.now() > run_at: waiting.append(task_meta) return waiting
7da7354668187167ccc6143ee0f469158c858ed3
lib/economic/proxies/actions/creditor_contact_proxy/find_by_name.rb
lib/economic/proxies/actions/creditor_contact_proxy/find_by_name.rb
module Economic module Proxies module Actions module CreditorContactProxy class FindByName attr_reader :name def initialize(caller, name) @caller = caller @name = name end def call # Get a list of CreditorContactHandles from e-conomic response = request('FindByName', { 'name' => name }) # Make sure we always have an array of handles even if the result only contains one handles = [response[:creditor_contact_handle]].flatten.reject(&:blank?) # Create partial CreditorContact entities contacts = handles.collect do |handle| creditor_contact = build creditor_contact.partial = true creditor_contact.persisted = true creditor_contact.handle = handle creditor_contact.id = handle[:id] creditor_contact.number = handle[:number] creditor_contact end if owner.is_a?(Creditor) # Scope to the owner contacts.select do |creditor_contact| creditor_contact.get_data creditor_contact.creditor.handle == owner.handle end else contacts end end private def build(*options) @caller.build(options) end def owner @caller.owner end def request(action, data) @caller.request(action, data) end end end end end end
module Economic module Proxies module Actions module CreditorContactProxy class FindByName attr_reader :name def initialize(caller, name) @caller = caller @name = name end def call # Get a list of CreditorContactHandles from e-conomic handles = [response[:creditor_contact_handle]].flatten.reject(&:blank?) contacts = build_partial_contact_entities(handles) scope_to_owner(contacts) end private def build(*options) @caller.build(options) end def build_partial_contact_entities(handles) handles.collect do |handle| creditor_contact = build creditor_contact.partial = true creditor_contact.persisted = true creditor_contact.handle = handle creditor_contact.id = handle[:id] creditor_contact.number = handle[:number] creditor_contact end end def owner @caller.owner end def request(action, data) @caller.request(action, data) end def response request('FindByName', {'name' => name}) end def scope_to_owner(contacts) if owner.is_a?(Creditor) # Scope to the owner contacts.select do |creditor_contact| creditor_contact.get_data creditor_contact.creditor.handle == owner.handle end else contacts end end end end end end end
Clean up FindByName a bit
Clean up FindByName a bit
Ruby
mit
anderslime/rconomic,kongens-net/rconomic,substancelab/rconomic,lokalebasen/rconomic
ruby
## Code Before: module Economic module Proxies module Actions module CreditorContactProxy class FindByName attr_reader :name def initialize(caller, name) @caller = caller @name = name end def call # Get a list of CreditorContactHandles from e-conomic response = request('FindByName', { 'name' => name }) # Make sure we always have an array of handles even if the result only contains one handles = [response[:creditor_contact_handle]].flatten.reject(&:blank?) # Create partial CreditorContact entities contacts = handles.collect do |handle| creditor_contact = build creditor_contact.partial = true creditor_contact.persisted = true creditor_contact.handle = handle creditor_contact.id = handle[:id] creditor_contact.number = handle[:number] creditor_contact end if owner.is_a?(Creditor) # Scope to the owner contacts.select do |creditor_contact| creditor_contact.get_data creditor_contact.creditor.handle == owner.handle end else contacts end end private def build(*options) @caller.build(options) end def owner @caller.owner end def request(action, data) @caller.request(action, data) end end end end end end ## Instruction: Clean up FindByName a bit ## Code After: module Economic module Proxies module Actions module CreditorContactProxy class FindByName attr_reader :name def initialize(caller, name) @caller = caller @name = name end def call # Get a list of CreditorContactHandles from e-conomic handles = [response[:creditor_contact_handle]].flatten.reject(&:blank?) contacts = build_partial_contact_entities(handles) scope_to_owner(contacts) end private def build(*options) @caller.build(options) end def build_partial_contact_entities(handles) handles.collect do |handle| creditor_contact = build creditor_contact.partial = true creditor_contact.persisted = true creditor_contact.handle = handle creditor_contact.id = handle[:id] creditor_contact.number = handle[:number] creditor_contact end end def owner @caller.owner end def request(action, data) @caller.request(action, data) end def response request('FindByName', {'name' => name}) end def scope_to_owner(contacts) if owner.is_a?(Creditor) # Scope to the owner contacts.select do |creditor_contact| creditor_contact.get_data creditor_contact.creditor.handle == owner.handle end else contacts end end end end end end end
870723483416ffe81a47facbbc68440735b2dfea
package.json
package.json
{ "name": "heroku-dyno", "version": "0.1.1", "description": "Unbaffled heroku dyno management", "repository": { "type": "git", "url": "git://github.com/visionmobile/heroku-dyno.git" }, "main": "src/Dyno.js", "scripts": { "test": "node test/index.js" }, "keywords": [ "heroku", "dyno", "process", "scale", "promises", "simple", "bluebird" ], "author": "Dimitrios C. Michalakos <[email protected]>", "license": "MIT", "dependencies": { "bluebird": "^2.9.24", "heroku-client": "1.9.1", "lodash": "^3.7.0", "type-of": "^2.0.1" }, "devDependencies": { "chai": "^2.2.0", "dotenv": "^1.1.0", "mocha": "^2.2.4" } }
{ "name": "heroku-dyno", "version": "0.1.1", "description": "Unbaffled heroku dyno management", "repository": { "type": "git", "url": "git://github.com/visionmobile/heroku-dyno.git" }, "main": "src/Dyno.js", "scripts": { "test": "node test/index.js" }, "keywords": [ "heroku", "dyno", "process", "scale", "promises", "simple", "bluebird" ], "author": { "name": "VisionMobile Ltd", "email": "[email protected]", "url": "https://github.com/visionmobile" }, "maintainers": [ { "name": "Dimitrios C. Michalakos", "email": "[email protected]", "url": "https://github.com/jmike" } ], "license": "MIT", "dependencies": { "bluebird": "^2.9.24", "heroku-client": "1.9.1", "lodash": "^3.7.0", "type-of": "^2.0.1" }, "devDependencies": { "chai": "^2.2.0", "dotenv": "^1.1.0", "mocha": "^2.2.4" } }
Change author to visionmobile; add jmike as maintainer
Change author to visionmobile; add jmike as maintainer
JSON
mit
visionmobile/heroku-dyno
json
## Code Before: { "name": "heroku-dyno", "version": "0.1.1", "description": "Unbaffled heroku dyno management", "repository": { "type": "git", "url": "git://github.com/visionmobile/heroku-dyno.git" }, "main": "src/Dyno.js", "scripts": { "test": "node test/index.js" }, "keywords": [ "heroku", "dyno", "process", "scale", "promises", "simple", "bluebird" ], "author": "Dimitrios C. Michalakos <[email protected]>", "license": "MIT", "dependencies": { "bluebird": "^2.9.24", "heroku-client": "1.9.1", "lodash": "^3.7.0", "type-of": "^2.0.1" }, "devDependencies": { "chai": "^2.2.0", "dotenv": "^1.1.0", "mocha": "^2.2.4" } } ## Instruction: Change author to visionmobile; add jmike as maintainer ## Code After: { "name": "heroku-dyno", "version": "0.1.1", "description": "Unbaffled heroku dyno management", "repository": { "type": "git", "url": "git://github.com/visionmobile/heroku-dyno.git" }, "main": "src/Dyno.js", "scripts": { "test": "node test/index.js" }, "keywords": [ "heroku", "dyno", "process", "scale", "promises", "simple", "bluebird" ], "author": { "name": "VisionMobile Ltd", "email": "[email protected]", "url": "https://github.com/visionmobile" }, "maintainers": [ { "name": "Dimitrios C. Michalakos", "email": "[email protected]", "url": "https://github.com/jmike" } ], "license": "MIT", "dependencies": { "bluebird": "^2.9.24", "heroku-client": "1.9.1", "lodash": "^3.7.0", "type-of": "^2.0.1" }, "devDependencies": { "chai": "^2.2.0", "dotenv": "^1.1.0", "mocha": "^2.2.4" } }
7ece85d1ab3d789ee1f91ce77d1934927339ae16
react/frontpage/components/PromoBanner.scss
react/frontpage/components/PromoBanner.scss
// Banner @import '../common'; .promoBanner { margin-bottom: 20pt; cursor: pointer; } .promoBanner .overlay { width: 100%; height: 100%; top: 0; left: 0; position: absolute; } .promoBanner:hover .overlay { background-color: $link-image-active-color; }
// Banner @import '../common'; .promoBanner { margin-bottom: 20pt; } .promoBanner .overlay { width: 100%; height: 100%; top: 0; left: 0; position: absolute; cursor: pointer; z-index: 1; // So overlay is on top of all other promo banner divs (need this to trigger :hover) &:hover { background-color: $link-image-active-color; } }
Fix some hover weirdness in the promo banner
Fix some hover weirdness in the promo banner
SCSS
agpl-3.0
uclaradio/uclaradio,uclaradio/uclaradio,uclaradio/uclaradio
scss
## Code Before: // Banner @import '../common'; .promoBanner { margin-bottom: 20pt; cursor: pointer; } .promoBanner .overlay { width: 100%; height: 100%; top: 0; left: 0; position: absolute; } .promoBanner:hover .overlay { background-color: $link-image-active-color; } ## Instruction: Fix some hover weirdness in the promo banner ## Code After: // Banner @import '../common'; .promoBanner { margin-bottom: 20pt; } .promoBanner .overlay { width: 100%; height: 100%; top: 0; left: 0; position: absolute; cursor: pointer; z-index: 1; // So overlay is on top of all other promo banner divs (need this to trigger :hover) &:hover { background-color: $link-image-active-color; } }
878c3dc00f31e2d35d6f26b99b3e45be772ba1fd
app/controllers/property_elements_controller.rb
app/controllers/property_elements_controller.rb
class PropertyElementsController < ApplicationController before_filter :check_for_cancel, :only => [:create] def index @user = current_user @property = Property.find(params[:property_id]) @property_elements = @property.property_elements if @property_elements.empty? flash[:error] = "La propiedad no tiene elementos" redirect_to user_property_path(@user.id, @property.id) end end def new @property_element = PropertyElement.new end # def show # @property = Property.find(params[:id]) # end def create @property_element = PropertyElement.new(property_element_params) if @property_element.save flash[:success] = 'Elemento registrado exitosamente.' redirect_to user_property_property_elements_path(current_user.id, params[:property_id]) else flash[:error] = @property_element.errors.full_messages.join(',') render 'new' end end def property_element_params params.require(:property_element).permit(:name, :description, :image, :element_type_id, :property_id) end def check_for_cancel if(params.size == 5) flash[:notice] = "Registro cancelado." redirect_to user_property_property_elements_path(current_user.id, params[:property_id]) end end end
class PropertyElementsController < ApplicationController before_filter :check_for_cancel, :only => [:create] def index @user = current_user @property = Property.find(params[:property_id]) @property_elements = @property.property_elements if @property_elements.empty? flash[:error] = "La propiedad no tiene elementos" redirect_to new_user_property_property_element_path(@user.id, @property.id) end end def new @property_element = PropertyElement.new end # def show # @property = Property.find(params[:id]) # end def create @property_element = PropertyElement.new(property_element_params) if @property_element.save flash[:success] = 'Elemento registrado exitosamente.' redirect_to user_property_property_elements_path(current_user.id, params[:property_id]) else flash[:error] = @property_element.errors.full_messages.join(',') render 'new' end end def property_element_params params.require(:property_element).permit(:name, :description, :image, :element_type_id, :property_id) end def check_for_cancel if(params.size == 5) flash[:notice] = "Registro cancelado." redirect_to user_property_property_elements_path(current_user.id, params[:property_id]) end end end
Fix bug redirecting to new when property has no elements.
Fix bug redirecting to new when property has no elements.
Ruby
apache-2.0
gtrujillop/RealEstateWeb,gtrujillop/RealEstateWeb,gtrujillop/RealEstateWeb
ruby
## Code Before: class PropertyElementsController < ApplicationController before_filter :check_for_cancel, :only => [:create] def index @user = current_user @property = Property.find(params[:property_id]) @property_elements = @property.property_elements if @property_elements.empty? flash[:error] = "La propiedad no tiene elementos" redirect_to user_property_path(@user.id, @property.id) end end def new @property_element = PropertyElement.new end # def show # @property = Property.find(params[:id]) # end def create @property_element = PropertyElement.new(property_element_params) if @property_element.save flash[:success] = 'Elemento registrado exitosamente.' redirect_to user_property_property_elements_path(current_user.id, params[:property_id]) else flash[:error] = @property_element.errors.full_messages.join(',') render 'new' end end def property_element_params params.require(:property_element).permit(:name, :description, :image, :element_type_id, :property_id) end def check_for_cancel if(params.size == 5) flash[:notice] = "Registro cancelado." redirect_to user_property_property_elements_path(current_user.id, params[:property_id]) end end end ## Instruction: Fix bug redirecting to new when property has no elements. ## Code After: class PropertyElementsController < ApplicationController before_filter :check_for_cancel, :only => [:create] def index @user = current_user @property = Property.find(params[:property_id]) @property_elements = @property.property_elements if @property_elements.empty? flash[:error] = "La propiedad no tiene elementos" redirect_to new_user_property_property_element_path(@user.id, @property.id) end end def new @property_element = PropertyElement.new end # def show # @property = Property.find(params[:id]) # end def create @property_element = PropertyElement.new(property_element_params) if @property_element.save flash[:success] = 'Elemento registrado exitosamente.' redirect_to user_property_property_elements_path(current_user.id, params[:property_id]) else flash[:error] = @property_element.errors.full_messages.join(',') render 'new' end end def property_element_params params.require(:property_element).permit(:name, :description, :image, :element_type_id, :property_id) end def check_for_cancel if(params.size == 5) flash[:notice] = "Registro cancelado." redirect_to user_property_property_elements_path(current_user.id, params[:property_id]) end end end
0a48250ec2d7350f11bc864a010954078f6e1588
vdb-bench-assembly/plugins/vdb-bench-widgets/fileImportControl.html
vdb-bench-assembly/plugins/vdb-bench-widgets/fileImportControl.html
<div> <!-- Show progress bar on loading --> <div ng-show="vm.inProgress"> <uib-progressbar class="progress-striped active" value="dynamic" type="info"></uib-progressbar> </div> <!-- Import control --> <div ng-show="! vm.inProgress"> <div class="import-control-title" translate="fileImportControl.selectFileMsg" /> <div class="import-control-button"> <label class="btn btn-primary" translate="shared.BrowseMsg"> <input ng-model="fileToImport" type="file" style="display: none;" file-input-handler="vm.importFile"> </label> <button class="btn" ng-click="vm.cancel()" ng-show="vm.displayCancel()" translate="shared.Cancel"></button> </div> <div class="import-control-response"> <div ng-class="vm.responseStyleClass">{{vm.response}}</div> </div> <div class="import-control-error"> {{vm.error}} </div> </div> </div>
<div> <!-- Show progress bar on loading --> <div ng-show="vm.inProgress"> <uib-progressbar class="progress-striped active" value="dynamic" type="info"></uib-progressbar> </div> <!-- Import control --> <div ng-show="! vm.inProgress"> <div class="import-control-title" translate="fileImportControl.selectFileMsg" /> <div class="import-control-button"> <label class="btn btn-primary"> <span translate="shared.BrowseMsg"></span> <input ng-model="fileToImport" type="file" style="display: none;" file-input-handler="vm.importFile"> </label> <button class="btn" ng-click="vm.cancel()" ng-show="vm.displayCancel()" translate="shared.Cancel"></button> </div> <div class="import-control-response"> <div ng-class="vm.responseStyleClass">{{vm.response}}</div> </div> <div class="import-control-error"> {{vm.error}} </div> </div> </div>
Fix for file import control button
Fix for file import control button * Seems that translate directive does not transclude the contents of the element but replaces it with its translated content. This is a problem for elements that already contain child-elements.
HTML
apache-2.0
phantomjinx/vdb-bench,Teiid-Designer/vdb-bench,Teiid-Designer/vdb-bench,mdrillin/vdb-bench,mdrillin/vdb-bench,phantomjinx/vdb-bench
html
## Code Before: <div> <!-- Show progress bar on loading --> <div ng-show="vm.inProgress"> <uib-progressbar class="progress-striped active" value="dynamic" type="info"></uib-progressbar> </div> <!-- Import control --> <div ng-show="! vm.inProgress"> <div class="import-control-title" translate="fileImportControl.selectFileMsg" /> <div class="import-control-button"> <label class="btn btn-primary" translate="shared.BrowseMsg"> <input ng-model="fileToImport" type="file" style="display: none;" file-input-handler="vm.importFile"> </label> <button class="btn" ng-click="vm.cancel()" ng-show="vm.displayCancel()" translate="shared.Cancel"></button> </div> <div class="import-control-response"> <div ng-class="vm.responseStyleClass">{{vm.response}}</div> </div> <div class="import-control-error"> {{vm.error}} </div> </div> </div> ## Instruction: Fix for file import control button * Seems that translate directive does not transclude the contents of the element but replaces it with its translated content. This is a problem for elements that already contain child-elements. ## Code After: <div> <!-- Show progress bar on loading --> <div ng-show="vm.inProgress"> <uib-progressbar class="progress-striped active" value="dynamic" type="info"></uib-progressbar> </div> <!-- Import control --> <div ng-show="! vm.inProgress"> <div class="import-control-title" translate="fileImportControl.selectFileMsg" /> <div class="import-control-button"> <label class="btn btn-primary"> <span translate="shared.BrowseMsg"></span> <input ng-model="fileToImport" type="file" style="display: none;" file-input-handler="vm.importFile"> </label> <button class="btn" ng-click="vm.cancel()" ng-show="vm.displayCancel()" translate="shared.Cancel"></button> </div> <div class="import-control-response"> <div ng-class="vm.responseStyleClass">{{vm.response}}</div> </div> <div class="import-control-error"> {{vm.error}} </div> </div> </div>
5ae47158073bc0f2f43b95c5e5a343e736537f98
delegatednews/delegatednews-config/src/main/config/dbRepository/oracle/delegatedNews/003/create_table.sql
delegatednews/delegatednews-config/src/main/config/dbRepository/oracle/delegatedNews/003/create_table.sql
create table sc_delegatednews_news ( pubId int not null, instanceId varchar(50) not null, status varchar(100) not null, contributorId varchar(50) not null, validatorId varchar(50) null, validationDate timestamp(0) null, beginDate timestamp(0) null, endDate timestamp(0) null, newsOrder int not null DEFAULT (0) );
create table sc_delegatednews_news ( pubId int not null, instanceId varchar(50) not null, status varchar(100) not null, contributorId varchar(50) not null, validatorId varchar(50) null, validationDate timestamp(0) null, beginDate timestamp(0) null, endDate timestamp(0) null, newsOrder int DEFAULT (0) not null );
Fix the creation of the table sc_delegatednews_news for Oracle: the declaration of the column newsOrder was incorrect
Fix the creation of the table sc_delegatednews_news for Oracle: the declaration of the column newsOrder was incorrect
SQL
agpl-3.0
ebonnet/Silverpeas-Components,ebonnet/Silverpeas-Components,SilverTeamWork/Silverpeas-Components,SilverYoCha/Silverpeas-Components,CecileBONIN/Silverpeas-Components,CecileBONIN/Silverpeas-Components,SilverYoCha/Silverpeas-Components,CecileBONIN/Silverpeas-Components,Silverpeas/Silverpeas-Components,auroreallibe/Silverpeas-Components,ebonnet/Silverpeas-Components,SilverTeamWork/Silverpeas-Components,auroreallibe/Silverpeas-Components,ebonnet/Silverpeas-Components,CecileBONIN/Silverpeas-Components,mmoqui/Silverpeas-Components,auroreallibe/Silverpeas-Components,Silverpeas/Silverpeas-Components,CecileBONIN/Silverpeas-Components,ebonnet/Silverpeas-Components,CecileBONIN/Silverpeas-Components,mmoqui/Silverpeas-Components,ebonnet/Silverpeas-Components,ebonnet/Silverpeas-Components,Silverpeas/Silverpeas-Components,SilverTeamWork/Silverpeas-Components,mmoqui/Silverpeas-Components,SilverYoCha/Silverpeas-Components
sql
## Code Before: create table sc_delegatednews_news ( pubId int not null, instanceId varchar(50) not null, status varchar(100) not null, contributorId varchar(50) not null, validatorId varchar(50) null, validationDate timestamp(0) null, beginDate timestamp(0) null, endDate timestamp(0) null, newsOrder int not null DEFAULT (0) ); ## Instruction: Fix the creation of the table sc_delegatednews_news for Oracle: the declaration of the column newsOrder was incorrect ## Code After: create table sc_delegatednews_news ( pubId int not null, instanceId varchar(50) not null, status varchar(100) not null, contributorId varchar(50) not null, validatorId varchar(50) null, validationDate timestamp(0) null, beginDate timestamp(0) null, endDate timestamp(0) null, newsOrder int DEFAULT (0) not null );
8445a067ae4b8158c108145cc9715b4ae6e5a88c
.travis.yml
.travis.yml
sudo: required dist: trusty language: cpp services: - docker branches: only: - master - develop before_install: - docker build -t travis-build -f infrastructure/tools/Dockerfile . script: # Common style check - docker run --rm travis-build infrastructure/tools/checkstyle.sh $TRAVIS_COMMIT_RANGE # Run tests - docker run --rm travis-build /bin/bash -c "fab test_all"
sudo: required dist: trusty language: cpp services: - docker branches: only: - master - develop before_install: - docker build -t travis-build -f infrastructure/tools/Dockerfile . script: # Common style check - docker run --rm travis-build infrastructure/tools/checkstyle.sh $TRAVIS_COMMIT_RANGE # Test whole project with pylint (can find issues in multiple files) - docker run --rm travis-build /bin/bash -c "pylint --rcfile=.pylintrc ./pylua" # Run tests - docker run --rm travis-build /bin/bash -c "fab test_all"
Check whole project with pylint on CI
Check whole project with pylint on CI
YAML
mit
malirod/pylua,malirod/pylua
yaml
## Code Before: sudo: required dist: trusty language: cpp services: - docker branches: only: - master - develop before_install: - docker build -t travis-build -f infrastructure/tools/Dockerfile . script: # Common style check - docker run --rm travis-build infrastructure/tools/checkstyle.sh $TRAVIS_COMMIT_RANGE # Run tests - docker run --rm travis-build /bin/bash -c "fab test_all" ## Instruction: Check whole project with pylint on CI ## Code After: sudo: required dist: trusty language: cpp services: - docker branches: only: - master - develop before_install: - docker build -t travis-build -f infrastructure/tools/Dockerfile . script: # Common style check - docker run --rm travis-build infrastructure/tools/checkstyle.sh $TRAVIS_COMMIT_RANGE # Test whole project with pylint (can find issues in multiple files) - docker run --rm travis-build /bin/bash -c "pylint --rcfile=.pylintrc ./pylua" # Run tests - docker run --rm travis-build /bin/bash -c "fab test_all"
d31abe5a6f6fc1045837ebe863e6a0ff1307c2df
requirements.txt
requirements.txt
doctr ipython jsonschema matplotlib numpydoc path.py sphinx sphinx-gallery sphinx_rtd_theme sphinxcontrib-napoleon
bluesky # for gallery example doctr ipython jsonschema matplotlib numpy # for gallery example numpydoc path.py sphinx sphinx-gallery sphinx_rtd_theme sphinxcontrib-napoleon
Add dependencies for gallery examples.
FIX: Add dependencies for gallery examples.
Text
bsd-2-clause
NSLS-II/docs,tacaswell/docs,NSLS-II/docs
text
## Code Before: doctr ipython jsonschema matplotlib numpydoc path.py sphinx sphinx-gallery sphinx_rtd_theme sphinxcontrib-napoleon ## Instruction: FIX: Add dependencies for gallery examples. ## Code After: bluesky # for gallery example doctr ipython jsonschema matplotlib numpy # for gallery example numpydoc path.py sphinx sphinx-gallery sphinx_rtd_theme sphinxcontrib-napoleon
c7c1c94635bc499b8c33fc4c9b65033ef2c24c89
source/popup/containers/EntriesPage.js
source/popup/containers/EntriesPage.js
import { connect } from "react-redux"; import EntriesPage from "../components/EntriesPage.js"; import { clearSearchResults } from "../library/messaging.js"; import { searchEntriesForTerm } from "../../shared/library/messaging.js"; export default connect( (state, ownProps) => ({}), { onPrepare: () => () => { clearSearchResults(); }, onSearchTermChange: searchTerm => () => { if (searchTerm.trim().length > 0) { searchEntriesForTerm(searchTerm); } else { clearSearchResults(); } } } )(EntriesPage);
import { connect } from "react-redux"; import EntriesPage from "../components/EntriesPage.js"; import { clearSearchResults } from "../library/messaging.js"; import { searchEntriesForTerm } from "../../shared/library/messaging.js"; export default connect( (state, ownProps) => ({}), { onPrepare: () => () => {}, onSearchTermChange: searchTerm => () => { if (searchTerm.trim().length > 0) { searchEntriesForTerm(searchTerm); } else { clearSearchResults(); } } } )(EntriesPage);
Stop clearing search results on popup init
Stop clearing search results on popup init
JavaScript
mit
buttercup-pw/buttercup-browser-extension,perry-mitchell/buttercup-chrome,buttercup-pw/buttercup-browser-extension,perry-mitchell/buttercup-chrome
javascript
## Code Before: import { connect } from "react-redux"; import EntriesPage from "../components/EntriesPage.js"; import { clearSearchResults } from "../library/messaging.js"; import { searchEntriesForTerm } from "../../shared/library/messaging.js"; export default connect( (state, ownProps) => ({}), { onPrepare: () => () => { clearSearchResults(); }, onSearchTermChange: searchTerm => () => { if (searchTerm.trim().length > 0) { searchEntriesForTerm(searchTerm); } else { clearSearchResults(); } } } )(EntriesPage); ## Instruction: Stop clearing search results on popup init ## Code After: import { connect } from "react-redux"; import EntriesPage from "../components/EntriesPage.js"; import { clearSearchResults } from "../library/messaging.js"; import { searchEntriesForTerm } from "../../shared/library/messaging.js"; export default connect( (state, ownProps) => ({}), { onPrepare: () => () => {}, onSearchTermChange: searchTerm => () => { if (searchTerm.trim().length > 0) { searchEntriesForTerm(searchTerm); } else { clearSearchResults(); } } } )(EntriesPage);
2ed5f3f6567cd7cc541fa58a49b6cf03f07d23e5
playbooks/roles/jenkins_worker/defaults/main.yml
playbooks/roles/jenkins_worker/defaults/main.yml
--- jenkins_user: "jenkins" # the main group for the jenkins user jenkins_group: "jenkins" # the groups that the jenkins user must belong to jenkins_groups: "jenkins,docker" jenkins_home: /home/jenkins jenkins_edx_platform_version: master # System packages jenkins_debian_pkgs: - pkg-config - libffi-dev - python-dev - libsqlite3-dev - libfreetype6-dev # packer direct download URL packer_url: "https://releases.hashicorp.com/packer/1.4.4/packer_1.4.4_linux_amd64.zip" jenkins_node_version: "12" ansible_distribution_release: "xenial" jenkins_worker_python_versions: - 2.7 - 3.5 - 3.6
--- jenkins_user: "jenkins" # the main group for the jenkins user jenkins_group: "jenkins" # the groups that the jenkins user must belong to jenkins_groups: "jenkins,docker" jenkins_home: /home/jenkins jenkins_edx_platform_version: master # System packages jenkins_debian_pkgs: - pkg-config - libffi-dev - python-dev - libsqlite3-dev - libfreetype6-dev - libpq-dev # packer direct download URL packer_url: "https://releases.hashicorp.com/packer/1.4.4/packer_1.4.4_linux_amd64.zip" jenkins_node_version: "12" ansible_distribution_release: "xenial" jenkins_worker_python_versions: - 2.7 - 3.5 - 3.6
Add postgres library to handle openedxstats upgrades.
Add postgres library to handle openedxstats upgrades.
YAML
agpl-3.0
EDUlib/configuration,arbrandes/edx-configuration,mitodl/configuration,edx/configuration,EDUlib/configuration,mitodl/configuration,arbrandes/edx-configuration,EDUlib/configuration,arbrandes/edx-configuration,mitodl/configuration,appsembler/configuration,edx/configuration,EDUlib/configuration,edx/configuration,stvstnfrd/configuration,EDUlib/configuration,stvstnfrd/configuration,edx/configuration,appsembler/configuration,appsembler/configuration,stvstnfrd/configuration,arbrandes/edx-configuration,mitodl/configuration,appsembler/configuration,arbrandes/edx-configuration,stvstnfrd/configuration,stvstnfrd/configuration
yaml
## Code Before: --- jenkins_user: "jenkins" # the main group for the jenkins user jenkins_group: "jenkins" # the groups that the jenkins user must belong to jenkins_groups: "jenkins,docker" jenkins_home: /home/jenkins jenkins_edx_platform_version: master # System packages jenkins_debian_pkgs: - pkg-config - libffi-dev - python-dev - libsqlite3-dev - libfreetype6-dev # packer direct download URL packer_url: "https://releases.hashicorp.com/packer/1.4.4/packer_1.4.4_linux_amd64.zip" jenkins_node_version: "12" ansible_distribution_release: "xenial" jenkins_worker_python_versions: - 2.7 - 3.5 - 3.6 ## Instruction: Add postgres library to handle openedxstats upgrades. ## Code After: --- jenkins_user: "jenkins" # the main group for the jenkins user jenkins_group: "jenkins" # the groups that the jenkins user must belong to jenkins_groups: "jenkins,docker" jenkins_home: /home/jenkins jenkins_edx_platform_version: master # System packages jenkins_debian_pkgs: - pkg-config - libffi-dev - python-dev - libsqlite3-dev - libfreetype6-dev - libpq-dev # packer direct download URL packer_url: "https://releases.hashicorp.com/packer/1.4.4/packer_1.4.4_linux_amd64.zip" jenkins_node_version: "12" ansible_distribution_release: "xenial" jenkins_worker_python_versions: - 2.7 - 3.5 - 3.6
3d22a51134963c3f188edeacb937d2a6f17a29e5
lib/env.js
lib/env.js
/** * @file env * @author treelite([email protected]) */ /** * 修复全局的异步函数 * 使之能在异步回调中恢复请求请求上下文 * * @inner * @param {Object} app server */ function fixGlobalMethods(app) { var methods = ['setTimeout', 'setInterval', 'setImmediate']; methods.forEach(function (name) { var method = global[name]; global[name] = function (callback) { var args = Array.prototype.slice.call(arguments, 1); var id = app.stashContext(); args.unshift(function () { app.revertContext(id); callback(); }); method.apply(null, args); }; }); } /** * 运行环境设置 * * @public * @param {Object} app server */ module.exports = function (app) { // 修复全局异步函数 fixGlobalMethods(app); };
/** * @file env * @author treelite([email protected]) */ /** * 修复异步函数 * 使之能在异步回调中恢复请求请求上下文 * * @inner * @param {Object} app server */ function fixAsyncMethods(app) { /** * 包裹异步函数 * * @param {Object} obj 模块 * @param {string} name 方法名 */ function wrap(obj, name) { var method = obj[name]; obj[name] = function (callback) { var args = Array.prototype.slice.call(arguments, 1); var id = app.stashContext(); args.unshift(function () { app.revertContext(id); callback(); }); method.apply(null, args); }; } var methods = ['setTimeout', 'setInterval', 'setImmediate']; methods.forEach(function (name) { wrap(global, name); }); wrap(process, 'nextTick'); } /** * 运行环境设置 * * @public * @param {Object} app server */ module.exports = function (app) { // 修复异步函数 fixAsyncMethods(app); };
Fix process.nextTick, try to save context
Fix process.nextTick, try to save context
JavaScript
mit
ecomfe/rebas,ecomfe/rebas,noikiy/rebas
javascript
## Code Before: /** * @file env * @author treelite([email protected]) */ /** * 修复全局的异步函数 * 使之能在异步回调中恢复请求请求上下文 * * @inner * @param {Object} app server */ function fixGlobalMethods(app) { var methods = ['setTimeout', 'setInterval', 'setImmediate']; methods.forEach(function (name) { var method = global[name]; global[name] = function (callback) { var args = Array.prototype.slice.call(arguments, 1); var id = app.stashContext(); args.unshift(function () { app.revertContext(id); callback(); }); method.apply(null, args); }; }); } /** * 运行环境设置 * * @public * @param {Object} app server */ module.exports = function (app) { // 修复全局异步函数 fixGlobalMethods(app); }; ## Instruction: Fix process.nextTick, try to save context ## Code After: /** * @file env * @author treelite([email protected]) */ /** * 修复异步函数 * 使之能在异步回调中恢复请求请求上下文 * * @inner * @param {Object} app server */ function fixAsyncMethods(app) { /** * 包裹异步函数 * * @param {Object} obj 模块 * @param {string} name 方法名 */ function wrap(obj, name) { var method = obj[name]; obj[name] = function (callback) { var args = Array.prototype.slice.call(arguments, 1); var id = app.stashContext(); args.unshift(function () { app.revertContext(id); callback(); }); method.apply(null, args); }; } var methods = ['setTimeout', 'setInterval', 'setImmediate']; methods.forEach(function (name) { wrap(global, name); }); wrap(process, 'nextTick'); } /** * 运行环境设置 * * @public * @param {Object} app server */ module.exports = function (app) { // 修复异步函数 fixAsyncMethods(app); };