commit
stringlengths 40
40
| old_file
stringlengths 4
237
| new_file
stringlengths 4
237
| old_contents
stringlengths 1
4.24k
| new_contents
stringlengths 5
4.84k
| subject
stringlengths 15
778
| message
stringlengths 16
6.86k
| lang
stringlengths 1
30
| license
stringclasses 13
values | repos
stringlengths 5
116k
| config
stringlengths 1
30
| content
stringlengths 105
8.72k
|
---|---|---|---|---|---|---|---|---|---|---|---|
51575888408e4bf08e67bcefc463547d31874b23 | test/CodeGen/X86/2009-06-07-ExpandMMXBitcast.ll | test/CodeGen/X86/2009-06-07-ExpandMMXBitcast.ll | ; RUN: llc < %s -mtriple=x86_64-linux -mattr=+mmx | grep movd | count 2
define i64 @a(i32 %a, i32 %b) nounwind readnone {
entry:
%0 = insertelement <2 x i32> undef, i32 %a, i32 0 ; <<2 x i32>> [#uses=1]
%1 = insertelement <2 x i32> %0, i32 %b, i32 1 ; <<2 x i32>> [#uses=1]
%conv = bitcast <2 x i32> %1 to i64 ; <i64> [#uses=1]
ret i64 %conv
}
| ; RUN: llc < %s -mtriple=x86_64-linux -mcpu=corei7 -mattr=+mmx | grep movd | count 2
define i64 @a(i32 %a, i32 %b) nounwind readnone {
entry:
%0 = insertelement <2 x i32> undef, i32 %a, i32 0 ; <<2 x i32>> [#uses=1]
%1 = insertelement <2 x i32> %0, i32 %b, i32 1 ; <<2 x i32>> [#uses=1]
%conv = bitcast <2 x i32> %1 to i64 ; <i64> [#uses=1]
ret i64 %conv
}
| Stabilize the test by specifying an exact cpu target | Stabilize the test by specifying an exact cpu target
git-svn-id: 0ff597fd157e6f4fc38580e8d64ab130330d2411@143307 91177308-0d34-0410-b5e6-96231b3b80d8
| LLVM | bsd-2-clause | dslab-epfl/asap,apple/swift-llvm,GPUOpen-Drivers/llvm,GPUOpen-Drivers/llvm,apple/swift-llvm,llvm-mirror/llvm,chubbymaggie/asap,llvm-mirror/llvm,chubbymaggie/asap,GPUOpen-Drivers/llvm,dslab-epfl/asap,llvm-mirror/llvm,llvm-mirror/llvm,llvm-mirror/llvm,apple/swift-llvm,dslab-epfl/asap,apple/swift-llvm,llvm-mirror/llvm,apple/swift-llvm,GPUOpen-Drivers/llvm,chubbymaggie/asap,apple/swift-llvm,dslab-epfl/asap,llvm-mirror/llvm,dslab-epfl/asap,GPUOpen-Drivers/llvm,GPUOpen-Drivers/llvm,chubbymaggie/asap,GPUOpen-Drivers/llvm,chubbymaggie/asap,dslab-epfl/asap,dslab-epfl/asap,llvm-mirror/llvm,GPUOpen-Drivers/llvm,chubbymaggie/asap,llvm-mirror/llvm,apple/swift-llvm,apple/swift-llvm | llvm | ## Code Before:
; RUN: llc < %s -mtriple=x86_64-linux -mattr=+mmx | grep movd | count 2
define i64 @a(i32 %a, i32 %b) nounwind readnone {
entry:
%0 = insertelement <2 x i32> undef, i32 %a, i32 0 ; <<2 x i32>> [#uses=1]
%1 = insertelement <2 x i32> %0, i32 %b, i32 1 ; <<2 x i32>> [#uses=1]
%conv = bitcast <2 x i32> %1 to i64 ; <i64> [#uses=1]
ret i64 %conv
}
## Instruction:
Stabilize the test by specifying an exact cpu target
git-svn-id: 0ff597fd157e6f4fc38580e8d64ab130330d2411@143307 91177308-0d34-0410-b5e6-96231b3b80d8
## Code After:
; RUN: llc < %s -mtriple=x86_64-linux -mcpu=corei7 -mattr=+mmx | grep movd | count 2
define i64 @a(i32 %a, i32 %b) nounwind readnone {
entry:
%0 = insertelement <2 x i32> undef, i32 %a, i32 0 ; <<2 x i32>> [#uses=1]
%1 = insertelement <2 x i32> %0, i32 %b, i32 1 ; <<2 x i32>> [#uses=1]
%conv = bitcast <2 x i32> %1 to i64 ; <i64> [#uses=1]
ret i64 %conv
}
|
3460e5b8ef4042ba05560b6f9a0343e8c4f4f918 | bootstrap/docker.json | bootstrap/docker.json | {
"data-root": "/mnt/stateful_partition/docker",
"experimental": true,
"deprecated-key-path": "/var/lib/docker/key.json",
"hosts": [
"tcp://127.0.0.1:2376",
"unix:///var/run/docker.sock"
],
"storage-driver": "devicemapper",
"storage-opts": [
"dm.override_udev_sync_check=true"
]
}
| {
"bip": "10.250.0.0/16",
"data-root": "/mnt/stateful_partition/docker",
"experimental": true,
"deprecated-key-path": "/var/lib/docker/key.json",
"default-address-pools":[
{ "base":"10.251.0.0/16", "size":24 },
{ "base":"10.252.0.0/16", "size":24 }
],
"hosts": [
"tcp://127.0.0.1:2376",
"unix:///var/run/docker.sock"
],
"storage-driver": "devicemapper",
"storage-opts": [
"dm.override_udev_sync_check=true"
]
}
| Add custom addresses for Docker containers | Add custom addresses for Docker containers
| JSON | mit | cartolari/dotfiles,cartolari/dotfiles,cartolari/dotfiles,cartolari/dotfiles | json | ## Code Before:
{
"data-root": "/mnt/stateful_partition/docker",
"experimental": true,
"deprecated-key-path": "/var/lib/docker/key.json",
"hosts": [
"tcp://127.0.0.1:2376",
"unix:///var/run/docker.sock"
],
"storage-driver": "devicemapper",
"storage-opts": [
"dm.override_udev_sync_check=true"
]
}
## Instruction:
Add custom addresses for Docker containers
## Code After:
{
"bip": "10.250.0.0/16",
"data-root": "/mnt/stateful_partition/docker",
"experimental": true,
"deprecated-key-path": "/var/lib/docker/key.json",
"default-address-pools":[
{ "base":"10.251.0.0/16", "size":24 },
{ "base":"10.252.0.0/16", "size":24 }
],
"hosts": [
"tcp://127.0.0.1:2376",
"unix:///var/run/docker.sock"
],
"storage-driver": "devicemapper",
"storage-opts": [
"dm.override_udev_sync_check=true"
]
}
|
3bf2779805eeb2fff9fdda563e9188814c0859cf | cookbooks/dns/templates/default/creds.json.erb | cookbooks/dns/templates/default/creds.json.erb | {
"gandi_v5": {
"apikey": "<%= @passwords["gandi"] %>",
"sharing_id": "7028b616-ba65-11e7-8343-00163ec31f40"
},
"cloudflare": {
"accountid": "049c95aba02c95fc1e78a9d255282e0f",
"accountname": "OpenStreetMap",
"apitoken": "<%= @passwords["cloudflare"] %>"
}
}
| {
"gandi_v5": {
"TYPE": "GANDI_V5",
"apikey": "<%= @passwords["gandi"] %>",
"sharing_id": "7028b616-ba65-11e7-8343-00163ec31f40"
},
"cloudflare": {
"TYPE": "CLOUDFLAREAPI",
"accountid": "049c95aba02c95fc1e78a9d255282e0f",
"accountname": "OpenStreetMap",
"apitoken": "<%= @passwords["cloudflare"] %>"
}
}
| Update dnscontrol auth to v3.16+ syntax | Update dnscontrol auth to v3.16+ syntax
https://stackexchange.github.io/dnscontrol/creds-json#missing
| HTML+ERB | apache-2.0 | tomhughes/openstreetmap-chef,tomhughes/openstreetmap-chef,tomhughes/openstreetmap-chef,tomhughes/openstreetmap-chef,openstreetmap/chef,tomhughes/openstreetmap-chef,openstreetmap/chef,tomhughes/openstreetmap-chef,openstreetmap/chef,openstreetmap/chef,openstreetmap/chef,tomhughes/openstreetmap-chef,openstreetmap/chef,tomhughes/openstreetmap-chef,openstreetmap/chef,openstreetmap/chef | html+erb | ## Code Before:
{
"gandi_v5": {
"apikey": "<%= @passwords["gandi"] %>",
"sharing_id": "7028b616-ba65-11e7-8343-00163ec31f40"
},
"cloudflare": {
"accountid": "049c95aba02c95fc1e78a9d255282e0f",
"accountname": "OpenStreetMap",
"apitoken": "<%= @passwords["cloudflare"] %>"
}
}
## Instruction:
Update dnscontrol auth to v3.16+ syntax
https://stackexchange.github.io/dnscontrol/creds-json#missing
## Code After:
{
"gandi_v5": {
"TYPE": "GANDI_V5",
"apikey": "<%= @passwords["gandi"] %>",
"sharing_id": "7028b616-ba65-11e7-8343-00163ec31f40"
},
"cloudflare": {
"TYPE": "CLOUDFLAREAPI",
"accountid": "049c95aba02c95fc1e78a9d255282e0f",
"accountname": "OpenStreetMap",
"apitoken": "<%= @passwords["cloudflare"] %>"
}
}
|
8e275a22aa419297eee0acc5ef65fc0cce4aa3cb | clients/web/test/spec/swaggerSpec.js | clients/web/test/spec/swaggerSpec.js | function jasmineTests() {
var jasmineEnv = jasmine.getEnv();
var consoleReporter = new jasmine.ConsoleReporter();
window.jasmine_phantom_reporter = consoleReporter;
jasmineEnv.addReporter(consoleReporter);
function waitAndExecute() {
if (!jasmineEnv.currentRunner().suites_.length) {
window.setTimeout(waitAndExecute, 10);
return;
}
jasmineEnv.execute();
}
waitAndExecute();
describe('Test the swagger pages', function () {
it('Test swagger', function () {
waitsFor(function () {
return $('li#resource_system.resource').length > 0;
}, 'swagger docs to appear');
runs(function () {
$('li#resource_system.resource .heading h2 a').click();
});
waitsFor(function () {
return $('#system_getVersion:visible').length > 0;
}, 'end points to be visible');
runs(function () {
$('#system_getVersion h3 a').click();
});
waitsFor(function () {
return $('#system_getVersion .sandbox_header input.submit[name="commit"]:visible').length > 0;
}, 'version try out button to be visible');
runs(function () {
$('#system_getVersion .sandbox_header input.submit[name="commit"]').click();
});
waitsFor(function () {
return $('#system_getVersion .response_body.json').text().indexOf('apiVersion') >= 0;
}, 'version information was returned');
});
});
}
$(function () {
$.getScript('/static/built/testing-no-cover.min.js', jasmineTests);
});
| function jasmineTests() {
var jasmineEnv = jasmine.getEnv();
var consoleReporter = new jasmine.ConsoleReporter();
window.jasmine_phantom_reporter = consoleReporter;
jasmineEnv.addReporter(consoleReporter);
function waitAndExecute() {
if (!jasmineEnv.currentRunner().suites_.length) {
window.setTimeout(waitAndExecute, 10);
return;
}
jasmineEnv.execute();
}
waitAndExecute();
describe('Test the swagger pages', function () {
it('Test swagger', function () {
waitsFor(function () {
return $('li#resource_system.resource').length > 0;
}, 'swagger docs to appear');
runs(function () {
$('li#resource_system.resource .heading h2 a').click();
});
waitsFor(function () {
return $('#system_system_getVersion:visible').length > 0;
}, 'end points to be visible');
runs(function () {
$('#system_system_getVersion h3 a').click();
});
waitsFor(function () {
return $('#system_system_getVersion .sandbox_header input.submit:visible').length > 0;
}, 'version try out button to be visible');
runs(function () {
$('#system_system_getVersion .sandbox_header input.submit').click();
});
waitsFor(function () {
return $('#system_system_getVersion .response_body.json').text().indexOf('apiVersion') >= 0;
}, 'version information was returned');
});
});
}
$(function () {
$.getScript('/static/built/testing-no-cover.min.js', jasmineTests);
});
| Update web client swagger test for swagger-ui 2.1.4 | Update web client swagger test for swagger-ui 2.1.4
| JavaScript | apache-2.0 | adsorensen/girder,kotfic/girder,data-exp-lab/girder,jbeezley/girder,Kitware/girder,jbeezley/girder,adsorensen/girder,RafaelPalomar/girder,kotfic/girder,girder/girder,sutartmelson/girder,Xarthisius/girder,sutartmelson/girder,sutartmelson/girder,manthey/girder,data-exp-lab/girder,RafaelPalomar/girder,data-exp-lab/girder,manthey/girder,data-exp-lab/girder,RafaelPalomar/girder,Xarthisius/girder,adsorensen/girder,kotfic/girder,data-exp-lab/girder,adsorensen/girder,RafaelPalomar/girder,RafaelPalomar/girder,girder/girder,sutartmelson/girder,jbeezley/girder,adsorensen/girder,manthey/girder,Xarthisius/girder,kotfic/girder,Kitware/girder,girder/girder,Kitware/girder,Kitware/girder,sutartmelson/girder,Xarthisius/girder,girder/girder,manthey/girder,jbeezley/girder,kotfic/girder,Xarthisius/girder | javascript | ## Code Before:
function jasmineTests() {
var jasmineEnv = jasmine.getEnv();
var consoleReporter = new jasmine.ConsoleReporter();
window.jasmine_phantom_reporter = consoleReporter;
jasmineEnv.addReporter(consoleReporter);
function waitAndExecute() {
if (!jasmineEnv.currentRunner().suites_.length) {
window.setTimeout(waitAndExecute, 10);
return;
}
jasmineEnv.execute();
}
waitAndExecute();
describe('Test the swagger pages', function () {
it('Test swagger', function () {
waitsFor(function () {
return $('li#resource_system.resource').length > 0;
}, 'swagger docs to appear');
runs(function () {
$('li#resource_system.resource .heading h2 a').click();
});
waitsFor(function () {
return $('#system_getVersion:visible').length > 0;
}, 'end points to be visible');
runs(function () {
$('#system_getVersion h3 a').click();
});
waitsFor(function () {
return $('#system_getVersion .sandbox_header input.submit[name="commit"]:visible').length > 0;
}, 'version try out button to be visible');
runs(function () {
$('#system_getVersion .sandbox_header input.submit[name="commit"]').click();
});
waitsFor(function () {
return $('#system_getVersion .response_body.json').text().indexOf('apiVersion') >= 0;
}, 'version information was returned');
});
});
}
$(function () {
$.getScript('/static/built/testing-no-cover.min.js', jasmineTests);
});
## Instruction:
Update web client swagger test for swagger-ui 2.1.4
## Code After:
function jasmineTests() {
var jasmineEnv = jasmine.getEnv();
var consoleReporter = new jasmine.ConsoleReporter();
window.jasmine_phantom_reporter = consoleReporter;
jasmineEnv.addReporter(consoleReporter);
function waitAndExecute() {
if (!jasmineEnv.currentRunner().suites_.length) {
window.setTimeout(waitAndExecute, 10);
return;
}
jasmineEnv.execute();
}
waitAndExecute();
describe('Test the swagger pages', function () {
it('Test swagger', function () {
waitsFor(function () {
return $('li#resource_system.resource').length > 0;
}, 'swagger docs to appear');
runs(function () {
$('li#resource_system.resource .heading h2 a').click();
});
waitsFor(function () {
return $('#system_system_getVersion:visible').length > 0;
}, 'end points to be visible');
runs(function () {
$('#system_system_getVersion h3 a').click();
});
waitsFor(function () {
return $('#system_system_getVersion .sandbox_header input.submit:visible').length > 0;
}, 'version try out button to be visible');
runs(function () {
$('#system_system_getVersion .sandbox_header input.submit').click();
});
waitsFor(function () {
return $('#system_system_getVersion .response_body.json').text().indexOf('apiVersion') >= 0;
}, 'version information was returned');
});
});
}
$(function () {
$.getScript('/static/built/testing-no-cover.min.js', jasmineTests);
});
|
17a18f1f8765bda183de2d52834faaa310ca086a | java/build.xml | java/build.xml | <project name="LibOpenConnect" default="dist" basedir=".">
<property name="build" location="build"/>
<property name="dist" location="dist"/>
<property name="src" location="src"/>
<property name="jar_lib" location="${dist}/openconnect-wrapper.jar"/>
<property name="jar_app" location="${dist}/example.jar"/>
<target name="init">
<mkdir dir="${build}"/>
</target>
<target name="compile" depends="init">
<javac srcdir="${src}" destdir="${build}"
includeantruntime="false"/>
</target>
<target name="dist" depends="compile">
<jar jarfile="${jar_lib}" basedir="${build}"
includes="org/infradead/libopenconnect/*" />
<jar jarfile="${jar_app}" basedir="${build}">
<manifest>
<attribute name="Main-Class" value="com.example.LibTest" />
</manifest>
</jar>
</target>
<target name="clean">
<delete dir="${build}"/>
<delete dir="${dist}"/>
</target>
</project>
| <project name="LibOpenConnect" default="dist" basedir=".">
<property name="build" location="build"/>
<property name="dist" location="dist"/>
<property name="src" location="src"/>
<property name="jar_lib" location="${dist}/openconnect-wrapper.jar"/>
<property name="jar_app" location="${dist}/example.jar"/>
<target name="init">
<mkdir dir="${build}"/>
</target>
<target name="compile" depends="init">
<javac srcdir="${src}" destdir="${build}"
source="1.5" target="1.5" includeantruntime="false"/>
</target>
<target name="dist" depends="compile">
<jar jarfile="${jar_lib}" basedir="${build}"
includes="org/infradead/libopenconnect/*" />
<jar jarfile="${jar_app}" basedir="${build}">
<manifest>
<attribute name="Main-Class" value="com.example.LibTest" />
</manifest>
</jar>
</target>
<target name="clean">
<delete dir="${build}"/>
<delete dir="${dist}"/>
</target>
</project>
| Build for Java 1.5 to avoid compatibility issues | Build for Java 1.5 to avoid compatibility issues
cf. https://code.google.com/p/android/issues/detail?id=22970
Signed-off-by: David Woodhouse <[email protected]>
| XML | lgpl-2.1 | cernekee/openconnect,matlockx/openconnect,nmav/openconnect-mine,lins05/openconnect,lins05/openconnect,mtmiller/openconnect,brandt/openconnect,atyndall/openconnect,mveplus/openconnect-client,don-johnny/openconnect,don-johnny/openconnect,mtmiller/openconnect,mtmiller/openconnect,liiir1985/openconnect,cernekee/openconnect,brandt/openconnect,lins05/openconnect,atyndall/openconnect,lins05/openconnect,mveplus/openconnect-client,matlockx/openconnect,cernekee/openconnect,liiir1985/openconnect,liiir1985/openconnect,liiir1985/openconnect,nmav/openconnect-mine,nmav/openconnect-mine,brandt/openconnect,don-johnny/openconnect,mveplus/openconnect-client,mtmiller/openconnect,matlockx/openconnect,mveplus/openconnect-client,atyndall/openconnect,brandt/openconnect,atyndall/openconnect,nmav/openconnect-mine,don-johnny/openconnect,cernekee/openconnect,matlockx/openconnect | xml | ## Code Before:
<project name="LibOpenConnect" default="dist" basedir=".">
<property name="build" location="build"/>
<property name="dist" location="dist"/>
<property name="src" location="src"/>
<property name="jar_lib" location="${dist}/openconnect-wrapper.jar"/>
<property name="jar_app" location="${dist}/example.jar"/>
<target name="init">
<mkdir dir="${build}"/>
</target>
<target name="compile" depends="init">
<javac srcdir="${src}" destdir="${build}"
includeantruntime="false"/>
</target>
<target name="dist" depends="compile">
<jar jarfile="${jar_lib}" basedir="${build}"
includes="org/infradead/libopenconnect/*" />
<jar jarfile="${jar_app}" basedir="${build}">
<manifest>
<attribute name="Main-Class" value="com.example.LibTest" />
</manifest>
</jar>
</target>
<target name="clean">
<delete dir="${build}"/>
<delete dir="${dist}"/>
</target>
</project>
## Instruction:
Build for Java 1.5 to avoid compatibility issues
cf. https://code.google.com/p/android/issues/detail?id=22970
Signed-off-by: David Woodhouse <[email protected]>
## Code After:
<project name="LibOpenConnect" default="dist" basedir=".">
<property name="build" location="build"/>
<property name="dist" location="dist"/>
<property name="src" location="src"/>
<property name="jar_lib" location="${dist}/openconnect-wrapper.jar"/>
<property name="jar_app" location="${dist}/example.jar"/>
<target name="init">
<mkdir dir="${build}"/>
</target>
<target name="compile" depends="init">
<javac srcdir="${src}" destdir="${build}"
source="1.5" target="1.5" includeantruntime="false"/>
</target>
<target name="dist" depends="compile">
<jar jarfile="${jar_lib}" basedir="${build}"
includes="org/infradead/libopenconnect/*" />
<jar jarfile="${jar_app}" basedir="${build}">
<manifest>
<attribute name="Main-Class" value="com.example.LibTest" />
</manifest>
</jar>
</target>
<target name="clean">
<delete dir="${build}"/>
<delete dir="${dist}"/>
</target>
</project>
|
e358e2b6fc2b6923b37a88f1b3fddb5a2f96fe37 | .travis.yml | .travis.yml | language: C
addons:
apt:
packages:
- libgmp-dev
- libmpfr-dev
cache:
apt: true
#directories:
# - compiler/
#install: bash scripts/build_cross_compiler.sh
script:
- make run-tests
| sudo: false
language: c
addons:
apt:
packages:
# Libraries needed to build cross compiler
- libgmp-dev
- libmpfr-dev
cache:
apt: true
directories:
- compiler/
# Compiles the cross compiler into the projects root
before_install: bash scripts/build_cross_compiler.sh
script:
- make
- make run-tests
| Build the CC, build the kernel, and run libk tests | Build the CC, build the kernel, and run libk tests | YAML | mit | iankronquist/kernel-of-truth,Herbstein/kernel-of-truth,Herbstein/kernel-of-truth,awensaunders/kernel-of-truth,iankronquist/kernel-of-truth,awensaunders/kernel-of-truth,iankronquist/kernel-of-truth,awensaunders/kernel-of-truth,Herbstein/kernel-of-truth,iankronquist/kernel-of-truth,iankronquist/kernel-of-truth | yaml | ## Code Before:
language: C
addons:
apt:
packages:
- libgmp-dev
- libmpfr-dev
cache:
apt: true
#directories:
# - compiler/
#install: bash scripts/build_cross_compiler.sh
script:
- make run-tests
## Instruction:
Build the CC, build the kernel, and run libk tests
## Code After:
sudo: false
language: c
addons:
apt:
packages:
# Libraries needed to build cross compiler
- libgmp-dev
- libmpfr-dev
cache:
apt: true
directories:
- compiler/
# Compiles the cross compiler into the projects root
before_install: bash scripts/build_cross_compiler.sh
script:
- make
- make run-tests
|
efb3b418eba8316b42ac6ee92249ba13598cbc9e | app.sh | app.sh | cat << EOS
AkkeyLab
The elapsed time does not matter.
Because speed is important.
EOS
#
# Install web apps.
#
echo " ----- Install web apps ------"
brew install --cask 1password
brew tap AdoptOpenJDK/openjdk
brew install --cask adoptopenjdk
brew install --cask android-studio
brew install --cask cyberduck
brew install --cask docker
brew install --cask google-chrome
brew install --cask iterm2
brew install --cask sourcetree
brew install --cask vlc
brew install --cask google-japanese-ime
brew install --cask swimat
brew install --cask spotify
brew install --cask imageoptim
brew install --cask soundflower
brew install --cask visual-studio-code
brew install --cask adobe-creative-cloud
brew install --cask gyazo
brew install --cask zoomus
brew install --cask tandem
brew install --cask discord
brew install --cask notion
brew install --cask figma
while true; do
read -p 'Add "need license" apps? [Y/n]' Answer
case $Answer in
'' | [Yy]* )
brew cask install microsoft-office
brew cask install intellij-idea
brew cask install clip-studio-paint
break;
;;
[Nn]* )
echo "Skip install"
break;
;;
* )
echo Please answer YES or NO.
esac
done;
echo " ------------ END ------------"
| cat << EOS
AkkeyLab
The elapsed time does not matter.
Because speed is important.
EOS
#
# Install web apps.
#
echo " ----- Install web apps ------"
brew install --cask 1password
brew tap AdoptOpenJDK/openjdk
brew install --cask adoptopenjdk
brew install --cask android-studio
brew install --cask cyberduck
brew install --cask docker
brew install --cask google-chrome
brew install --cask iterm2
brew install --cask vlc
brew install --cask google-japanese-ime
brew install --cask spotify
brew install --cask imageoptim
brew install --cask visual-studio-code
brew install --cask adobe-creative-cloud
brew install --cask gyazo
brew install --cask zoomus
brew install --cask discord
brew install --cask notion
brew install --cask figma
brew install --cask tableplus
brew install --cask postman
while true; do
read -p 'Add "need license" apps? [Y/n]' Answer
case $Answer in
'' | [Yy]* )
brew cask install microsoft-office
brew cask install intellij-idea
brew cask install clip-studio-paint
break;
;;
[Nn]* )
echo "Skip install"
break;
;;
* )
echo Please answer YES or NO.
esac
done;
echo " ------------ END ------------"
| Update brew cask install commands | :bulb: Update brew cask install commands
| Shell | mit | AkkeyLab/mac-auto-setup | shell | ## Code Before:
cat << EOS
AkkeyLab
The elapsed time does not matter.
Because speed is important.
EOS
#
# Install web apps.
#
echo " ----- Install web apps ------"
brew install --cask 1password
brew tap AdoptOpenJDK/openjdk
brew install --cask adoptopenjdk
brew install --cask android-studio
brew install --cask cyberduck
brew install --cask docker
brew install --cask google-chrome
brew install --cask iterm2
brew install --cask sourcetree
brew install --cask vlc
brew install --cask google-japanese-ime
brew install --cask swimat
brew install --cask spotify
brew install --cask imageoptim
brew install --cask soundflower
brew install --cask visual-studio-code
brew install --cask adobe-creative-cloud
brew install --cask gyazo
brew install --cask zoomus
brew install --cask tandem
brew install --cask discord
brew install --cask notion
brew install --cask figma
while true; do
read -p 'Add "need license" apps? [Y/n]' Answer
case $Answer in
'' | [Yy]* )
brew cask install microsoft-office
brew cask install intellij-idea
brew cask install clip-studio-paint
break;
;;
[Nn]* )
echo "Skip install"
break;
;;
* )
echo Please answer YES or NO.
esac
done;
echo " ------------ END ------------"
## Instruction:
:bulb: Update brew cask install commands
## Code After:
cat << EOS
AkkeyLab
The elapsed time does not matter.
Because speed is important.
EOS
#
# Install web apps.
#
echo " ----- Install web apps ------"
brew install --cask 1password
brew tap AdoptOpenJDK/openjdk
brew install --cask adoptopenjdk
brew install --cask android-studio
brew install --cask cyberduck
brew install --cask docker
brew install --cask google-chrome
brew install --cask iterm2
brew install --cask vlc
brew install --cask google-japanese-ime
brew install --cask spotify
brew install --cask imageoptim
brew install --cask visual-studio-code
brew install --cask adobe-creative-cloud
brew install --cask gyazo
brew install --cask zoomus
brew install --cask discord
brew install --cask notion
brew install --cask figma
brew install --cask tableplus
brew install --cask postman
while true; do
read -p 'Add "need license" apps? [Y/n]' Answer
case $Answer in
'' | [Yy]* )
brew cask install microsoft-office
brew cask install intellij-idea
brew cask install clip-studio-paint
break;
;;
[Nn]* )
echo "Skip install"
break;
;;
* )
echo Please answer YES or NO.
esac
done;
echo " ------------ END ------------"
|
7a403519539146fc50c0897576f3453f11960032 | cli/README.md | cli/README.md |
The GitHub Actions for [Google Cloud Platform](https://cloud.google.com/) and wraps the [gcloud SDK](https://cloud.google.com/sdk/) to enable common Google Cloud commands. This is a thin wrapper around the `gcloud` utility.
## Usage
An example workflow to list clusters on Google Cloud Platform:
```
workflow "Run gcloud command" {
on = "push"
resolves = "GCP Authenticate"
}
action "GCP Authenticate" {
uses = "actions/gcloud/auth@master"
secrets = ["GCLOUD_AUTH"]
}
action "GCP List Clusters" {
needs = ["GCP Authenticate"]
uses = "actions/gcloud/cli@master"
args = "container clusters list"
}
```
For more information about the authentication step, [see `auth`](/auth).
## License
The Dockerfile and associated scripts and documentation in this project are released under the [MIT License](LICENSE).
Container images built with this project include third party materials. See [THIRD_PARTY_NOTICE.md](THIRD_PARTY_NOTICE.md) for details.
|
The GitHub Actions for [Google Cloud Platform](https://cloud.google.com/) and wraps the [gcloud SDK](https://cloud.google.com/sdk/) to enable common Google Cloud commands. This is a thin wrapper around the `gcloud` utility.
## Usage
An example workflow to list clusters on Google Cloud Platform:
```
workflow "Run gcloud command" {
on = "push"
resolves = "GCP List Clusters"
}
action "GCP Authenticate" {
uses = "actions/gcloud/auth@master"
secrets = ["GCLOUD_AUTH"]
}
action "GCP List Clusters" {
needs = ["GCP Authenticate"]
uses = "actions/gcloud/cli@master"
args = "container clusters list"
}
```
For more information about the authentication step, [see `auth`](/auth).
## License
The Dockerfile and associated scripts and documentation in this project are released under the [MIT License](LICENSE).
Container images built with this project include third party materials. See [THIRD_PARTY_NOTICE.md](THIRD_PARTY_NOTICE.md) for details.
| Update text to GCP List Clusters | Update text to GCP List Clusters
Addresses PR feedback, thanks @parkr! | Markdown | apache-2.0 | google-github-actions/setup-gcloud,google-github-actions/setup-gcloud,google-github-actions/setup-gcloud | markdown | ## Code Before:
The GitHub Actions for [Google Cloud Platform](https://cloud.google.com/) and wraps the [gcloud SDK](https://cloud.google.com/sdk/) to enable common Google Cloud commands. This is a thin wrapper around the `gcloud` utility.
## Usage
An example workflow to list clusters on Google Cloud Platform:
```
workflow "Run gcloud command" {
on = "push"
resolves = "GCP Authenticate"
}
action "GCP Authenticate" {
uses = "actions/gcloud/auth@master"
secrets = ["GCLOUD_AUTH"]
}
action "GCP List Clusters" {
needs = ["GCP Authenticate"]
uses = "actions/gcloud/cli@master"
args = "container clusters list"
}
```
For more information about the authentication step, [see `auth`](/auth).
## License
The Dockerfile and associated scripts and documentation in this project are released under the [MIT License](LICENSE).
Container images built with this project include third party materials. See [THIRD_PARTY_NOTICE.md](THIRD_PARTY_NOTICE.md) for details.
## Instruction:
Update text to GCP List Clusters
Addresses PR feedback, thanks @parkr!
## Code After:
The GitHub Actions for [Google Cloud Platform](https://cloud.google.com/) and wraps the [gcloud SDK](https://cloud.google.com/sdk/) to enable common Google Cloud commands. This is a thin wrapper around the `gcloud` utility.
## Usage
An example workflow to list clusters on Google Cloud Platform:
```
workflow "Run gcloud command" {
on = "push"
resolves = "GCP List Clusters"
}
action "GCP Authenticate" {
uses = "actions/gcloud/auth@master"
secrets = ["GCLOUD_AUTH"]
}
action "GCP List Clusters" {
needs = ["GCP Authenticate"]
uses = "actions/gcloud/cli@master"
args = "container clusters list"
}
```
For more information about the authentication step, [see `auth`](/auth).
## License
The Dockerfile and associated scripts and documentation in this project are released under the [MIT License](LICENSE).
Container images built with this project include third party materials. See [THIRD_PARTY_NOTICE.md](THIRD_PARTY_NOTICE.md) for details.
|
d84d5f831424806c82033f660205cbfb2c1e7341 | src/app/header/header.component.html | src/app/header/header.component.html | <nav class="navbar fixed-top navbar-dark bg-inverse">
<button class="navbar-toggler navbar-toggler-right" type="button" data-toggle="collapse" data-target="#navbarNav" aria-controls="navbarNav" aria-expanded="false" aria-label="Toggle navigation">
<span class="navbar-toggler-icon"></span>
</button>
<a class="navbar-brand" [routerLink]="['/home']">Catalogo Ricette</a>
<div class="collapse navbar-collapse" id="navbarNav">
<ul class="navbar-nav">
<li class="nav-item active">
<a class="nav-link" [routerLink]="['/home']">Home <span class="sr-only">(current)</span></a>
</li>
<li class="nav-item">
<a class="nav-link disabled">Ingredienti</a>
</li>
<li class="nav-item">
<a class="nav-link" [routerLink]="['/dishTypes']">Categorie</a>
</li>
<li class="nav-item">
<a class="nav-link disabled">Libri</a>
</li>
</ul>
</div>
</nav> | <nav class="navbar fixed-top navbar-toggleable-md navbar-inverse bg-inverse">
<button class="navbar-toggler navbar-toggler-right" type="button" data-toggle="collapse" data-target="#navbarSupportedContent" aria-controls="navbarSupportedContent" aria-expanded="false" aria-label="Toggle navigation">
<span class="navbar-toggler-icon"></span>
</button>
<a class="navbar-brand" [routerLink]="['/home']">Catalogo Ricette</a>
<div class="collapse navbar-collapse" id="navbarSupportedContent">
<ul class="navbar-nav">
<li class="nav-item">
<a class="nav-link" [routerLink]="['/home']">Home <span class="sr-only">(current)</span></a>
</li>
<li class="nav-item">
<a class="nav-link disabled">Ingredienti</a>
</li>
<li class="nav-item">
<a class="nav-link" [routerLink]="['/dishTypes']">Categorie</a>
</li>
<li class="nav-item">
<a class="nav-link disabled">Libri</a>
</li>
</ul>
</div>
</nav> | Fix navigation bar to work with latest bootstrap | Fix navigation bar to work with latest bootstrap
| HTML | mit | riccardobellini/recipecatalog-frontend-angular,riccardobellini/recipecatalog-frontend-angular,riccardobellini/recipecatalog-frontend-angular | html | ## Code Before:
<nav class="navbar fixed-top navbar-dark bg-inverse">
<button class="navbar-toggler navbar-toggler-right" type="button" data-toggle="collapse" data-target="#navbarNav" aria-controls="navbarNav" aria-expanded="false" aria-label="Toggle navigation">
<span class="navbar-toggler-icon"></span>
</button>
<a class="navbar-brand" [routerLink]="['/home']">Catalogo Ricette</a>
<div class="collapse navbar-collapse" id="navbarNav">
<ul class="navbar-nav">
<li class="nav-item active">
<a class="nav-link" [routerLink]="['/home']">Home <span class="sr-only">(current)</span></a>
</li>
<li class="nav-item">
<a class="nav-link disabled">Ingredienti</a>
</li>
<li class="nav-item">
<a class="nav-link" [routerLink]="['/dishTypes']">Categorie</a>
</li>
<li class="nav-item">
<a class="nav-link disabled">Libri</a>
</li>
</ul>
</div>
</nav>
## Instruction:
Fix navigation bar to work with latest bootstrap
## Code After:
<nav class="navbar fixed-top navbar-toggleable-md navbar-inverse bg-inverse">
<button class="navbar-toggler navbar-toggler-right" type="button" data-toggle="collapse" data-target="#navbarSupportedContent" aria-controls="navbarSupportedContent" aria-expanded="false" aria-label="Toggle navigation">
<span class="navbar-toggler-icon"></span>
</button>
<a class="navbar-brand" [routerLink]="['/home']">Catalogo Ricette</a>
<div class="collapse navbar-collapse" id="navbarSupportedContent">
<ul class="navbar-nav">
<li class="nav-item">
<a class="nav-link" [routerLink]="['/home']">Home <span class="sr-only">(current)</span></a>
</li>
<li class="nav-item">
<a class="nav-link disabled">Ingredienti</a>
</li>
<li class="nav-item">
<a class="nav-link" [routerLink]="['/dishTypes']">Categorie</a>
</li>
<li class="nav-item">
<a class="nav-link disabled">Libri</a>
</li>
</ul>
</div>
</nav> |
2ad9290db1bdc9a155a6c327f92088209f13d141 | lib/pdoc.rb | lib/pdoc.rb | DIR = File.dirname(__FILE__)
OUTPUT_DIR = File.expand_path(File.join(DIR, "..", "output"))
TEMPLATES_DIR = File.expand_path(File.join(DIR, "..", "templates"))
PARSER_DIR = File.expand_path(File.join(DIR, "pdoc", "parser"))
VENDOR_DIR = File.expand_path(File.join(DIR, "..", "vendor"))
[DIR, VENDOR_DIR, PARSER_DIR, OUTPUT_DIR, TEMPLATES_DIR].each do |c|
$:.unshift File.expand_path(c)
end
require 'erb'
require 'fileutils'
require File.expand_path(File.join(DIR, "pdoc", "runner"))
require File.expand_path(File.join(DIR, "pdoc", "generators"))
require File.expand_path(File.join(DIR, "pdoc", "parser"))
module PDoc
def self.copy_templates(template_type, destination)
dir = File.expand_path(destination)
raise "File already exists: #{destination}" if File.exist?(dir)
FileUtils.cp_r("#{TEMPLATES_DIR}/#{template_type}", dir)
end
end
| DIR = File.expand_path(File.dirname(__FILE__))
OUTPUT_DIR = File.join(DIR, '..', 'output')
TEMPLATES_DIR = File.join(DIR, '..', 'templates')
VENDOR_DIR = File.join(DIR, '..', 'vendor')
PARSER_DIR = File.join(DIR, 'pdoc', 'parser')
[DIR, VENDOR_DIR, PARSER_DIR, OUTPUT_DIR, TEMPLATES_DIR].each do |c|
$:.unshift(c)
end
require 'rubygems'
require 'erb'
require 'fileutils'
require 'pdoc/runner'
require 'pdoc/generators'
require 'pdoc/parser'
module PDoc
def self.copy_templates(template_type, destination)
dir = File.expand_path(destination)
raise "File already exists: #{destination}" if File.exist?(dir)
FileUtils.cp_r("#{TEMPLATES_DIR}/#{template_type}", dir)
end
end
| Clean up internal requires and external dependencies. | Clean up internal requires and external dependencies.
| Ruby | mit | tobie/pdoc,tobie/pdoc | ruby | ## Code Before:
DIR = File.dirname(__FILE__)
OUTPUT_DIR = File.expand_path(File.join(DIR, "..", "output"))
TEMPLATES_DIR = File.expand_path(File.join(DIR, "..", "templates"))
PARSER_DIR = File.expand_path(File.join(DIR, "pdoc", "parser"))
VENDOR_DIR = File.expand_path(File.join(DIR, "..", "vendor"))
[DIR, VENDOR_DIR, PARSER_DIR, OUTPUT_DIR, TEMPLATES_DIR].each do |c|
$:.unshift File.expand_path(c)
end
require 'erb'
require 'fileutils'
require File.expand_path(File.join(DIR, "pdoc", "runner"))
require File.expand_path(File.join(DIR, "pdoc", "generators"))
require File.expand_path(File.join(DIR, "pdoc", "parser"))
module PDoc
def self.copy_templates(template_type, destination)
dir = File.expand_path(destination)
raise "File already exists: #{destination}" if File.exist?(dir)
FileUtils.cp_r("#{TEMPLATES_DIR}/#{template_type}", dir)
end
end
## Instruction:
Clean up internal requires and external dependencies.
## Code After:
DIR = File.expand_path(File.dirname(__FILE__))
OUTPUT_DIR = File.join(DIR, '..', 'output')
TEMPLATES_DIR = File.join(DIR, '..', 'templates')
VENDOR_DIR = File.join(DIR, '..', 'vendor')
PARSER_DIR = File.join(DIR, 'pdoc', 'parser')
[DIR, VENDOR_DIR, PARSER_DIR, OUTPUT_DIR, TEMPLATES_DIR].each do |c|
$:.unshift(c)
end
require 'rubygems'
require 'erb'
require 'fileutils'
require 'pdoc/runner'
require 'pdoc/generators'
require 'pdoc/parser'
module PDoc
def self.copy_templates(template_type, destination)
dir = File.expand_path(destination)
raise "File already exists: #{destination}" if File.exist?(dir)
FileUtils.cp_r("#{TEMPLATES_DIR}/#{template_type}", dir)
end
end
|
44e5a3a2338907b975a7042372739bf88de81c6e | gcloud/aliases.zsh | gcloud/aliases.zsh | alias gcrli='gcloud container images list --repository=asia.gcr.io/greensync-container-registry'
gcrlt() {
gcloud container images list-tags asia.gcr.io/greensync-container-registry/"$1"
}
| alias gcrli='gcloud container images list --repository=asia.gcr.io/greensync-container-registry'
gcrlt() {
local full_image_name
if [[ "$1" =~ ^asia.gcr.io/greensync-container-registry/ ]]; then
full_image_name="$1"
else
full_image_name="asia.gcr.io/greensync-container-registry/$1"
fi
gcloud container images list-tags "$full_image_name" --sort-by=TIMESTAMP
}
gcrl() {
if [[ -n "$1" ]]; then
gcrlt "$1"
else
gcrli
fi
}
| Make gcrlt support either image name or full image name; Add gcrl | Make gcrlt support either image name or full image name; Add gcrl
| Shell | mit | ZimbiX/dotfiles,ZimbiX/dotfiles | shell | ## Code Before:
alias gcrli='gcloud container images list --repository=asia.gcr.io/greensync-container-registry'
gcrlt() {
gcloud container images list-tags asia.gcr.io/greensync-container-registry/"$1"
}
## Instruction:
Make gcrlt support either image name or full image name; Add gcrl
## Code After:
alias gcrli='gcloud container images list --repository=asia.gcr.io/greensync-container-registry'
gcrlt() {
local full_image_name
if [[ "$1" =~ ^asia.gcr.io/greensync-container-registry/ ]]; then
full_image_name="$1"
else
full_image_name="asia.gcr.io/greensync-container-registry/$1"
fi
gcloud container images list-tags "$full_image_name" --sort-by=TIMESTAMP
}
gcrl() {
if [[ -n "$1" ]]; then
gcrlt "$1"
else
gcrli
fi
}
|
95329797ba6e76fa2d0aecc348805ac06b6e3e0f | scripts/after.sh | scripts/after.sh | echo "Running after.sh"
#!/bin/sh
| echo "Running after.sh"
#!/bin/sh
apt-get -y install phantomjs screen
mysql -u root -pvagrant -e "GRANT ALL PRIVILEGES ON homestead.* TO 'homestead'@'%' IDENTIFIED BY 'secret'"
echo "Set Travis privs"
(crontab -l 2>/dev/null; echo "@reboot screen -S server -d -m phantomjs --webdriver=4444") | crontab -
echo "Added scree phantomjs to crontab"
screen -S server -d -m phantomjs --webdriver=4444
echo "Started phantomjs is screen"
perl -pi -e 's/upload_max_filesize = 2M/upload_max_filesize = 50M/g' /etc/php5/fpm/php.ini
perl -pi -e 's/post_max_size = 8M/post_max_size = 51M/g' /etc/php5/fpm/php.ini
service nginx restart
service php5-fpm reload
/usr/local/bin/composer self-update
echo "" >> /home/vagrant/.bashrc
echo "PATH=$PATH:/home/vagrant/memvents.dev/vendor/bin" >> /home/vagrant/.bashrc
echo "export PATH" >> /home/vagrant/.bashrc
echo Done! | Add codeception testing stuff to vagrant | Add codeception testing stuff to vagrant
| Shell | mit | svpernova09/memvents,svpernova09/memvents,svpernova09/memvents,svpernova09/memvents | shell | ## Code Before:
echo "Running after.sh"
#!/bin/sh
## Instruction:
Add codeception testing stuff to vagrant
## Code After:
echo "Running after.sh"
#!/bin/sh
apt-get -y install phantomjs screen
mysql -u root -pvagrant -e "GRANT ALL PRIVILEGES ON homestead.* TO 'homestead'@'%' IDENTIFIED BY 'secret'"
echo "Set Travis privs"
(crontab -l 2>/dev/null; echo "@reboot screen -S server -d -m phantomjs --webdriver=4444") | crontab -
echo "Added scree phantomjs to crontab"
screen -S server -d -m phantomjs --webdriver=4444
echo "Started phantomjs is screen"
perl -pi -e 's/upload_max_filesize = 2M/upload_max_filesize = 50M/g' /etc/php5/fpm/php.ini
perl -pi -e 's/post_max_size = 8M/post_max_size = 51M/g' /etc/php5/fpm/php.ini
service nginx restart
service php5-fpm reload
/usr/local/bin/composer self-update
echo "" >> /home/vagrant/.bashrc
echo "PATH=$PATH:/home/vagrant/memvents.dev/vendor/bin" >> /home/vagrant/.bashrc
echo "export PATH" >> /home/vagrant/.bashrc
echo Done! |
58d56bd6b7608349fb491e9f281e10c3aea838ba | lib/mforwd/deliver/slack.rb | lib/mforwd/deliver/slack.rb | class MForwd::Deliver::Slack < MForwd::Deliver::HttpPost
private
# @param message [MForwd::Message]
def create_body message
{ text: message.body }.to_json
end
end
| class MForwd::Deliver::Slack < MForwd::Deliver::HttpPost
@@color_of_kind = {
'good' => /^(green$|ok$|good$|recover)/i,
'warning' => /^(yellow$|warn)/i,
'danger' => /^(red$|fatal$|crit|error$|danger$)/i,
}
private
# @param message [MForwd::Message]
def create_body message
param = {
attachments: [{
fallback: "#{message.title}\n#{message.body}",
fields: [{
title: message.title,
value: message.body,
}],
}],
}
@@color_of_kind.find { |col, regexp|
regexp.match message.kind
}.tap do |col, regexp|
param[:attachments][0][:color] = col
end
param.to_json
end
end
| Set rich format for Slack POST | Set rich format for Slack POST
| Ruby | mit | key-amb/poloxy,key-amb/poloxy | ruby | ## Code Before:
class MForwd::Deliver::Slack < MForwd::Deliver::HttpPost
private
# @param message [MForwd::Message]
def create_body message
{ text: message.body }.to_json
end
end
## Instruction:
Set rich format for Slack POST
## Code After:
class MForwd::Deliver::Slack < MForwd::Deliver::HttpPost
@@color_of_kind = {
'good' => /^(green$|ok$|good$|recover)/i,
'warning' => /^(yellow$|warn)/i,
'danger' => /^(red$|fatal$|crit|error$|danger$)/i,
}
private
# @param message [MForwd::Message]
def create_body message
param = {
attachments: [{
fallback: "#{message.title}\n#{message.body}",
fields: [{
title: message.title,
value: message.body,
}],
}],
}
@@color_of_kind.find { |col, regexp|
regexp.match message.kind
}.tap do |col, regexp|
param[:attachments][0][:color] = col
end
param.to_json
end
end
|
6f2b4ccdfa297605fd3d7be63039a8fcbda714d1 | routes/titles.rb | routes/titles.rb | module Sinatra
module TreeStats
module Routing
module Titles
def self.registered(app)
app.get '/titles/?' do
@titles = TitleHelper::TITLES
@titles = @titles.select { |i,t| i > 0 }.sort_by { |i,t| t }
not_found("No titles found.") if @titles.length < 1
haml :titles
end
app.get '/title/:title' do |title_name|
titles = TitleHelper::TITLES
title = titles.select { |i,title| title == title_name }
not_found("Title #{title} not found.") if title.length == 0
title_id = title.first[0]
@title_name = title.first[1]
@characters = Character.where(:titles.in => [title_id]).limit(100).project()
haml :title
end
end
end
end
end
end
| module Sinatra
module TreeStats
module Routing
module Titles
def self.registered(app)
app.get '/titles/?' do
@titles = TitleHelper::TITLES
@titles = @titles.select { |i,t| i > 0 }.sort_by { |i,t| t }
not_found("No titles found.") if @titles.length < 1
haml :titles
end
app.get '/title/:title' do |title_name|
titles = TitleHelper::TITLES
title = titles.select { |i,title| title == title_name }
not_found("Title #{title} not found.") if title.length == 0
title_id = title.first[0]
@title_name = title.first[1]
@characters = Character.where(:titles.in => [title_id]).sort(:s => 1, :n => 1).limit(100)
haml :title
end
end
end
end
end
end
| Sort retrieved chars with title by server then player name | Sort retrieved chars with title by server then player name
| Ruby | mit | amoeba/treestats.net,amoeba/treestats.net,amoeba/treestats.net | ruby | ## Code Before:
module Sinatra
module TreeStats
module Routing
module Titles
def self.registered(app)
app.get '/titles/?' do
@titles = TitleHelper::TITLES
@titles = @titles.select { |i,t| i > 0 }.sort_by { |i,t| t }
not_found("No titles found.") if @titles.length < 1
haml :titles
end
app.get '/title/:title' do |title_name|
titles = TitleHelper::TITLES
title = titles.select { |i,title| title == title_name }
not_found("Title #{title} not found.") if title.length == 0
title_id = title.first[0]
@title_name = title.first[1]
@characters = Character.where(:titles.in => [title_id]).limit(100).project()
haml :title
end
end
end
end
end
end
## Instruction:
Sort retrieved chars with title by server then player name
## Code After:
module Sinatra
module TreeStats
module Routing
module Titles
def self.registered(app)
app.get '/titles/?' do
@titles = TitleHelper::TITLES
@titles = @titles.select { |i,t| i > 0 }.sort_by { |i,t| t }
not_found("No titles found.") if @titles.length < 1
haml :titles
end
app.get '/title/:title' do |title_name|
titles = TitleHelper::TITLES
title = titles.select { |i,title| title == title_name }
not_found("Title #{title} not found.") if title.length == 0
title_id = title.first[0]
@title_name = title.first[1]
@characters = Character.where(:titles.in => [title_id]).sort(:s => 1, :n => 1).limit(100)
haml :title
end
end
end
end
end
end
|
d8eb5c32f51407b905c99ab43fcf65d441cd9f84 | app/controllers/financials_controller.rb | app/controllers/financials_controller.rb | class FinancialsController < ProductController
before_action :find_product!
def index
unless params[:product_id] == 'coderwall'
redirect_to product_path(@product)
end
end
end
| class FinancialsController < ProductController
before_action :find_product!
def index
unless @product.slug == 'coderwall'
redirect_to product_path(@product)
end
end
end
| Check slug instead of path param | Check slug instead of path param
| Ruby | agpl-3.0 | lachlanjc/meta,assemblymade/meta,lachlanjc/meta,lachlanjc/meta,assemblymade/meta,lachlanjc/meta,assemblymade/meta,assemblymade/meta | ruby | ## Code Before:
class FinancialsController < ProductController
before_action :find_product!
def index
unless params[:product_id] == 'coderwall'
redirect_to product_path(@product)
end
end
end
## Instruction:
Check slug instead of path param
## Code After:
class FinancialsController < ProductController
before_action :find_product!
def index
unless @product.slug == 'coderwall'
redirect_to product_path(@product)
end
end
end
|
49fb91671dc758d0d692c6d8ac90bbb63678ad75 | mc/Voyage-Model-Tests.package/VOArangoRepositoryTest.class/instance/setUp.st | mc/Voyage-Model-Tests.package/VOArangoRepositoryTest.class/instance/setUp.st | running
setUp
repository := VOArangoRepository uri: 'http://localhost:8529' | running
setUp
repository := VOArangoRepository uri: 'http://localhost:8529'.
PQQueryBuilder generatorClass: PQArangoQueryGenerator. | Set up generator class for arango | Set up generator class for arango | Smalltalk | mit | pharo-nosql/voyage,tinchodias/voyage,tinchodias/voyage,noha/voyage,noha/voyage | smalltalk | ## Code Before:
running
setUp
repository := VOArangoRepository uri: 'http://localhost:8529'
## Instruction:
Set up generator class for arango
## Code After:
running
setUp
repository := VOArangoRepository uri: 'http://localhost:8529'.
PQQueryBuilder generatorClass: PQArangoQueryGenerator. |
f822b890d91d20ccb5a7da2935a1d4981e524c0f | app/views/home/show.html.erb | app/views/home/show.html.erb | <%= render 'shared/header' %>
<main class="mainPageContent">
<div class="l-container">
<h1><%= t('.heading_html') %></h1>
<ul class="specialList"><%= t('.summary_list_html') %></ul>
<div id="introductionText"><%= t('.introduction_text_html') %></div>
</div>
</main>
<%= render 'shared/footer_site_links', locals: alternative_locales %>
<%= render 'shared/footer_social_links' %>
| <%= render 'shared/header' %>
<main class="mainPageContent">
<div class="l-container">
<h1><%= t('.heading_html') %></h1>
<ul class="list--iconBullets"><%= t('.summary_list_html') %></ul>
<div id="introductionText"><%= t('.introduction_text_html') %></div>
</div>
</main>
<%= render 'shared/footer_site_links', locals: alternative_locales %>
<%= render 'shared/footer_social_links' %>
| Replace specialList with a more semantic name. | Replace specialList with a more semantic name.
| HTML+ERB | mit | moneyadviceservice/frontend,moneyadviceservice/frontend,moneyadviceservice/frontend,moneyadviceservice/frontend | html+erb | ## Code Before:
<%= render 'shared/header' %>
<main class="mainPageContent">
<div class="l-container">
<h1><%= t('.heading_html') %></h1>
<ul class="specialList"><%= t('.summary_list_html') %></ul>
<div id="introductionText"><%= t('.introduction_text_html') %></div>
</div>
</main>
<%= render 'shared/footer_site_links', locals: alternative_locales %>
<%= render 'shared/footer_social_links' %>
## Instruction:
Replace specialList with a more semantic name.
## Code After:
<%= render 'shared/header' %>
<main class="mainPageContent">
<div class="l-container">
<h1><%= t('.heading_html') %></h1>
<ul class="list--iconBullets"><%= t('.summary_list_html') %></ul>
<div id="introductionText"><%= t('.introduction_text_html') %></div>
</div>
</main>
<%= render 'shared/footer_site_links', locals: alternative_locales %>
<%= render 'shared/footer_social_links' %>
|
faa739924994acc3d61459db9b75ffa81cd87bcb | examples/reset_password.html | examples/reset_password.html | <!DOCTYPE html>
<head>
<meta name="viewport" content="width=device-width, initial-scale=1">
</head>
{% if error %}
<p>{{ error }}</p>
{% endif %}
<form method="POST" action="/reset-password">
<label for="password">New Password</label>
<input type="password" name="password"><br>
<label for="confirm">Confirm Password</label>
<input type="password" name="confirm"><br>
<input type="hidden" name="code" value="{{ code }}">
<input type="hidden" name="user_id" value="{{ user_id }}">
<input type="hidden" name="expire_at" value="{{ expire_at }}">
<input type="submit" value="Submit">
</form>
| <!DOCTYPE html>
<head>
<meta name="viewport" content="width=device-width, initial-scale=1">
</head>
{% if error %}
<p>{{ error }}</p>
{% endif %}
<form method="POST" action="/auth/fogot_password/reset_password">
<label for="password">New Password</label>
<input type="password" name="password"><br>
<label for="confirm">Confirm Password</label>
<input type="password" name="confirm"><br>
<input type="hidden" name="code" value="{{ code }}">
<input type="hidden" name="user_id" value="{{ user_id }}">
<input type="hidden" name="expire_at" value="{{ expire_at }}">
<input type="submit" value="Submit">
</form>
| Update template reset password url | Update template reset password url
| HTML | apache-2.0 | SkygearIO/skygear-server,SkygearIO/skygear-server,SkygearIO/skygear-server,SkygearIO/skygear-server | html | ## Code Before:
<!DOCTYPE html>
<head>
<meta name="viewport" content="width=device-width, initial-scale=1">
</head>
{% if error %}
<p>{{ error }}</p>
{% endif %}
<form method="POST" action="/reset-password">
<label for="password">New Password</label>
<input type="password" name="password"><br>
<label for="confirm">Confirm Password</label>
<input type="password" name="confirm"><br>
<input type="hidden" name="code" value="{{ code }}">
<input type="hidden" name="user_id" value="{{ user_id }}">
<input type="hidden" name="expire_at" value="{{ expire_at }}">
<input type="submit" value="Submit">
</form>
## Instruction:
Update template reset password url
## Code After:
<!DOCTYPE html>
<head>
<meta name="viewport" content="width=device-width, initial-scale=1">
</head>
{% if error %}
<p>{{ error }}</p>
{% endif %}
<form method="POST" action="/auth/fogot_password/reset_password">
<label for="password">New Password</label>
<input type="password" name="password"><br>
<label for="confirm">Confirm Password</label>
<input type="password" name="confirm"><br>
<input type="hidden" name="code" value="{{ code }}">
<input type="hidden" name="user_id" value="{{ user_id }}">
<input type="hidden" name="expire_at" value="{{ expire_at }}">
<input type="submit" value="Submit">
</form>
|
e53ecef685569dfad2c62cd38c53190a9b2012d0 | metpy/gridding/__init__.py | metpy/gridding/__init__.py |
from .gridding_functions import * # noqa: F403
from .points import * # noqa: F403
from .triangles import * # noqa: F403
from .polygons import * # noqa: F403
from .interpolation import * # noqa: F403
__all__ = gridding_functions.__all__[:] # pylint: disable=undefined-variable
__all__.extend(interpolation.__all__) # pylint: disable=undefined-variable
|
from .gridding_functions import * # noqa: F403
from .interpolation import * # noqa: F403
__all__ = gridding_functions.__all__[:] # pylint: disable=undefined-variable
__all__.extend(interpolation.__all__) # pylint: disable=undefined-variable
| Remove some imported modules from gridding | MNT: Remove some imported modules from gridding
We never intended these to be part of the supported API. The fact that
these were available from metpy.gridding was revealed by the new
autosummary docs.
| Python | bsd-3-clause | ShawnMurd/MetPy,dopplershift/MetPy,ahaberlie/MetPy,jrleeman/MetPy,ahaberlie/MetPy,dopplershift/MetPy,Unidata/MetPy,jrleeman/MetPy,Unidata/MetPy | python | ## Code Before:
from .gridding_functions import * # noqa: F403
from .points import * # noqa: F403
from .triangles import * # noqa: F403
from .polygons import * # noqa: F403
from .interpolation import * # noqa: F403
__all__ = gridding_functions.__all__[:] # pylint: disable=undefined-variable
__all__.extend(interpolation.__all__) # pylint: disable=undefined-variable
## Instruction:
MNT: Remove some imported modules from gridding
We never intended these to be part of the supported API. The fact that
these were available from metpy.gridding was revealed by the new
autosummary docs.
## Code After:
from .gridding_functions import * # noqa: F403
from .interpolation import * # noqa: F403
__all__ = gridding_functions.__all__[:] # pylint: disable=undefined-variable
__all__.extend(interpolation.__all__) # pylint: disable=undefined-variable
|
6bdc335ebb8ed009cbc926e681ee6be8f475f323 | reviewboard/reviews/features.py | reviewboard/reviews/features.py | """Feature definitions for reviews."""
from __future__ import unicode_literals
from django.utils.translation import ugettext_lazy as _
from djblets.features import Feature, FeatureLevel
class GeneralCommentsFeature(Feature):
"""A feature for general comments.
General comments allow comments to be created directly on a review request
without accompanying file attachment or diff. These can be used to raise
issues with the review request itself, such as its summary or description,
or general implementation issues.
"""
feature_id = 'reviews.general_comments'
name = _('General Comments')
level = FeatureLevel.STABLE
summary = _('Allow comments on review requests without an associated file '
'attachment or diff.')
class StatusUpdatesFeature(Feature):
"""A feature for status updates.
A status update is a way for third-party tools to provide feedback on a
review request. In the past, this was done just as a normal review. Status
updates allow those tools (via some integration like Review Bot) to mark
their state (such as pending, success, failure, or error) and then
associate that with a review.
"""
feature_id = 'reviews.status_updates'
name = _('Status Updates')
summary = _('A way for external tools to do checks on a review request '
'and report the results of those checks.')
general_comments_feature = GeneralCommentsFeature()
status_updates_feature = StatusUpdatesFeature()
| """Feature definitions for reviews."""
from __future__ import unicode_literals
from django.utils.translation import ugettext_lazy as _
from djblets.features import Feature, FeatureLevel
class GeneralCommentsFeature(Feature):
"""A feature for general comments.
General comments allow comments to be created directly on a review request
without accompanying file attachment or diff. These can be used to raise
issues with the review request itself, such as its summary or description,
or general implementation issues.
"""
feature_id = 'reviews.general_comments'
name = _('General Comments')
level = FeatureLevel.STABLE
summary = _('Allow comments on review requests without an associated file '
'attachment or diff.')
class StatusUpdatesFeature(Feature):
"""A feature for status updates.
A status update is a way for third-party tools to provide feedback on a
review request. In the past, this was done just as a normal review. Status
updates allow those tools (via some integration like Review Bot) to mark
their state (such as pending, success, failure, or error) and then
associate that with a review.
"""
feature_id = 'reviews.status_updates'
name = _('Status Updates')
level = FeatureLevel.STABLE
summary = _('A way for external tools to do checks on a review request '
'and report the results of those checks.')
general_comments_feature = GeneralCommentsFeature()
status_updates_feature = StatusUpdatesFeature()
| Mark status updates as stable. | Mark status updates as stable.
This allows people to actually use the feature without turning it on in their
`settings_local.py`.
Reviewed at https://reviews.reviewboard.org/r/8569/
| Python | mit | sgallagher/reviewboard,brennie/reviewboard,chipx86/reviewboard,chipx86/reviewboard,chipx86/reviewboard,davidt/reviewboard,reviewboard/reviewboard,davidt/reviewboard,reviewboard/reviewboard,chipx86/reviewboard,brennie/reviewboard,sgallagher/reviewboard,davidt/reviewboard,sgallagher/reviewboard,davidt/reviewboard,brennie/reviewboard,sgallagher/reviewboard,reviewboard/reviewboard,brennie/reviewboard,reviewboard/reviewboard | python | ## Code Before:
"""Feature definitions for reviews."""
from __future__ import unicode_literals
from django.utils.translation import ugettext_lazy as _
from djblets.features import Feature, FeatureLevel
class GeneralCommentsFeature(Feature):
"""A feature for general comments.
General comments allow comments to be created directly on a review request
without accompanying file attachment or diff. These can be used to raise
issues with the review request itself, such as its summary or description,
or general implementation issues.
"""
feature_id = 'reviews.general_comments'
name = _('General Comments')
level = FeatureLevel.STABLE
summary = _('Allow comments on review requests without an associated file '
'attachment or diff.')
class StatusUpdatesFeature(Feature):
"""A feature for status updates.
A status update is a way for third-party tools to provide feedback on a
review request. In the past, this was done just as a normal review. Status
updates allow those tools (via some integration like Review Bot) to mark
their state (such as pending, success, failure, or error) and then
associate that with a review.
"""
feature_id = 'reviews.status_updates'
name = _('Status Updates')
summary = _('A way for external tools to do checks on a review request '
'and report the results of those checks.')
general_comments_feature = GeneralCommentsFeature()
status_updates_feature = StatusUpdatesFeature()
## Instruction:
Mark status updates as stable.
This allows people to actually use the feature without turning it on in their
`settings_local.py`.
Reviewed at https://reviews.reviewboard.org/r/8569/
## Code After:
"""Feature definitions for reviews."""
from __future__ import unicode_literals
from django.utils.translation import ugettext_lazy as _
from djblets.features import Feature, FeatureLevel
class GeneralCommentsFeature(Feature):
"""A feature for general comments.
General comments allow comments to be created directly on a review request
without accompanying file attachment or diff. These can be used to raise
issues with the review request itself, such as its summary or description,
or general implementation issues.
"""
feature_id = 'reviews.general_comments'
name = _('General Comments')
level = FeatureLevel.STABLE
summary = _('Allow comments on review requests without an associated file '
'attachment or diff.')
class StatusUpdatesFeature(Feature):
"""A feature for status updates.
A status update is a way for third-party tools to provide feedback on a
review request. In the past, this was done just as a normal review. Status
updates allow those tools (via some integration like Review Bot) to mark
their state (such as pending, success, failure, or error) and then
associate that with a review.
"""
feature_id = 'reviews.status_updates'
name = _('Status Updates')
level = FeatureLevel.STABLE
summary = _('A way for external tools to do checks on a review request '
'and report the results of those checks.')
general_comments_feature = GeneralCommentsFeature()
status_updates_feature = StatusUpdatesFeature()
|
0a3495e269ff97b7f8e83d6f23c2bab81c88f99f | travis-ci.sh | travis-ci.sh | set -ex
set | sort
curl -L -o release-1.8.0.zip \
https://github.com/google/googletest/archive/release-1.8.0.zip
unzip release-1.8.0.zip
cd googletest-release-1.8.0
cmake -DCMAKE_PREFIX_PATH=.. .
make
make install
cd ..
mkdir build
cd build
cmake ..
make
make test
| set -ex
set | sort
curl -L -o release-1.8.0.zip \
https://github.com/google/googletest/archive/release-1.8.0.zip
unzip release-1.8.0.zip
cd googletest-release-1.8.0
cmake -DCMAKE_INSTALL_PREFIX=.. .
make
make install
cd ..
mkdir build
cd build
cmake ..
make
make test
| Fix CMake install paths in travis | Fix CMake install paths in travis
| Shell | isc | DemiMarie/SlipRock,DemiMarie/SlipRock,DemiMarie/SlipRock,DemiMarie/SlipRock,DemiMarie/SlipRock | shell | ## Code Before:
set -ex
set | sort
curl -L -o release-1.8.0.zip \
https://github.com/google/googletest/archive/release-1.8.0.zip
unzip release-1.8.0.zip
cd googletest-release-1.8.0
cmake -DCMAKE_PREFIX_PATH=.. .
make
make install
cd ..
mkdir build
cd build
cmake ..
make
make test
## Instruction:
Fix CMake install paths in travis
## Code After:
set -ex
set | sort
curl -L -o release-1.8.0.zip \
https://github.com/google/googletest/archive/release-1.8.0.zip
unzip release-1.8.0.zip
cd googletest-release-1.8.0
cmake -DCMAKE_INSTALL_PREFIX=.. .
make
make install
cd ..
mkdir build
cd build
cmake ..
make
make test
|
2d8cfb3e35d0728314cd262c26157e835c2f8369 | quickstart/src/main/resources/archetype-catalog.xml | quickstart/src/main/resources/archetype-catalog.xml | <?xml version="1.0" encoding="UTF-8"?>
<archetype-catalog>
<archetypes>
<archetype>
<groupId>org.apache.tapestry</groupId>
<artifactId>quickstart</artifactId>
<version>@project.version@</version>
<repository>http://repository.apache.org/snapshots</repository>
<description>Tapestry @project.version@ Quickstart Project</description>
</archetype>
</archetypes>
</archetype-catalog>
| <?xml version="1.0" encoding="UTF-8"?>
<archetype-catalog>
<archetypes>
<archetype>
<groupId>org.apache.tapestry</groupId>
<artifactId>quickstart</artifactId>
<version>@project.version@</version>
<repository>http://repository.apache.org/snapshots</repository>
<description>Tapestry 5 Quickstart Project</description>
</archetype>
</archetypes>
</archetype-catalog>
| Change the description for the quickstart archetype to just "Tapestry 5 Quickstart Project" | TAP5-1265: Change the description for the quickstart archetype to just "Tapestry 5 Quickstart Project"
git-svn-id: d9b8539636d91aff9cd33ed5cd52a0cf73394897@992517 13f79535-47bb-0310-9956-ffa450edef68
| XML | apache-2.0 | agileowl/tapestry-5,apache/tapestry-5,apache/tapestry-5,agileowl/tapestry-5,apache/tapestry-5,apache/tapestry-5,apache/tapestry-5,agileowl/tapestry-5 | xml | ## Code Before:
<?xml version="1.0" encoding="UTF-8"?>
<archetype-catalog>
<archetypes>
<archetype>
<groupId>org.apache.tapestry</groupId>
<artifactId>quickstart</artifactId>
<version>@project.version@</version>
<repository>http://repository.apache.org/snapshots</repository>
<description>Tapestry @project.version@ Quickstart Project</description>
</archetype>
</archetypes>
</archetype-catalog>
## Instruction:
TAP5-1265: Change the description for the quickstart archetype to just "Tapestry 5 Quickstart Project"
git-svn-id: d9b8539636d91aff9cd33ed5cd52a0cf73394897@992517 13f79535-47bb-0310-9956-ffa450edef68
## Code After:
<?xml version="1.0" encoding="UTF-8"?>
<archetype-catalog>
<archetypes>
<archetype>
<groupId>org.apache.tapestry</groupId>
<artifactId>quickstart</artifactId>
<version>@project.version@</version>
<repository>http://repository.apache.org/snapshots</repository>
<description>Tapestry 5 Quickstart Project</description>
</archetype>
</archetypes>
</archetype-catalog>
|
040c2f0f52a606daf9092175363193cc4aa2cee2 | lib/multitenancy/rack/filter.rb | lib/multitenancy/rack/filter.rb | module Multitenancy
class Filter
def initialize(app)
@app = app
end
def call(env)
tenant = Tenant.new tenant_id, sub_tenant_id
Multitenancy.with_tenant tenant do
@app.call env
end
end
private
def tenant_id
# rack converts X_FOO to HTTP_X_FOO
id = env["HTTP_#{Multitenancy.tenant_header}"]
if Multitenancy.tenant_header_regexp.blank?
id
else
match = Multitenancy.tenant_header_regexp.match(id)
match ? match[0] : id
end
end
def subtenant_id
id = env["HTTP_#{Multitenancy.sub_tenant_header}"]
if Multitenancy.sub_tenant_header_regexp.blank?
id
else
match = Multitenancy.sub_tenant_header_regexp.match(id)
match ? match[0] : id
end
end
end
end
| module Multitenancy
class Filter
def initialize(app)
@app = app
end
def call(env)
tenant = Tenant.new tenant_id(env), sub_tenant_id(env)
Multitenancy.with_tenant tenant do
@app.call env
end
end
private
def tenant_id(env)
# rack converts X_FOO to HTTP_X_FOO
id = env["HTTP_#{Multitenancy.tenant_header}"]
if Multitenancy.tenant_header_regexp.blank?
id
else
match = Multitenancy.tenant_header_regexp.match(id)
match ? match[0] : id
end
end
def sub_tenant_id(env)
id = env["HTTP_#{Multitenancy.sub_tenant_header}"]
if Multitenancy.sub_tenant_header_regexp.blank?
id
else
match = Multitenancy.sub_tenant_header_regexp.match(id)
match ? match[0] : id
end
end
end
end
| Fix syntax errors in regexp support. | Fix syntax errors in regexp support.
| Ruby | mit | 85x14/multitenancy | ruby | ## Code Before:
module Multitenancy
class Filter
def initialize(app)
@app = app
end
def call(env)
tenant = Tenant.new tenant_id, sub_tenant_id
Multitenancy.with_tenant tenant do
@app.call env
end
end
private
def tenant_id
# rack converts X_FOO to HTTP_X_FOO
id = env["HTTP_#{Multitenancy.tenant_header}"]
if Multitenancy.tenant_header_regexp.blank?
id
else
match = Multitenancy.tenant_header_regexp.match(id)
match ? match[0] : id
end
end
def subtenant_id
id = env["HTTP_#{Multitenancy.sub_tenant_header}"]
if Multitenancy.sub_tenant_header_regexp.blank?
id
else
match = Multitenancy.sub_tenant_header_regexp.match(id)
match ? match[0] : id
end
end
end
end
## Instruction:
Fix syntax errors in regexp support.
## Code After:
module Multitenancy
class Filter
def initialize(app)
@app = app
end
def call(env)
tenant = Tenant.new tenant_id(env), sub_tenant_id(env)
Multitenancy.with_tenant tenant do
@app.call env
end
end
private
def tenant_id(env)
# rack converts X_FOO to HTTP_X_FOO
id = env["HTTP_#{Multitenancy.tenant_header}"]
if Multitenancy.tenant_header_regexp.blank?
id
else
match = Multitenancy.tenant_header_regexp.match(id)
match ? match[0] : id
end
end
def sub_tenant_id(env)
id = env["HTTP_#{Multitenancy.sub_tenant_header}"]
if Multitenancy.sub_tenant_header_regexp.blank?
id
else
match = Multitenancy.sub_tenant_header_regexp.match(id)
match ? match[0] : id
end
end
end
end
|
75b5fdd2b6627081bd1eb095169ed5f8ac6f3090 | about_me/index.html | about_me/index.html | <!doctype html>
<html lang="en">
<head>
<meta charset="utf-8">
<title>CS290: Homework</title>
</head>
<body>
<main>
<img alt="Andy Chase" src="portrait.jpeg" /><br />
<a href="mailto:[email protected]">[email protected]</a>
</main>
</body>
</html> | <!doctype html>
<html lang="en">
<head>
<meta charset="utf-8">
<title>CS290: Homework</title>
</head>
<body>
<main>
<img alt="Andy Chase" src="portrait.jpeg"/><br/>
<a href="mailto:[email protected]">[email protected]</a>
<div class="schedule">
<h2>My schedule</h2>
<table border="1">
<thead>
<tr>
<th></th>
<th>Mon</th>
<th>Tue</th>
<th>Wed</th>
<th>Thurs</th>
<th>Fri</th>
</tr>
</thead>
<tbody>
<tr>
<th scope="row">10a</th>
<td></td>
<td>CS 419</td>
<td></td>
<td>CS 419
</td>
<td></td>
</tr>
<tr>
<th>11a</th>
<td></td>
<td></td>
<td></td>
<td></td>
<td></td>
</tr>
<tr>
<th>12p</th>
<td></td>
<td>CS 461</td>
<td></td>
<td>CS 461</td>
<td></td>
</tr>
<tr>
<th>1p</th>
<td></td>
<td></td>
<td></td>
<td></td>
<td></td>
</tr>
<tr>
<th>2p</th>
<td>CS 290</td>
<td></td>
<td>CS 290</td>
<td></td>
<td>CS 290</td>
</tr>
</tbody>
</table>
</div>
</main>
</body>
</html> | Add schedule to "about me" | Add schedule to "about me"
| HTML | mit | andychase/classwork,andychase/classwork,andychase/classwork,andychase/classwork,andychase/classwork,andychase/classwork,andychase/classwork,andychase/classwork | html | ## Code Before:
<!doctype html>
<html lang="en">
<head>
<meta charset="utf-8">
<title>CS290: Homework</title>
</head>
<body>
<main>
<img alt="Andy Chase" src="portrait.jpeg" /><br />
<a href="mailto:[email protected]">[email protected]</a>
</main>
</body>
</html>
## Instruction:
Add schedule to "about me"
## Code After:
<!doctype html>
<html lang="en">
<head>
<meta charset="utf-8">
<title>CS290: Homework</title>
</head>
<body>
<main>
<img alt="Andy Chase" src="portrait.jpeg"/><br/>
<a href="mailto:[email protected]">[email protected]</a>
<div class="schedule">
<h2>My schedule</h2>
<table border="1">
<thead>
<tr>
<th></th>
<th>Mon</th>
<th>Tue</th>
<th>Wed</th>
<th>Thurs</th>
<th>Fri</th>
</tr>
</thead>
<tbody>
<tr>
<th scope="row">10a</th>
<td></td>
<td>CS 419</td>
<td></td>
<td>CS 419
</td>
<td></td>
</tr>
<tr>
<th>11a</th>
<td></td>
<td></td>
<td></td>
<td></td>
<td></td>
</tr>
<tr>
<th>12p</th>
<td></td>
<td>CS 461</td>
<td></td>
<td>CS 461</td>
<td></td>
</tr>
<tr>
<th>1p</th>
<td></td>
<td></td>
<td></td>
<td></td>
<td></td>
</tr>
<tr>
<th>2p</th>
<td>CS 290</td>
<td></td>
<td>CS 290</td>
<td></td>
<td>CS 290</td>
</tr>
</tbody>
</table>
</div>
</main>
</body>
</html> |
a887308a086c3cfa32e0802b3f3c950499347399 | js/locales/bootstrap-datepicker.es.js | js/locales/bootstrap-datepicker.es.js | /**
* Spanish translation for bootstrap-datepicker
* Bruno Bonamin <[email protected]>
*/
;(function($){
$.fn.datepicker.dates['es'] = {
days: ["Domingo", "Lunes", "Martes", "Miércoles", "Jueves", "Viernes", "Sábado", "Domingo"],
daysShort: ["Dom", "Lun", "Mar", "Mié", "Jue", "Vie", "Sáb", "Dom"],
daysMin: ["Do", "Lu", "Ma", "Mi", "Ju", "Vi", "Sa", "Do"],
months: ["Enero", "Febrero", "Marzo", "Abril", "Mayo", "Junio", "Julio", "Agosto", "Septiembre", "Octubre", "Noviembre", "Diciembre"],
monthsShort: ["Ene", "Feb", "Mar", "Abr", "May", "Jun", "Jul", "Ago", "Sep", "Oct", "Nov", "Dic"],
today: "Hoy"
};
}(jQuery));
| /**
* Spanish translation for bootstrap-datepicker
* Bruno Bonamin <[email protected]>
*/
;(function($){
$.fn.datepicker.dates['es'] = {
days: ["Domingo", "Lunes", "Martes", "Miércoles", "Jueves", "Viernes", "Sábado", "Domingo"],
daysShort: ["Dom", "Lun", "Mar", "Mié", "Jue", "Vie", "Sáb", "Dom"],
daysMin: ["Do", "Lu", "Ma", "Mi", "Ju", "Vi", "Sa", "Do"],
months: ["Enero", "Febrero", "Marzo", "Abril", "Mayo", "Junio", "Julio", "Agosto", "Septiembre", "Octubre", "Noviembre", "Diciembre"],
monthsShort: ["Ene", "Feb", "Mar", "Abr", "May", "Jun", "Jul", "Ago", "Sep", "Oct", "Nov", "Dic"],
today: "Hoy",
clear: "Limpiar",
weekStart: 1,
format: "dd.mm.yyyy"
};
}(jQuery));
| Add clear, week start and format. | Add clear, week start and format. | JavaScript | apache-2.0 | NFC-DITO/bootstrap-datepicker,nerionavea/bootstrap-datepicker,kiwiupover/bootstrap-datepicker,aldano/bootstrap-datepicker,inway/bootstrap-datepicker,janusnic/bootstrap-datepicker,Tekzenit/bootstrap-datepicker,josegomezr/bootstrap-datepicker,defrian8/bootstrap-datepicker,vegardok/bootstrap-datepicker,otnavle/bootstrap-datepicker,xutongtong/bootstrap-datepicker,jhalak/bootstrap-datepicker,inway/bootstrap-datepicker,Invulner/bootstrap-datepicker,gn0st1k4m/bootstrap-datepicker,acrobat/bootstrap-datepicker,kiwiupover/bootstrap-datepicker,CherryDT/bootstrap-datepicker,wearespindle/datepicker-js,hebbet/bootstrap-datepicker,Azaret/bootstrap-datepicker,SimpleUpdates/bootstrap-datepicker,christophercaburog/bootstrap-datepicker,kevintvh/bootstrap-datepicker,inukshuk/bootstrap-datepicker,mreiden/bootstrap-datepicker,Invulner/bootstrap-datepicker,yangyichen/bootstrap-datepicker,rocLv/bootstrap-datepicker,champierre/bootstrap-datepicker,dckesler/bootstrap-datepicker,vsn4ik/bootstrap-datepicker,steffendietz/bootstrap-datepicker,menatoric59/bootstrap-datepicker,osama9/bootstrap-datepicker,saurabh27125/bootstrap-datepicker,jhalak/bootstrap-datepicker,dswitzer/bootstrap-datepicker,bitzesty/bootstrap-datepicker,stenmuchow/bootstrap-datepicker,rocLv/bootstrap-datepicker,wetet2/bootstrap-datepicker,saurabh27125/bootstrap-datepicker,i-am-kenny/bootstrap-datepicker,huang-x-h/bootstrap-datepicker,eternicode/bootstrap-datepicker,inukshuk/bootstrap-datepicker,MartijnDwars/bootstrap-datepicker,pacozaa/bootstrap-datepicker,parkeugene/bootstrap-datepicker,thetimbanks/bootstrap-datepicker,gn0st1k4m/bootstrap-datepicker,joubertredrat/bootstrap-datepicker,zhaoyan158567/bootstrap-datepicker,josegomezr/bootstrap-datepicker,vegardok/bootstrap-datepicker,dckesler/bootstrap-datepicker,MartijnDwars/bootstrap-datepicker,martinRjs/bootstrap-datepicker,Mteuahasan/bootstrap-datepicker,Mteuahasan/bootstrap-datepicker,inspur-tobacco/bootstrap-datepicker,acrobat/bootstrap-datepicker,wulinmengzhu/bootstrap-datepicker,thetimbanks/bootstrap-datepicker,ibcooley/bootstrap-datepicker,darluc/bootstrap-datepicker,vgrish/bootstrap-datepicker,dswitzer/bootstrap-datepicker,HuBandiT/bootstrap-datepicker,riyan8250/bootstrap-datepicker,keiwerkgvr/bootstrap-datepicker,NFC-DITO/bootstrap-datepicker,daniyel/bootstrap-datepicker,cherylyan/bootstrap-datepicker,nilbus/bootstrap-datepicker,zhaoyan158567/bootstrap-datepicker,daniyel/bootstrap-datepicker,mreiden/bootstrap-datepicker,Doublemedia/bootstrap-datepicker,ibcooley/bootstrap-datepicker,cbryer/bootstrap-datepicker,stenmuchow/bootstrap-datepicker,christophercaburog/bootstrap-datepicker,ashleymcdonald/bootstrap-datepicker,chengky/bootstrap-datepicker,SimpleUpdates/bootstrap-datepicker,HuBandiT/bootstrap-datepicker,josegomezr/bootstrap-datepicker,defrian8/bootstrap-datepicker,bitzesty/bootstrap-datepicker,aldano/bootstrap-datepicker,nilbus/bootstrap-datepicker,CherryDT/bootstrap-datepicker,rstone770/bootstrap-datepicker,cbryer/bootstrap-datepicker,Sprinkle7/bootstrap-datepicker,chengky/bootstrap-datepicker,janusnic/bootstrap-datepicker,i-am-kenny/bootstrap-datepicker,uxsolutions/bootstrap-datepicker,Azaret/bootstrap-datepicker,Sprinkle7/bootstrap-datepicker,CheapSk8/bootstrap-datepicker,ashleymcdonald/bootstrap-datepicker,vgrish/bootstrap-datepicker,menatoric59/bootstrap-datepicker,joubertredrat/bootstrap-datepicker,otnavle/bootstrap-datepicker,jesperronn/bootstrap-datepicker,WeiLend/bootstrap-datepicker,huang-x-h/bootstrap-datepicker,wetet2/bootstrap-datepicker,pacozaa/bootstrap-datepicker,champierre/bootstrap-datepicker,wulinmengzhu/bootstrap-datepicker,Tekzenit/bootstrap-datepicker,osama9/bootstrap-datepicker,yangyichen/bootstrap-datepicker,tcrossland/bootstrap-datepicker,Doublemedia/bootstrap-datepicker,HNygard/bootstrap-datepicker,uxsolutions/bootstrap-datepicker,riyan8250/bootstrap-datepicker,parkeugene/bootstrap-datepicker,Youraiseme/bootstrap-datepicker,keiwerkgvr/bootstrap-datepicker,WeiLend/bootstrap-datepicker,martinRjs/bootstrap-datepicker,steffendietz/bootstrap-datepicker,kevintvh/bootstrap-datepicker,HNygard/bootstrap-datepicker,rstone770/bootstrap-datepicker,xutongtong/bootstrap-datepicker,Youraiseme/bootstrap-datepicker,cherylyan/bootstrap-datepicker,darluc/bootstrap-datepicker,eternicode/bootstrap-datepicker,vsn4ik/bootstrap-datepicker,jesperronn/bootstrap-datepicker,nerionavea/bootstrap-datepicker,hemp/bootstrap-datepicker,CheapSk8/bootstrap-datepicker,hemp/bootstrap-datepicker,inspur-tobacco/bootstrap-datepicker,tcrossland/bootstrap-datepicker,hebbet/bootstrap-datepicker | javascript | ## Code Before:
/**
* Spanish translation for bootstrap-datepicker
* Bruno Bonamin <[email protected]>
*/
;(function($){
$.fn.datepicker.dates['es'] = {
days: ["Domingo", "Lunes", "Martes", "Miércoles", "Jueves", "Viernes", "Sábado", "Domingo"],
daysShort: ["Dom", "Lun", "Mar", "Mié", "Jue", "Vie", "Sáb", "Dom"],
daysMin: ["Do", "Lu", "Ma", "Mi", "Ju", "Vi", "Sa", "Do"],
months: ["Enero", "Febrero", "Marzo", "Abril", "Mayo", "Junio", "Julio", "Agosto", "Septiembre", "Octubre", "Noviembre", "Diciembre"],
monthsShort: ["Ene", "Feb", "Mar", "Abr", "May", "Jun", "Jul", "Ago", "Sep", "Oct", "Nov", "Dic"],
today: "Hoy"
};
}(jQuery));
## Instruction:
Add clear, week start and format.
## Code After:
/**
* Spanish translation for bootstrap-datepicker
* Bruno Bonamin <[email protected]>
*/
;(function($){
$.fn.datepicker.dates['es'] = {
days: ["Domingo", "Lunes", "Martes", "Miércoles", "Jueves", "Viernes", "Sábado", "Domingo"],
daysShort: ["Dom", "Lun", "Mar", "Mié", "Jue", "Vie", "Sáb", "Dom"],
daysMin: ["Do", "Lu", "Ma", "Mi", "Ju", "Vi", "Sa", "Do"],
months: ["Enero", "Febrero", "Marzo", "Abril", "Mayo", "Junio", "Julio", "Agosto", "Septiembre", "Octubre", "Noviembre", "Diciembre"],
monthsShort: ["Ene", "Feb", "Mar", "Abr", "May", "Jun", "Jul", "Ago", "Sep", "Oct", "Nov", "Dic"],
today: "Hoy",
clear: "Limpiar",
weekStart: 1,
format: "dd.mm.yyyy"
};
}(jQuery));
|
ef156b7e8fed483f33be8243a627363984aee96c | blueprints/in-repo-addon/files/lib/__name__/index.js | blueprints/in-repo-addon/files/lib/__name__/index.js | /*jshint node:true*/
module.exports = {
name: '<%= dasherizedModuleName %>',
isDevelopingAddon: function() {
return true;
}
};
| /*jshint node:true*/
'use strict';
module.exports = {
name: '<%= dasherizedModuleName %>',
isDevelopingAddon: function() {
return true;
}
};
| Make in-repo-addon blueprint 'use strict'. | Make in-repo-addon blueprint 'use strict'. | JavaScript | mit | kellyselden/ember-cli,Turbo87/ember-cli,jrjohnson/ember-cli,mike-north/ember-cli,HeroicEric/ember-cli,Turbo87/ember-cli,ef4/ember-cli,calderas/ember-cli,seawatts/ember-cli,sivakumar-kailasam/ember-cli,kategengler/ember-cli,kriswill/ember-cli,asakusuma/ember-cli,thoov/ember-cli,akatov/ember-cli,gfvcastro/ember-cli,patocallaghan/ember-cli,buschtoens/ember-cli,ef4/ember-cli,pzuraq/ember-cli,jgwhite/ember-cli,sivakumar-kailasam/ember-cli,twokul/ember-cli,kellyselden/ember-cli,kellyselden/ember-cli,HeroicEric/ember-cli,mike-north/ember-cli,Turbo87/ember-cli,calderas/ember-cli,rtablada/ember-cli,seawatts/ember-cli,patocallaghan/ember-cli,patocallaghan/ember-cli,kellyselden/ember-cli,cibernox/ember-cli,Turbo87/ember-cli,jgwhite/ember-cli,akatov/ember-cli,pzuraq/ember-cli,fpauser/ember-cli,romulomachado/ember-cli,patocallaghan/ember-cli,trentmwillis/ember-cli,calderas/ember-cli,romulomachado/ember-cli,rtablada/ember-cli,twokul/ember-cli,gfvcastro/ember-cli,asakusuma/ember-cli,mike-north/ember-cli,cibernox/ember-cli,romulomachado/ember-cli,mike-north/ember-cli,thoov/ember-cli,kriswill/ember-cli,ef4/ember-cli,fpauser/ember-cli,twokul/ember-cli,fpauser/ember-cli,trentmwillis/ember-cli,trentmwillis/ember-cli,raycohen/ember-cli,ember-cli/ember-cli,trentmwillis/ember-cli,elwayman02/ember-cli,calderas/ember-cli,kanongil/ember-cli,ember-cli/ember-cli,ember-cli/ember-cli,jrjohnson/ember-cli,kanongil/ember-cli,seawatts/ember-cli,balinterdi/ember-cli,balinterdi/ember-cli,buschtoens/ember-cli,pzuraq/ember-cli,kriswill/ember-cli,HeroicEric/ember-cli,rtablada/ember-cli,raycohen/ember-cli,fpauser/ember-cli,seawatts/ember-cli,twokul/ember-cli,pzuraq/ember-cli,kriswill/ember-cli,jgwhite/ember-cli,gfvcastro/ember-cli,thoov/ember-cli,kategengler/ember-cli,jgwhite/ember-cli,akatov/ember-cli,kanongil/ember-cli,thoov/ember-cli,ef4/ember-cli,romulomachado/ember-cli,gfvcastro/ember-cli,sivakumar-kailasam/ember-cli,HeroicEric/ember-cli,elwayman02/ember-cli,cibernox/ember-cli,kanongil/ember-cli,rtablada/ember-cli,akatov/ember-cli,cibernox/ember-cli,sivakumar-kailasam/ember-cli | javascript | ## Code Before:
/*jshint node:true*/
module.exports = {
name: '<%= dasherizedModuleName %>',
isDevelopingAddon: function() {
return true;
}
};
## Instruction:
Make in-repo-addon blueprint 'use strict'.
## Code After:
/*jshint node:true*/
'use strict';
module.exports = {
name: '<%= dasherizedModuleName %>',
isDevelopingAddon: function() {
return true;
}
};
|
9f79c567d0f46f02f77ea516180801b25fcafd6a | lib/nmap/run_stat.rb | lib/nmap/run_stat.rb | module Nmap
#
# Represents the runstats of a scan.
#
class RunStat < Struct.new(:end_time, :elapsed, :summary, :exit_status)
#
# Converts the stats to a String.
#
# @return [String]
# The String form of the scan.
#
def to_s
"#{self.end_time} #{self.elapsed} #{self.exit_status}"
end
end
end
| module Nmap
#
# Represents the runstats of a scan.
#
# @since 0.7.0
#
class RunStat < Struct.new(:end_time, :elapsed, :summary, :exit_status)
#
# Converts the stats to a String.
#
# @return [String]
# The String form of the scan.
#
def to_s
"#{self.end_time} #{self.elapsed} #{self.exit_status}"
end
end
end
| Mark RunStat as being added in 0.7.0. | Mark RunStat as being added in 0.7.0.
| Ruby | mit | squid108/ruby-nmap,sophsec/ruby-nmap | ruby | ## Code Before:
module Nmap
#
# Represents the runstats of a scan.
#
class RunStat < Struct.new(:end_time, :elapsed, :summary, :exit_status)
#
# Converts the stats to a String.
#
# @return [String]
# The String form of the scan.
#
def to_s
"#{self.end_time} #{self.elapsed} #{self.exit_status}"
end
end
end
## Instruction:
Mark RunStat as being added in 0.7.0.
## Code After:
module Nmap
#
# Represents the runstats of a scan.
#
# @since 0.7.0
#
class RunStat < Struct.new(:end_time, :elapsed, :summary, :exit_status)
#
# Converts the stats to a String.
#
# @return [String]
# The String form of the scan.
#
def to_s
"#{self.end_time} #{self.elapsed} #{self.exit_status}"
end
end
end
|
c1330b3e685ca1ab4bc41ab16bc47631afab887a | lib/mpd-child-process.js | lib/mpd-child-process.js | var Q = require('Q'),
childProcess = require('child_process');
function processPath(exec) {
var exec = exec || Q.denodeify(childProcess.exec);
var deferred = Q.defer();
function handleDone(stdout, stderr) {
var location = stdout[0].trim().replace(/\n$/, '');
deferred.resolve(location);
}
function handleError(error) {
deferred.reject(error);
}
exec('which mpd').done(handleDone, handleError);
return deferred.promise;
}
function create(configFilepath, spawn) {
var spawn = spawn || childProcess.spawn;
var dfd = Q.defer();
this.processPath().then(function (path) {
var mpdProcess = spawn(path, [configFilepath, '--no-daemon']);
mpdProcess.stdout.on('data', function (data) {
dfd.resolve();
});
mpdProcess.on('error', function (error) {
dfd.reject();
});
}, function() { dfd.reject(); });
return dfd.promise;
}
module.exports = {
processPath: processPath,
create: create
};
| var Q = require('Q'),
childProcess = require('child_process');
function processPath(exec) {
var exec = exec || Q.denodeify(childProcess.exec);
var deferred = Q.defer();
function handleDone(stdout, stderr) {
var location = stdout[0].trim().replace(/\n$/, '');
deferred.resolve(location);
}
function handleError(error) {
deferred.reject(error);
}
exec('which mpd').done(handleDone, handleError);
return deferred.promise;
}
function create(configFilepath, spawn) {
var spawn = spawn || childProcess.spawn;
var dfd = Q.defer();
this.processPath().then(function (path) {
var mpdProcess = spawn(path, [configFilepath, '--no-daemon']);
mpdProcess.stdout.on('data', function (data) {
dfd.resolve();
});
mpdProcess.on('error', function (error) {
dfd.reject();
});
}, function() { dfd.reject('Could not find MPD binary'); });
return dfd.promise;
}
module.exports = {
processPath: processPath,
create: create
};
| Add failure message when MPD path is not found | Add failure message when MPD path is not found
| JavaScript | apache-2.0 | radiodan/radiodan.js,radiodan/radiodan.js | javascript | ## Code Before:
var Q = require('Q'),
childProcess = require('child_process');
function processPath(exec) {
var exec = exec || Q.denodeify(childProcess.exec);
var deferred = Q.defer();
function handleDone(stdout, stderr) {
var location = stdout[0].trim().replace(/\n$/, '');
deferred.resolve(location);
}
function handleError(error) {
deferred.reject(error);
}
exec('which mpd').done(handleDone, handleError);
return deferred.promise;
}
function create(configFilepath, spawn) {
var spawn = spawn || childProcess.spawn;
var dfd = Q.defer();
this.processPath().then(function (path) {
var mpdProcess = spawn(path, [configFilepath, '--no-daemon']);
mpdProcess.stdout.on('data', function (data) {
dfd.resolve();
});
mpdProcess.on('error', function (error) {
dfd.reject();
});
}, function() { dfd.reject(); });
return dfd.promise;
}
module.exports = {
processPath: processPath,
create: create
};
## Instruction:
Add failure message when MPD path is not found
## Code After:
var Q = require('Q'),
childProcess = require('child_process');
function processPath(exec) {
var exec = exec || Q.denodeify(childProcess.exec);
var deferred = Q.defer();
function handleDone(stdout, stderr) {
var location = stdout[0].trim().replace(/\n$/, '');
deferred.resolve(location);
}
function handleError(error) {
deferred.reject(error);
}
exec('which mpd').done(handleDone, handleError);
return deferred.promise;
}
function create(configFilepath, spawn) {
var spawn = spawn || childProcess.spawn;
var dfd = Q.defer();
this.processPath().then(function (path) {
var mpdProcess = spawn(path, [configFilepath, '--no-daemon']);
mpdProcess.stdout.on('data', function (data) {
dfd.resolve();
});
mpdProcess.on('error', function (error) {
dfd.reject();
});
}, function() { dfd.reject('Could not find MPD binary'); });
return dfd.promise;
}
module.exports = {
processPath: processPath,
create: create
};
|
bf514f23138cf1fbf4971341ad4d9aff915884cc | wiki/README.md | wiki/README.md |
You may have to create a docker volume for MySQL first:
docker volume create mysql-data
To start the services, just run this command inside the `wiki` directory:
docker-compose up
|
You may have to create a docker volume for MySQL first:
docker volume create mysql-data
To start the services, just run this command inside the `wiki` directory:
docker-compose up
To view the wiki page locally on mac:
In another terminal window, inside the `wiki` directory, connect to
the running docker-machine and then list its IP address:
docker-machine env
eval $(docker-machine env)
Locate the IP address in the DOCKER_HOST variable. Enter this IP address in
your browser, followed by port 8080. Eg:
192.168.99.100:8080
| Add instructions to view wiki on mac | Add instructions to view wiki on mac
| Markdown | mit | oc4d/main,oc4d/main,oc4d/main | markdown | ## Code Before:
You may have to create a docker volume for MySQL first:
docker volume create mysql-data
To start the services, just run this command inside the `wiki` directory:
docker-compose up
## Instruction:
Add instructions to view wiki on mac
## Code After:
You may have to create a docker volume for MySQL first:
docker volume create mysql-data
To start the services, just run this command inside the `wiki` directory:
docker-compose up
To view the wiki page locally on mac:
In another terminal window, inside the `wiki` directory, connect to
the running docker-machine and then list its IP address:
docker-machine env
eval $(docker-machine env)
Locate the IP address in the DOCKER_HOST variable. Enter this IP address in
your browser, followed by port 8080. Eg:
192.168.99.100:8080
|
c15a83118c98a5746136167768975ad97c733205 | app/views/characters/show.haml | app/views/characters/show.haml | - content_for :breadcrumbs do
= link_to "Characters", characters_path
»
%b= @character.name
- if @character.icon
= icon_tag @character.icon
%br
%b= @character.screenname
%br
= @character.name
- if @character.template
= surround '(', ')' do
= link_to @character.template.name, template_path(@character.template)
- if @character.pb
%br
Played by:
= @character.pb
- if logged_in? && current_user.id == @character.user_id
%br
= link_to 'Edit character', edit_character_path(@character)
%br
%br
- if @character.recent_posts.present?
= render partial: 'posts/list', locals: {posts: @character.recent_posts, title: 'Recent Threads' }
| - content_for :breadcrumbs do
= link_to "Characters", characters_path
»
%b= @character.name
- if @character.icon
= icon_tag @character.icon
%br
%b= @character.screenname
%br
= @character.name
- if @character.template
= surround '(', ')' do
= link_to @character.template.name, template_path(@character.template)
- if @character.template_name
%br
Nickname: #{@character.template_name}
- if @character.pb
%br
Played by:
= @character.pb
- if logged_in? && current_user.id == @character.user_id
%br
= link_to 'Edit character', edit_character_path(@character)
%br
%br
- if @character.recent_posts.present?
= render partial: 'posts/list', locals: {posts: @character.recent_posts, title: 'Recent Threads' }
| Add nickname to character page | Add nickname to character page
| Haml | mit | Marri/glowfic,Marri/glowfic,Marri/glowfic,Marri/glowfic | haml | ## Code Before:
- content_for :breadcrumbs do
= link_to "Characters", characters_path
»
%b= @character.name
- if @character.icon
= icon_tag @character.icon
%br
%b= @character.screenname
%br
= @character.name
- if @character.template
= surround '(', ')' do
= link_to @character.template.name, template_path(@character.template)
- if @character.pb
%br
Played by:
= @character.pb
- if logged_in? && current_user.id == @character.user_id
%br
= link_to 'Edit character', edit_character_path(@character)
%br
%br
- if @character.recent_posts.present?
= render partial: 'posts/list', locals: {posts: @character.recent_posts, title: 'Recent Threads' }
## Instruction:
Add nickname to character page
## Code After:
- content_for :breadcrumbs do
= link_to "Characters", characters_path
»
%b= @character.name
- if @character.icon
= icon_tag @character.icon
%br
%b= @character.screenname
%br
= @character.name
- if @character.template
= surround '(', ')' do
= link_to @character.template.name, template_path(@character.template)
- if @character.template_name
%br
Nickname: #{@character.template_name}
- if @character.pb
%br
Played by:
= @character.pb
- if logged_in? && current_user.id == @character.user_id
%br
= link_to 'Edit character', edit_character_path(@character)
%br
%br
- if @character.recent_posts.present?
= render partial: 'posts/list', locals: {posts: @character.recent_posts, title: 'Recent Threads' }
|
4b91cb368de4eaa86f5ca1c7bf03784252c68077 | spec/support/generator_spec_support.rb | spec/support/generator_spec_support.rb | module GeneratorSpecSupport
def assert_expectation_file(path, expectation_filepath = nil)
expectation_filepath ||= File.expand_path("../../generators/expectations/#{path}", __FILE__)
expectation_filepath.sub!(/_spec\.rb\z/, '_spek.rb')
assert_file(path, File.read(expectation_filepath))
end
end
RSpec.configure do |config|
config.include GeneratorSpecSupport, type: :generator
end
| module GeneratorSpecSupport
def assert_expectation_file(path, expectation_filepath = nil)
expectation_filepath ||= File.expand_path("../../../examples/#{path}", __FILE__)
expectation_filepath.sub!(/_spec\.rb\z/, '_spek.rb')
assert_file(path, File.read(expectation_filepath))
end
end
RSpec.configure do |config|
config.include GeneratorSpecSupport, type: :generator
end
| Use examples instead of spec/generators/expectations | Use examples instead of spec/generators/expectations
| Ruby | mit | akm/model_base_generators,akm/model_base_generators,akm/model_base_generators | ruby | ## Code Before:
module GeneratorSpecSupport
def assert_expectation_file(path, expectation_filepath = nil)
expectation_filepath ||= File.expand_path("../../generators/expectations/#{path}", __FILE__)
expectation_filepath.sub!(/_spec\.rb\z/, '_spek.rb')
assert_file(path, File.read(expectation_filepath))
end
end
RSpec.configure do |config|
config.include GeneratorSpecSupport, type: :generator
end
## Instruction:
Use examples instead of spec/generators/expectations
## Code After:
module GeneratorSpecSupport
def assert_expectation_file(path, expectation_filepath = nil)
expectation_filepath ||= File.expand_path("../../../examples/#{path}", __FILE__)
expectation_filepath.sub!(/_spec\.rb\z/, '_spek.rb')
assert_file(path, File.read(expectation_filepath))
end
end
RSpec.configure do |config|
config.include GeneratorSpecSupport, type: :generator
end
|
aaf80ab5b3700b32b20e0b3d069462c50d1a0aca | scgi.gemspec | scgi.gemspec | spec = Gem::Specification.new do |s|
s.name = "scgi"
s.version = "0.9.0"
s.author = "Jeremy Evans"
s.email = "[email protected]"
s.platform = Gem::Platform::RUBY
s.summary = "Simple support for using SCGI in ruby apps, such as Rails"
s.files = %w'LICENSE README lib/scgi.rb'
s.extra_rdoc_files = ['README']
s.require_paths = ["lib"]
s.has_rdoc = true
s.rdoc_options = %w'--inline-source --line-numbers'
s.rubyforge_project = 'scgi'
end
| spec = Gem::Specification.new do |s|
s.name = "scgi"
s.version = "0.9.0"
s.author = "Jeremy Evans"
s.email = "[email protected]"
s.platform = Gem::Platform::RUBY
s.summary = "Simple support for using SCGI in ruby apps, such as Rails"
s.files = %w'LICENSE README lib/scgi.rb'
s.require_paths = ["lib"]
s.has_rdoc = true
s.rdoc_options = %w'--inline-source --line-numbers README lib'
s.rubyforge_project = 'scgi'
end
| Make README the default RDoc page | Make README the default RDoc page
| Ruby | mit | jeremyevans/ruby-scgi | ruby | ## Code Before:
spec = Gem::Specification.new do |s|
s.name = "scgi"
s.version = "0.9.0"
s.author = "Jeremy Evans"
s.email = "[email protected]"
s.platform = Gem::Platform::RUBY
s.summary = "Simple support for using SCGI in ruby apps, such as Rails"
s.files = %w'LICENSE README lib/scgi.rb'
s.extra_rdoc_files = ['README']
s.require_paths = ["lib"]
s.has_rdoc = true
s.rdoc_options = %w'--inline-source --line-numbers'
s.rubyforge_project = 'scgi'
end
## Instruction:
Make README the default RDoc page
## Code After:
spec = Gem::Specification.new do |s|
s.name = "scgi"
s.version = "0.9.0"
s.author = "Jeremy Evans"
s.email = "[email protected]"
s.platform = Gem::Platform::RUBY
s.summary = "Simple support for using SCGI in ruby apps, such as Rails"
s.files = %w'LICENSE README lib/scgi.rb'
s.require_paths = ["lib"]
s.has_rdoc = true
s.rdoc_options = %w'--inline-source --line-numbers README lib'
s.rubyforge_project = 'scgi'
end
|
461281865ccd7521f25b2d4e893eb55c036f56dc | static/js/crowdsource.interceptor.js | static/js/crowdsource.interceptor.js | /**
* Date: 1/9/15
* Author: Shirish Goyal
*/
(function () {
'use strict';
angular
.module('crowdsource.interceptor', [])
.factory('AuthHttpResponseInterceptor', AuthHttpResponseInterceptor);
AuthHttpResponseInterceptor.$inject = ['$location', '$log', '$injector', '$q'];
function AuthHttpResponseInterceptor($location, $log, $injector, $q) {
return {
responseError: function (rejection) {
if (rejection.status === 403) {
if (rejection.hasOwnProperty('data')
&& rejection.data.hasOwnProperty('detail')
&& (rejection.data.detail.indexOf("Authentication credentials were not provided") != -1)
) {
var $http = $injector.get('$http');
var Authentication = $injector.get('Authentication');
if (!Authentication.isAuthenticated()) {
$location.path('/login');
}else{
deferred.resolve(data);
}
}
}
return $q.reject(rejection);
}
}
}
})(); | /**
* Date: 1/9/15
* Author: Shirish Goyal
*/
(function () {
'use strict';
angular
.module('crowdsource.interceptor', [])
.factory('AuthHttpResponseInterceptor', AuthHttpResponseInterceptor);
AuthHttpResponseInterceptor.$inject = ['$location', '$log', '$injector', '$q'];
function AuthHttpResponseInterceptor($location, $log, $injector, $q) {
return {
responseError: function (rejection) {
if (rejection.status === 403) {
if (rejection.hasOwnProperty('data')
&& rejection.data.hasOwnProperty('detail')
&& (rejection.data.detail.indexOf("Authentication credentials were not provided") != -1)
) {
var $http = $injector.get('$http');
var Authentication = $injector.get('Authentication');
if (!Authentication.isAuthenticated()) {
$location.path('/login');
}
}
}
return $q.reject(rejection);
}
}
}
})(); | Remove deferred as not needed | Remove deferred as not needed
| JavaScript | mit | crowdresearch/crowdsource-platform,crowdresearch/crowdsource-platform,crowdresearch/daemo,rakshit-agrawal/crowdsource-platform,Milstein/crowdsource-platform,DESHRAJ/crowdsource-platform,Macbull/crowdsource-platform,crowdresearch/daemo,xasos/crowdsource-platform,crowdresearch/crowdsource-platform,shirishgoyal/crowdsource-platform,xasos/crowdsource-platform,xasos/crowdsource-platform,rakshit-agrawal/crowdsource-platform,aginzberg/crowdsource-platform,DESHRAJ/crowdsource-platform,aginzberg/crowdsource-platform,crowdresearch/daemo,Macbull/crowdsource-platform,shirishgoyal/crowdsource-platform,crowdresearch/crowdsource-platform,aginzberg/crowdsource-platform,Milstein/crowdsource-platform,tanujs22/crowdsource-platform,crowdresearch/daemo,DESHRAJ/crowdsource-platform,aginzberg/crowdsource-platform,shirishgoyal/crowdsource-platform,Macbull/crowdsource-platform,DESHRAJ/crowdsource-platform,rakshit-agrawal/crowdsource-platform,tanujs22/crowdsource-platform,rakshit-agrawal/crowdsource-platform,tanujs22/crowdsource-platform,Milstein/crowdsource-platform,Macbull/crowdsource-platform,xasos/crowdsource-platform,Milstein/crowdsource-platform,shirishgoyal/crowdsource-platform,Milstein/crowdsource-platform,tanujs22/crowdsource-platform | javascript | ## Code Before:
/**
* Date: 1/9/15
* Author: Shirish Goyal
*/
(function () {
'use strict';
angular
.module('crowdsource.interceptor', [])
.factory('AuthHttpResponseInterceptor', AuthHttpResponseInterceptor);
AuthHttpResponseInterceptor.$inject = ['$location', '$log', '$injector', '$q'];
function AuthHttpResponseInterceptor($location, $log, $injector, $q) {
return {
responseError: function (rejection) {
if (rejection.status === 403) {
if (rejection.hasOwnProperty('data')
&& rejection.data.hasOwnProperty('detail')
&& (rejection.data.detail.indexOf("Authentication credentials were not provided") != -1)
) {
var $http = $injector.get('$http');
var Authentication = $injector.get('Authentication');
if (!Authentication.isAuthenticated()) {
$location.path('/login');
}else{
deferred.resolve(data);
}
}
}
return $q.reject(rejection);
}
}
}
})();
## Instruction:
Remove deferred as not needed
## Code After:
/**
* Date: 1/9/15
* Author: Shirish Goyal
*/
(function () {
'use strict';
angular
.module('crowdsource.interceptor', [])
.factory('AuthHttpResponseInterceptor', AuthHttpResponseInterceptor);
AuthHttpResponseInterceptor.$inject = ['$location', '$log', '$injector', '$q'];
function AuthHttpResponseInterceptor($location, $log, $injector, $q) {
return {
responseError: function (rejection) {
if (rejection.status === 403) {
if (rejection.hasOwnProperty('data')
&& rejection.data.hasOwnProperty('detail')
&& (rejection.data.detail.indexOf("Authentication credentials were not provided") != -1)
) {
var $http = $injector.get('$http');
var Authentication = $injector.get('Authentication');
if (!Authentication.isAuthenticated()) {
$location.path('/login');
}
}
}
return $q.reject(rejection);
}
}
}
})(); |
e07d7bba31ffe9cca422417891fee20bf07f462b | lib/lokka/models/tag.rb | lib/lokka/models/tag.rb |
class Tag < ActiveRecord::Base
has_many :taggings,
dependent: :destroy
has_many :entries,
through: :taggings,
source: :tag,
class_name: 'Entry'
validates :name,
presence: true,
uniqueness: true
scope :any,
->(list) {
where(list.map {|name| sanitize_sql(['name = ?', name]) }.join(' OR '))
}
def link
"/tags/#{name}"
end
end
def Tag(name)
Tag.where(name: name).first
end
|
class Tag < ActiveRecord::Base
has_many :taggings,
dependent: :destroy
has_many :entries,
through: :taggings,
source: :taggable,
source_type: 'Entry'
validates :name,
presence: true,
uniqueness: true
scope :any,
->(list) {
where(list.map {|name| sanitize_sql(['name = ?', name]) }.join(' OR '))
}
def link
"/tags/#{name}"
end
end
def Tag(name)
Tag.where(name: name).first
end
| Fix Tag to Entry association | Fix Tag to Entry association
| Ruby | mit | morygonzalez/portalshit.net,morygonzalez/portalshit.net,morygonzalez/portalshit.net,morygonzalez/portalshit.net | ruby | ## Code Before:
class Tag < ActiveRecord::Base
has_many :taggings,
dependent: :destroy
has_many :entries,
through: :taggings,
source: :tag,
class_name: 'Entry'
validates :name,
presence: true,
uniqueness: true
scope :any,
->(list) {
where(list.map {|name| sanitize_sql(['name = ?', name]) }.join(' OR '))
}
def link
"/tags/#{name}"
end
end
def Tag(name)
Tag.where(name: name).first
end
## Instruction:
Fix Tag to Entry association
## Code After:
class Tag < ActiveRecord::Base
has_many :taggings,
dependent: :destroy
has_many :entries,
through: :taggings,
source: :taggable,
source_type: 'Entry'
validates :name,
presence: true,
uniqueness: true
scope :any,
->(list) {
where(list.map {|name| sanitize_sql(['name = ?', name]) }.join(' OR '))
}
def link
"/tags/#{name}"
end
end
def Tag(name)
Tag.where(name: name).first
end
|
24105a13247aba8509c6a0f6ee74a039bf7f77ae | README.md | README.md |
![travis-badge](https://api.travis-ci.org/arthurmco/ghostwriter.svg?branch=master)
Ghostwriter is a software for those that want a platform for article/blog
editing and management, but don't want a predefined way to display it,
with limited choice for themes, complicated APIs and a difficult configuration
This software gives *you* the control of how the posts are shown, offering
a simple REST API for creating and showing them, along a simple interface
for managing them. We don't offer a way to show them, we let you do it at
your way.
**THIS SOFTWARE IS UNDER DEVELOPMENT**
This software is licensed under the MIT License
## Installing
- Clone this repository
- Create a virtualenv and enter it
- Run `pip install -e . `
- Run `FLASK_APP=ghostwriter flask initdb`. This will create the database and
the needed tables
## Running
To test, simply run flask with the `FLASK_APP` environment variable as `ghostwriter`.
You may want to run ghostwriter in a WSGI environment. Although isn't supported yet, I'd love some testing.
The '/' route contains a test message. In the future, that message will be removable
The default login is 'admin' with password 'admin'
|
[https://travis-ci.org/arthurmco/ghostwriter](![Travis Status Badge](https://api.travis-ci.org/arthurmco/ghostwriter.svg?branch=master))
Ghostwriter is a software for those that want a platform for article/blog
editing and management, but don't want a predefined way to display it,
with limited choice for themes, complicated APIs and a difficult configuration
This software gives *you* the control of how the posts are shown, offering
a simple REST API for creating and showing them, along with a simple interface
for managing them. We don't offer a way to show them, we let you do it at
your way.
**THIS SOFTWARE IS UNDER DEVELOPMENT**
This software is licensed under the MIT License
## Installing
- Clone this repository
- Create a virtualenv and enter it
- Run `pip install -e . `
- Run `FLASK_APP=ghostwriter flask initdb`. This will create the database and
the needed tables
## Running
To test, simply run flask with the `FLASK_APP` environment variable as `ghostwriter`.
You may want to run ghostwriter in a WSGI environment. Although isn't supported yet, I'd love some testing.
The '/' route contains a test message. In the future, that message will be removable
The default login is 'admin' with password 'admin'. You can access the admin interface through the '/admin' entry point
| Add link to Travis page in readme | Add link to Travis page in readme
| Markdown | mit | arthurmco/ghostwriter,arthurmco/ghostwriter | markdown | ## Code Before:
![travis-badge](https://api.travis-ci.org/arthurmco/ghostwriter.svg?branch=master)
Ghostwriter is a software for those that want a platform for article/blog
editing and management, but don't want a predefined way to display it,
with limited choice for themes, complicated APIs and a difficult configuration
This software gives *you* the control of how the posts are shown, offering
a simple REST API for creating and showing them, along a simple interface
for managing them. We don't offer a way to show them, we let you do it at
your way.
**THIS SOFTWARE IS UNDER DEVELOPMENT**
This software is licensed under the MIT License
## Installing
- Clone this repository
- Create a virtualenv and enter it
- Run `pip install -e . `
- Run `FLASK_APP=ghostwriter flask initdb`. This will create the database and
the needed tables
## Running
To test, simply run flask with the `FLASK_APP` environment variable as `ghostwriter`.
You may want to run ghostwriter in a WSGI environment. Although isn't supported yet, I'd love some testing.
The '/' route contains a test message. In the future, that message will be removable
The default login is 'admin' with password 'admin'
## Instruction:
Add link to Travis page in readme
## Code After:
[https://travis-ci.org/arthurmco/ghostwriter](![Travis Status Badge](https://api.travis-ci.org/arthurmco/ghostwriter.svg?branch=master))
Ghostwriter is a software for those that want a platform for article/blog
editing and management, but don't want a predefined way to display it,
with limited choice for themes, complicated APIs and a difficult configuration
This software gives *you* the control of how the posts are shown, offering
a simple REST API for creating and showing them, along with a simple interface
for managing them. We don't offer a way to show them, we let you do it at
your way.
**THIS SOFTWARE IS UNDER DEVELOPMENT**
This software is licensed under the MIT License
## Installing
- Clone this repository
- Create a virtualenv and enter it
- Run `pip install -e . `
- Run `FLASK_APP=ghostwriter flask initdb`. This will create the database and
the needed tables
## Running
To test, simply run flask with the `FLASK_APP` environment variable as `ghostwriter`.
You may want to run ghostwriter in a WSGI environment. Although isn't supported yet, I'd love some testing.
The '/' route contains a test message. In the future, that message will be removable
The default login is 'admin' with password 'admin'. You can access the admin interface through the '/admin' entry point
|
98c9ddd43d7503044ce135680b9a2966d64d8096 | source/_splash.erb | source/_splash.erb | <div class="row">
<div class="col-6 col-lg-6">
<img src="images/foodcritic.png" />
</div>
<div class="col-6 col-lg-6">
<h1>About Foodcritic</h1>
<p>
Foodcritic is a helpful lint tool you can use to check your Chef
cookbooks for common problems.
</p>
<p>
It comes with <b><%= rule_count %></b> built-in rules that identify problems ranging
from simple style inconsistencies to difficult to diagnose issues that will
hurt in production.
</p>
<p>
Various nice people in the Chef community have also written extra rules
for foodcritic that you can install and run. Or write your own!
</p>
<p>
It's easy to get started:
</p>
<p>
<code>$ gem install foodcritic</code><br />
<code>$ foodcritic my_cookbook_dir</code>
</p>
</div>
</div>
| <div class="row">
<div class="col-6 col-lg-6">
<img src="images/foodcritic.png" />
</div>
<div class="col-6 col-lg-6">
<h1>About Foodcritic</h1>
<p>
Foodcritic is a helpful lint tool you can use to check your Chef
cookbooks for common problems.
</p>
<p>
It comes with <b><%= rule_count %></b> built-in rules that identify problems ranging
from simple style inconsistencies to difficult to diagnose issues that will
hurt in production.
</p>
<p>
Various nice people in the Chef community have also written extra rules
for foodcritic that you can install and run. Or write your own!
</p>
<p>
It's easy to get started with Foodcritic in Chef-DK already
</p>
<p>
<code>$ foodcritic my_cookbook_dir</code>
</p>
</div>
</div>
| Update for Foodcritic in ChefDK | Update for Foodcritic in ChefDK
Signed-off-by: Tim Smith <[email protected]>
| HTML+ERB | mit | acrmp/foodcritic-site,Foodcritic/foodcritic-site,acrmp/foodcritic-site,Foodcritic/foodcritic-site | html+erb | ## Code Before:
<div class="row">
<div class="col-6 col-lg-6">
<img src="images/foodcritic.png" />
</div>
<div class="col-6 col-lg-6">
<h1>About Foodcritic</h1>
<p>
Foodcritic is a helpful lint tool you can use to check your Chef
cookbooks for common problems.
</p>
<p>
It comes with <b><%= rule_count %></b> built-in rules that identify problems ranging
from simple style inconsistencies to difficult to diagnose issues that will
hurt in production.
</p>
<p>
Various nice people in the Chef community have also written extra rules
for foodcritic that you can install and run. Or write your own!
</p>
<p>
It's easy to get started:
</p>
<p>
<code>$ gem install foodcritic</code><br />
<code>$ foodcritic my_cookbook_dir</code>
</p>
</div>
</div>
## Instruction:
Update for Foodcritic in ChefDK
Signed-off-by: Tim Smith <[email protected]>
## Code After:
<div class="row">
<div class="col-6 col-lg-6">
<img src="images/foodcritic.png" />
</div>
<div class="col-6 col-lg-6">
<h1>About Foodcritic</h1>
<p>
Foodcritic is a helpful lint tool you can use to check your Chef
cookbooks for common problems.
</p>
<p>
It comes with <b><%= rule_count %></b> built-in rules that identify problems ranging
from simple style inconsistencies to difficult to diagnose issues that will
hurt in production.
</p>
<p>
Various nice people in the Chef community have also written extra rules
for foodcritic that you can install and run. Or write your own!
</p>
<p>
It's easy to get started with Foodcritic in Chef-DK already
</p>
<p>
<code>$ foodcritic my_cookbook_dir</code>
</p>
</div>
</div>
|
1e1c164fb1eae8d873c4aba244d89cb8fe42dbc1 | integration-tests/spec/spec_helper_integ.rb | integration-tests/spec/spec_helper_integ.rb |
require 'ostruct'
require 'capybara/dsl'
require 'jmx4r'
require 'stilts-stomp-client'
require 'driver_helper'
def mbean(name)
url = JMX::JDKHelper.find_local_url(/org.jboss.as.standalone/i)
puts "default jmx url: " + url
url.gsub!(/\d+\.\d+\.\d+\.\d+/, "127.0.0.1")
puts "using jmx url: " + url
JMX::MBean.establish_connection :url => url #:command => /org.jboss.as.standalone/i
JMX::MBean.find_by_name(name)
end
def get_msc_service_state(service_name)
@msc ||= mbean('jboss.msc:type=container,name=jboss-as')
@msc.getServiceStatus(service_name).get('stateName')
end
def set_msc_service_mode(service_name, mode)
@msc ||= mbean('jboss.msc:type=container,name=jboss-as')
@msc.setServiceMode(service_name, mode)
end
def verify_msc_service_state(service_name, state, options={})
options[:timeout] ||= 30 # default to wait 30 seconds
service_state = nil
elapsed_seconds = 0
until service_state == state || elapsed_seconds > options[:timeout] do
service_state = get_msc_service_state(service_name)
sleep(0.5)
end
service_state.should == state
end
|
require 'ostruct'
require 'capybara/dsl'
require 'jmx4r'
require 'stilts-stomp-client'
require 'driver_helper'
def mbean(name)
JMX::MBean.establish_connection :command => /org.jboss.as.standalone/i
JMX::MBean.find_by_name(name)
end
def get_msc_service_state(service_name)
@msc ||= mbean('jboss.msc:type=container,name=jboss-as')
@msc.getServiceStatus(service_name).get('stateName')
end
def set_msc_service_mode(service_name, mode)
@msc ||= mbean('jboss.msc:type=container,name=jboss-as')
@msc.setServiceMode(service_name, mode)
end
def verify_msc_service_state(service_name, state, options={})
options[:timeout] ||= 30 # default to wait 30 seconds
service_state = nil
elapsed_seconds = 0
until service_state == state || elapsed_seconds > options[:timeout] do
service_state = get_msc_service_state(service_name)
sleep(0.5)
end
service_state.should == state
end
| Revert "A total f'ing shot in the dark." | Revert "A total f'ing shot in the dark."
This reverts commit c04c63fd9d0160bb5b22dc475e64eb7e64a196ef.
| Ruby | apache-2.0 | mje113/torquebox,torquebox/torquebox,vaskoz/torquebox,torquebox/torquebox,vaskoz/torquebox,ksw2599/torquebox,torquebox/torquebox-release,torquebox/torquebox-release,mje113/torquebox,mje113/torquebox,ksw2599/torquebox,torquebox/torquebox-release,torquebox/torquebox,samwgoldman/torquebox,vaskoz/torquebox,torquebox/torquebox,samwgoldman/torquebox,mje113/torquebox,samwgoldman/torquebox,ksw2599/torquebox,vaskoz/torquebox,samwgoldman/torquebox,torquebox/torquebox-release,ksw2599/torquebox | ruby | ## Code Before:
require 'ostruct'
require 'capybara/dsl'
require 'jmx4r'
require 'stilts-stomp-client'
require 'driver_helper'
def mbean(name)
url = JMX::JDKHelper.find_local_url(/org.jboss.as.standalone/i)
puts "default jmx url: " + url
url.gsub!(/\d+\.\d+\.\d+\.\d+/, "127.0.0.1")
puts "using jmx url: " + url
JMX::MBean.establish_connection :url => url #:command => /org.jboss.as.standalone/i
JMX::MBean.find_by_name(name)
end
def get_msc_service_state(service_name)
@msc ||= mbean('jboss.msc:type=container,name=jboss-as')
@msc.getServiceStatus(service_name).get('stateName')
end
def set_msc_service_mode(service_name, mode)
@msc ||= mbean('jboss.msc:type=container,name=jboss-as')
@msc.setServiceMode(service_name, mode)
end
def verify_msc_service_state(service_name, state, options={})
options[:timeout] ||= 30 # default to wait 30 seconds
service_state = nil
elapsed_seconds = 0
until service_state == state || elapsed_seconds > options[:timeout] do
service_state = get_msc_service_state(service_name)
sleep(0.5)
end
service_state.should == state
end
## Instruction:
Revert "A total f'ing shot in the dark."
This reverts commit c04c63fd9d0160bb5b22dc475e64eb7e64a196ef.
## Code After:
require 'ostruct'
require 'capybara/dsl'
require 'jmx4r'
require 'stilts-stomp-client'
require 'driver_helper'
def mbean(name)
JMX::MBean.establish_connection :command => /org.jboss.as.standalone/i
JMX::MBean.find_by_name(name)
end
def get_msc_service_state(service_name)
@msc ||= mbean('jboss.msc:type=container,name=jboss-as')
@msc.getServiceStatus(service_name).get('stateName')
end
def set_msc_service_mode(service_name, mode)
@msc ||= mbean('jboss.msc:type=container,name=jboss-as')
@msc.setServiceMode(service_name, mode)
end
def verify_msc_service_state(service_name, state, options={})
options[:timeout] ||= 30 # default to wait 30 seconds
service_state = nil
elapsed_seconds = 0
until service_state == state || elapsed_seconds > options[:timeout] do
service_state = get_msc_service_state(service_name)
sleep(0.5)
end
service_state.should == state
end
|
2e397e534b4ad6732e57b9d7638df7e494339d85 | source/stylesheets/site.scss | source/stylesheets/site.scss | @charset "utf-8";
@import "vendor/normalize";
@import "vendor/animate";
// http://bourbon.io
@import "bourbon/bourbon";
@import "base/base";
@import "neat/neat";
// Partials
@import "partials/layout";
@import "partials/nav-transparent";
@import "partials/nav";
@import "partials/alpha";
@import "partials/earthscape";
@import "partials/lessons";
@import "partials/lessonx";
@import "partials/faith-cometh";
@import "partials/talking-points";
@import "partials/call-to-action";
@import "partials/video";
@import "partials/training";
@import "partials/downloads";
@import "partials/contact";
@import "partials/purchase";
@import "partials/bulk-purchase";
@import "partials/thanks";
@import "partials/callout-box";
@import "partials/breadcrumbs";
@import "partials/copyright-footer";
@import "partials/universal-footer";
@import "partials/404";
//////////////////////// SANDBOX ////////////////////////
//// Try out new code here, then extract to partial. ////
/////////////////////////////////////////////////////////
| @charset "utf-8";
@import "vendor/normalize";
@import "vendor/animate";
// http://bourbon.io
@import "bourbon/bourbon";
@import "base/grid-settings";
@import "neat/neat";
@import "base/base";
// Partials
@import "partials/layout";
@import "partials/nav-transparent";
@import "partials/nav";
@import "partials/alpha";
@import "partials/earthscape";
@import "partials/lessons";
@import "partials/lessonx";
@import "partials/faith-cometh";
@import "partials/talking-points";
@import "partials/call-to-action";
@import "partials/video";
@import "partials/training";
@import "partials/downloads";
@import "partials/contact";
@import "partials/purchase";
@import "partials/bulk-purchase";
@import "partials/thanks";
@import "partials/callout-box";
@import "partials/breadcrumbs";
@import "partials/copyright-footer";
@import "partials/universal-footer";
@import "partials/404";
//////////////////////// SANDBOX ////////////////////////
//// Try out new code here, then extract to partial. ////
/////////////////////////////////////////////////////////
| Update sass imports order for new bourbon/neat/bitters requirements. | Update sass imports order for new bourbon/neat/bitters requirements.
| SCSS | mit | euroteamoutreach/getbiblefirst.com,joshukraine/getbiblefirst.com,joshukraine/getbiblefirst.com,euroteamoutreach/getbiblefirst.com,joshukraine/getbiblefirst.com,joshukraine/getbiblefirst.com,euroteamoutreach/getbiblefirst.com | scss | ## Code Before:
@charset "utf-8";
@import "vendor/normalize";
@import "vendor/animate";
// http://bourbon.io
@import "bourbon/bourbon";
@import "base/base";
@import "neat/neat";
// Partials
@import "partials/layout";
@import "partials/nav-transparent";
@import "partials/nav";
@import "partials/alpha";
@import "partials/earthscape";
@import "partials/lessons";
@import "partials/lessonx";
@import "partials/faith-cometh";
@import "partials/talking-points";
@import "partials/call-to-action";
@import "partials/video";
@import "partials/training";
@import "partials/downloads";
@import "partials/contact";
@import "partials/purchase";
@import "partials/bulk-purchase";
@import "partials/thanks";
@import "partials/callout-box";
@import "partials/breadcrumbs";
@import "partials/copyright-footer";
@import "partials/universal-footer";
@import "partials/404";
//////////////////////// SANDBOX ////////////////////////
//// Try out new code here, then extract to partial. ////
/////////////////////////////////////////////////////////
## Instruction:
Update sass imports order for new bourbon/neat/bitters requirements.
## Code After:
@charset "utf-8";
@import "vendor/normalize";
@import "vendor/animate";
// http://bourbon.io
@import "bourbon/bourbon";
@import "base/grid-settings";
@import "neat/neat";
@import "base/base";
// Partials
@import "partials/layout";
@import "partials/nav-transparent";
@import "partials/nav";
@import "partials/alpha";
@import "partials/earthscape";
@import "partials/lessons";
@import "partials/lessonx";
@import "partials/faith-cometh";
@import "partials/talking-points";
@import "partials/call-to-action";
@import "partials/video";
@import "partials/training";
@import "partials/downloads";
@import "partials/contact";
@import "partials/purchase";
@import "partials/bulk-purchase";
@import "partials/thanks";
@import "partials/callout-box";
@import "partials/breadcrumbs";
@import "partials/copyright-footer";
@import "partials/universal-footer";
@import "partials/404";
//////////////////////// SANDBOX ////////////////////////
//// Try out new code here, then extract to partial. ////
/////////////////////////////////////////////////////////
|
befb57176f20692ce514ebce37282592e6bfe882 | hammerspoon/window-expand.lua | hammerspoon/window-expand.lua | local function pressFn(mods, key)
if key == nil then
key = mods
mods = {}
end
return function() hs.eventtap.keyStroke(mods, key, 1000) end
end
local function remap(mods, key, pressFn)
hs.hotkey.bind(mods, key, pressFn, nil, pressFn)
end
remap({"cmd", "alt", "ctrl", "shift"}, "delete", pressFn({"cmd", "shift"}, "Return"))
hs.hotkey.bind({"cmd", "alt", "ctrl", "shift"}, "Return", function()
local win = hs.window.focusedWindow()
local f = win:frame()
local max = win:screen():frame()
f.x = max.x
f.y = max.y
f.w = max.w
f.h = max.h
win:setFrame(f)
end)
hs.hotkey.bind({"cmd", "alt", "ctrl", "shift"}, "F", function()
local win = hs.window.focusedWindow()
win:toggleFullScreen()
end)
| local function pressFn(mods, key)
if key == nil then
key = mods
mods = {}
end
return function() hs.eventtap.keyStroke(mods, key, 1000) end
end
local function remap(mods, key, pressFn)
hs.hotkey.bind(mods, key, pressFn, nil, pressFn)
end
remap({"cmd", "alt", "ctrl", "shift"}, "delete", pressFn({"cmd", "shift"}, "Return"))
hs.hotkey.bind({"cmd", "alt", "ctrl", "shift"}, "Return", function()
local win = hs.window.focusedWindow()
local f = win:frame()
local max = win:screen():frame()
f.x = max.x
f.y = max.y
f.w = max.w
f.h = max.h
win:setFrame(f)
end)
hs.hotkey.bind({"cmd", "alt", "ctrl", "shift"}, "F", function()
local win = hs.window.focusedWindow()
win:toggleFullScreen()
end)
-- cycle through all windows of the current application
hs.hotkey.bind({"cmd", "alt", "ctrl", "shift"}, "space", function()
local app = hs.application.frontmostApplication()
local current = 1
local next = current
for i, window in pairs(app:allWindows()) do
if app:focusedWindow() == app:allWindows()[i] then
current = i
end
if i > current then
next = i
end
end
app:allWindows()[next]:focus()
end)
| Add hyper+space for window cycling | Add hyper+space for window cycling
| Lua | mit | bsingr/dotfiles,bsingr/dotfiles | lua | ## Code Before:
local function pressFn(mods, key)
if key == nil then
key = mods
mods = {}
end
return function() hs.eventtap.keyStroke(mods, key, 1000) end
end
local function remap(mods, key, pressFn)
hs.hotkey.bind(mods, key, pressFn, nil, pressFn)
end
remap({"cmd", "alt", "ctrl", "shift"}, "delete", pressFn({"cmd", "shift"}, "Return"))
hs.hotkey.bind({"cmd", "alt", "ctrl", "shift"}, "Return", function()
local win = hs.window.focusedWindow()
local f = win:frame()
local max = win:screen():frame()
f.x = max.x
f.y = max.y
f.w = max.w
f.h = max.h
win:setFrame(f)
end)
hs.hotkey.bind({"cmd", "alt", "ctrl", "shift"}, "F", function()
local win = hs.window.focusedWindow()
win:toggleFullScreen()
end)
## Instruction:
Add hyper+space for window cycling
## Code After:
local function pressFn(mods, key)
if key == nil then
key = mods
mods = {}
end
return function() hs.eventtap.keyStroke(mods, key, 1000) end
end
local function remap(mods, key, pressFn)
hs.hotkey.bind(mods, key, pressFn, nil, pressFn)
end
remap({"cmd", "alt", "ctrl", "shift"}, "delete", pressFn({"cmd", "shift"}, "Return"))
hs.hotkey.bind({"cmd", "alt", "ctrl", "shift"}, "Return", function()
local win = hs.window.focusedWindow()
local f = win:frame()
local max = win:screen():frame()
f.x = max.x
f.y = max.y
f.w = max.w
f.h = max.h
win:setFrame(f)
end)
hs.hotkey.bind({"cmd", "alt", "ctrl", "shift"}, "F", function()
local win = hs.window.focusedWindow()
win:toggleFullScreen()
end)
-- cycle through all windows of the current application
hs.hotkey.bind({"cmd", "alt", "ctrl", "shift"}, "space", function()
local app = hs.application.frontmostApplication()
local current = 1
local next = current
for i, window in pairs(app:allWindows()) do
if app:focusedWindow() == app:allWindows()[i] then
current = i
end
if i > current then
next = i
end
end
app:allWindows()[next]:focus()
end)
|
8af7e9a2e63a155289c2840dd0c8950c95f9153b | tests/spec/storm/storm_spec.rb | tests/spec/storm/storm_spec.rb | require 'faraday'
require 'json'
describe 'Apache Storm Mesos release' do
context 'Storm UI is running' do
let(:storm_ui_url) { 'http://10.244.1.22:8080/' }
it 'has an accessible home page' do
status_code = Faraday.head(storm_ui_url).status
expect(status_code).to eq(200)
end
end
context 'Storm Nimbus is running' do
let(:mesos_master_state) { 'http://10.244.1.6:5050/master/state.json' }
it 'is registered with Mesos master' do
raw_state = Faraday.get mesos_master_state
state_json = JSON.parse(raw_state.body)
storm = state_json.fetch('frameworks').find do |framework|
framework.fetch('name') == 'Storm!!!'
end
expect(storm).not_to be_nil
expect(storm['name']).to eql('Storm!!!')
end
xit 'accepts a tuple' do
end
end
end
| require 'spec_helper'
describe 'Apache Storm Mesos release' do
context 'Storm UI is running' do
let(:storm_ui_url) { 'http://10.244.1.22:8080/' }
it 'has an accessible home page' do
status_code = Faraday.head(storm_ui_url).status
expect(status_code).to eq(200)
end
end
context 'Storm Nimbus is running' do
it 'is registered with Mesos master' do
storm = mesos_master_state_json.fetch('frameworks').find do |framework|
framework.fetch('name') == 'Storm!!!'
end
expect(storm).not_to be_nil
expect(storm['name']).to eql('Storm!!!')
end
xit 'accepts a tuple' do
end
end
end
| Move shared Mesos and Bosh logic to spec helper | Move shared Mesos and Bosh logic to spec helper
| Ruby | apache-2.0 | CloudCredo/mesos-boshrelease,CloudCredo/mesos-boshrelease | ruby | ## Code Before:
require 'faraday'
require 'json'
describe 'Apache Storm Mesos release' do
context 'Storm UI is running' do
let(:storm_ui_url) { 'http://10.244.1.22:8080/' }
it 'has an accessible home page' do
status_code = Faraday.head(storm_ui_url).status
expect(status_code).to eq(200)
end
end
context 'Storm Nimbus is running' do
let(:mesos_master_state) { 'http://10.244.1.6:5050/master/state.json' }
it 'is registered with Mesos master' do
raw_state = Faraday.get mesos_master_state
state_json = JSON.parse(raw_state.body)
storm = state_json.fetch('frameworks').find do |framework|
framework.fetch('name') == 'Storm!!!'
end
expect(storm).not_to be_nil
expect(storm['name']).to eql('Storm!!!')
end
xit 'accepts a tuple' do
end
end
end
## Instruction:
Move shared Mesos and Bosh logic to spec helper
## Code After:
require 'spec_helper'
describe 'Apache Storm Mesos release' do
context 'Storm UI is running' do
let(:storm_ui_url) { 'http://10.244.1.22:8080/' }
it 'has an accessible home page' do
status_code = Faraday.head(storm_ui_url).status
expect(status_code).to eq(200)
end
end
context 'Storm Nimbus is running' do
it 'is registered with Mesos master' do
storm = mesos_master_state_json.fetch('frameworks').find do |framework|
framework.fetch('name') == 'Storm!!!'
end
expect(storm).not_to be_nil
expect(storm['name']).to eql('Storm!!!')
end
xit 'accepts a tuple' do
end
end
end
|
fbd0ce006326e2e410084297e64afdfb9052d2a1 | .travis.yml | .travis.yml | language: node_js
node_js:
- "6.9.2"
addons:
firefox: "latest"
before_install:
- ./tools/travis/init-env
- ./tools/travis/install-geckodriver
- ./tools/travis/start-xvfb
- . ./tools/travis/env.sh
- export DISPLAY=$ENV_DISPLAY
- export PATH=$PATH:$ENV_LOCAL_BIN_DIR
install:
- npm install --global gulp-cli
- npm run bootstrap
script:
- gulp unit-test-with-coverage
- gulp dist
- gulp start-server
- gulp acceptance-test
- gulp stop-server
- gulp lint
- gulp docs
after_script:
- gulp publish-coverage
| language: node_js
node_js:
- "6.9.2"
addons:
firefox: "latest"
before_install:
- ./tools/travis/init-env
- ./tools/travis/install-geckodriver
- ./tools/travis/start-xvfb
- . ./tools/travis/env.sh
- export DISPLAY=$ENV_DISPLAY
- export PATH=$PATH:$ENV_LOCAL_BIN_DIR
install:
- npm install --global gulp-cli phantomjs
- npm run bootstrap
script:
- gulp unit-test-with-coverage
- gulp dist
- gulp start-server
- gulp acceptance-test:server
- ENV_WEBDRIVER_BROWSER_NAME=phantomjs gulp acceptance-test:client
- ENV_WEBDRIVER_BROWSER_NAME=firefox gulp acceptance-test:client
- gulp stop-server
- gulp lint
- gulp docs
after_script:
- gulp publish-coverage
| Add PhantomJS to list of browsers to be tested on Travis | Add PhantomJS to list of browsers to be tested on Travis
| YAML | mit | ssoloff/dice-server-js,ssoloff/dice-server-js,ssoloff/dice-server-js,ssoloff/dice-server-js | yaml | ## Code Before:
language: node_js
node_js:
- "6.9.2"
addons:
firefox: "latest"
before_install:
- ./tools/travis/init-env
- ./tools/travis/install-geckodriver
- ./tools/travis/start-xvfb
- . ./tools/travis/env.sh
- export DISPLAY=$ENV_DISPLAY
- export PATH=$PATH:$ENV_LOCAL_BIN_DIR
install:
- npm install --global gulp-cli
- npm run bootstrap
script:
- gulp unit-test-with-coverage
- gulp dist
- gulp start-server
- gulp acceptance-test
- gulp stop-server
- gulp lint
- gulp docs
after_script:
- gulp publish-coverage
## Instruction:
Add PhantomJS to list of browsers to be tested on Travis
## Code After:
language: node_js
node_js:
- "6.9.2"
addons:
firefox: "latest"
before_install:
- ./tools/travis/init-env
- ./tools/travis/install-geckodriver
- ./tools/travis/start-xvfb
- . ./tools/travis/env.sh
- export DISPLAY=$ENV_DISPLAY
- export PATH=$PATH:$ENV_LOCAL_BIN_DIR
install:
- npm install --global gulp-cli phantomjs
- npm run bootstrap
script:
- gulp unit-test-with-coverage
- gulp dist
- gulp start-server
- gulp acceptance-test:server
- ENV_WEBDRIVER_BROWSER_NAME=phantomjs gulp acceptance-test:client
- ENV_WEBDRIVER_BROWSER_NAME=firefox gulp acceptance-test:client
- gulp stop-server
- gulp lint
- gulp docs
after_script:
- gulp publish-coverage
|
edebe37458da391723e3206c63102cbb69606c5b | ideascube/conf/idb_irq_bardarash.py | ideascube/conf/idb_irq_bardarash.py | """Bardarash in Kurdistan, Iraq"""
from .idb_jor_azraq import * # pragma: no flakes
from django.utils.translation import ugettext_lazy as _
USER_FORM_FIELDS = (
('Ideasbox', ['serial', 'box_awareness']),
(_('Personal informations'), ['short_name', 'full_name', 'latin_name', 'birth_year', 'gender', 'country_of_origin_occupation', 'school_level']), # noqa
(_('Family'), ['marital_status', 'family_status', 'children_under_12', 'children_under_18', 'children_above_18']), # noqa
(_('In the camp'), ['camp_entry_date', 'camp_activities', 'current_occupation', 'camp_address']), # noqa
(_('Language skills'), ['ar_level', 'ku_level', 'sdb_level', 'en_level']),
)
MONITORING_ENTRY_EXPORT_FIELDS = ['serial', 'birth_year', 'gender']
ENTRY_ACTIVITY_CHOICES = []
| """Bardarash in Kurdistan, Iraq"""
from .idb_jor_azraq import * # pragma: no flakes
from django.utils.translation import ugettext_lazy as _
USER_FORM_FIELDS = (
('Ideasbox', ['serial', 'box_awareness']),
(_('Personal informations'), ['short_name', 'full_name', 'latin_name', 'birth_year', 'gender', 'country_of_origin_occupation', 'school_level']), # noqa
(_('Family'), ['marital_status', 'family_status', 'children_under_12', 'children_under_18', 'children_above_18']), # noqa
(_('In the camp'), ['camp_entry_date', 'camp_activities', 'current_occupation', 'camp_address']), # noqa
(_('Language skills'), ['ar_level', 'ku_level', 'sdb_level', 'en_level']),
)
MONITORING_ENTRY_EXPORT_FIELDS = ['serial', 'birth_year', 'gender']
ENTRY_ACTIVITY_CHOICES = []
HOME_CARDS = STAFF_HOME_CARDS + [
{
'id': 'blog',
},
{
'id': 'mediacenter',
},
{
'id': 'library',
}
]
| Remove Kalite until Arabic language is available | Remove Kalite until Arabic language is available
| Python | agpl-3.0 | ideascube/ideascube,ideascube/ideascube,ideascube/ideascube,ideascube/ideascube | python | ## Code Before:
"""Bardarash in Kurdistan, Iraq"""
from .idb_jor_azraq import * # pragma: no flakes
from django.utils.translation import ugettext_lazy as _
USER_FORM_FIELDS = (
('Ideasbox', ['serial', 'box_awareness']),
(_('Personal informations'), ['short_name', 'full_name', 'latin_name', 'birth_year', 'gender', 'country_of_origin_occupation', 'school_level']), # noqa
(_('Family'), ['marital_status', 'family_status', 'children_under_12', 'children_under_18', 'children_above_18']), # noqa
(_('In the camp'), ['camp_entry_date', 'camp_activities', 'current_occupation', 'camp_address']), # noqa
(_('Language skills'), ['ar_level', 'ku_level', 'sdb_level', 'en_level']),
)
MONITORING_ENTRY_EXPORT_FIELDS = ['serial', 'birth_year', 'gender']
ENTRY_ACTIVITY_CHOICES = []
## Instruction:
Remove Kalite until Arabic language is available
## Code After:
"""Bardarash in Kurdistan, Iraq"""
from .idb_jor_azraq import * # pragma: no flakes
from django.utils.translation import ugettext_lazy as _
USER_FORM_FIELDS = (
('Ideasbox', ['serial', 'box_awareness']),
(_('Personal informations'), ['short_name', 'full_name', 'latin_name', 'birth_year', 'gender', 'country_of_origin_occupation', 'school_level']), # noqa
(_('Family'), ['marital_status', 'family_status', 'children_under_12', 'children_under_18', 'children_above_18']), # noqa
(_('In the camp'), ['camp_entry_date', 'camp_activities', 'current_occupation', 'camp_address']), # noqa
(_('Language skills'), ['ar_level', 'ku_level', 'sdb_level', 'en_level']),
)
MONITORING_ENTRY_EXPORT_FIELDS = ['serial', 'birth_year', 'gender']
ENTRY_ACTIVITY_CHOICES = []
HOME_CARDS = STAFF_HOME_CARDS + [
{
'id': 'blog',
},
{
'id': 'mediacenter',
},
{
'id': 'library',
}
]
|
9eafe8385c8134d60241ef653a9342884c04d3e3 | tox.ini | tox.ini | [tox]
envlist = py26, py27, py33, py34, py35
[testenv]
setenv =
PYTHONPATH = {toxinidir}:{toxinidir}/regressors
commands =
python setup.py test
deps =
pip==7.0.2
-r{toxinidir}/requirements_dev.txt
[testenv:py26]
; Python 2.6 tests need to be run with nose to work with unittest2
commands =
nosetests tests/test_regressors.py
[testenv:py35]
; Output coverage report in last environment tested
commands =
coverage erase
coverage run --source regressors setup.py test
coverage combine
coverage report
[flake8]
max-line-length = 80
| [tox]
envlist = py26, py27, py33, py34, py35
[testenv]
setenv =
PYTHONPATH = {toxinidir}:{toxinidir}/regressors
commands =
python setup.py test
deps =
pip==7.0.2
-r{toxinidir}/requirements_dev.txt
[testenv:py26]
; Python 2.6 tests need to be run with nose to work with unittest2
commands =
nosetests tests/test_regressors.py
[testenv:py33]
; Need to discover tests with unittest to properly pick them up in Python 3
commands =
coverage erase
coverage run --source regressors -m unittest discover
[testenv:py34]
; Need to discover tests with unittest to properly pick them up in Python 3
commands =
coverage erase
coverage run --source regressors -m unittest discover
[testenv:py35]
; Output coverage report in last environment tested
commands =
coverage erase
coverage run --source regressors -m unittest discover
coverage combine
coverage report
[flake8]
max-line-length = 80
| Fix tests not getting picked up in Python 3 | Fix tests not getting picked up in Python 3
| INI | isc | nsh87/regressors | ini | ## Code Before:
[tox]
envlist = py26, py27, py33, py34, py35
[testenv]
setenv =
PYTHONPATH = {toxinidir}:{toxinidir}/regressors
commands =
python setup.py test
deps =
pip==7.0.2
-r{toxinidir}/requirements_dev.txt
[testenv:py26]
; Python 2.6 tests need to be run with nose to work with unittest2
commands =
nosetests tests/test_regressors.py
[testenv:py35]
; Output coverage report in last environment tested
commands =
coverage erase
coverage run --source regressors setup.py test
coverage combine
coverage report
[flake8]
max-line-length = 80
## Instruction:
Fix tests not getting picked up in Python 3
## Code After:
[tox]
envlist = py26, py27, py33, py34, py35
[testenv]
setenv =
PYTHONPATH = {toxinidir}:{toxinidir}/regressors
commands =
python setup.py test
deps =
pip==7.0.2
-r{toxinidir}/requirements_dev.txt
[testenv:py26]
; Python 2.6 tests need to be run with nose to work with unittest2
commands =
nosetests tests/test_regressors.py
[testenv:py33]
; Need to discover tests with unittest to properly pick them up in Python 3
commands =
coverage erase
coverage run --source regressors -m unittest discover
[testenv:py34]
; Need to discover tests with unittest to properly pick them up in Python 3
commands =
coverage erase
coverage run --source regressors -m unittest discover
[testenv:py35]
; Output coverage report in last environment tested
commands =
coverage erase
coverage run --source regressors -m unittest discover
coverage combine
coverage report
[flake8]
max-line-length = 80
|
2ed094d6902272f95dfd556c2244b13f9fbbc76b | build.sh | build.sh | cd $(dirname $0)
file=$(perl -e '
use JSON::PP; local $/;
$info = decode_json <>;
print $info->{name}."_".$info->{version}' \
src/info.json)
mkdir -p build
rm -rf build/*
git archive HEAD:src --prefix=$file/ -9 -o build/$file.zip
| cd $(dirname $0)
file=$(perl -e '
use JSON::PP; local $/;
$info = decode_json <>;
print $info->{name}."_".$info->{version}' \
src/info.json)
mkdir -p build
rm -rf build/*
git archive HEAD:src --prefix=$file/ -9 -o build/$file.zip
# Include LICENSE.md with release
cd build
mkdir -p $file
cp ../LICENSE.md $file/LICENSE.md
zip -q -9 $file.zip $file/LICENSE.md
rm -rf $file
| Add LICENSE.md to built .zip | Add LICENSE.md to built .zip
| Shell | mit | jpas/factorio-expanded-tools | shell | ## Code Before:
cd $(dirname $0)
file=$(perl -e '
use JSON::PP; local $/;
$info = decode_json <>;
print $info->{name}."_".$info->{version}' \
src/info.json)
mkdir -p build
rm -rf build/*
git archive HEAD:src --prefix=$file/ -9 -o build/$file.zip
## Instruction:
Add LICENSE.md to built .zip
## Code After:
cd $(dirname $0)
file=$(perl -e '
use JSON::PP; local $/;
$info = decode_json <>;
print $info->{name}."_".$info->{version}' \
src/info.json)
mkdir -p build
rm -rf build/*
git archive HEAD:src --prefix=$file/ -9 -o build/$file.zip
# Include LICENSE.md with release
cd build
mkdir -p $file
cp ../LICENSE.md $file/LICENSE.md
zip -q -9 $file.zip $file/LICENSE.md
rm -rf $file
|
dcf0f8aaf8a45ea5be82b3de8b57e9f9580feab1 | spec/support/backend.rb | spec/support/backend.rb | shared_examples "a Que backend" do
it "should be able to drop and create the jobs table" do
DB.table_exists?(:que_jobs).should be true
Que.drop!
DB.table_exists?(:que_jobs).should be false
Que.create!
DB.table_exists?(:que_jobs).should be true
end
it "should be able to clear the jobs table" do
DB[:que_jobs].insert :type => "Que::Job"
DB[:que_jobs].count.should be 1
Que.clear!
DB[:que_jobs].count.should be 0
end
end
| shared_examples "a Que backend" do
it "should be able to drop and create the jobs table" do
DB.table_exists?(:que_jobs).should be true
Que.drop!
DB.table_exists?(:que_jobs).should be false
Que.create!
DB.table_exists?(:que_jobs).should be true
end
it "should be able to clear the jobs table" do
DB[:que_jobs].insert :type => "Que::Job"
DB[:que_jobs].count.should be 1
Que.clear!
DB[:que_jobs].count.should be 0
end
describe "when queueing jobs" do
it "should be able to queue a job with arguments" do
pending
end
it "should be able to queue a job with a specific time to run" do
pending
end
it "should be able to queue a job with a specific priority" do
pending
end
it "should be able to queue a job with queueing options in addition to argument options" do
pending
end
it "should respect a default (but overridable) priority for the job class" do
pending
end
it "should respect a default (but overridable) run_at for the job class" do
pending
end
it "should raise an error if given arguments that can't convert to and from JSON unambiguously" do
pending
end
end
end
| Add some pending specs on queueing jobs. | Add some pending specs on queueing jobs.
| Ruby | mit | chanks/que,zloydadka/que,hardbap/que,gocardless/que,goddamnhippie/que,joevandyk/que,brandur/que,gocardless/que,godfat/que,heroku/que | ruby | ## Code Before:
shared_examples "a Que backend" do
it "should be able to drop and create the jobs table" do
DB.table_exists?(:que_jobs).should be true
Que.drop!
DB.table_exists?(:que_jobs).should be false
Que.create!
DB.table_exists?(:que_jobs).should be true
end
it "should be able to clear the jobs table" do
DB[:que_jobs].insert :type => "Que::Job"
DB[:que_jobs].count.should be 1
Que.clear!
DB[:que_jobs].count.should be 0
end
end
## Instruction:
Add some pending specs on queueing jobs.
## Code After:
shared_examples "a Que backend" do
it "should be able to drop and create the jobs table" do
DB.table_exists?(:que_jobs).should be true
Que.drop!
DB.table_exists?(:que_jobs).should be false
Que.create!
DB.table_exists?(:que_jobs).should be true
end
it "should be able to clear the jobs table" do
DB[:que_jobs].insert :type => "Que::Job"
DB[:que_jobs].count.should be 1
Que.clear!
DB[:que_jobs].count.should be 0
end
describe "when queueing jobs" do
it "should be able to queue a job with arguments" do
pending
end
it "should be able to queue a job with a specific time to run" do
pending
end
it "should be able to queue a job with a specific priority" do
pending
end
it "should be able to queue a job with queueing options in addition to argument options" do
pending
end
it "should respect a default (but overridable) priority for the job class" do
pending
end
it "should respect a default (but overridable) run_at for the job class" do
pending
end
it "should raise an error if given arguments that can't convert to and from JSON unambiguously" do
pending
end
end
end
|
c8918e5ee06c3f7ec84124ebfccddb738dca2bfb | test/python_api/default-constructor/sb_communication.py | test/python_api/default-constructor/sb_communication.py |
import sys
import lldb
def fuzz_obj(obj):
broadcaster = obj.GetBroadcaster()
# Do fuzz testing on the broadcaster obj, it should not crash lldb.
import sb_broadcaster
sb_broadcaster.fuzz_obj(broadcaster)
obj.AdoptFileDesriptor(0, False)
obj.AdoptFileDesriptor(1, False)
obj.AdoptFileDesriptor(2, False)
obj.Connect("file:/tmp/myfile")
obj.Disconnect()
obj.IsConnected()
obj.GetCloseOnEOF()
obj.SetCloseOnEOF(True)
obj.SetCloseOnEOF(False)
#obj.Write(None, sys.maxint, None)
#obj.Read(None, sys.maxint, 0xffffffff, None)
obj.ReadThreadStart()
obj.ReadThreadStop()
obj.ReadThreadIsRunning()
obj.SetReadThreadBytesReceivedCallback(None, None)
|
import sys
import lldb
def fuzz_obj(obj):
broadcaster = obj.GetBroadcaster()
# Do fuzz testing on the broadcaster obj, it should not crash lldb.
import sb_broadcaster
sb_broadcaster.fuzz_obj(broadcaster)
obj.AdoptFileDesriptor(0, False)
obj.AdoptFileDesriptor(1, False)
obj.AdoptFileDesriptor(2, False)
obj.Connect("file:/tmp/myfile")
obj.Connect(None)
obj.Disconnect()
obj.IsConnected()
obj.GetCloseOnEOF()
obj.SetCloseOnEOF(True)
obj.SetCloseOnEOF(False)
#obj.Write(None, sys.maxint, None)
#obj.Read(None, sys.maxint, 0xffffffff, None)
obj.ReadThreadStart()
obj.ReadThreadStop()
obj.ReadThreadIsRunning()
obj.SetReadThreadBytesReceivedCallback(None, None)
| Add a fuzz call for SBCommunication: obj.connect(None). | Add a fuzz call for SBCommunication: obj.connect(None).
git-svn-id: 4c4cc70b1ef44ba2b7963015e681894188cea27e@146912 91177308-0d34-0410-b5e6-96231b3b80d8
| Python | apache-2.0 | llvm-mirror/lldb,apple/swift-lldb,llvm-mirror/lldb,apple/swift-lldb,llvm-mirror/lldb,llvm-mirror/lldb,apple/swift-lldb,apple/swift-lldb,apple/swift-lldb,llvm-mirror/lldb,apple/swift-lldb | python | ## Code Before:
import sys
import lldb
def fuzz_obj(obj):
broadcaster = obj.GetBroadcaster()
# Do fuzz testing on the broadcaster obj, it should not crash lldb.
import sb_broadcaster
sb_broadcaster.fuzz_obj(broadcaster)
obj.AdoptFileDesriptor(0, False)
obj.AdoptFileDesriptor(1, False)
obj.AdoptFileDesriptor(2, False)
obj.Connect("file:/tmp/myfile")
obj.Disconnect()
obj.IsConnected()
obj.GetCloseOnEOF()
obj.SetCloseOnEOF(True)
obj.SetCloseOnEOF(False)
#obj.Write(None, sys.maxint, None)
#obj.Read(None, sys.maxint, 0xffffffff, None)
obj.ReadThreadStart()
obj.ReadThreadStop()
obj.ReadThreadIsRunning()
obj.SetReadThreadBytesReceivedCallback(None, None)
## Instruction:
Add a fuzz call for SBCommunication: obj.connect(None).
git-svn-id: 4c4cc70b1ef44ba2b7963015e681894188cea27e@146912 91177308-0d34-0410-b5e6-96231b3b80d8
## Code After:
import sys
import lldb
def fuzz_obj(obj):
broadcaster = obj.GetBroadcaster()
# Do fuzz testing on the broadcaster obj, it should not crash lldb.
import sb_broadcaster
sb_broadcaster.fuzz_obj(broadcaster)
obj.AdoptFileDesriptor(0, False)
obj.AdoptFileDesriptor(1, False)
obj.AdoptFileDesriptor(2, False)
obj.Connect("file:/tmp/myfile")
obj.Connect(None)
obj.Disconnect()
obj.IsConnected()
obj.GetCloseOnEOF()
obj.SetCloseOnEOF(True)
obj.SetCloseOnEOF(False)
#obj.Write(None, sys.maxint, None)
#obj.Read(None, sys.maxint, 0xffffffff, None)
obj.ReadThreadStart()
obj.ReadThreadStop()
obj.ReadThreadIsRunning()
obj.SetReadThreadBytesReceivedCallback(None, None)
|
ddef37f7e6fc75b314fdf64ec74929f3e305390b | developer/js/tests/test-compile-trie.ts | developer/js/tests/test-compile-trie.ts | import LexicalModelCompiler from '../';
import {assert} from 'chai';
import 'mocha';
import {makePathToFixture, compileModelSourceCode} from './helpers';
describe('LexicalModelCompiler', function () {
describe('#generateLexicalModelCode', function () {
const MODEL_ID = 'example.qaa.trivial';
const PATH = makePathToFixture(MODEL_ID);
it('should compile a trivial word list', function () {
let compiler = new LexicalModelCompiler;
let code = compiler.generateLexicalModelCode(MODEL_ID, {
format: 'trie-1.0',
sources: ['wordlist.tsv']
}, PATH) as string;
let result = compileModelSourceCode(code);
assert.isFalse(result.hasSyntaxError);
assert.isNotNull(result.exportedModel);
assert.equal(result.modelConstructorName, 'TrieModel');
// Sanity check: the word list has three total unweighted words, with a
// total weight of 3!
assert.match(code, /\btotalWeight\b["']?:\s*3\b/);
});
})
});
| import LexicalModelCompiler from '../';
import {assert} from 'chai';
import 'mocha';
import {makePathToFixture, compileModelSourceCode} from './helpers';
describe('LexicalModelCompiler', function () {
describe('#generateLexicalModelCode', function () {
it('should compile a trivial word list', function () {
const MODEL_ID = 'example.qaa.trivial';
const PATH = makePathToFixture(MODEL_ID);
let compiler = new LexicalModelCompiler;
let code = compiler.generateLexicalModelCode(MODEL_ID, {
format: 'trie-1.0',
sources: ['wordlist.tsv']
}, PATH) as string;
let result = compileModelSourceCode(code);
assert.isFalse(result.hasSyntaxError);
assert.isNotNull(result.exportedModel);
assert.equal(result.modelConstructorName, 'TrieModel');
// Sanity check: the word list has three total unweighted words, with a
// total weight of 3!
assert.match(code, /\btotalWeight\b["']?:\s*3\b/);
});
it('should compile a word list exported by Microsoft Excel', function () {
const MODEL_ID = 'example.qaa.utf16le';
const PATH = makePathToFixture(MODEL_ID);
let compiler = new LexicalModelCompiler;
let code = compiler.generateLexicalModelCode(MODEL_ID, {
format: 'trie-1.0',
sources: ['wordlist.txt']
}, PATH) as string;
let result = compileModelSourceCode(code);
assert.isFalse(result.hasSyntaxError);
assert.isNotNull(result.exportedModel);
assert.equal(result.modelConstructorName, 'TrieModel');
// Sanity check: the word list has three total unweighted words, with a
// total weight of 44,103!
assert.match(code, /\btotalWeight\b["']?:\s*44103\b/);
});
})
});
| Test a UTF-16LE file compiles properly. | Test a UTF-16LE file compiles properly.
| TypeScript | apache-2.0 | tavultesoft/keymanweb,tavultesoft/keymanweb | typescript | ## Code Before:
import LexicalModelCompiler from '../';
import {assert} from 'chai';
import 'mocha';
import {makePathToFixture, compileModelSourceCode} from './helpers';
describe('LexicalModelCompiler', function () {
describe('#generateLexicalModelCode', function () {
const MODEL_ID = 'example.qaa.trivial';
const PATH = makePathToFixture(MODEL_ID);
it('should compile a trivial word list', function () {
let compiler = new LexicalModelCompiler;
let code = compiler.generateLexicalModelCode(MODEL_ID, {
format: 'trie-1.0',
sources: ['wordlist.tsv']
}, PATH) as string;
let result = compileModelSourceCode(code);
assert.isFalse(result.hasSyntaxError);
assert.isNotNull(result.exportedModel);
assert.equal(result.modelConstructorName, 'TrieModel');
// Sanity check: the word list has three total unweighted words, with a
// total weight of 3!
assert.match(code, /\btotalWeight\b["']?:\s*3\b/);
});
})
});
## Instruction:
Test a UTF-16LE file compiles properly.
## Code After:
import LexicalModelCompiler from '../';
import {assert} from 'chai';
import 'mocha';
import {makePathToFixture, compileModelSourceCode} from './helpers';
describe('LexicalModelCompiler', function () {
describe('#generateLexicalModelCode', function () {
it('should compile a trivial word list', function () {
const MODEL_ID = 'example.qaa.trivial';
const PATH = makePathToFixture(MODEL_ID);
let compiler = new LexicalModelCompiler;
let code = compiler.generateLexicalModelCode(MODEL_ID, {
format: 'trie-1.0',
sources: ['wordlist.tsv']
}, PATH) as string;
let result = compileModelSourceCode(code);
assert.isFalse(result.hasSyntaxError);
assert.isNotNull(result.exportedModel);
assert.equal(result.modelConstructorName, 'TrieModel');
// Sanity check: the word list has three total unweighted words, with a
// total weight of 3!
assert.match(code, /\btotalWeight\b["']?:\s*3\b/);
});
it('should compile a word list exported by Microsoft Excel', function () {
const MODEL_ID = 'example.qaa.utf16le';
const PATH = makePathToFixture(MODEL_ID);
let compiler = new LexicalModelCompiler;
let code = compiler.generateLexicalModelCode(MODEL_ID, {
format: 'trie-1.0',
sources: ['wordlist.txt']
}, PATH) as string;
let result = compileModelSourceCode(code);
assert.isFalse(result.hasSyntaxError);
assert.isNotNull(result.exportedModel);
assert.equal(result.modelConstructorName, 'TrieModel');
// Sanity check: the word list has three total unweighted words, with a
// total weight of 44,103!
assert.match(code, /\btotalWeight\b["']?:\s*44103\b/);
});
})
});
|
416bdc29aa6cf1ab764623f5f226e96e4b0ae081 | file-processing/ignore-script-comments/ignore-script-comments-commands.sh | file-processing/ignore-script-comments/ignore-script-comments-commands.sh | sed '/^[ ]*#/d' "${sourceFile}"
sed '/^[ ]*#/d' < "${sourceFile}"
grep -v '^[ ]*#' "${sourceFile}"
grep -v '^[ ]*#' < "${sourceFile}"
egrep -v '^[ ]*#' "${sourceFile}"
egrep -v '^[ ]*#' < "${sourceFile}"
awk '$0 !~ /^[ ]*#/ { print $0 }' "${sourceFile}"
awk '$0 !~ /^[ ]*#/ { print $0 }' < "${sourceFile}"
awk '{ if ( $0 !~ "^[ ]*#" ) { print $0 } }' "${sourceFile}"
awk '{ if ( $0 !~ "^[ ]*#" ) { print $0 } }' < "${sourceFile}"
| sed '/^[ ]*#/d' "${sourceFile}"
sed '/^[ ]*#/d' < "${sourceFile}"
grep -v '^[ ]*#' "${sourceFile}"
grep -v '^[ ]*#' < "${sourceFile}"
grep -v '^\\s*#' "${sourceFile}"
grep -v '^\\s*#' < "${sourceFile}"
egrep -v '^[ ]*#' "${sourceFile}"
egrep -v '^[ ]*#' < "${sourceFile}"
egrep -v '^\\s*#' "${sourceFile}"
egrep -v '^\\s*#' < "${sourceFile}"
awk '$0 !~ /^[ ]*#/ { print $0 }' "${sourceFile}"
awk '$0 !~ /^[ ]*#/ { print $0 }' < "${sourceFile}"
awk '{ if ( $0 !~ "^[ ]*#" ) { print $0 } }' "${sourceFile}"
awk '{ if ( $0 !~ "^[ ]*#" ) { print $0 } }' < "${sourceFile}"
| Add \s for space/tab variant in grep/egrep | Add \s for space/tab variant in grep/egrep
| Shell | mit | Sylvain-Bugat/linux-commands-reference,Sylvain-Bugat/linux-commands-reference | shell | ## Code Before:
sed '/^[ ]*#/d' "${sourceFile}"
sed '/^[ ]*#/d' < "${sourceFile}"
grep -v '^[ ]*#' "${sourceFile}"
grep -v '^[ ]*#' < "${sourceFile}"
egrep -v '^[ ]*#' "${sourceFile}"
egrep -v '^[ ]*#' < "${sourceFile}"
awk '$0 !~ /^[ ]*#/ { print $0 }' "${sourceFile}"
awk '$0 !~ /^[ ]*#/ { print $0 }' < "${sourceFile}"
awk '{ if ( $0 !~ "^[ ]*#" ) { print $0 } }' "${sourceFile}"
awk '{ if ( $0 !~ "^[ ]*#" ) { print $0 } }' < "${sourceFile}"
## Instruction:
Add \s for space/tab variant in grep/egrep
## Code After:
sed '/^[ ]*#/d' "${sourceFile}"
sed '/^[ ]*#/d' < "${sourceFile}"
grep -v '^[ ]*#' "${sourceFile}"
grep -v '^[ ]*#' < "${sourceFile}"
grep -v '^\\s*#' "${sourceFile}"
grep -v '^\\s*#' < "${sourceFile}"
egrep -v '^[ ]*#' "${sourceFile}"
egrep -v '^[ ]*#' < "${sourceFile}"
egrep -v '^\\s*#' "${sourceFile}"
egrep -v '^\\s*#' < "${sourceFile}"
awk '$0 !~ /^[ ]*#/ { print $0 }' "${sourceFile}"
awk '$0 !~ /^[ ]*#/ { print $0 }' < "${sourceFile}"
awk '{ if ( $0 !~ "^[ ]*#" ) { print $0 } }' "${sourceFile}"
awk '{ if ( $0 !~ "^[ ]*#" ) { print $0 } }' < "${sourceFile}"
|
b0f7a0474f84b9ce7652c93d8e7821075f20b2df | config/deploy.rb | config/deploy.rb | require "bundler/capistrano"
set :deploy_via, :remote_cache
set :application, "devcon"
set :repository, "git://github.com/devcon-ph/devcon.git"
set :deploy_to, "/home/deploy/apps/devcon-release/"
set :scm, :git
set :branch, "release"
default_run_options[:pty] = true
default_run_options[:shell] = "/bin/bash --login"
server "devcon.ph", :app, :web, :db, :primary => true
set :user, "deploy"
set :use_sudo, false
depend :remote, :gem, "bundler"
set :rails_env, "production"
# for carrierwave
set :shared_children, shared_children + %w{public/uploads}
before "deploy:finalize_update", :copy_production_database_configuration, :replace_secret_token
task :copy_production_database_configuration do
run "cp #{shared_path}/config/database.yml #{release_path}/config/database.yml"
end
task :replace_secret_token do
run "cp #{shared_path}/config/secret_token.rb #{release_path}/config/initializers/secret_token.rb"
end
after "deploy:update", "deploy:cleanup", "deploy:migrate"
require "capistrano-unicorn"
after 'deploy:restart', 'unicorn:reload'
| require "bundler/capistrano"
set :deploy_via, :remote_cache
set :application, "devcon"
set :repository, "git://github.com/devcon-ph/devcon.git"
set :deploy_to, "/home/deploy/apps/devcon-release/"
set :scm, :git
set :branch, "release"
default_run_options[:pty] = true
default_run_options[:shell] = "/bin/bash --login"
server "devcon.ph", :app, :web, :db, :primary => true
set :user, "deploy"
set :use_sudo, false
depend :remote, :gem, "bundler"
set :rails_env, "production"
# for carrierwave
set :shared_children, shared_children + %w{public/uploads tmp/sockets}
before "deploy:finalize_update", :copy_production_database_configuration, :replace_secret_token
task :copy_production_database_configuration do
run "cp #{shared_path}/config/database.yml #{release_path}/config/database.yml"
end
task :replace_secret_token do
run "cp #{shared_path}/config/secret_token.rb #{release_path}/config/initializers/secret_token.rb"
end
after "deploy:update", "deploy:cleanup", "deploy:migrate"
require "capistrano-unicorn"
after 'deploy:restart', 'unicorn:reload'
| Fix missing shared sockets folder | Fix missing shared sockets folder [ci skip]
| Ruby | mit | devcon-ph/devcon,devcon-ph/devcon,devcon-ph/devcon | ruby | ## Code Before:
require "bundler/capistrano"
set :deploy_via, :remote_cache
set :application, "devcon"
set :repository, "git://github.com/devcon-ph/devcon.git"
set :deploy_to, "/home/deploy/apps/devcon-release/"
set :scm, :git
set :branch, "release"
default_run_options[:pty] = true
default_run_options[:shell] = "/bin/bash --login"
server "devcon.ph", :app, :web, :db, :primary => true
set :user, "deploy"
set :use_sudo, false
depend :remote, :gem, "bundler"
set :rails_env, "production"
# for carrierwave
set :shared_children, shared_children + %w{public/uploads}
before "deploy:finalize_update", :copy_production_database_configuration, :replace_secret_token
task :copy_production_database_configuration do
run "cp #{shared_path}/config/database.yml #{release_path}/config/database.yml"
end
task :replace_secret_token do
run "cp #{shared_path}/config/secret_token.rb #{release_path}/config/initializers/secret_token.rb"
end
after "deploy:update", "deploy:cleanup", "deploy:migrate"
require "capistrano-unicorn"
after 'deploy:restart', 'unicorn:reload'
## Instruction:
Fix missing shared sockets folder [ci skip]
## Code After:
require "bundler/capistrano"
set :deploy_via, :remote_cache
set :application, "devcon"
set :repository, "git://github.com/devcon-ph/devcon.git"
set :deploy_to, "/home/deploy/apps/devcon-release/"
set :scm, :git
set :branch, "release"
default_run_options[:pty] = true
default_run_options[:shell] = "/bin/bash --login"
server "devcon.ph", :app, :web, :db, :primary => true
set :user, "deploy"
set :use_sudo, false
depend :remote, :gem, "bundler"
set :rails_env, "production"
# for carrierwave
set :shared_children, shared_children + %w{public/uploads tmp/sockets}
before "deploy:finalize_update", :copy_production_database_configuration, :replace_secret_token
task :copy_production_database_configuration do
run "cp #{shared_path}/config/database.yml #{release_path}/config/database.yml"
end
task :replace_secret_token do
run "cp #{shared_path}/config/secret_token.rb #{release_path}/config/initializers/secret_token.rb"
end
after "deploy:update", "deploy:cleanup", "deploy:migrate"
require "capistrano-unicorn"
after 'deploy:restart', 'unicorn:reload'
|
27375ed871235c8f03a6f641201eea8400451bac | .github/workflows/main.yml | .github/workflows/main.yml | name: main
on:
push:
pull_request:
jobs:
build:
runs-on: windows-latest
strategy:
fail-fast: false
matrix:
include: [
{ msystem: MINGW64, arch: x86_64 },
{ msystem: MINGW32, arch: i686 }
]
defaults:
run:
shell: msys2 {0}
steps:
- uses: actions/checkout@v2
with:
fetch-depth: 0
- uses: msys2/setup-msys2@v2
with:
msystem: ${{ matrix.msystem }}
install: git base-devel mingw-w64-${{ matrix.arch }}-toolchain
update: true
- name: CI-Build
run: MINGW_INSTALLS=${{ matrix.msystem }} ./ci-build.sh
- name: "Upload binaries"
uses: actions/upload-artifact@v2
with:
name: ${{ matrix.msystem }}-packages
path: artifacts/*.pkg.tar.*
- name: "Upload sources"
uses: actions/upload-artifact@v2
with:
name: sources
path: artifacts/*.src.tar.*
| name: main
on:
push:
pull_request:
jobs:
build:
runs-on: windows-latest
strategy:
fail-fast: false
matrix:
include: [
{ msystem: MINGW64, arch: x86_64 },
{ msystem: MINGW32, arch: i686 }
]
steps:
- uses: actions/checkout@v2
with:
path: temp
fetch-depth: 0
- uses: msys2/setup-msys2@v2
with:
msystem: ${{ matrix.msystem }}
install: git base-devel mingw-w64-${{ matrix.arch }}-toolchain
update: true
- name: Move Checkout
run: |
Copy-Item -Path ".\temp" -Destination "C:\_" -Recurse
- name: CI-Build
shell: msys2 {0}
run: |
cd /C/_
MINGW_INSTALLS=${{ matrix.msystem }} ./ci-build.sh
- name: "Upload binaries"
uses: actions/upload-artifact@v2
with:
name: ${{ matrix.msystem }}-packages
path: artifacts/*.pkg.tar.*
- name: "Upload sources"
uses: actions/upload-artifact@v2
with:
name: sources
path: artifacts/*.src.tar.*
| Move the checkout to C:/_ | CI: Move the checkout to C:/_
To (a) get more space and (b) reduce the path lengths
| YAML | bsd-3-clause | Ede123/MINGW-packages,Optiligence/MINGW-packages,JPeterMugaas/MINGW-packages,JPeterMugaas/MINGW-packages,stahta01/MINGW-packages,oroppas/MINGW-packages,lhmouse/MINGW-packages,mati865/MINGW-packages,Chocobo1/MINGW-packages,stahta01/MINGW-packages,oroppas/MINGW-packages,fabiangreffrath/MINGW-packages,RickLeaf/MINGW-packages,RickLeaf/MINGW-packages,nazar554/MINGW-packages,Alexpux/MINGW-packages,peterbud/MINGW-packages,jtanx/MINGW-packages,stahta01/MINGW-packages,bjlemke/MINGW-packages,Ede123/MINGW-packages,JPeterMugaas/MINGW-packages,Alexpux/MINGW-packages,lhmouse/MINGW-packages,Ede123/MINGW-packages,oroppas/MINGW-packages,nazar554/MINGW-packages,Chocobo1/MINGW-packages,JPeterMugaas/MINGW-packages,oscarfv/MINGW-packages,fabiangreffrath/MINGW-packages,oscarfv/MINGW-packages,peterbud/MINGW-packages,InitialForce/MINGW-packages,Ede123/MINGW-packages,Alexpux/MINGW-packages,jtanx/MINGW-packages,stahta01/MINGW-packages,Optiligence/MINGW-packages,fabiangreffrath/MINGW-packages,bjlemke/MINGW-packages,stahta01/MINGW-packages,lhmouse/MINGW-packages,jtanx/MINGW-packages,JPeterMugaas/MINGW-packages,bjlemke/MINGW-packages,oscarfv/MINGW-packages,Alexpux/MINGW-packages,lhmouse/MINGW-packages,Optiligence/MINGW-packages,InitialForce/MINGW-packages,RickLeaf/MINGW-packages,mati865/MINGW-packages,oscarfv/MINGW-packages,JPeterMugaas/MINGW-packages,oscarfv/MINGW-packages,nazar554/MINGW-packages,oroppas/MINGW-packages,stahta01/MINGW-packages,oroppas/MINGW-packages,InitialForce/MINGW-packages,mati865/MINGW-packages,oscarfv/MINGW-packages,Optiligence/MINGW-packages,peterbud/MINGW-packages,RickLeaf/MINGW-packages,mati865/MINGW-packages,oroppas/MINGW-packages,fabiangreffrath/MINGW-packages,mati865/MINGW-packages,nazar554/MINGW-packages,Alexpux/MINGW-packages,bjlemke/MINGW-packages,lhmouse/MINGW-packages,fabiangreffrath/MINGW-packages,peterbud/MINGW-packages,fabiangreffrath/MINGW-packages,Chocobo1/MINGW-packages,lhmouse/MINGW-packages,jtanx/MINGW-packages,Alexpux/MINGW-packages | yaml | ## Code Before:
name: main
on:
push:
pull_request:
jobs:
build:
runs-on: windows-latest
strategy:
fail-fast: false
matrix:
include: [
{ msystem: MINGW64, arch: x86_64 },
{ msystem: MINGW32, arch: i686 }
]
defaults:
run:
shell: msys2 {0}
steps:
- uses: actions/checkout@v2
with:
fetch-depth: 0
- uses: msys2/setup-msys2@v2
with:
msystem: ${{ matrix.msystem }}
install: git base-devel mingw-w64-${{ matrix.arch }}-toolchain
update: true
- name: CI-Build
run: MINGW_INSTALLS=${{ matrix.msystem }} ./ci-build.sh
- name: "Upload binaries"
uses: actions/upload-artifact@v2
with:
name: ${{ matrix.msystem }}-packages
path: artifacts/*.pkg.tar.*
- name: "Upload sources"
uses: actions/upload-artifact@v2
with:
name: sources
path: artifacts/*.src.tar.*
## Instruction:
CI: Move the checkout to C:/_
To (a) get more space and (b) reduce the path lengths
## Code After:
name: main
on:
push:
pull_request:
jobs:
build:
runs-on: windows-latest
strategy:
fail-fast: false
matrix:
include: [
{ msystem: MINGW64, arch: x86_64 },
{ msystem: MINGW32, arch: i686 }
]
steps:
- uses: actions/checkout@v2
with:
path: temp
fetch-depth: 0
- uses: msys2/setup-msys2@v2
with:
msystem: ${{ matrix.msystem }}
install: git base-devel mingw-w64-${{ matrix.arch }}-toolchain
update: true
- name: Move Checkout
run: |
Copy-Item -Path ".\temp" -Destination "C:\_" -Recurse
- name: CI-Build
shell: msys2 {0}
run: |
cd /C/_
MINGW_INSTALLS=${{ matrix.msystem }} ./ci-build.sh
- name: "Upload binaries"
uses: actions/upload-artifact@v2
with:
name: ${{ matrix.msystem }}-packages
path: artifacts/*.pkg.tar.*
- name: "Upload sources"
uses: actions/upload-artifact@v2
with:
name: sources
path: artifacts/*.src.tar.*
|
1bf81b51ad340d9b29c7d8f0593f2b911598fad3 | install.sh | install.sh |
function bootstrap {
mkdir -p ~/.vim/{backups,swaps,undo}
echo '" This is where your personal configuration goes.
" Whatever you write in here will override the default.
' > ~/.vim/config.vim
cd ~/.vim/
git clone https://github.com/umayr/vimrc.git runtime
}
function extended {
cd ~/.vim/runtime
echo 'set runtimepath+=~/.vim/runtime
source ~/.vim/runtime/config/basic.vim
source ~/.vim/runtime/config/filetypes.vim
source ~/.vim/runtime/config/plugins_config.vim
source ~/.vim/runtime/config/extended.vim
try
source ~/.vim/runtime/config.vim
catch
endtry' > ~/.vimrc
echo "Extended Vim configurations has been installed."
exit 0;
}
function basic {
cd ~/.vim/runtime
cat ~/.vim/runtime/config/basic.vim > ~/.vimrc
echo "Basic Vim configurations has been installed."
exit 0;
}
bootstrap; [[ $* == *--extended* ]] && extended; basic; |
function bootstrap {
mkdir -p ~/.vim/{backups,swaps,undo}
echo '" This is where your personal configuration goes.
" Whatever you write in here will override the default.
' > ~/.vim/config.vim
cd ~/.vim/
git clone --recursive https://github.com/umayr/vimrc.git runtime
}
function extended {
cd ~/.vim/runtime
echo 'set runtimepath+=~/.vim/runtime
source ~/.vim/runtime/config/basic.vim
source ~/.vim/runtime/config/filetypes.vim
source ~/.vim/runtime/config/plugins_config.vim
source ~/.vim/runtime/config/extended.vim
try
source ~/.vim/runtime/config.vim
catch
endtry' > ~/.vimrc
echo "Extended Vim configurations has been installed."
exit 0;
}
function basic {
cd ~/.vim/runtime
cat ~/.vim/runtime/config/basic.vim > ~/.vimrc
echo "Basic Vim configurations has been installed."
exit 0;
}
bootstrap; [[ $* == *--extended* ]] && extended; basic;
| Add --recursive flag to clone repo with sub-modules | Add --recursive flag to clone repo with sub-modules | Shell | mit | umayr/vimrc | shell | ## Code Before:
function bootstrap {
mkdir -p ~/.vim/{backups,swaps,undo}
echo '" This is where your personal configuration goes.
" Whatever you write in here will override the default.
' > ~/.vim/config.vim
cd ~/.vim/
git clone https://github.com/umayr/vimrc.git runtime
}
function extended {
cd ~/.vim/runtime
echo 'set runtimepath+=~/.vim/runtime
source ~/.vim/runtime/config/basic.vim
source ~/.vim/runtime/config/filetypes.vim
source ~/.vim/runtime/config/plugins_config.vim
source ~/.vim/runtime/config/extended.vim
try
source ~/.vim/runtime/config.vim
catch
endtry' > ~/.vimrc
echo "Extended Vim configurations has been installed."
exit 0;
}
function basic {
cd ~/.vim/runtime
cat ~/.vim/runtime/config/basic.vim > ~/.vimrc
echo "Basic Vim configurations has been installed."
exit 0;
}
bootstrap; [[ $* == *--extended* ]] && extended; basic;
## Instruction:
Add --recursive flag to clone repo with sub-modules
## Code After:
function bootstrap {
mkdir -p ~/.vim/{backups,swaps,undo}
echo '" This is where your personal configuration goes.
" Whatever you write in here will override the default.
' > ~/.vim/config.vim
cd ~/.vim/
git clone --recursive https://github.com/umayr/vimrc.git runtime
}
function extended {
cd ~/.vim/runtime
echo 'set runtimepath+=~/.vim/runtime
source ~/.vim/runtime/config/basic.vim
source ~/.vim/runtime/config/filetypes.vim
source ~/.vim/runtime/config/plugins_config.vim
source ~/.vim/runtime/config/extended.vim
try
source ~/.vim/runtime/config.vim
catch
endtry' > ~/.vimrc
echo "Extended Vim configurations has been installed."
exit 0;
}
function basic {
cd ~/.vim/runtime
cat ~/.vim/runtime/config/basic.vim > ~/.vimrc
echo "Basic Vim configurations has been installed."
exit 0;
}
bootstrap; [[ $* == *--extended* ]] && extended; basic;
|
b21854fe55dd70c284570648a8c8fa431551c178 | src/test/java/by/triumgroup/recourse/supplier/entity/model/EntityIntegerPKSupplier.java | src/test/java/by/triumgroup/recourse/supplier/entity/model/EntityIntegerPKSupplier.java | package by.triumgroup.recourse.supplier.entity.model;
import by.triumgroup.recourse.entity.model.BaseEntity;
import org.springframework.data.util.Pair;
public interface EntityIntegerPKSupplier<E extends BaseEntity<Integer>> extends EntitySupplier<E, Integer>{
@Override
default Integer getAnyId(){
return 1;
}
@Override
default Pair<Integer, Integer> getDifferentIds(){
return Pair.of(1, 2);
}
}
| package by.triumgroup.recourse.supplier.entity.model;
import by.triumgroup.recourse.entity.model.BaseEntity;
import org.springframework.data.util.Pair;
import java.util.Random;
public interface EntityIntegerPKSupplier<E extends BaseEntity<Integer>> extends EntitySupplier<E, Integer>{
@Override
default Integer getAnyId(){
return new Random().ints(1, Integer.MAX_VALUE).findFirst().getAsInt();
}
@Override
default Pair<Integer, Integer> getDifferentIds(){
Integer first = getAnyId();
Integer second = first + 1;
return Pair.of(first, second);
}
}
| Add random id generating for tests | Add random id generating for tests
| Java | mit | TriumGroup/ReCourse,TriumGroup/ReCourse,TriumGroup/ReCourse | java | ## Code Before:
package by.triumgroup.recourse.supplier.entity.model;
import by.triumgroup.recourse.entity.model.BaseEntity;
import org.springframework.data.util.Pair;
public interface EntityIntegerPKSupplier<E extends BaseEntity<Integer>> extends EntitySupplier<E, Integer>{
@Override
default Integer getAnyId(){
return 1;
}
@Override
default Pair<Integer, Integer> getDifferentIds(){
return Pair.of(1, 2);
}
}
## Instruction:
Add random id generating for tests
## Code After:
package by.triumgroup.recourse.supplier.entity.model;
import by.triumgroup.recourse.entity.model.BaseEntity;
import org.springframework.data.util.Pair;
import java.util.Random;
public interface EntityIntegerPKSupplier<E extends BaseEntity<Integer>> extends EntitySupplier<E, Integer>{
@Override
default Integer getAnyId(){
return new Random().ints(1, Integer.MAX_VALUE).findFirst().getAsInt();
}
@Override
default Pair<Integer, Integer> getDifferentIds(){
Integer first = getAnyId();
Integer second = first + 1;
return Pair.of(first, second);
}
}
|
df41be792be5b5b44f75d3848bd01809a05eefd2 | README.md | README.md | tacopajen.se
===================
| [tacopajen.se](http://tacopajen.se)
===================
##### What is this?
[tacopajen.se](http://tacopajen.se) is a simple webpage by and for taco pie lovers to share the best taco pie recipe.
##### The code and the commit message look awfully!
Yeah, I know.. and I'm sorry for that but [tacopajen.se](http://tacopajen.se) was made on airports, bus stops etc. and in extremely short time periods, meaning each development round lasted only for a few minutes.
##### Why?
This project is one of the contestants in a SEO contest. The word "tacopaj" (taco pie) is one of the most googled word in Sweden so the goal is to get as good position as possible on the SERP with the word "tacopaj" on Google.
##### Did you win?
Too early to say. The competition will end February 20.
##### Can I help?
Absolute! Share the URL [tacopajen.se](http://tacopajen.se), visit the site, write comments. I will love you for this.
| Update readme with some valid information | Update readme with some valid information
Updated the readme file with some relevant information about the project | Markdown | mit | tjoskar/tacopajen.se,tjoskar/tacopajen.se | markdown | ## Code Before:
tacopajen.se
===================
## Instruction:
Update readme with some valid information
Updated the readme file with some relevant information about the project
## Code After:
[tacopajen.se](http://tacopajen.se)
===================
##### What is this?
[tacopajen.se](http://tacopajen.se) is a simple webpage by and for taco pie lovers to share the best taco pie recipe.
##### The code and the commit message look awfully!
Yeah, I know.. and I'm sorry for that but [tacopajen.se](http://tacopajen.se) was made on airports, bus stops etc. and in extremely short time periods, meaning each development round lasted only for a few minutes.
##### Why?
This project is one of the contestants in a SEO contest. The word "tacopaj" (taco pie) is one of the most googled word in Sweden so the goal is to get as good position as possible on the SERP with the word "tacopaj" on Google.
##### Did you win?
Too early to say. The competition will end February 20.
##### Can I help?
Absolute! Share the URL [tacopajen.se](http://tacopajen.se), visit the site, write comments. I will love you for this.
|
10a345c82df463291a60957e160bdea87f729ffd | README.md | README.md |
* `bower install angular-chart-donut`
* `npm install angular-chart-donut`
```js
angular.module('mymodule', ['chart.donut']);
```
```html
<chart-donut data="modeldata" title="Percentage awesomeness" size="150"></chart-donut>
```
<img src="https://nervetattoo.github.io/angular-chart-donut/images/shot.png">
### Element attribute options
* `data` Bind to an integer ranging from 0 to 100
* `title` Use this title for the drop-pin
* `color` Optionally set a color. This can also be achieved through CSS overrides
* `symbol` Defaults to `%` but can be set to other values, like empty string
* `size` Pixel size of chart, if not set the chart will attempt to size itself by its parent, but if parent does not have a size while the chart is rendered this will fail.
### Contributors
* [Raymond Julin](http://twitter.com/nervetattoo)
* [Christian Leon Christensen](http://twitter.com/chrleon) — Design
|
A simple donut chart for Angular built on d3
* `bower install angular-chart-donut`
* `npm install angular-chart-donut`
```js
angular.module('mymodule', ['chart.donut']);
```
```html
<chart-donut data="modeldata" title="Percentage awesomeness" size="150"></chart-donut>
```
<img src="https://nervetattoo.github.io/angular-chart-donut/images/shot.png">
### Element attribute options
* `data` Bind to an integer ranging from 0 to 100
* `title` Use this title for the drop-pin
* `color` Optionally set a color. This can also be achieved through CSS overrides
* `symbol` Defaults to `%` but can be set to other values, like empty string
* `size` Pixel size of chart, if not set the chart will attempt to size itself by its parent, but if parent does not have a size while the chart is rendered this will fail.
### Contributors
* [Raymond Julin](http://twitter.com/nervetattoo)
* [Christian Leon Christensen](http://twitter.com/chrleon) — Design
| Add simple description to readme | Add simple description to readme
| Markdown | mit | nervetattoo/angular-chart-donut | markdown | ## Code Before:
* `bower install angular-chart-donut`
* `npm install angular-chart-donut`
```js
angular.module('mymodule', ['chart.donut']);
```
```html
<chart-donut data="modeldata" title="Percentage awesomeness" size="150"></chart-donut>
```
<img src="https://nervetattoo.github.io/angular-chart-donut/images/shot.png">
### Element attribute options
* `data` Bind to an integer ranging from 0 to 100
* `title` Use this title for the drop-pin
* `color` Optionally set a color. This can also be achieved through CSS overrides
* `symbol` Defaults to `%` but can be set to other values, like empty string
* `size` Pixel size of chart, if not set the chart will attempt to size itself by its parent, but if parent does not have a size while the chart is rendered this will fail.
### Contributors
* [Raymond Julin](http://twitter.com/nervetattoo)
* [Christian Leon Christensen](http://twitter.com/chrleon) — Design
## Instruction:
Add simple description to readme
## Code After:
A simple donut chart for Angular built on d3
* `bower install angular-chart-donut`
* `npm install angular-chart-donut`
```js
angular.module('mymodule', ['chart.donut']);
```
```html
<chart-donut data="modeldata" title="Percentage awesomeness" size="150"></chart-donut>
```
<img src="https://nervetattoo.github.io/angular-chart-donut/images/shot.png">
### Element attribute options
* `data` Bind to an integer ranging from 0 to 100
* `title` Use this title for the drop-pin
* `color` Optionally set a color. This can also be achieved through CSS overrides
* `symbol` Defaults to `%` but can be set to other values, like empty string
* `size` Pixel size of chart, if not set the chart will attempt to size itself by its parent, but if parent does not have a size while the chart is rendered this will fail.
### Contributors
* [Raymond Julin](http://twitter.com/nervetattoo)
* [Christian Leon Christensen](http://twitter.com/chrleon) — Design
|
89b7a6bd1f19405be5426d32ac55fdb718b8868c | test/powershell/Tee-Object.Tests.ps1 | test/powershell/Tee-Object.Tests.ps1 | Describe "Tee-Object" {
Context "Validate Tee-Object is correctly forking output" {
$testfile = Join-Path -Path (Join-Path -Path $PSScriptRoot -ChildPath assets) -ChildPath testfile.txt
It "Should return the output to the screen and to the variable" {
$teefile = $testfile
echo teeobjecttest1 | Tee-Object -variable teeresults
$teeresults | Should Be "teeobjecttest1"
Remove-Item $teefile -ErrorAction SilentlyContinue
}
It "Should tee the output to a file" {
$teefile = $testfile
echo teeobjecttest3 | Tee-Object $teefile
Get-Content $teefile | Should Be "teeobjecttest3"
Remove-Item $teefile -ErrorAction SilentlyContinue
}
}
}
| Describe "Tee-Object" {
Context "Validate Tee-Object is correctly forking output" {
$testfile = Join-Path -Path (Join-Path -Path $PSScriptRoot -ChildPath assets) -ChildPath testfile.txt
It "Should return the output to the screen and to the variable" {
$teefile = $testfile
echo teeobjecttest1 | Tee-Object -variable teeresults
$teeresults | Should Be "teeobjecttest1"
Remove-Item $teefile -ErrorAction SilentlyContinue
}
It "Should tee the output to a file" {
$teefile = $testfile
echo teeobjecttest3 | Tee-Object $teefile
Get-Content $teefile | Should Be "teeobjecttest3"
Remove-Item $teefile -ErrorAction SilentlyContinue
}
}
}
Describe "Tee-Object DRT Unit Tests" -Tags DRT {
$tempDirectory = Join-Path $TestDrive -ChildPath "TeeObjectTestsTempDir"
$tempFile = "TeeObjectTestsTempFile"
New-Item $tempDirectory -ItemType Directory -Force
It "Positive File Test" {
$expected = "1", "2", "3"
$filePath = Join-Path $tempDirectory -ChildPath $tempFile
$results = $expected | Tee-Object -FilePath $filePath
$results.Length | Should be 3
$results | Should Be $expected
$content = Get-Content $filePath
$content | Should Be $expected
}
It "Positive Var Test" {
$expected = "1", "2", "3"
$varName = "teeObjectTestVar"
$results = $expected | Tee-Object -Variable $varName
$results.Length | Should be 3
$results | Should Be $expected
$results = Get-Variable -Name $varName -ValueOnly
$results.Length | Should be 3
$results | Should Be $expected
}
} | Add Unit Test For Tee-Object | Add Unit Test For Tee-Object
| PowerShell | mit | bingbing8/PowerShell,JamesWTruher/PowerShell-1,JamesWTruher/PowerShell-1,bmanikm/PowerShell,bmanikm/PowerShell,TravisEz13/PowerShell,daxian-dbw/PowerShell,bingbing8/PowerShell,kmosher/PowerShell,KarolKaczmarek/PowerShell,kmosher/PowerShell,KarolKaczmarek/PowerShell,PaulHigin/PowerShell,bingbing8/PowerShell,PaulHigin/PowerShell,bmanikm/PowerShell,PaulHigin/PowerShell,bingbing8/PowerShell,jsoref/PowerShell,JamesWTruher/PowerShell-1,jsoref/PowerShell,bingbing8/PowerShell,jsoref/PowerShell,PaulHigin/PowerShell,daxian-dbw/PowerShell,bmanikm/PowerShell,TravisEz13/PowerShell,bmanikm/PowerShell,daxian-dbw/PowerShell,jsoref/PowerShell,kmosher/PowerShell,daxian-dbw/PowerShell,JamesWTruher/PowerShell-1,kmosher/PowerShell,KarolKaczmarek/PowerShell,TravisEz13/PowerShell,kmosher/PowerShell,KarolKaczmarek/PowerShell,KarolKaczmarek/PowerShell,TravisEz13/PowerShell,jsoref/PowerShell | powershell | ## Code Before:
Describe "Tee-Object" {
Context "Validate Tee-Object is correctly forking output" {
$testfile = Join-Path -Path (Join-Path -Path $PSScriptRoot -ChildPath assets) -ChildPath testfile.txt
It "Should return the output to the screen and to the variable" {
$teefile = $testfile
echo teeobjecttest1 | Tee-Object -variable teeresults
$teeresults | Should Be "teeobjecttest1"
Remove-Item $teefile -ErrorAction SilentlyContinue
}
It "Should tee the output to a file" {
$teefile = $testfile
echo teeobjecttest3 | Tee-Object $teefile
Get-Content $teefile | Should Be "teeobjecttest3"
Remove-Item $teefile -ErrorAction SilentlyContinue
}
}
}
## Instruction:
Add Unit Test For Tee-Object
## Code After:
Describe "Tee-Object" {
Context "Validate Tee-Object is correctly forking output" {
$testfile = Join-Path -Path (Join-Path -Path $PSScriptRoot -ChildPath assets) -ChildPath testfile.txt
It "Should return the output to the screen and to the variable" {
$teefile = $testfile
echo teeobjecttest1 | Tee-Object -variable teeresults
$teeresults | Should Be "teeobjecttest1"
Remove-Item $teefile -ErrorAction SilentlyContinue
}
It "Should tee the output to a file" {
$teefile = $testfile
echo teeobjecttest3 | Tee-Object $teefile
Get-Content $teefile | Should Be "teeobjecttest3"
Remove-Item $teefile -ErrorAction SilentlyContinue
}
}
}
Describe "Tee-Object DRT Unit Tests" -Tags DRT {
$tempDirectory = Join-Path $TestDrive -ChildPath "TeeObjectTestsTempDir"
$tempFile = "TeeObjectTestsTempFile"
New-Item $tempDirectory -ItemType Directory -Force
It "Positive File Test" {
$expected = "1", "2", "3"
$filePath = Join-Path $tempDirectory -ChildPath $tempFile
$results = $expected | Tee-Object -FilePath $filePath
$results.Length | Should be 3
$results | Should Be $expected
$content = Get-Content $filePath
$content | Should Be $expected
}
It "Positive Var Test" {
$expected = "1", "2", "3"
$varName = "teeObjectTestVar"
$results = $expected | Tee-Object -Variable $varName
$results.Length | Should be 3
$results | Should Be $expected
$results = Get-Variable -Name $varName -ValueOnly
$results.Length | Should be 3
$results | Should Be $expected
}
} |
12ccfab25a885a6c69685bb48dc783ecc358c86b | demo/router-components/echo.tsx | demo/router-components/echo.tsx | import app, { Component } from '../../index-jsx';
var model = 'world';
const Hello = ({name}) => <div>Hello: {name}</div>;
const view = (val) => {
return <div>
<Hello name={val}/>
<input value={val} oninput={function() { app.run('render', this.value)}}/>
</div>
};
const update = {
'#echo': (model, pushState) => pushState || model,
'render': (_, val) => {
history.pushState(null, null, '#echo/' + val);
return val;
}
}
export default (element) => app.start(element, model, view, update);
| import app from '../../index-jsx';
var model = 'world';
const Hello = ({name}) => <div>Hello: {name}</div>;
const view = (val) => {
return <div>
<Hello name={val}/>
<input value={val} oninput={function() { app.run('render', this.value)}}/>
</div>
};
const update = {
'#echo': (model, pushState) => pushState || model,
'render': (_, val) => {
history.pushState(null, null, '#echo/' + val);
return val;
}
}
export default (element) => app.start(element, model, view, update);
| Fix demo to use app.start instead of new Component | Fix demo to use app.start instead of new Component
| TypeScript | mit | yysun/apprun,yysun/apprun,yysun/apprun | typescript | ## Code Before:
import app, { Component } from '../../index-jsx';
var model = 'world';
const Hello = ({name}) => <div>Hello: {name}</div>;
const view = (val) => {
return <div>
<Hello name={val}/>
<input value={val} oninput={function() { app.run('render', this.value)}}/>
</div>
};
const update = {
'#echo': (model, pushState) => pushState || model,
'render': (_, val) => {
history.pushState(null, null, '#echo/' + val);
return val;
}
}
export default (element) => app.start(element, model, view, update);
## Instruction:
Fix demo to use app.start instead of new Component
## Code After:
import app from '../../index-jsx';
var model = 'world';
const Hello = ({name}) => <div>Hello: {name}</div>;
const view = (val) => {
return <div>
<Hello name={val}/>
<input value={val} oninput={function() { app.run('render', this.value)}}/>
</div>
};
const update = {
'#echo': (model, pushState) => pushState || model,
'render': (_, val) => {
history.pushState(null, null, '#echo/' + val);
return val;
}
}
export default (element) => app.start(element, model, view, update);
|
cdc4160c8aff31bdb7859285d021f5a122ed755d | docs/sample_code/MYINST/MYINST.h | docs/sample_code/MYINST/MYINST.h |
class MYINST : public Instrument {
public:
MYINST();
virtual ~MYINST();
virtual int init(double *, int);
virtual int configure();
virtual int run();
private:
void doupdate();
int _nargs, _inchan, _branch;
float _amp, _pan;
float *_in;
};
|
class MYINST : public Instrument {
public:
MYINST();
virtual ~MYINST();
virtual int init(double *, int);
virtual int configure();
virtual int run();
private:
void doupdate();
float *_in;
int _nargs, _inchan, _branch;
float _amp, _pan;
};
| Change order of declarations to suppress compiler warning. | Change order of declarations to suppress compiler warning.
| C | apache-2.0 | RTcmix/RTcmix,RTcmix/RTcmix,RTcmix/RTcmix,RTcmix/RTcmix,RTcmix/RTcmix,RTcmix/RTcmix | c | ## Code Before:
class MYINST : public Instrument {
public:
MYINST();
virtual ~MYINST();
virtual int init(double *, int);
virtual int configure();
virtual int run();
private:
void doupdate();
int _nargs, _inchan, _branch;
float _amp, _pan;
float *_in;
};
## Instruction:
Change order of declarations to suppress compiler warning.
## Code After:
class MYINST : public Instrument {
public:
MYINST();
virtual ~MYINST();
virtual int init(double *, int);
virtual int configure();
virtual int run();
private:
void doupdate();
float *_in;
int _nargs, _inchan, _branch;
float _amp, _pan;
};
|
762d803b88da17766b8f0347f0dd6b09cab8160f | modules/analyse/src/main/TimeChart.scala | modules/analyse/src/main/TimeChart.scala | package lila.analyse
import scala.concurrent.Future
import chess.Color
import play.api.libs.json.Json
import lila.game.{ Game, Namer }
final class TimeChart(game: Game, moves: List[String]) {
private val pgnMoves = moves.toIndexedSeq
def series = Json stringify {
Json.obj(
"white" -> points(true),
"black" -> points(false)
)
}
private def points(white: Boolean) = indexedMoveTimes collect {
case (m, ply) if (white ^ (ply % 2 == 1)) =>
val mt = if (m < 0.5) 0 else m
val san = ~(pgnMoves lift ply)
val turn = 1 + (ply - 1) / 2
val dots = if (ply % 2 == 1) "..." else "."
Json.obj(
"name" -> s"$turn$dots $san",
"x" -> ply,
"y" -> (if (white) mt else -mt)
)
}
def maxTime = moveTimes.foldLeft(0f) {
case (x, y) => if (y > x) y else x
}
private val moveTimes = game.moveTimesInSeconds
private val indexedMoveTimes = game.moveTimesInSeconds.zipWithIndex
}
| package lila.analyse
import scala.concurrent.Future
import chess.Color
import play.api.libs.json.Json
import lila.game.{ Game, Namer }
final class TimeChart(game: Game, moves: List[String]) {
private val pgnMoves = moves.toIndexedSeq
def series = Json stringify {
Json.obj(
"white" -> points(true),
"black" -> points(false)
)
}
private def points(white: Boolean) = indexedMoveTimes collect {
case (m, ply) if (white ^ (ply % 2 == 1)) =>
val mt = if (m < 0.5) 0 else m
val san = ~(pgnMoves lift ply)
val turn = ((ply/2).floor + 1).toInt
val dots = if (ply % 2 == 1) "..." else "."
Json.obj(
"name" -> s"$turn$dots $san",
"x" -> ply,
"y" -> (if (white) mt else -mt)
)
}
def maxTime = moveTimes.foldLeft(0f) {
case (x, y) => if (y > x) y else x
}
private val moveTimes = game.moveTimesInSeconds
private val indexedMoveTimes = game.moveTimesInSeconds.zipWithIndex
}
| Fix move numbering in timeChart | Fix move numbering in timeChart
| Scala | agpl-3.0 | r0k3/lila,pawank/lila,TangentialAlan/lila,r0k3/lila,pawank/lila,bjhaid/lila,Happy0/lila,TangentialAlan/lila,Unihedro/lila,pavelo65/lila,elioair/lila,clarkerubber/lila,Unihedro/lila,ccampo133/lila,clarkerubber/lila,arex1337/lila,pavelo65/lila,r0k3/lila,terokinnunen/lila,systemovich/lila,clarkerubber/lila,pawank/lila,systemovich/lila,luanlv/lila,pawank/lila,elioair/lila,arex1337/lila,elioair/lila,luanlv/lila,pavelo65/lila,pavelo65/lila,r0k3/lila,bjhaid/lila,bjhaid/lila,pawank/lila,Unihedro/lila,Unihedro/lila,elioair/lila,TangentialAlan/lila,arex1337/lila,danilovsergey/i-bur,Enigmahack/lila,samuel-soubeyran/lila,systemovich/lila,JimmyMow/lila,ccampo133/lila,clarkerubber/lila,Happy0/lila,abougouffa/lila,pawank/lila,Unihedro/lila,TangentialAlan/lila,JimmyMow/lila,arex1337/lila,Happy0/lila,abougouffa/lila,ccampo133/lila,elioair/lila,samuel-soubeyran/lila,JimmyMow/lila,danilovsergey/i-bur,luanlv/lila,terokinnunen/lila,clarkerubber/lila,samuel-soubeyran/lila,r0k3/lila,samuel-soubeyran/lila,bjhaid/lila,ccampo133/lila,r0k3/lila,terokinnunen/lila,JimmyMow/lila,danilovsergey/i-bur,abougouffa/lila,JimmyMow/lila,terokinnunen/lila,bjhaid/lila,terokinnunen/lila,systemovich/lila,pavelo65/lila,clarkerubber/lila,Enigmahack/lila,ccampo133/lila,Enigmahack/lila,clarkerubber/lila,arex1337/lila,luanlv/lila,TangentialAlan/lila,abougouffa/lila,Happy0/lila,pavelo65/lila,Enigmahack/lila,Enigmahack/lila,TangentialAlan/lila,samuel-soubeyran/lila,danilovsergey/i-bur,Happy0/lila,systemovich/lila,arex1337/lila,luanlv/lila,bjhaid/lila,Happy0/lila,Unihedro/lila,systemovich/lila,luanlv/lila,Unihedro/lila,arex1337/lila,abougouffa/lila,ccampo133/lila,pavelo65/lila,danilovsergey/i-bur,JimmyMow/lila,bjhaid/lila,Enigmahack/lila,JimmyMow/lila,Enigmahack/lila,danilovsergey/i-bur,samuel-soubeyran/lila,TangentialAlan/lila,r0k3/lila,pawank/lila,luanlv/lila,terokinnunen/lila,Happy0/lila,systemovich/lila,abougouffa/lila,ccampo133/lila,danilovsergey/i-bur,terokinnunen/lila,abougouffa/lila,elioair/lila,elioair/lila,samuel-soubeyran/lila | scala | ## Code Before:
package lila.analyse
import scala.concurrent.Future
import chess.Color
import play.api.libs.json.Json
import lila.game.{ Game, Namer }
final class TimeChart(game: Game, moves: List[String]) {
private val pgnMoves = moves.toIndexedSeq
def series = Json stringify {
Json.obj(
"white" -> points(true),
"black" -> points(false)
)
}
private def points(white: Boolean) = indexedMoveTimes collect {
case (m, ply) if (white ^ (ply % 2 == 1)) =>
val mt = if (m < 0.5) 0 else m
val san = ~(pgnMoves lift ply)
val turn = 1 + (ply - 1) / 2
val dots = if (ply % 2 == 1) "..." else "."
Json.obj(
"name" -> s"$turn$dots $san",
"x" -> ply,
"y" -> (if (white) mt else -mt)
)
}
def maxTime = moveTimes.foldLeft(0f) {
case (x, y) => if (y > x) y else x
}
private val moveTimes = game.moveTimesInSeconds
private val indexedMoveTimes = game.moveTimesInSeconds.zipWithIndex
}
## Instruction:
Fix move numbering in timeChart
## Code After:
package lila.analyse
import scala.concurrent.Future
import chess.Color
import play.api.libs.json.Json
import lila.game.{ Game, Namer }
final class TimeChart(game: Game, moves: List[String]) {
private val pgnMoves = moves.toIndexedSeq
def series = Json stringify {
Json.obj(
"white" -> points(true),
"black" -> points(false)
)
}
private def points(white: Boolean) = indexedMoveTimes collect {
case (m, ply) if (white ^ (ply % 2 == 1)) =>
val mt = if (m < 0.5) 0 else m
val san = ~(pgnMoves lift ply)
val turn = ((ply/2).floor + 1).toInt
val dots = if (ply % 2 == 1) "..." else "."
Json.obj(
"name" -> s"$turn$dots $san",
"x" -> ply,
"y" -> (if (white) mt else -mt)
)
}
def maxTime = moveTimes.foldLeft(0f) {
case (x, y) => if (y > x) y else x
}
private val moveTimes = game.moveTimesInSeconds
private val indexedMoveTimes = game.moveTimesInSeconds.zipWithIndex
}
|
69b2da9f8eab86b0e4bad789edcea0332e6d692b | requirements-dev.txt | requirements-dev.txt | pytest>=2.7.0,<=3.0.7
mock>=1.0.1
tox>=1.9.2
behave>=1.2.4
pexpect==3.3
pre-commit>=1.16.0
coverage==4.3.4
codecov>=1.5.1
docutils>=0.13.1
autopep8==1.3.3
click==6.7
twine==1.11.0
| pytest>=2.7.0,<=3.0.7
mock>=1.0.1
tox>=1.9.2
behave>=1.2.4
pexpect==3.3
pre-commit>=1.16.0
coverage==4.3.4
codecov>=1.5.1
docutils>=0.13.1
autopep8==1.3.3
click==6.7
twine==1.11.0
wheel==0.33.6
| Add wheel to dev reqs (needed to release). | Add wheel to dev reqs (needed to release).
| Text | bsd-3-clause | dbcli/pgcli,dbcli/pgcli | text | ## Code Before:
pytest>=2.7.0,<=3.0.7
mock>=1.0.1
tox>=1.9.2
behave>=1.2.4
pexpect==3.3
pre-commit>=1.16.0
coverage==4.3.4
codecov>=1.5.1
docutils>=0.13.1
autopep8==1.3.3
click==6.7
twine==1.11.0
## Instruction:
Add wheel to dev reqs (needed to release).
## Code After:
pytest>=2.7.0,<=3.0.7
mock>=1.0.1
tox>=1.9.2
behave>=1.2.4
pexpect==3.3
pre-commit>=1.16.0
coverage==4.3.4
codecov>=1.5.1
docutils>=0.13.1
autopep8==1.3.3
click==6.7
twine==1.11.0
wheel==0.33.6
|
355286252d8fedb670da44dac191414d7a7e837e | about/index.html | about/index.html | ---
layout: default
title: About
---
<article class="page container">
{% include intro.html %}
<section class="post-list">
<div class="container">
<ul class="project">
<li>
<a href="http://dx.doi.org/10.14279/depositonce-4310"><img width="150px" src="../assets/icons/thesis.png"/>
<h3><span>PhD thesis</span></h3></a>
</li>
<li>
<a href="http://twoears.eu"><img width="150px" src="../assets/icons/twoears.png"/>
<h3><span>Two!Ears</span></h3></a>
</li>
<li>
<a href="http://sfstoolbox.org"><img width="150px" src="../assets/icons/sfs.png"/>
<h3><span>SFS Toolbox</span></h3></a>
</li>
<li>
<a href="http://gnuplotting.org"><img width="150px" src="../assets/icons/gnuplotting.png"/>
<h3><span>Gnuplotting</span></h3></a>
</li>
</ul>
</div>
</section>
</article>
| ---
layout: default
title: About
permalink: /about/
---
<article class="page container">
{% include intro.html %}
<section class="post-list">
<div class="container">
<ul class="project">
<li>
<a href="http://dx.doi.org/10.14279/depositonce-4310"><img width="150px" src="../assets/icons/thesis.png"/>
<h3><span>PhD thesis</span></h3></a>
</li>
<li>
<a href="http://twoears.eu"><img width="150px" src="../assets/icons/twoears.png"/>
<h3><span>Two!Ears</span></h3></a>
</li>
<li>
<a href="http://sfstoolbox.org"><img width="150px" src="../assets/icons/sfs.png"/>
<h3><span>SFS Toolbox</span></h3></a>
</li>
<li>
<a href="http://gnuplotting.org"><img width="150px" src="../assets/icons/gnuplotting.png"/>
<h3><span>Gnuplotting</span></h3></a>
</li>
</ul>
</div>
</section>
</article>
| Set permalink name for about | Set permalink name for about
| HTML | mit | hagenw/hagenw.github.io,hagenw/hagenw.github.io | html | ## Code Before:
---
layout: default
title: About
---
<article class="page container">
{% include intro.html %}
<section class="post-list">
<div class="container">
<ul class="project">
<li>
<a href="http://dx.doi.org/10.14279/depositonce-4310"><img width="150px" src="../assets/icons/thesis.png"/>
<h3><span>PhD thesis</span></h3></a>
</li>
<li>
<a href="http://twoears.eu"><img width="150px" src="../assets/icons/twoears.png"/>
<h3><span>Two!Ears</span></h3></a>
</li>
<li>
<a href="http://sfstoolbox.org"><img width="150px" src="../assets/icons/sfs.png"/>
<h3><span>SFS Toolbox</span></h3></a>
</li>
<li>
<a href="http://gnuplotting.org"><img width="150px" src="../assets/icons/gnuplotting.png"/>
<h3><span>Gnuplotting</span></h3></a>
</li>
</ul>
</div>
</section>
</article>
## Instruction:
Set permalink name for about
## Code After:
---
layout: default
title: About
permalink: /about/
---
<article class="page container">
{% include intro.html %}
<section class="post-list">
<div class="container">
<ul class="project">
<li>
<a href="http://dx.doi.org/10.14279/depositonce-4310"><img width="150px" src="../assets/icons/thesis.png"/>
<h3><span>PhD thesis</span></h3></a>
</li>
<li>
<a href="http://twoears.eu"><img width="150px" src="../assets/icons/twoears.png"/>
<h3><span>Two!Ears</span></h3></a>
</li>
<li>
<a href="http://sfstoolbox.org"><img width="150px" src="../assets/icons/sfs.png"/>
<h3><span>SFS Toolbox</span></h3></a>
</li>
<li>
<a href="http://gnuplotting.org"><img width="150px" src="../assets/icons/gnuplotting.png"/>
<h3><span>Gnuplotting</span></h3></a>
</li>
</ul>
</div>
</section>
</article>
|
babfc82613508e619e5debf044e68eaeae03ba11 | lib/node_modules/@stdlib/math/base/dist/degenerate/logcdf/lib/logcdf.js | lib/node_modules/@stdlib/math/base/dist/degenerate/logcdf/lib/logcdf.js | 'use strict';
// MODULES //
var isnan = require( '@stdlib/math/base/utils/is-nan' );
var NINF = require( '@stdlib/math/constants/float64-ninf' );
// MAIN //
/**
* Evaluates the logarithm of the cumulative distribution function (CDF) for a degenerate distribution with a point mass at `mu`.
*
* @param {number} x - input value
* @param {number} mu - constant value of distribution
* @returns {Probability} evaluated logarithm of cumulative distribution function
*
* @example
* var y = logcdf( 2.0, 3.0 );
* // returns Number.NEGATIVE_INFINITY
* @example
* var y = logcdf( 4.0, 3.0 );
* // returns 0
* @example
* var y = logcdf( 3.0, 3.0 );
* // returns 0
* @example
* var y = logcdf( NaN, 0.0 );
* // returns NaN
* @example
* var y = logcdf( 0.0, NaN );
* // returns NaN
*/
function logcdf( x, mu ) {
if ( isnan( x ) || isnan( mu ) ) {
return NaN;
}
return x < mu ? NINF : 0.0;
} // end FUNCTION logcdf()
// EXPORTS //
module.exports = logcdf;
| 'use strict';
// MODULES //
var isnan = require( '@stdlib/math/base/utils/is-nan' );
var NINF = require( '@stdlib/math/constants/float64-ninf' );
// MAIN //
/**
* Evaluates the natural logarithm of the cumulative distribution function (logCDF) for a degenerate distribution with mean `mu`.
*
* @param {number} x - input value
* @param {number} mu - constant value of distribution
* @returns {number} natural logarithm of cumulative distribution function
*
* @example
* var y = logcdf( 2.0, 3.0 );
* // returns Number.NEGATIVE_INFINITY
* @example
* var y = logcdf( 4.0, 3.0 );
* // returns 0.0
* @example
* var y = logcdf( 3.0, 3.0 );
* // returns 0.0
* @example
* var y = logcdf( NaN, 0.0 );
* // returns NaN
* @example
* var y = logcdf( 0.0, NaN );
* // returns NaN
*/
function logcdf( x, mu ) {
if ( isnan( x ) || isnan( mu ) ) {
return NaN;
}
return ( x < mu ) ? NINF : 0.0;
} // end FUNCTION logcdf()
// EXPORTS //
module.exports = logcdf;
| Fix return type and add decimal points | Fix return type and add decimal points
| JavaScript | apache-2.0 | stdlib-js/stdlib,stdlib-js/stdlib,stdlib-js/stdlib,stdlib-js/stdlib,stdlib-js/stdlib,stdlib-js/stdlib,stdlib-js/stdlib,stdlib-js/stdlib | javascript | ## Code Before:
'use strict';
// MODULES //
var isnan = require( '@stdlib/math/base/utils/is-nan' );
var NINF = require( '@stdlib/math/constants/float64-ninf' );
// MAIN //
/**
* Evaluates the logarithm of the cumulative distribution function (CDF) for a degenerate distribution with a point mass at `mu`.
*
* @param {number} x - input value
* @param {number} mu - constant value of distribution
* @returns {Probability} evaluated logarithm of cumulative distribution function
*
* @example
* var y = logcdf( 2.0, 3.0 );
* // returns Number.NEGATIVE_INFINITY
* @example
* var y = logcdf( 4.0, 3.0 );
* // returns 0
* @example
* var y = logcdf( 3.0, 3.0 );
* // returns 0
* @example
* var y = logcdf( NaN, 0.0 );
* // returns NaN
* @example
* var y = logcdf( 0.0, NaN );
* // returns NaN
*/
function logcdf( x, mu ) {
if ( isnan( x ) || isnan( mu ) ) {
return NaN;
}
return x < mu ? NINF : 0.0;
} // end FUNCTION logcdf()
// EXPORTS //
module.exports = logcdf;
## Instruction:
Fix return type and add decimal points
## Code After:
'use strict';
// MODULES //
var isnan = require( '@stdlib/math/base/utils/is-nan' );
var NINF = require( '@stdlib/math/constants/float64-ninf' );
// MAIN //
/**
* Evaluates the natural logarithm of the cumulative distribution function (logCDF) for a degenerate distribution with mean `mu`.
*
* @param {number} x - input value
* @param {number} mu - constant value of distribution
* @returns {number} natural logarithm of cumulative distribution function
*
* @example
* var y = logcdf( 2.0, 3.0 );
* // returns Number.NEGATIVE_INFINITY
* @example
* var y = logcdf( 4.0, 3.0 );
* // returns 0.0
* @example
* var y = logcdf( 3.0, 3.0 );
* // returns 0.0
* @example
* var y = logcdf( NaN, 0.0 );
* // returns NaN
* @example
* var y = logcdf( 0.0, NaN );
* // returns NaN
*/
function logcdf( x, mu ) {
if ( isnan( x ) || isnan( mu ) ) {
return NaN;
}
return ( x < mu ) ? NINF : 0.0;
} // end FUNCTION logcdf()
// EXPORTS //
module.exports = logcdf;
|
b9a8ad6ade66070f022f9b808df08cc9ff8d8ab2 | 05-ext/CMakeLists.txt | 05-ext/CMakeLists.txt | cmake_minimum_required(VERSION 3.5)
project(main)
include(ExternalProject)
ExternalProject_Add(fooProject
SOURCE_DIR ${CMAKE_CURRENT_SOURCE_DIR}/foo1
CMAKE_ARGS -DCMAKE_INSTALL_PREFIX=${CMAKE_BINARY_DIR}/ext
)
ExternalProject_Get_Property(fooProject install_dir)
message("fooProject install_dir=${install_dir}")
ExternalProject_Add(helloProject
SOURCE_DIR ${CMAKE_CURRENT_SOURCE_DIR}/hello
CMAKE_ARGS -DCMAKE_MODULE_PATH=${CMAKE_BINARY_DIR}/fooProject-prefix/src/fooProject-build
)
| cmake_minimum_required(VERSION 3.5)
project(main)
include(ExternalProject)
ExternalProject_Add(fooProject
SOURCE_DIR ${CMAKE_CURRENT_SOURCE_DIR}/foo1
CMAKE_ARGS -DCMAKE_INSTALL_PREFIX=${CMAKE_BINARY_DIR}/ext
)
ExternalProject_Get_Property(fooProject install_dir)
ExternalProject_Get_Property(fooProject source_dir)
ExternalProject_Get_Property(fooProject binary_dir)
message("fooProject install_dir=${install_dir}")
message("fooProject source_dir=${source_dir}")
message("fooProject binary_dir=${binary_dir}")
ExternalProject_Add(helloProject
SOURCE_DIR ${CMAKE_CURRENT_SOURCE_DIR}/hello
CMAKE_ARGS -DCMAKE_PREFIX_PATH=${binary_dir}
)
| Fix the cmake include directory | Fix the cmake include directory
Signed-off-by: Francis Giraldeau <[email protected]>
| Text | mit | giraldeau/cmake-tutorial,giraldeau/cmake-tutorial | text | ## Code Before:
cmake_minimum_required(VERSION 3.5)
project(main)
include(ExternalProject)
ExternalProject_Add(fooProject
SOURCE_DIR ${CMAKE_CURRENT_SOURCE_DIR}/foo1
CMAKE_ARGS -DCMAKE_INSTALL_PREFIX=${CMAKE_BINARY_DIR}/ext
)
ExternalProject_Get_Property(fooProject install_dir)
message("fooProject install_dir=${install_dir}")
ExternalProject_Add(helloProject
SOURCE_DIR ${CMAKE_CURRENT_SOURCE_DIR}/hello
CMAKE_ARGS -DCMAKE_MODULE_PATH=${CMAKE_BINARY_DIR}/fooProject-prefix/src/fooProject-build
)
## Instruction:
Fix the cmake include directory
Signed-off-by: Francis Giraldeau <[email protected]>
## Code After:
cmake_minimum_required(VERSION 3.5)
project(main)
include(ExternalProject)
ExternalProject_Add(fooProject
SOURCE_DIR ${CMAKE_CURRENT_SOURCE_DIR}/foo1
CMAKE_ARGS -DCMAKE_INSTALL_PREFIX=${CMAKE_BINARY_DIR}/ext
)
ExternalProject_Get_Property(fooProject install_dir)
ExternalProject_Get_Property(fooProject source_dir)
ExternalProject_Get_Property(fooProject binary_dir)
message("fooProject install_dir=${install_dir}")
message("fooProject source_dir=${source_dir}")
message("fooProject binary_dir=${binary_dir}")
ExternalProject_Add(helloProject
SOURCE_DIR ${CMAKE_CURRENT_SOURCE_DIR}/hello
CMAKE_ARGS -DCMAKE_PREFIX_PATH=${binary_dir}
)
|
6149b251296bac9e26bd81763cbf8379a743e3b5 | DEVELOPER.md | DEVELOPER.md |
1. Google HTML CSS Style Guide (https://google.github.io/styleguide/htmlcssguide.xml)
2.
...to be continue... |
1. Google HTML CSS Style Guide (https://google.github.io/styleguide/htmlcssguide.xml)
2.
Coding Style Guide
- Don't add the hard code style if it can be done by CSS.
...to be continue... | Update the coding style guide | Update the coding style guide
| Markdown | mit | GoEngin/GoEngin,GoEngin/GoEngin,GoEngin/GoEngin | markdown | ## Code Before:
1. Google HTML CSS Style Guide (https://google.github.io/styleguide/htmlcssguide.xml)
2.
...to be continue...
## Instruction:
Update the coding style guide
## Code After:
1. Google HTML CSS Style Guide (https://google.github.io/styleguide/htmlcssguide.xml)
2.
Coding Style Guide
- Don't add the hard code style if it can be done by CSS.
...to be continue... |
a08294c731f9bd190786911830cc636f00b2af7d | console/kubernetes/darzee-service.yml | console/kubernetes/darzee-service.yml | apiVersion: apps/v1
kind: Service
metadata:
labels:
app: darzee
name: darzee-svc
spec:
type: LoadBalancer
ports:
# NOTE: Exposting ports
- port: 8080
targetPort: 8080
- port: 9090
targetPort: 9090
selector:
app: darzee
sessionAffinity: None
| apiVersion: v1
kind: Service
metadata:
labels:
app: darzee
name: darzee-svc
spec:
type: LoadBalancer
ports:
# NOTE: Exposting ports
- port: 8080
name: nodejs-server-port
targetPort: 8080
- port: 9090
name: prometheus-port
targetPort: 9090
selector:
app: darzee
sessionAffinity: None
| Add name to ports. Fix API version. | Add name to ports. Fix API version.
| YAML | mit | emc-mongoose/console,emc-mongoose/console,emc-mongoose/console | yaml | ## Code Before:
apiVersion: apps/v1
kind: Service
metadata:
labels:
app: darzee
name: darzee-svc
spec:
type: LoadBalancer
ports:
# NOTE: Exposting ports
- port: 8080
targetPort: 8080
- port: 9090
targetPort: 9090
selector:
app: darzee
sessionAffinity: None
## Instruction:
Add name to ports. Fix API version.
## Code After:
apiVersion: v1
kind: Service
metadata:
labels:
app: darzee
name: darzee-svc
spec:
type: LoadBalancer
ports:
# NOTE: Exposting ports
- port: 8080
name: nodejs-server-port
targetPort: 8080
- port: 9090
name: prometheus-port
targetPort: 9090
selector:
app: darzee
sessionAffinity: None
|
1bac0ee98dc7150e51dd699da52f741924c09abb | .dotfiles/env.sh | .dotfiles/env.sh | export PATH=$PATH:/mingw64/bin
export PATH=/cmd:$PATH
export TERM=cygwin
export GOPATH=$HOME/go
export PATH=$PATH:$GOPATH/bin
export PATH="$PATH:$HOME/npm/bin"
export PATH="$PATH:$HOME/npm"
. ~/.dotfiles/env/docker.sh
. ~/.dotfiles/env/git-ssh-forwarding.sh
. ~/.dotfiles/env/git.sh
| export PATH=$PATH:/mingw64/bin
export PATH=/cmd:$PATH
export TERM=cygwin
export GOPATH=$HOME/go
export PATH=$PATH:$GOPATH/bin
export PATH="$PATH:$HOME/npm/bin"
export PATH="$PATH:$HOME/npm"
export ANDROID_HOME="$LOCALAPPDATA/Android/Sdk"
export PATH="$PATH:$ANDROID_HOME/platform-tools"
. ~/.dotfiles/env/docker.sh
. ~/.dotfiles/env/git-ssh-forwarding.sh
. ~/.dotfiles/env/git.sh
| Add android tools support - for react-native tools | Add android tools support - for react-native tools
| Shell | mit | micnigh/windows-dotfiles | shell | ## Code Before:
export PATH=$PATH:/mingw64/bin
export PATH=/cmd:$PATH
export TERM=cygwin
export GOPATH=$HOME/go
export PATH=$PATH:$GOPATH/bin
export PATH="$PATH:$HOME/npm/bin"
export PATH="$PATH:$HOME/npm"
. ~/.dotfiles/env/docker.sh
. ~/.dotfiles/env/git-ssh-forwarding.sh
. ~/.dotfiles/env/git.sh
## Instruction:
Add android tools support - for react-native tools
## Code After:
export PATH=$PATH:/mingw64/bin
export PATH=/cmd:$PATH
export TERM=cygwin
export GOPATH=$HOME/go
export PATH=$PATH:$GOPATH/bin
export PATH="$PATH:$HOME/npm/bin"
export PATH="$PATH:$HOME/npm"
export ANDROID_HOME="$LOCALAPPDATA/Android/Sdk"
export PATH="$PATH:$ANDROID_HOME/platform-tools"
. ~/.dotfiles/env/docker.sh
. ~/.dotfiles/env/git-ssh-forwarding.sh
. ~/.dotfiles/env/git.sh
|
c6f913f5c74c54ec839cf78ca6d42112246617af | docs/source/releases/v1.1.rst | docs/source/releases/v1.1.rst | ===========================================
Oscar 1.1 release notes - UNDER DEVELOPMENT
===========================================
:release: TBD
.. warning::
Since v1.1 has not been released yet, these release notes are still a work-in-progress.
Welcome to Oscar 1.1!
Table of contents:
.. contents::
:local:
:depth: 1
.. _compatibility_of_1.1:
Compatibility
-------------
.. _new_in_1.1:
What's new in Oscar 1.1?
------------------------
.. _minor_changes_in_1.1:
Minor changes
~~~~~~~~~~~~~
.. _incompatible_changes_in_1.1:
Backwards incompatible changes in 1.1
-------------------------------------
Misc
~~~~
Migrations
~~~~~~~~~~
.. _deprecated_features_in_1.1:
Deprecated features
~~~~~~~~~~~~~~~~~~~
The following features have been deprecated in this release:
Removal of deprecated features
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
These methods have been removed:
These classes have been removed:
| ===========================================
Oscar 1.1 release notes - UNDER DEVELOPMENT
===========================================
:release: TBD
.. warning::
Since v1.1 has not been released yet, these release notes are still a work-in-progress.
Welcome to Oscar 1.1!
Table of contents:
.. contents::
:local:
:depth: 1
.. _compatibility_of_1.1:
Compatibility
-------------
.. _new_in_1.1:
What's new in Oscar 1.1?
------------------------
.. _minor_changes_in_1.1:
Minor changes
~~~~~~~~~~~~~
.. _incompatible_changes_in_1.1:
Backwards incompatible changes in 1.1
-------------------------------------
Misc
~~~~
* The :class:`~oscar.apps.shipping.abstract_models.AbstractWeightBased` shipping
method now allows zero-weight baskets to have a non-zero shipping cost
(`#1565`_). This means that sites that rely on zero-weight baskets having no
change will need to introduce a new weight band that covers this edge case.
.. _`#1565`: https://github.com/django-oscar/django-oscar/pull/1565
Migrations
~~~~~~~~~~
.. _deprecated_features_in_1.1:
Deprecated features
~~~~~~~~~~~~~~~~~~~
The following features have been deprecated in this release:
Removal of deprecated features
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
These methods have been removed:
These classes have been removed:
| Update release note for change to shipping method | Update release note for change to shipping method
| reStructuredText | bsd-3-clause | pasqualguerrero/django-oscar,WillisXChen/django-oscar,spartonia/django-oscar,adamend/django-oscar,thechampanurag/django-oscar,sasha0/django-oscar,john-parton/django-oscar,bnprk/django-oscar,nickpack/django-oscar,binarydud/django-oscar,MatthewWilkes/django-oscar,ademuk/django-oscar,sonofatailor/django-oscar,jlmadurga/django-oscar,saadatqadri/django-oscar,sonofatailor/django-oscar,eddiep1101/django-oscar,ka7eh/django-oscar,rocopartners/django-oscar,bschuon/django-oscar,nickpack/django-oscar,michaelkuty/django-oscar,rocopartners/django-oscar,pasqualguerrero/django-oscar,jinnykoo/christmas,dongguangming/django-oscar,michaelkuty/django-oscar,eddiep1101/django-oscar,saadatqadri/django-oscar,manevant/django-oscar,kapari/django-oscar,jmt4/django-oscar,sasha0/django-oscar,anentropic/django-oscar,thechampanurag/django-oscar,vovanbo/django-oscar,WadeYuChen/django-oscar,pasqualguerrero/django-oscar,spartonia/django-oscar,spartonia/django-oscar,bschuon/django-oscar,QLGu/django-oscar,mexeniz/django-oscar,monikasulik/django-oscar,jlmadurga/django-oscar,kapari/django-oscar,MatthewWilkes/django-oscar,ademuk/django-oscar,lijoantony/django-oscar,ka7eh/django-oscar,jlmadurga/django-oscar,pdonadeo/django-oscar,Jannes123/django-oscar,nfletton/django-oscar,rocopartners/django-oscar,amirrpp/django-oscar,thechampanurag/django-oscar,Jannes123/django-oscar,kapari/django-oscar,WillisXChen/django-oscar,okfish/django-oscar,itbabu/django-oscar,ademuk/django-oscar,josesanch/django-oscar,jinnykoo/wuyisj.com,jlmadurga/django-oscar,adamend/django-oscar,sasha0/django-oscar,jinnykoo/wuyisj.com,lijoantony/django-oscar,Jannes123/django-oscar,anentropic/django-oscar,WillisXChen/django-oscar,WillisXChen/django-oscar,rocopartners/django-oscar,spartonia/django-oscar,adamend/django-oscar,vovanbo/django-oscar,nickpack/django-oscar,dongguangming/django-oscar,pdonadeo/django-oscar,Bogh/django-oscar,pasqualguerrero/django-oscar,john-parton/django-oscar,Bogh/django-oscar,anentropic/django-oscar,taedori81/django-oscar,monikasulik/django-oscar,jinnykoo/wuyisj,itbabu/django-oscar,manevant/django-oscar,okfish/django-oscar,django-oscar/django-oscar,monikasulik/django-oscar,jmt4/django-oscar,solarissmoke/django-oscar,pdonadeo/django-oscar,jinnykoo/wuyisj.com,mexeniz/django-oscar,adamend/django-oscar,kapari/django-oscar,bschuon/django-oscar,okfish/django-oscar,taedori81/django-oscar,ka7eh/django-oscar,machtfit/django-oscar,josesanch/django-oscar,anentropic/django-oscar,sonofatailor/django-oscar,binarydud/django-oscar,WillisXChen/django-oscar,WadeYuChen/django-oscar,mexeniz/django-oscar,django-oscar/django-oscar,amirrpp/django-oscar,jmt4/django-oscar,QLGu/django-oscar,john-parton/django-oscar,django-oscar/django-oscar,michaelkuty/django-oscar,machtfit/django-oscar,nfletton/django-oscar,nfletton/django-oscar,josesanch/django-oscar,jinnykoo/wuyisj,eddiep1101/django-oscar,solarissmoke/django-oscar,binarydud/django-oscar,solarissmoke/django-oscar,faratro/django-oscar,WadeYuChen/django-oscar,jinnykoo/wuyisj,django-oscar/django-oscar,vovanbo/django-oscar,binarydud/django-oscar,MatthewWilkes/django-oscar,sonofatailor/django-oscar,solarissmoke/django-oscar,nickpack/django-oscar,thechampanurag/django-oscar,dongguangming/django-oscar,mexeniz/django-oscar,nfletton/django-oscar,jinnykoo/wuyisj.com,WillisXChen/django-oscar,QLGu/django-oscar,john-parton/django-oscar,pdonadeo/django-oscar,machtfit/django-oscar,monikasulik/django-oscar,manevant/django-oscar,lijoantony/django-oscar,bnprk/django-oscar,saadatqadri/django-oscar,ka7eh/django-oscar,jinnykoo/christmas,ademuk/django-oscar,itbabu/django-oscar,Jannes123/django-oscar,itbabu/django-oscar,amirrpp/django-oscar,MatthewWilkes/django-oscar,okfish/django-oscar,WadeYuChen/django-oscar,faratro/django-oscar,taedori81/django-oscar,jmt4/django-oscar,saadatqadri/django-oscar,Bogh/django-oscar,jinnykoo/wuyisj,vovanbo/django-oscar,faratro/django-oscar,dongguangming/django-oscar,faratro/django-oscar,taedori81/django-oscar,michaelkuty/django-oscar,sasha0/django-oscar,bnprk/django-oscar,eddiep1101/django-oscar,Bogh/django-oscar,amirrpp/django-oscar,manevant/django-oscar,lijoantony/django-oscar,bschuon/django-oscar,QLGu/django-oscar,jinnykoo/christmas,bnprk/django-oscar | restructuredtext | ## Code Before:
===========================================
Oscar 1.1 release notes - UNDER DEVELOPMENT
===========================================
:release: TBD
.. warning::
Since v1.1 has not been released yet, these release notes are still a work-in-progress.
Welcome to Oscar 1.1!
Table of contents:
.. contents::
:local:
:depth: 1
.. _compatibility_of_1.1:
Compatibility
-------------
.. _new_in_1.1:
What's new in Oscar 1.1?
------------------------
.. _minor_changes_in_1.1:
Minor changes
~~~~~~~~~~~~~
.. _incompatible_changes_in_1.1:
Backwards incompatible changes in 1.1
-------------------------------------
Misc
~~~~
Migrations
~~~~~~~~~~
.. _deprecated_features_in_1.1:
Deprecated features
~~~~~~~~~~~~~~~~~~~
The following features have been deprecated in this release:
Removal of deprecated features
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
These methods have been removed:
These classes have been removed:
## Instruction:
Update release note for change to shipping method
## Code After:
===========================================
Oscar 1.1 release notes - UNDER DEVELOPMENT
===========================================
:release: TBD
.. warning::
Since v1.1 has not been released yet, these release notes are still a work-in-progress.
Welcome to Oscar 1.1!
Table of contents:
.. contents::
:local:
:depth: 1
.. _compatibility_of_1.1:
Compatibility
-------------
.. _new_in_1.1:
What's new in Oscar 1.1?
------------------------
.. _minor_changes_in_1.1:
Minor changes
~~~~~~~~~~~~~
.. _incompatible_changes_in_1.1:
Backwards incompatible changes in 1.1
-------------------------------------
Misc
~~~~
* The :class:`~oscar.apps.shipping.abstract_models.AbstractWeightBased` shipping
method now allows zero-weight baskets to have a non-zero shipping cost
(`#1565`_). This means that sites that rely on zero-weight baskets having no
change will need to introduce a new weight band that covers this edge case.
.. _`#1565`: https://github.com/django-oscar/django-oscar/pull/1565
Migrations
~~~~~~~~~~
.. _deprecated_features_in_1.1:
Deprecated features
~~~~~~~~~~~~~~~~~~~
The following features have been deprecated in this release:
Removal of deprecated features
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
These methods have been removed:
These classes have been removed:
|
190a71eed1e158ad61b44f8f7c51b764f52264fe | src/js/plugins/bandcamp.js | src/js/plugins/bandcamp.js | "use strict";
var $ = require("jquery");
var Plugin = require("../modules/Plugin");
var Utils = require("../modules/Utilities");
var bandcamp = Object.create(Plugin);
bandcamp.init("bandcamp", "Bandcamp");
bandcamp.scrape = function () {
var discover, info, isTrack, pageTitle;
discover = window.location.pathname.slice(1) === "discover";
info = {
stopped: !$(".inline_player .playbutton").hasClass("playing")
};
if (!info.stopped) {
if (discover) {
info.artist = $("#detail_body_container .itemsubtext a").text();
} else {
info.artist = $("span[itemprop=byArtist]").text();
}
info.title = $(".track_info .title").first().text();
info.duration = Utils.calculateDuration($(".inline_player .track_info .time_total").text());
info.elapsed = Utils.calculateDuration($(".inline_player .track_info .time_elapsed").text());
if (!info.title) {
info.title = $(".trackTitle").first().text();
}
}
return info;
};
module.exports = bandcamp; | "use strict";
var $ = require("jquery");
var Plugin = require("../modules/Plugin");
var Utils = require("../modules/Utilities");
var bandcamp = Object.create(Plugin);
bandcamp.init("bandcamp", "Bandcamp");
bandcamp.scrape = function () {
var discover, info, isTrack, isAlbum, pageTitle;
discover = window.location.hash === "#discover";
isTrack = window.location.pathname.split("/")[1] === "track";
isAlbum = window.location.pathname.split("/")[1] === "album";
info = {
stopped: !$(".inline_player .playbutton").hasClass("playing")
};
if (!info.stopped) {
if (discover) {
info.artist = $("#detail_body_container .itemsubtext a").text();
} else {
info.artist = $("span[itemprop=byArtist]").text();
}
if (isTrack) {
info.album = $("#name-section h3 span a span").text();
} else if (isAlbum) {
info.album = $("#name-section > h2").text().trim();
} else if (discover) {
info.album = $("div.itemtext a").filter(function (i, e) {
return e.href.match(/album/)}).text().trim();
}
info.title = $(".track_info .title").first().text();
info.duration = Utils.calculateDuration($(".inline_player .track_info .time_total").text());
info.elapsed = Utils.calculateDuration($(".inline_player .track_info .time_elapsed").text());
if (!info.title) {
info.title = $(".trackTitle").first().text();
}
}
return info;
};
module.exports = bandcamp; | Add album support, fix Discover detection | Add album support, fix Discover detection
| JavaScript | mit | m9x/np2file,nsue/scroblr,nwalker/scroblr,Nicholi/scroblr,Amir-Saeed/scroblr,Nicholi/scroblr,nsue/scroblr,m9x/np2file,mziyut/scroblr,cgravolet/scroblr,Amir-Saeed/scroblr,cgravolet/scroblr,nwalker/scroblr,bdukes/scroblr,mziyut/scroblr,bdukes/scroblr | javascript | ## Code Before:
"use strict";
var $ = require("jquery");
var Plugin = require("../modules/Plugin");
var Utils = require("../modules/Utilities");
var bandcamp = Object.create(Plugin);
bandcamp.init("bandcamp", "Bandcamp");
bandcamp.scrape = function () {
var discover, info, isTrack, pageTitle;
discover = window.location.pathname.slice(1) === "discover";
info = {
stopped: !$(".inline_player .playbutton").hasClass("playing")
};
if (!info.stopped) {
if (discover) {
info.artist = $("#detail_body_container .itemsubtext a").text();
} else {
info.artist = $("span[itemprop=byArtist]").text();
}
info.title = $(".track_info .title").first().text();
info.duration = Utils.calculateDuration($(".inline_player .track_info .time_total").text());
info.elapsed = Utils.calculateDuration($(".inline_player .track_info .time_elapsed").text());
if (!info.title) {
info.title = $(".trackTitle").first().text();
}
}
return info;
};
module.exports = bandcamp;
## Instruction:
Add album support, fix Discover detection
## Code After:
"use strict";
var $ = require("jquery");
var Plugin = require("../modules/Plugin");
var Utils = require("../modules/Utilities");
var bandcamp = Object.create(Plugin);
bandcamp.init("bandcamp", "Bandcamp");
bandcamp.scrape = function () {
var discover, info, isTrack, isAlbum, pageTitle;
discover = window.location.hash === "#discover";
isTrack = window.location.pathname.split("/")[1] === "track";
isAlbum = window.location.pathname.split("/")[1] === "album";
info = {
stopped: !$(".inline_player .playbutton").hasClass("playing")
};
if (!info.stopped) {
if (discover) {
info.artist = $("#detail_body_container .itemsubtext a").text();
} else {
info.artist = $("span[itemprop=byArtist]").text();
}
if (isTrack) {
info.album = $("#name-section h3 span a span").text();
} else if (isAlbum) {
info.album = $("#name-section > h2").text().trim();
} else if (discover) {
info.album = $("div.itemtext a").filter(function (i, e) {
return e.href.match(/album/)}).text().trim();
}
info.title = $(".track_info .title").first().text();
info.duration = Utils.calculateDuration($(".inline_player .track_info .time_total").text());
info.elapsed = Utils.calculateDuration($(".inline_player .track_info .time_elapsed").text());
if (!info.title) {
info.title = $(".trackTitle").first().text();
}
}
return info;
};
module.exports = bandcamp; |
4870b09309293fb053f743068f603747177088eb | org.metaborg.spoofax.core/src/main/java/org/metaborg/spoofax/core/messages/ISourceRegion.java | org.metaborg.spoofax.core/src/main/java/org/metaborg/spoofax/core/messages/ISourceRegion.java | package org.metaborg.spoofax.core.messages;
public interface ISourceRegion {
public int startOffset();
public int startRow();
public int startColumn();
public int endOffset();
public int endRow();
public int endColumn();
}
| package org.metaborg.spoofax.core.messages;
/**
* Interface for representing a finite region in source code text. A region has:
* <ul>
* <li>Offset - number of characters from the beginning of the source text, with interval [0,#chars).</li>
* <li>Row - row or line in the source text, with interval [0,#rows).</li>
* <li>Column - column in the source text, with interval [0,#columns@row).</li>
* </ul>
* Both the starting and ending locations for these numbers are available.
*/
public interface ISourceRegion {
public int startOffset();
public int startRow();
public int startColumn();
public int endOffset();
public int endRow();
public int endColumn();
}
| Clarify intervals of source region. | Clarify intervals of source region.
| Java | apache-2.0 | metaborg/spoofax-eclipse,metaborg/spoofax,metaborg/spoofax,metaborg/spoofax,metaborg/spoofax | java | ## Code Before:
package org.metaborg.spoofax.core.messages;
public interface ISourceRegion {
public int startOffset();
public int startRow();
public int startColumn();
public int endOffset();
public int endRow();
public int endColumn();
}
## Instruction:
Clarify intervals of source region.
## Code After:
package org.metaborg.spoofax.core.messages;
/**
* Interface for representing a finite region in source code text. A region has:
* <ul>
* <li>Offset - number of characters from the beginning of the source text, with interval [0,#chars).</li>
* <li>Row - row or line in the source text, with interval [0,#rows).</li>
* <li>Column - column in the source text, with interval [0,#columns@row).</li>
* </ul>
* Both the starting and ending locations for these numbers are available.
*/
public interface ISourceRegion {
public int startOffset();
public int startRow();
public int startColumn();
public int endOffset();
public int endRow();
public int endColumn();
}
|
cdc915a97a99f28332e24c4b1cac9a4c87466233 | README.md | README.md |
Query the host networking when working in a Docker environment on Rackspace
```
$ docker run --net=host racknet/ip public
104.130.0.127
```
```
$ docker run --net=host racknet/ip service
10.184.13.252
```
## Usage
```
racknet public [ipv4|ipv6]
racknet service [ipv4|ipv6]
Examples:
$ racknet public
104.130.0.127
$ racknet service ipv6
fe80::be76:4eff:fe20:b452
Examples when run with Docker:
$ docker run --net=host racknet/ip public
104.130.0.127
```
|
Query the host networking when working in a Docker environment on Rackspace
```
$ docker run --net=host --rm racknet/ip public
104.130.0.127
```
```
$ docker run --net=host --rm racknet/ip service
10.184.13.252
```
## Usage
```
racknet public [ipv4|ipv6]
racknet service [ipv4|ipv6]
Examples:
$ racknet public
104.130.0.127
$ racknet service ipv6
fe80::be76:4eff:fe20:b452
Examples when run with Docker:
$ docker run --net=host racknet/ip public
104.130.0.127
```
| Add --rm flag, courtesy decaturdev | Add --rm flag, courtesy decaturdev | Markdown | apache-2.0 | rgbkrk/docker-rackspace-net | markdown | ## Code Before:
Query the host networking when working in a Docker environment on Rackspace
```
$ docker run --net=host racknet/ip public
104.130.0.127
```
```
$ docker run --net=host racknet/ip service
10.184.13.252
```
## Usage
```
racknet public [ipv4|ipv6]
racknet service [ipv4|ipv6]
Examples:
$ racknet public
104.130.0.127
$ racknet service ipv6
fe80::be76:4eff:fe20:b452
Examples when run with Docker:
$ docker run --net=host racknet/ip public
104.130.0.127
```
## Instruction:
Add --rm flag, courtesy decaturdev
## Code After:
Query the host networking when working in a Docker environment on Rackspace
```
$ docker run --net=host --rm racknet/ip public
104.130.0.127
```
```
$ docker run --net=host --rm racknet/ip service
10.184.13.252
```
## Usage
```
racknet public [ipv4|ipv6]
racknet service [ipv4|ipv6]
Examples:
$ racknet public
104.130.0.127
$ racknet service ipv6
fe80::be76:4eff:fe20:b452
Examples when run with Docker:
$ docker run --net=host racknet/ip public
104.130.0.127
```
|
1e22a392f818a5b309342500b5a7f81fa063ff3b | cp-content-shell.sh | cp-content-shell.sh | rm -rf ~/content-shell
rm -f ~/content-shell.tar.gz
mkdir ~/content-shell
cd ~/src/out/Default
cp icudtl.dat ~/content-shell
cp *.bin ~/content-shell
cp *.pak ~/content-shell
cp *.so ~/content-shell
cp content_shell ~/content-shell
cd ~
tar -zvcf content-shell.tar.gz content-shell
| rm -rf ~/content-shell
rm -f ~/content-shell.tar.gz
mkdir ~/content-shell
cd ~/src/out/Default
cp icudtl.dat ~/content-shell
cp *.bin ~/content-shell
cp *.pak ~/content-shell
cp *.so ~/content-shell
cp content_shell ~/content-shell
cp chrome ~/content-shell
cp -r locales ~/content-shell
cp -r resources ~/content-shell
cp -r ui ~/content-shell
cd ~
tar -zvcf content-shell.tar.gz content-shell
| Add needed files for chrome binary | Add needed files for chrome binary
| Shell | mit | bjwbell/ironframe,bjwbell/ironframe,bjwbell/ironframe,bjwbell/ironframe,bjwbell/ironframe | shell | ## Code Before:
rm -rf ~/content-shell
rm -f ~/content-shell.tar.gz
mkdir ~/content-shell
cd ~/src/out/Default
cp icudtl.dat ~/content-shell
cp *.bin ~/content-shell
cp *.pak ~/content-shell
cp *.so ~/content-shell
cp content_shell ~/content-shell
cd ~
tar -zvcf content-shell.tar.gz content-shell
## Instruction:
Add needed files for chrome binary
## Code After:
rm -rf ~/content-shell
rm -f ~/content-shell.tar.gz
mkdir ~/content-shell
cd ~/src/out/Default
cp icudtl.dat ~/content-shell
cp *.bin ~/content-shell
cp *.pak ~/content-shell
cp *.so ~/content-shell
cp content_shell ~/content-shell
cp chrome ~/content-shell
cp -r locales ~/content-shell
cp -r resources ~/content-shell
cp -r ui ~/content-shell
cd ~
tar -zvcf content-shell.tar.gz content-shell
|
8f1c475347ca637417ede6b573a81aa4e8145c0b | .travis.yml | .travis.yml | language: generic
services:
- docker
cache:
directories:
- $HOME/.ccache
env:
global:
- CCACHE_DIR=$HOME/.ccache
matrix:
- ROS_DISTRO="kinetic" ROS_REPO=ros UPSTREAM_WORKSPACE=file CATKIN_LINT=true
- ROS_DISTRO="melodic" ROS_REPO=ros UPSTREAM_WORKSPACE=file CATKIN_LINT=true
- ROS_DISTRO="kinetic" CLANG_FORMAT_CHECK=file CLANG_FORMAT_VERSION=6.0
install:
- git clone --depth=1 https://github.com/ros-industrial/industrial_ci.git .ci_config
script:
- .ci_config/travis.sh
| language: generic
services:
- docker
cache:
directories:
- $HOME/.ccache
env:
global:
- CCACHE_DIR=$HOME/.ccache
matrix:
- ROS_DISTRO="kinetic" ROS_REPO=ros UPSTREAM_WORKSPACE=file
- ROS_DISTRO="melodic" ROS_REPO=ros UPSTREAM_WORKSPACE=file
- ROS_DISTRO="kinetic" CLANG_FORMAT_CHECK=file CLANG_FORMAT_VERSION=6.0
install:
- git clone --depth=1 https://github.com/ros-industrial/industrial_ci.git .ci_config
script:
- .ci_config/travis.sh
| Revert "explicitly add catkin_lint to the ci pipeline" | Revert "explicitly add catkin_lint to the ci pipeline"
This reverts commit 3108a45a6ea3fa81533e7e50f62d3b1bfdc8f864.
| YAML | apache-2.0 | UniversalRobots/Universal_Robots_Client_Library,UniversalRobots/Universal_Robots_Client_Library | yaml | ## Code Before:
language: generic
services:
- docker
cache:
directories:
- $HOME/.ccache
env:
global:
- CCACHE_DIR=$HOME/.ccache
matrix:
- ROS_DISTRO="kinetic" ROS_REPO=ros UPSTREAM_WORKSPACE=file CATKIN_LINT=true
- ROS_DISTRO="melodic" ROS_REPO=ros UPSTREAM_WORKSPACE=file CATKIN_LINT=true
- ROS_DISTRO="kinetic" CLANG_FORMAT_CHECK=file CLANG_FORMAT_VERSION=6.0
install:
- git clone --depth=1 https://github.com/ros-industrial/industrial_ci.git .ci_config
script:
- .ci_config/travis.sh
## Instruction:
Revert "explicitly add catkin_lint to the ci pipeline"
This reverts commit 3108a45a6ea3fa81533e7e50f62d3b1bfdc8f864.
## Code After:
language: generic
services:
- docker
cache:
directories:
- $HOME/.ccache
env:
global:
- CCACHE_DIR=$HOME/.ccache
matrix:
- ROS_DISTRO="kinetic" ROS_REPO=ros UPSTREAM_WORKSPACE=file
- ROS_DISTRO="melodic" ROS_REPO=ros UPSTREAM_WORKSPACE=file
- ROS_DISTRO="kinetic" CLANG_FORMAT_CHECK=file CLANG_FORMAT_VERSION=6.0
install:
- git clone --depth=1 https://github.com/ros-industrial/industrial_ci.git .ci_config
script:
- .ci_config/travis.sh
|
b064a09395534c553a385fb81ca7548359747ba1 | tools/Setup.hx | tools/Setup.hx | import com.haxepunk.utils.HaxelibInfo;
class Setup
{
public static function setup()
{
installDependencies();
Sys.command('haxelib run $OPENFL setup');
}
public static function update()
{
Sys.command("haxelib update HaxePunk");
installDependencies();
}
public static function installDependencies()
{
for (field in Reflect.fields(HaxelibInfo.install))
{
Sys.command("haxelib install " + field + " " + Reflect.field(HaxelibInfo.install, field));
}
}
private static inline var OPENFL = "lime";
}
| import com.haxepunk.utils.HaxelibInfo;
class Setup
{
public static function setup()
{
installDependencies();
Sys.command("haxelib", ["run", OPENFL, "setup"]);
}
public static function update()
{
Sys.command("haxelib" ["update", "HaxePunk"]);
installDependencies();
}
public static function installDependencies()
{
for (field in Reflect.fields(HaxelibInfo.install))
{
Sys.command("haxelib" ["install", field, Reflect.field(HaxelibInfo.install, field)]);
}
}
private static inline var OPENFL = "lime";
}
| Fix command call in setup tool | Fix command call in setup tool | Haxe | mit | scriptorum/HaxePunk,wselken/HaxePunk,scriptorum/HaxePunk,HaxePunk/HaxePunk,MattTuttle/HaxePunk | haxe | ## Code Before:
import com.haxepunk.utils.HaxelibInfo;
class Setup
{
public static function setup()
{
installDependencies();
Sys.command('haxelib run $OPENFL setup');
}
public static function update()
{
Sys.command("haxelib update HaxePunk");
installDependencies();
}
public static function installDependencies()
{
for (field in Reflect.fields(HaxelibInfo.install))
{
Sys.command("haxelib install " + field + " " + Reflect.field(HaxelibInfo.install, field));
}
}
private static inline var OPENFL = "lime";
}
## Instruction:
Fix command call in setup tool
## Code After:
import com.haxepunk.utils.HaxelibInfo;
class Setup
{
public static function setup()
{
installDependencies();
Sys.command("haxelib", ["run", OPENFL, "setup"]);
}
public static function update()
{
Sys.command("haxelib" ["update", "HaxePunk"]);
installDependencies();
}
public static function installDependencies()
{
for (field in Reflect.fields(HaxelibInfo.install))
{
Sys.command("haxelib" ["install", field, Reflect.field(HaxelibInfo.install, field)]);
}
}
private static inline var OPENFL = "lime";
}
|
8658ce434085f0262a2e3e72e7014ac8c459c180 | HISTORY.md | HISTORY.md | 3.2.0 / unreleased
==================
* new: Ukrainian support
3.1.0 / 2015-06-25
==================
* new: Greek support
3.0.0 / 2015-05-29
==================
* new: `largest` option to get largest units
* update: languages can change decimal point (which can be overridden)
* update: fix some unintuitive `round` behavior
* update: specify units in shorthand rather than longhand
* update: use the Unlicense
* update: `nob` is now `no`
* update: use underscores instead of dashes for Chinese
* remove: half units
2.8.0 / 2015-05-10
==================
* new: `getSupportedLanguages`
2.7.0 / 2015-05-08
==================
* new: changelog
* new: Arabic support
* new: Italian support
* remove: (spoken) languages from package.json keywords
| 3.2.0 / 2015-07-15
==================
* new: Ukrainian support
* fix: things would break in global strict mode
3.1.0 / 2015-06-25
==================
* new: Greek support
3.0.0 / 2015-05-29
==================
* new: `largest` option to get largest units
* update: languages can change decimal point (which can be overridden)
* update: fix some unintuitive `round` behavior
* update: specify units in shorthand rather than longhand
* update: use the Unlicense
* update: `nob` is now `no`
* update: use underscores instead of dashes for Chinese
* remove: half units
2.8.0 / 2015-05-10
==================
* new: `getSupportedLanguages`
2.7.0 / 2015-05-08
==================
* new: changelog
* new: Arabic support
* new: Italian support
* remove: (spoken) languages from package.json keywords
| Update changelog in preparation for 3.2.0 release | Update changelog in preparation for 3.2.0 release
| Markdown | unlicense | EvanHahn/HumanizeDuration.js,smnbbrv/HumanizeDuration.js,TioNoob/HumanizeDuration.js | markdown | ## Code Before:
3.2.0 / unreleased
==================
* new: Ukrainian support
3.1.0 / 2015-06-25
==================
* new: Greek support
3.0.0 / 2015-05-29
==================
* new: `largest` option to get largest units
* update: languages can change decimal point (which can be overridden)
* update: fix some unintuitive `round` behavior
* update: specify units in shorthand rather than longhand
* update: use the Unlicense
* update: `nob` is now `no`
* update: use underscores instead of dashes for Chinese
* remove: half units
2.8.0 / 2015-05-10
==================
* new: `getSupportedLanguages`
2.7.0 / 2015-05-08
==================
* new: changelog
* new: Arabic support
* new: Italian support
* remove: (spoken) languages from package.json keywords
## Instruction:
Update changelog in preparation for 3.2.0 release
## Code After:
3.2.0 / 2015-07-15
==================
* new: Ukrainian support
* fix: things would break in global strict mode
3.1.0 / 2015-06-25
==================
* new: Greek support
3.0.0 / 2015-05-29
==================
* new: `largest` option to get largest units
* update: languages can change decimal point (which can be overridden)
* update: fix some unintuitive `round` behavior
* update: specify units in shorthand rather than longhand
* update: use the Unlicense
* update: `nob` is now `no`
* update: use underscores instead of dashes for Chinese
* remove: half units
2.8.0 / 2015-05-10
==================
* new: `getSupportedLanguages`
2.7.0 / 2015-05-08
==================
* new: changelog
* new: Arabic support
* new: Italian support
* remove: (spoken) languages from package.json keywords
|
eee0f7795d1da18931b27b15abcf17287b02b3af | _data/messages.yaml | _data/messages.yaml | - content: Everything up and running
status: green
timestamp: 2015-04-09 12:00 +0200
| - content: This is a test of rss feed hooks
status: green
timestamp: 2015-04-09 17:40 +0200
- content: Everything up and running
status: green
timestamp: 2015-04-09 12:00 +0200
| Add rss feed test message | Add rss feed test message
| YAML | mit | frigg/status.frigg.io | yaml | ## Code Before:
- content: Everything up and running
status: green
timestamp: 2015-04-09 12:00 +0200
## Instruction:
Add rss feed test message
## Code After:
- content: This is a test of rss feed hooks
status: green
timestamp: 2015-04-09 17:40 +0200
- content: Everything up and running
status: green
timestamp: 2015-04-09 12:00 +0200
|
f319b5ed12e04827aeed9d72d225eb227e948328 | README.md | README.md | DevOps/DX for Linux. Continuous integration snippets and common tasks.
* [ ] Adding automatic build to Linux projects through Travis - [![Travis](https://img.shields.io/travis/yakshaveinc/linux.svg)](https://travis-ci.org/yakshaveinc/linux)
* [ ] Adding build status indicators to your development dashboard
---
* [ ] Adding static site pages through Netlify
* [ ] design previews
* [ ] Django (https://filmfest.netlify.com)
* [ ] host documentation, JAMStack sites (https://jamstack.org/)
* [ ] automatic deployment
* [ ] PR previews
| Linux DevOps snippets, recipes and services.
#### Elements
* [x] GitHub (https://github.com/yakshaveinc/linux)
* [x] Travis CI (https://travis-ci.org/yakshaveinc/linux)
#### Ops
* [x] Adding automatic builds to Linux projects through Travis - [![Travis](https://img.shields.io/travis/yakshaveinc/linux.svg)](https://travis-ci.org/yakshaveinc/linux)
* [ ] Adding build status indicators to your development dashboard
* [ ] Automatic image uploads from Travis CI to DockerHub
---
* [ ] Adding static site pages through Netlify
* [ ] design previews
* [ ] Django (https://filmfest.netlify.com)
* [ ] host documentation, JAMStack sites (https://jamstack.org/)
* [ ] automatic deployment
* [ ] PR previews
| Add DockerHub to the list of services | Add DockerHub to the list of services | Markdown | unlicense | yakshaveinc/linux | markdown | ## Code Before:
DevOps/DX for Linux. Continuous integration snippets and common tasks.
* [ ] Adding automatic build to Linux projects through Travis - [![Travis](https://img.shields.io/travis/yakshaveinc/linux.svg)](https://travis-ci.org/yakshaveinc/linux)
* [ ] Adding build status indicators to your development dashboard
---
* [ ] Adding static site pages through Netlify
* [ ] design previews
* [ ] Django (https://filmfest.netlify.com)
* [ ] host documentation, JAMStack sites (https://jamstack.org/)
* [ ] automatic deployment
* [ ] PR previews
## Instruction:
Add DockerHub to the list of services
## Code After:
Linux DevOps snippets, recipes and services.
#### Elements
* [x] GitHub (https://github.com/yakshaveinc/linux)
* [x] Travis CI (https://travis-ci.org/yakshaveinc/linux)
#### Ops
* [x] Adding automatic builds to Linux projects through Travis - [![Travis](https://img.shields.io/travis/yakshaveinc/linux.svg)](https://travis-ci.org/yakshaveinc/linux)
* [ ] Adding build status indicators to your development dashboard
* [ ] Automatic image uploads from Travis CI to DockerHub
---
* [ ] Adding static site pages through Netlify
* [ ] design previews
* [ ] Django (https://filmfest.netlify.com)
* [ ] host documentation, JAMStack sites (https://jamstack.org/)
* [ ] automatic deployment
* [ ] PR previews
|
329ab24aed5a9848e98d6a504ad8b2997142c23b | lib/main.js | lib/main.js | var buffer = require("./buffer");
/**
* @namespace Splat
*/
module.exports = {
makeBuffer: buffer.makeBuffer,
flipBufferHorizontally: buffer.flipBufferHorizontally,
flipBufferVertically: buffer.flipBufferVertically,
ads: require("./ads"),
AStar: require("./astar"),
BinaryHeap: require("./binary-heap"),
Game: require("./game"),
iap: require("./iap"),
Input: require("./input"),
leaderboards: require("./leaderboards"),
math: require("./math"),
openUrl: require("./openUrl"),
NinePatch: require("./ninepatch"),
Particles: require("./particles"),
saveData: require("./save-data"),
Scene: require("./scene"),
components: {
animation: require("./components/animation"),
camera: require("./components/camera"),
friction: require("./components/friction"),
image: require("./components/image"),
movement2d: require("./components/movement-2d"),
playableArea: require("./components/playable-area"),
playerController2d: require("./components/player-controller-2d"),
position: require("./components/position"),
size: require("./components/size"),
timers: require("./components/timers"),
velocity: require("./components/velocity")
}
};
| var buffer = require("./buffer");
/**
* @namespace Splat
*/
module.exports = {
makeBuffer: buffer.makeBuffer,
flipBufferHorizontally: buffer.flipBufferHorizontally,
flipBufferVertically: buffer.flipBufferVertically,
ads: require("./ads"),
AStar: require("./astar"),
BinaryHeap: require("./binary-heap"),
Game: require("./game"),
iap: require("./iap"),
Input: require("./input"),
leaderboards: require("./leaderboards"),
math: require("./math"),
openUrl: require("./openUrl"),
NinePatch: require("./ninepatch"),
Particles: require("./particles"),
saveData: require("./save-data"),
Scene: require("./scene")
};
| Remove big export of components | Remove big export of components
| JavaScript | mit | SplatJS/splat-ecs | javascript | ## Code Before:
var buffer = require("./buffer");
/**
* @namespace Splat
*/
module.exports = {
makeBuffer: buffer.makeBuffer,
flipBufferHorizontally: buffer.flipBufferHorizontally,
flipBufferVertically: buffer.flipBufferVertically,
ads: require("./ads"),
AStar: require("./astar"),
BinaryHeap: require("./binary-heap"),
Game: require("./game"),
iap: require("./iap"),
Input: require("./input"),
leaderboards: require("./leaderboards"),
math: require("./math"),
openUrl: require("./openUrl"),
NinePatch: require("./ninepatch"),
Particles: require("./particles"),
saveData: require("./save-data"),
Scene: require("./scene"),
components: {
animation: require("./components/animation"),
camera: require("./components/camera"),
friction: require("./components/friction"),
image: require("./components/image"),
movement2d: require("./components/movement-2d"),
playableArea: require("./components/playable-area"),
playerController2d: require("./components/player-controller-2d"),
position: require("./components/position"),
size: require("./components/size"),
timers: require("./components/timers"),
velocity: require("./components/velocity")
}
};
## Instruction:
Remove big export of components
## Code After:
var buffer = require("./buffer");
/**
* @namespace Splat
*/
module.exports = {
makeBuffer: buffer.makeBuffer,
flipBufferHorizontally: buffer.flipBufferHorizontally,
flipBufferVertically: buffer.flipBufferVertically,
ads: require("./ads"),
AStar: require("./astar"),
BinaryHeap: require("./binary-heap"),
Game: require("./game"),
iap: require("./iap"),
Input: require("./input"),
leaderboards: require("./leaderboards"),
math: require("./math"),
openUrl: require("./openUrl"),
NinePatch: require("./ninepatch"),
Particles: require("./particles"),
saveData: require("./save-data"),
Scene: require("./scene")
};
|
d312446805ee548188bcdf423a854b5a780f8749 | .travis.yml | .travis.yml | language: csharp
dist: trusty
dotnet: 1.0.1
mono: none
env:
global:
- DOTNET_SKIP_FIRST_TIME_EXPERIENCE: 1
- DOTNET_CLI_TELEMETRY_OPTOUT: 1
script:
- ./bin/fetch-configlet
- ./bin/configlet lint .
- ./build.sh
| language: csharp
dist: trusty
dotnet: 1.0.1
mono: none
env:
global:
- DOTNET_SKIP_FIRST_TIME_EXPERIENCE: 1
- DOTNET_CLI_TELEMETRY_OPTOUT: 1
script:
- ./bin/fetch-configlet
- ./bin/configlet lint .
- ./build.sh
cache:
directories:
- $HOME/.nuget/packages
| Add NuGet package caching to Travis script | Add NuGet package caching to Travis script | YAML | mit | ErikSchierboom/xcsharp,GKotfis/csharp,robkeim/xcsharp,robkeim/xcsharp,GKotfis/csharp,exercism/xcsharp,exercism/xcsharp,ErikSchierboom/xcsharp | yaml | ## Code Before:
language: csharp
dist: trusty
dotnet: 1.0.1
mono: none
env:
global:
- DOTNET_SKIP_FIRST_TIME_EXPERIENCE: 1
- DOTNET_CLI_TELEMETRY_OPTOUT: 1
script:
- ./bin/fetch-configlet
- ./bin/configlet lint .
- ./build.sh
## Instruction:
Add NuGet package caching to Travis script
## Code After:
language: csharp
dist: trusty
dotnet: 1.0.1
mono: none
env:
global:
- DOTNET_SKIP_FIRST_TIME_EXPERIENCE: 1
- DOTNET_CLI_TELEMETRY_OPTOUT: 1
script:
- ./bin/fetch-configlet
- ./bin/configlet lint .
- ./build.sh
cache:
directories:
- $HOME/.nuget/packages
|
7b57656633836b48a0565a2559a9c3709d93ce5b | composer.json | composer.json | {
"name": "tomkyle/database-servicelocator",
"replace": {"tomkyle/databases": "self.version"},
"type": "library",
"description": "A nifty Database ServiceLocator, Pimple-style",
"keywords": ["dependency injection", "database", "service locator", "mysql", "aura", "sql"],
"license": "MIT",
"authors": [
{
"name": "Carsten Witt",
"email": "[email protected]"
}
],
"require": {
"php": ">=5.4.0",
"pimple/pimple": "dev-master"
},
"suggest": {
"aura/sql": "Aura.SQL – Provides adapters for SQL database access around PDO"
},
"autoload": {
"psr-4": {
"tomkyle\\Databases\\": "src/"
}
}
}
| {
"name": "tomkyle/databases",
"replace": {"tomkyle/databases": "self.version"},
"type": "library",
"description": "A nifty Database ServiceLocator, Pimple-style",
"keywords": ["dependency injection", "database", "service locator", "mysql", "aura", "sql"],
"license": "MIT",
"authors": [
{
"name": "Carsten Witt",
"email": "[email protected]"
}
],
"require": {
"php": ">=5.4.0",
"pimple/pimple": "dev-master"
},
"suggest": {
"aura/sql": "Aura.SQL – Provides adapters for SQL database access around PDO"
},
"autoload": {
"psr-4": {
"tomkyle\\Databases\\": "src/"
}
}
}
| Update with 'new old' package name… | Update with 'new old' package name… | JSON | mit | tomkyle/Databases | json | ## Code Before:
{
"name": "tomkyle/database-servicelocator",
"replace": {"tomkyle/databases": "self.version"},
"type": "library",
"description": "A nifty Database ServiceLocator, Pimple-style",
"keywords": ["dependency injection", "database", "service locator", "mysql", "aura", "sql"],
"license": "MIT",
"authors": [
{
"name": "Carsten Witt",
"email": "[email protected]"
}
],
"require": {
"php": ">=5.4.0",
"pimple/pimple": "dev-master"
},
"suggest": {
"aura/sql": "Aura.SQL – Provides adapters for SQL database access around PDO"
},
"autoload": {
"psr-4": {
"tomkyle\\Databases\\": "src/"
}
}
}
## Instruction:
Update with 'new old' package name…
## Code After:
{
"name": "tomkyle/databases",
"replace": {"tomkyle/databases": "self.version"},
"type": "library",
"description": "A nifty Database ServiceLocator, Pimple-style",
"keywords": ["dependency injection", "database", "service locator", "mysql", "aura", "sql"],
"license": "MIT",
"authors": [
{
"name": "Carsten Witt",
"email": "[email protected]"
}
],
"require": {
"php": ">=5.4.0",
"pimple/pimple": "dev-master"
},
"suggest": {
"aura/sql": "Aura.SQL – Provides adapters for SQL database access around PDO"
},
"autoload": {
"psr-4": {
"tomkyle\\Databases\\": "src/"
}
}
}
|
24faff8b034a8d57a9e67249dbf182c2c1c8ca9c | .github/workflows/build_msw_xml_libs.yml | .github/workflows/build_msw_xml_libs.yml | name: XML MSW binaries
on:
workflow_dispatch:
jobs:
build-xml-libs:
name: Build XML binaries
runs-on: ubuntu-latest
steps:
- name: Checkout
uses: actions/checkout@v2
- name: Install Compiler
run: |
sudo apt-get update -qq
sudo apt-get install -qq --no-install-recommends g++-mingw-w64-i686 g++-mingw-w64-x86-64
- name: Build 32-bit Libraries
run: |
HOST=i686-w64-mingw32 ./scripts/ci/install.sh
- name: Build 64-bit Libraries
run: |
HOST=x86_64-w64-mingw32 ./scripts/ci/install.sh
- name: Upload
uses: actions/upload-artifact@v3
with:
name: xml-libs
path: /usr/local
| name: XML MSW binaries
on:
workflow_dispatch:
jobs:
build-xml-libs:
name: Build XML binaries
runs-on: ubuntu-latest
steps:
- name: Checkout
uses: actions/checkout@v2
- name: Install Compiler
run: |
sudo apt-get update -qq
sudo apt-get install -qq --no-install-recommends g++-mingw-w64-i686 g++-mingw-w64-x86-64
- name: Build 32-bit Libraries
run: |
HOST=i686-w64-mingw32 ./scripts/install_deps.sh
- name: Build 64-bit Libraries
run: |
HOST=x86_64-w64-mingw32 ./scripts/install_deps.sh
- name: Upload
uses: actions/upload-artifact@v3
with:
name: xml-libs
path: /usr/local
| Fix wrong script name used in the workflow file | Fix wrong script name used in the workflow file
| YAML | bsd-3-clause | vslavik/xmlwrapp,vslavik/xmlwrapp,vslavik/xmlwrapp | yaml | ## Code Before:
name: XML MSW binaries
on:
workflow_dispatch:
jobs:
build-xml-libs:
name: Build XML binaries
runs-on: ubuntu-latest
steps:
- name: Checkout
uses: actions/checkout@v2
- name: Install Compiler
run: |
sudo apt-get update -qq
sudo apt-get install -qq --no-install-recommends g++-mingw-w64-i686 g++-mingw-w64-x86-64
- name: Build 32-bit Libraries
run: |
HOST=i686-w64-mingw32 ./scripts/ci/install.sh
- name: Build 64-bit Libraries
run: |
HOST=x86_64-w64-mingw32 ./scripts/ci/install.sh
- name: Upload
uses: actions/upload-artifact@v3
with:
name: xml-libs
path: /usr/local
## Instruction:
Fix wrong script name used in the workflow file
## Code After:
name: XML MSW binaries
on:
workflow_dispatch:
jobs:
build-xml-libs:
name: Build XML binaries
runs-on: ubuntu-latest
steps:
- name: Checkout
uses: actions/checkout@v2
- name: Install Compiler
run: |
sudo apt-get update -qq
sudo apt-get install -qq --no-install-recommends g++-mingw-w64-i686 g++-mingw-w64-x86-64
- name: Build 32-bit Libraries
run: |
HOST=i686-w64-mingw32 ./scripts/install_deps.sh
- name: Build 64-bit Libraries
run: |
HOST=x86_64-w64-mingw32 ./scripts/install_deps.sh
- name: Upload
uses: actions/upload-artifact@v3
with:
name: xml-libs
path: /usr/local
|
8d1929c2b56630889e64a8671ac635633a015632 | README.rst | README.rst | Swoopi
======
.. image:: https://img.shields.io/travis/swoopi/swoopi/master.svg?maxAge=3600?style=flat-square :target: https://travis-ci.org/swoopi/swoopi
.. image:: https://img.shields.io/coveralls/swoopi/swoopi/master.svg?maxAge=3600?style=flat-square :target: https://coveralls.io/github/swoopi/swoopi?branch=master
Camera interactions for the Raspberry Pi.
| Swoopi
======
|Travis|_ |Coveralls|_
Camera interactions for the Raspberry Pi.
.. |Travis| image:: https://img.shields.io/travis/swoopi/swoopi/master.svg?maxAge=3600?style=flat-square
.. _Travis: https://travis-ci.org/swoopi/swoopi
.. |Coveralls| image:: https://img.shields.io/coveralls/swoopi/swoopi/master.svg?maxAge=3600?style=flat-square
.. _Coveralls: https://coveralls.io/github/swoopi/swoopi?branch=master
| Change shields to rst references | Change shields to rst references
| reStructuredText | apache-2.0 | swoopi/swoopi,YouLookFamiliar/swoop | restructuredtext | ## Code Before:
Swoopi
======
.. image:: https://img.shields.io/travis/swoopi/swoopi/master.svg?maxAge=3600?style=flat-square :target: https://travis-ci.org/swoopi/swoopi
.. image:: https://img.shields.io/coveralls/swoopi/swoopi/master.svg?maxAge=3600?style=flat-square :target: https://coveralls.io/github/swoopi/swoopi?branch=master
Camera interactions for the Raspberry Pi.
## Instruction:
Change shields to rst references
## Code After:
Swoopi
======
|Travis|_ |Coveralls|_
Camera interactions for the Raspberry Pi.
.. |Travis| image:: https://img.shields.io/travis/swoopi/swoopi/master.svg?maxAge=3600?style=flat-square
.. _Travis: https://travis-ci.org/swoopi/swoopi
.. |Coveralls| image:: https://img.shields.io/coveralls/swoopi/swoopi/master.svg?maxAge=3600?style=flat-square
.. _Coveralls: https://coveralls.io/github/swoopi/swoopi?branch=master
|
02b60cc05265429b90ac7289086661a7d8bd7814 | pac_script.js | pac_script.js | var BOARD_HEIGHT = 288;
var BOARD_WIDTH = 224;
var VERT_TILES = BOARD_HEIGHT / 8;
var HORIZ_TILES = BOARD_WIDTH / 8;
gameBoard = new Array(VERT_TILES);
for(var y = 0; y < VERT_TILES; y++) {
gameBoard[y] = new Array(HORIZ_TILES);
for(var x = 0; x < HORIZ_TILES; x++) {
gameBoard[y][x] = 0;
}
}
var canvas, context;
var ready = function(fun) {
if(document.readyState != "loading") {
fun();
}
else if(document.addEventListener) {
document.addEventListener("DOMContentLoaded", fun);
}
else {
document.attachEvent("onreadystatechange", function() {
if(document.readyState != "loading") {
fun();
}
});
}
}
ready(function() {
canvas = document.getElementById("board");
context = canvas.getContext("2d");
});
| var BOARD_HEIGHT = 288;
var BOARD_WIDTH = 224;
var VERT_TILES = BOARD_HEIGHT / 8;
var HORIZ_TILES = BOARD_WIDTH / 8;
gameBoard = new Array(VERT_TILES);
for(var y = 0; y < VERT_TILES; y++) {
gameBoard[y] = new Array(HORIZ_TILES);
for(var x = 0; x < HORIZ_TILES; x++) {
gameBoard[y][x] = "";
}
}
var canvas, context;
var drawObject = function(args) {
args.context.clearRect(0, 0, BOARD_WIDTH, BOARD_HEIGHT);
args.context.fillStyle = args.color;
args.objectArr.forEach(function(row, rIndex) {
row.forEach(function(col, cIndex) {
if(col == 1) {
args.context.fillRect(args.x + cIndex, args.y + rIndex, 1, 1);
}
});
});
}
var ready = function(fun) {
if(document.readyState != "loading") {
fun();
}
else if(document.addEventListener) {
document.addEventListener("DOMContentLoaded", fun);
}
else {
document.attachEvent("onreadystatechange", function() {
if(document.readyState != "loading") {
fun();
}
});
}
}
ready(function() {
canvas = document.getElementById("board");
context = canvas.getContext("2d");
drawObject({
x: 80,
y: 80,
objectArr: charset["A"],
color: "#FFF",
context: context
});
});
| Add drawObject function to draw matrix of pixels | Add drawObject function to draw matrix of pixels
| JavaScript | mit | peternatewood/pac-man-replica,peternatewood/pac-man-replica | javascript | ## Code Before:
var BOARD_HEIGHT = 288;
var BOARD_WIDTH = 224;
var VERT_TILES = BOARD_HEIGHT / 8;
var HORIZ_TILES = BOARD_WIDTH / 8;
gameBoard = new Array(VERT_TILES);
for(var y = 0; y < VERT_TILES; y++) {
gameBoard[y] = new Array(HORIZ_TILES);
for(var x = 0; x < HORIZ_TILES; x++) {
gameBoard[y][x] = 0;
}
}
var canvas, context;
var ready = function(fun) {
if(document.readyState != "loading") {
fun();
}
else if(document.addEventListener) {
document.addEventListener("DOMContentLoaded", fun);
}
else {
document.attachEvent("onreadystatechange", function() {
if(document.readyState != "loading") {
fun();
}
});
}
}
ready(function() {
canvas = document.getElementById("board");
context = canvas.getContext("2d");
});
## Instruction:
Add drawObject function to draw matrix of pixels
## Code After:
var BOARD_HEIGHT = 288;
var BOARD_WIDTH = 224;
var VERT_TILES = BOARD_HEIGHT / 8;
var HORIZ_TILES = BOARD_WIDTH / 8;
gameBoard = new Array(VERT_TILES);
for(var y = 0; y < VERT_TILES; y++) {
gameBoard[y] = new Array(HORIZ_TILES);
for(var x = 0; x < HORIZ_TILES; x++) {
gameBoard[y][x] = "";
}
}
var canvas, context;
var drawObject = function(args) {
args.context.clearRect(0, 0, BOARD_WIDTH, BOARD_HEIGHT);
args.context.fillStyle = args.color;
args.objectArr.forEach(function(row, rIndex) {
row.forEach(function(col, cIndex) {
if(col == 1) {
args.context.fillRect(args.x + cIndex, args.y + rIndex, 1, 1);
}
});
});
}
var ready = function(fun) {
if(document.readyState != "loading") {
fun();
}
else if(document.addEventListener) {
document.addEventListener("DOMContentLoaded", fun);
}
else {
document.attachEvent("onreadystatechange", function() {
if(document.readyState != "loading") {
fun();
}
});
}
}
ready(function() {
canvas = document.getElementById("board");
context = canvas.getContext("2d");
drawObject({
x: 80,
y: 80,
objectArr: charset["A"],
color: "#FFF",
context: context
});
});
|
9504529dd4b9140be0026d0b30a0e88e5dea5e25 | rtrss/config.py | rtrss/config.py | import os
import logging
import importlib
# All configuration defaults are set in this module
TRACKER_HOST = 'rutracker.org'
# Timeone for the tracker times
TZNAME = 'Europe/Moscow'
LOGLEVEL = logging.INFO
LOG_FORMAT_LOGENTRIES = '%(levelname)s %(name)s %(message)s'
LOG_FORMAT_BRIEF = '%(asctime)s %(levelname)s %(name)s %(message)s'
ADMIN_LOGIN = os.environ.get('ADMIN_LOGIN', 'admin')
ADMIN_PASSWORD = os.environ.get('ADMIN_PASSWORD', 'admin')
ADMIN_EMAIL = os.environ.get('ADMIN_EMAIL', 'admin@localhost')
# path to save torrent files
TORRENT_PATH_PATTERN = 'torrents/{}.torrent'
APP_ENVIRONMENT = os.environ.get('RTRSS_ENVIRONMENT')
if not APP_ENVIRONMENT:
raise EnvironmentError('RTRSS_ENVIRONMENT must be set')
_mod = importlib.import_module('rtrss.config_{}'.format(APP_ENVIRONMENT))
_envconf = {k: v for k, v in _mod.__dict__.items() if k == k.upper()}
globals().update(_envconf)
| import os
import logging
import importlib
# All configuration defaults are set in this module
TRACKER_HOST = 'rutracker.org'
# Timeone for the tracker times
TZNAME = 'Europe/Moscow'
LOGLEVEL = logging.INFO
LOG_FORMAT_LOGENTRIES = '%(levelname)s %(name)s %(message)s'
LOG_FORMAT_BRIEF = '%(asctime)s %(levelname)s %(name)s %(message)s'
ADMIN_LOGIN = os.environ.get('ADMIN_LOGIN', 'admin')
ADMIN_PASSWORD = os.environ.get('ADMIN_PASSWORD', 'admin')
ADMIN_EMAIL = os.environ.get('ADMIN_EMAIL', 'admin@localhost')
# path to save torrent files
TORRENT_PATH_PATTERN = 'torrents/{}.torrent'
APP_ENVIRONMENT = os.environ.get('RTRSS_ENVIRONMENT')
if not APP_ENVIRONMENT:
raise EnvironmentError('RTRSS_ENVIRONMENT must be set')
IP = '0.0.0.0'
PORT = 8080
_mod = importlib.import_module('rtrss.config_{}'.format(APP_ENVIRONMENT))
_envconf = {k: v for k, v in _mod.__dict__.items() if k == k.upper()}
globals().update(_envconf)
| Add default IP and PORT | Add default IP and PORT
| Python | apache-2.0 | notapresent/rtrss,notapresent/rtrss,notapresent/rtrss,notapresent/rtrss | python | ## Code Before:
import os
import logging
import importlib
# All configuration defaults are set in this module
TRACKER_HOST = 'rutracker.org'
# Timeone for the tracker times
TZNAME = 'Europe/Moscow'
LOGLEVEL = logging.INFO
LOG_FORMAT_LOGENTRIES = '%(levelname)s %(name)s %(message)s'
LOG_FORMAT_BRIEF = '%(asctime)s %(levelname)s %(name)s %(message)s'
ADMIN_LOGIN = os.environ.get('ADMIN_LOGIN', 'admin')
ADMIN_PASSWORD = os.environ.get('ADMIN_PASSWORD', 'admin')
ADMIN_EMAIL = os.environ.get('ADMIN_EMAIL', 'admin@localhost')
# path to save torrent files
TORRENT_PATH_PATTERN = 'torrents/{}.torrent'
APP_ENVIRONMENT = os.environ.get('RTRSS_ENVIRONMENT')
if not APP_ENVIRONMENT:
raise EnvironmentError('RTRSS_ENVIRONMENT must be set')
_mod = importlib.import_module('rtrss.config_{}'.format(APP_ENVIRONMENT))
_envconf = {k: v for k, v in _mod.__dict__.items() if k == k.upper()}
globals().update(_envconf)
## Instruction:
Add default IP and PORT
## Code After:
import os
import logging
import importlib
# All configuration defaults are set in this module
TRACKER_HOST = 'rutracker.org'
# Timeone for the tracker times
TZNAME = 'Europe/Moscow'
LOGLEVEL = logging.INFO
LOG_FORMAT_LOGENTRIES = '%(levelname)s %(name)s %(message)s'
LOG_FORMAT_BRIEF = '%(asctime)s %(levelname)s %(name)s %(message)s'
ADMIN_LOGIN = os.environ.get('ADMIN_LOGIN', 'admin')
ADMIN_PASSWORD = os.environ.get('ADMIN_PASSWORD', 'admin')
ADMIN_EMAIL = os.environ.get('ADMIN_EMAIL', 'admin@localhost')
# path to save torrent files
TORRENT_PATH_PATTERN = 'torrents/{}.torrent'
APP_ENVIRONMENT = os.environ.get('RTRSS_ENVIRONMENT')
if not APP_ENVIRONMENT:
raise EnvironmentError('RTRSS_ENVIRONMENT must be set')
IP = '0.0.0.0'
PORT = 8080
_mod = importlib.import_module('rtrss.config_{}'.format(APP_ENVIRONMENT))
_envconf = {k: v for k, v in _mod.__dict__.items() if k == k.upper()}
globals().update(_envconf)
|
c937f502cd1efcf01eb4b9fe899912c75dae08f7 | .appveyor.yml | .appveyor.yml | version: 2.0.{build}
# select image with MinGW-w64 available
image: Visual Studio 2013
clone_folder: c:\projects\sif
environment:
global:
PATH: C:\mingw-w64\i686-5.3.0-posix-dwarf-rt_v4-rev0\mingw32\bin;C:\msys64\usr\bin;$(PATH)
CXX: g++
CC: gcc
GCOV: gcov
MAKE: make
install:
# show all env vars
#- set
# display available tools
- '%CC% --version'
- pwd
build_script:
- cmd /c build_all.bat
artifacts:
- path: obj32\*.exe
name: built sif2
| version: 2.0.{build}
# select image with MinGW-w64 available
image: Visual Studio 2013
environment:
global:
PATH: C:\mingw-w64\i686-5.3.0-posix-dwarf-rt_v4-rev0\mingw32\bin;C:\msys64\usr\bin;$(PATH)
CXX: g++
CC: gcc
GCOV: gcov
MAKE: make
install:
# show PATH env var
#- echo %PATH%
- '%CC% --version'
- pwd
build_script:
- cmd /c build_all.bat
artifacts:
- path: obj32\*.exe
name: built sif2
| Remove clone folder setting from AppVeyor config | Remove clone folder setting from AppVeyor config
| YAML | mit | goriy/sif,goriy/sif | yaml | ## Code Before:
version: 2.0.{build}
# select image with MinGW-w64 available
image: Visual Studio 2013
clone_folder: c:\projects\sif
environment:
global:
PATH: C:\mingw-w64\i686-5.3.0-posix-dwarf-rt_v4-rev0\mingw32\bin;C:\msys64\usr\bin;$(PATH)
CXX: g++
CC: gcc
GCOV: gcov
MAKE: make
install:
# show all env vars
#- set
# display available tools
- '%CC% --version'
- pwd
build_script:
- cmd /c build_all.bat
artifacts:
- path: obj32\*.exe
name: built sif2
## Instruction:
Remove clone folder setting from AppVeyor config
## Code After:
version: 2.0.{build}
# select image with MinGW-w64 available
image: Visual Studio 2013
environment:
global:
PATH: C:\mingw-w64\i686-5.3.0-posix-dwarf-rt_v4-rev0\mingw32\bin;C:\msys64\usr\bin;$(PATH)
CXX: g++
CC: gcc
GCOV: gcov
MAKE: make
install:
# show PATH env var
#- echo %PATH%
- '%CC% --version'
- pwd
build_script:
- cmd /c build_all.bat
artifacts:
- path: obj32\*.exe
name: built sif2
|
5d80a93196f1822f3688e0612fa6c73a7679235b | steer_bot_hardware_gazebo/config/joints.yaml | steer_bot_hardware_gazebo/config/joints.yaml | steer_bot_hardware_gazebo:
rear_wheel : 'base_to_wheel'
front_steer : 'base_to_steer'
virtual_rear_wheels : ['base_to_right_rear_wheel', 'base_to_left_rear_wheel']
virtual_front_wheels : ['base_to_right_front_wheel', 'base_to_left_front_wheel']
virtual_front_steers : ['base_to_right_front_steer', 'base_to_left_front_steer']
# ackermann link mechanism
enable_ackermann_link: 1
wheel_separation_w : 0.5
wheel_separation_h : 0.79
| steer_bot_hardware_gazebo:
rear_wheel : 'rear_wheel_joint'
front_steer : 'front_steer_joint'
virtual_rear_wheels : ['base_to_right_rear_wheel', 'base_to_left_rear_wheel']
virtual_front_wheels : ['base_to_right_front_wheel', 'base_to_left_front_wheel']
virtual_front_steers : ['base_to_right_front_steer', 'base_to_left_front_steer']
# ackermann link mechanism
enable_ackermann_link: 1
wheel_separation_w : 0.5
wheel_separation_h : 0.79
| Update joint setting of steer_bot for new steer_drive configulation. | Update joint setting of steer_bot for new steer_drive configulation.
| YAML | mit | CIR-KIT/steer_drive_ros,CIR-KIT/steer_drive_ros | yaml | ## Code Before:
steer_bot_hardware_gazebo:
rear_wheel : 'base_to_wheel'
front_steer : 'base_to_steer'
virtual_rear_wheels : ['base_to_right_rear_wheel', 'base_to_left_rear_wheel']
virtual_front_wheels : ['base_to_right_front_wheel', 'base_to_left_front_wheel']
virtual_front_steers : ['base_to_right_front_steer', 'base_to_left_front_steer']
# ackermann link mechanism
enable_ackermann_link: 1
wheel_separation_w : 0.5
wheel_separation_h : 0.79
## Instruction:
Update joint setting of steer_bot for new steer_drive configulation.
## Code After:
steer_bot_hardware_gazebo:
rear_wheel : 'rear_wheel_joint'
front_steer : 'front_steer_joint'
virtual_rear_wheels : ['base_to_right_rear_wheel', 'base_to_left_rear_wheel']
virtual_front_wheels : ['base_to_right_front_wheel', 'base_to_left_front_wheel']
virtual_front_steers : ['base_to_right_front_steer', 'base_to_left_front_steer']
# ackermann link mechanism
enable_ackermann_link: 1
wheel_separation_w : 0.5
wheel_separation_h : 0.79
|
3b879db6b81844cdd64c4f3fb34814daa18bb68d | _source/_sass/_notices.scss | _source/_sass/_notices.scss | // NOTICES
// --------------------------------------------------
@mixin notice($notice-color) {
position: relative;
padding: 1.5em;
font-family: $alt-font;
@include font-size(14,39);
color: $white;
background-color: $notice-color;
border-radius: $border-radius;
}
.notice {
@include notice($primary-color);
}
.notice-inverse {
@include notice($white);
color: $text-color;
}
.notice-info {
@include notice($info-color);
}
.notice-warning {
@include notice($warning-color);
}
.notice-success {
@include notice($success-color);
}
.notice-danger {
@include notice($danger-color);
} | // NOTICES
// --------------------------------------------------
@mixin notice($notice-color) {
position: relative;
padding: 1.5em;
font-family: $alt-font;
@include font-size(14,39);
color: $white;
background-color: $notice-color;
border-radius: $border-radius;
a {
color: $white;
border-bottom: 1px dotted $white !important;
}
}
.notice {
@include notice($primary-color);
}
.notice-inverse {
@include notice($white);
color: $text-color;
a {
color: $text-color;
border-bottom: 1px dotted $text-color !important;
}
}
.notice-info {
@include notice($info-color);
}
.notice-warning {
@include notice($warning-color);
}
.notice-success {
@include notice($success-color);
}
.notice-danger {
@include notice($danger-color);
} | Fix link colors on notices | Fix link colors on notices
| SCSS | mit | BrianLusina/brianlusina.github.io,QBayLogic/qbaylogic.com,ucuptimposu/ucuptimposu.github.io,hootnot/skinny-bones-jekyll,dgrtwo/rgallery-default,tzandrki/tzandrki.github.io,Shinobi881/skinny-bones-jekyll,CmdrKerfy/old_ind-volcanoes,dgrtwo/broom-gallery,tobbio/tobbio.github.io,frassati-new-haven/frassati-new-haven.github.io,chengdubuddy/chengdubuddy.github.io,l1905/l1905.github.io,jpfreeley/jpfreeley.github.io,dgrtwo/rgallery-default,piapandelakis/piapandelakis.github.io,teear/teear.github.com,plinioseniore/skinnybonesjeckyll,philippeowagner/skinny-bones-jekyll,cooldepartment/skinny-bones-jekyll,swat-ds/zombies-reimagined,cwigington3/cwigington3.github.io,pricelessbrewing/pricelessbrewing.github.io,jasonpchang/jasonpchang.github.io,gladtosee/skinny-bones-jekyll,nikolspencer/skinny-bones-jekyll,juvham/skinny-bones-jekyll,pricelessbrewing/pricelessbrewing.github.io,IncaProject/IncaProject.github.io,joehalliwell/joehalliwell.github.io,heeryung/skinny-bones-jekyll,iambsk1/iambsk1.github.io,open-learning/open-learning.github.io,cooldepartment/skinny-bones-jekyll,mightymetris/mightymetris.github.io,Shinobi881/skinny-bones-jekyll,warrenryu/warrenryu.github.io,iambsk1/iambsk1.github.io,hbc/hbc.github.io,massich/massich.github.io,BrianLusina/brianlusina.github.io,sgng/sgng.github.io,nagoya-workshop/skinny-bones-jekyll,GuyPaddock/guypaddock.github.io,adeldaoud/adeldaoud.github.io,teelahti/kerkesix.fi,BetsyHamlinArt/BetsyHamlinArt.github.io,zomeelee/skinny-bones-jekyll,warrenryu/warrenryu.github.io,iambsk1/iambsk1.github.io,ckochx/vapeblog,swat-ds/zombies-reimagined,AswiniC/skinny-bones-jekyll,BetsyHamlinArt/BetsyHamlinArt.github.io,chroni-blog/chroni,teear/teear.github.com,hootnot/skinny-bones-jekyll,cyproassist/cyproassist.github.io,AZURECITADEL/azurecitadel.github.io,Bigmonstercai/skinny-bones-jekyll,deaconblade/deaconblade.github.io,CmdrKerfy/old_ind-volcanoes,hsaghir/hsaghir.github.io,piapandelakis/piapandelakis.github.io,cyproassist/cyproassist.github.io,deaconblade/deaconblade.github.io,hsaghir/hsaghir.github.io,SusannStruwe/cyproassist.github.io,AswiniC/skinny-bones-jekyll,chengdubuddy/chengdubuddy.github.io,damienkilgannon/damienkilgannon.github.io,isbead/skinny-bones-jekyll,heeryung/skinny-bones-jekyll,alumnosdcao/alumnosdcao.github.io,SusannStruwe/cyproassist.github.io,emptymalei/planets-jekyll,SusannStruwe/cyproassist.github.io,joehalliwell/joehalliwell.github.io,mightymetris/mightymetris.github.io,vynci/phosabyss-dev,zomeelee/skinny-bones-jekyll,cooldepartment/skinny-bones-jekyll,ucuptimposu/ucuptimposu.github.io,philippeowagner/skinny-bones-jekyll,zomeelee/skinny-bones-jekyll,adeldaoud/adeldaoud.github.io,opengers/opengers.github.io,QN7Gardening/QN7Gardening.github.io,chengdubuddy/chengdubuddy.github.io,carlosdelfino/carlosdelfino.github.com,l1905/l1905.github.io,damienkilgannon/damienkilgannon.github.io,sgng/sgng.github.io,alumnosdcao/alumnosdcao.github.io,AswiniC/skinny-bones-jekyll,QN7Gardening/QN7Gardening.github.io,eliteconceptsmn/main-web,zbintliff/zbintliff.github.io,tobbio/tobbio.github.io,GuyPaddock/guypaddock.github.io,imran031387/mimran.github.io,vynci/phosabyss-dev,adeldaoud/adeldaoud.github.io,BetsyHamlinArt/BetsyHamlinArt.github.io,luislhl/luislhl.github.io,juvham/skinny-bones-jekyll,iyamegane/iyamegane.github.io,heeryung/skinny-bones-jekyll,IncaProject/IncaProject.github.io,CmdrKerfy/old_ind-volcanoes,emptymalei/planets-jekyll,hootnot/skinny-bones-jekyll,nagoya-workshop/skinny-bones-jekyll,piapandelakis/piapandelakis.github.io,vynci/phosabyss-dev,tobbio/tobbio.github.io,mikaelkaron/open-learning.github.io,hbc/hbc.github.io,deaconblade/deaconblade.github.io,teear/teear.github.com,eliteconceptsmn/main-web,Bigmonstercai/skinny-bones-jekyll,laredaccion/laredaccion.github.io,QBayLogic/qbaylogic.com,philippeowagner/skinny-bones-jekyll,sgng/sgng.github.io,mikaelkaron/open-learning.github.io,chroni-blog/chroni,jpfreeley/jpfreeley.github.io,carlosdelfino/carlosdelfino.github.com,mikaelkaron/open-learning.github.io,mmistakes/skinny-bones-jekyll,QBayLogic/qbaylogic.com,massich/massich.github.io,hbc/hbc.github.io,mmistakes/skinny-bones-jekyll,pricelessbrewing/pricelessbrewing.github.io,baumanab/baumanab.github.io,thecraftingcorollary/thecraftingcorollary.github.io,juvham/skinny-bones-jekyll,thecraftingcorollary/thecraftingcorollary.github.io,dgrtwo/rgallery-default,carlesgutierrez/carlesgutierrez.github.io,dgrtwo/broom-gallery,cwigington3/cwigington3.github.io,dgrtwo/broom-gallery,baumanab/baumanab.github.io,AZURECITADEL/azurecitadel.github.io,thecraftingcorollary/thecraftingcorollary.github.io,GuyPaddock/guypaddock.github.io,l1905/l1905.github.io,eliteconceptsmn/main-web,nikolspencer/skinny-bones-jekyll,cyproassist/cyproassist.github.io,swat-ds/zombies-reimagined,AlexPutz/alexputz.github.io,mmistakes/skinny-bones-jekyll,baumanab/baumanab.github.io,luislhl/luislhl.github.io,jasonpchang/jasonpchang.github.io,nagoya-workshop/skinny-bones-jekyll,luislhl/luislhl.github.io,jasonpchang/jasonpchang.github.io,open-learning/open-learning.github.io,IncaProject/IncaProject.github.io,IncaProject/IncaProject.github.io,AZURECITADEL/azurecitadel.github.io,damienkilgannon/damienkilgannon.github.io,frassati-new-haven/frassati-new-haven.github.io,warrenryu/warrenryu.github.io,isbead/skinny-bones-jekyll,isbead/skinny-bones-jekyll,bruce-lcf/bruce-lcf.github.io,iyamegane/iyamegane.github.io,carlosdelfino/carlosdelfino.github.com,carlosdelfino/carlosdelfino.github.com,open-learning/open-learning.github.io,ckochx/vapeblog,praxeos/praxeos.github.io,plinioseniore/skinnybonesjeckyll,frassati-new-haven/frassati-new-haven.github.io,jpfreeley/jpfreeley.github.io,emptymalei/planets-jekyll,laredaccion/laredaccion.github.io,carlesgutierrez/carlesgutierrez.github.io,AlexPutz/alexputz.github.io,BrianLusina/brianlusina.github.io,nikolspencer/skinny-bones-jekyll,massich/massich.github.io,rasnes/rasnes.github.io,chroni-blog/chroni,zbintliff/zbintliff.github.io,rasnes/rasnes.github.io,praxeos/praxeos.github.io,imran031387/mimran.github.io,Bigmonstercai/skinny-bones-jekyll,teelahti/kerkesix.fi,QN7Gardening/QN7Gardening.github.io,alumnosdcao/alumnosdcao.github.io,mightymetris/mightymetris.github.io,gladtosee/skinny-bones-jekyll,Shinobi881/skinny-bones-jekyll,hsaghir/hsaghir.github.io,gladtosee/skinny-bones-jekyll,cwigington3/cwigington3.github.io,imran031387/mimran.github.io,carlesgutierrez/carlesgutierrez.github.io,laredaccion/laredaccion.github.io,joehalliwell/joehalliwell.github.io,hsaghir/hsaghir.github.io,ucuptimposu/ucuptimposu.github.io,emptymalei/planets-jekyll,iyamegane/iyamegane.github.io,AlexPutz/alexputz.github.io,rasnes/rasnes.github.io,BrianLusina/brianlusina.github.io,plinioseniore/skinnybonesjeckyll,zbintliff/zbintliff.github.io | scss | ## Code Before:
// NOTICES
// --------------------------------------------------
@mixin notice($notice-color) {
position: relative;
padding: 1.5em;
font-family: $alt-font;
@include font-size(14,39);
color: $white;
background-color: $notice-color;
border-radius: $border-radius;
}
.notice {
@include notice($primary-color);
}
.notice-inverse {
@include notice($white);
color: $text-color;
}
.notice-info {
@include notice($info-color);
}
.notice-warning {
@include notice($warning-color);
}
.notice-success {
@include notice($success-color);
}
.notice-danger {
@include notice($danger-color);
}
## Instruction:
Fix link colors on notices
## Code After:
// NOTICES
// --------------------------------------------------
@mixin notice($notice-color) {
position: relative;
padding: 1.5em;
font-family: $alt-font;
@include font-size(14,39);
color: $white;
background-color: $notice-color;
border-radius: $border-radius;
a {
color: $white;
border-bottom: 1px dotted $white !important;
}
}
.notice {
@include notice($primary-color);
}
.notice-inverse {
@include notice($white);
color: $text-color;
a {
color: $text-color;
border-bottom: 1px dotted $text-color !important;
}
}
.notice-info {
@include notice($info-color);
}
.notice-warning {
@include notice($warning-color);
}
.notice-success {
@include notice($success-color);
}
.notice-danger {
@include notice($danger-color);
} |
d3fd194c82b2d8f701b52d67a95a54bb622eaa53 | multeval.sh | multeval.sh | set -eo pipefail
version=0.1
scriptDir=$(dirname $0)
$scriptDir/get_deps.sh
java -XX:+UseCompressedOops -Xmx2g -cp $scriptDir/lib/tercom-0.7.26.jar:$scriptDir/lib/meteor-1.2/meteor-1.2.jar:$scriptDir/multeval-${version}.jar multeval.MultEval $@
| set -eo pipefail
version=0.1
scriptDir=$(dirname $0)
$scriptDir/get_deps.sh
# Specify -Dfile.encoding so that the meteor paraphrase tables load properly
java -Dfile.encoding=UTF8 -XX:+UseCompressedOops -Xmx2g \
-cp $scriptDir/lib/tercom-0.7.26.jar:$scriptDir/lib/meteor-1.2/meteor-1.2.jar:$scriptDir/multeval-${version}.jar \
multeval.MultEval $@
| Use file.encoding to fix loading meteor paraphrase tables | Use file.encoding to fix loading meteor paraphrase tables
| Shell | mpl-2.0 | jhclark/multeval,jhclark/multeval,jhclark/multeval,jhclark/multeval | shell | ## Code Before:
set -eo pipefail
version=0.1
scriptDir=$(dirname $0)
$scriptDir/get_deps.sh
java -XX:+UseCompressedOops -Xmx2g -cp $scriptDir/lib/tercom-0.7.26.jar:$scriptDir/lib/meteor-1.2/meteor-1.2.jar:$scriptDir/multeval-${version}.jar multeval.MultEval $@
## Instruction:
Use file.encoding to fix loading meteor paraphrase tables
## Code After:
set -eo pipefail
version=0.1
scriptDir=$(dirname $0)
$scriptDir/get_deps.sh
# Specify -Dfile.encoding so that the meteor paraphrase tables load properly
java -Dfile.encoding=UTF8 -XX:+UseCompressedOops -Xmx2g \
-cp $scriptDir/lib/tercom-0.7.26.jar:$scriptDir/lib/meteor-1.2/meteor-1.2.jar:$scriptDir/multeval-${version}.jar \
multeval.MultEval $@
|
06a9c42fb2163156ff2147fccc88bb73db08e76b | exports.js | exports.js | // Matching the public exports in husl-colors/husl
module['exports'] = {};
module['exports']["fromRGB"] = husl.Husl.rgbToHusl;
module['exports']["fromHex"] = husl.Husl.hexToHusl;
module['exports']["toRGB"] = husl.Husl.huslToRgb;
module['exports']["toHex"] = husl.Husl.huslToHex;
module['exports']['p'] = {};
module['exports']['p']["fromRGB"] = husl.Husl.rgbToHuslp;
module['exports']['p']["fromHex"] = husl.Husl.hexToHuslp;
module['exports']['p']["toRGB"] = husl.Husl.huslpToRgb;
module['exports']['p']["toHex"] = husl.Husl.huslpToHex; | // Matching the public exports in husl-colors/husl
function expandParams(f) {
return function(c1, c2, c3) {
return f([c1, c2, c3])
}
}
module['exports'] = {};
module['exports']["fromRGB"] = expandParams(husl.Husl.rgbToHusl);
module['exports']["fromHex"] = husl.Husl.hexToHusl;
module['exports']["toRGB"] = expandParams(husl.Husl.huslToRgb);
module['exports']["toHex"] = expandParams(husl.Husl.huslToHex);
module['exports']['p'] = {};
module['exports']['p']["fromRGB"] = expandParams(husl.Husl.rgbToHuslp);
module['exports']['p']["fromHex"] = husl.Husl.hexToHuslp;
module['exports']['p']["toRGB"] = expandParams(husl.Husl.huslpToRgb);
module['exports']['p']["toHex"] = expandParams(husl.Husl.huslpToHex); | Add adaptor for multiple argument functions | Add adaptor for multiple argument functions
| JavaScript | mit | hsluv/hsluv,hsluv/hsluv | javascript | ## Code Before:
// Matching the public exports in husl-colors/husl
module['exports'] = {};
module['exports']["fromRGB"] = husl.Husl.rgbToHusl;
module['exports']["fromHex"] = husl.Husl.hexToHusl;
module['exports']["toRGB"] = husl.Husl.huslToRgb;
module['exports']["toHex"] = husl.Husl.huslToHex;
module['exports']['p'] = {};
module['exports']['p']["fromRGB"] = husl.Husl.rgbToHuslp;
module['exports']['p']["fromHex"] = husl.Husl.hexToHuslp;
module['exports']['p']["toRGB"] = husl.Husl.huslpToRgb;
module['exports']['p']["toHex"] = husl.Husl.huslpToHex;
## Instruction:
Add adaptor for multiple argument functions
## Code After:
// Matching the public exports in husl-colors/husl
function expandParams(f) {
return function(c1, c2, c3) {
return f([c1, c2, c3])
}
}
module['exports'] = {};
module['exports']["fromRGB"] = expandParams(husl.Husl.rgbToHusl);
module['exports']["fromHex"] = husl.Husl.hexToHusl;
module['exports']["toRGB"] = expandParams(husl.Husl.huslToRgb);
module['exports']["toHex"] = expandParams(husl.Husl.huslToHex);
module['exports']['p'] = {};
module['exports']['p']["fromRGB"] = expandParams(husl.Husl.rgbToHuslp);
module['exports']['p']["fromHex"] = husl.Husl.hexToHuslp;
module['exports']['p']["toRGB"] = expandParams(husl.Husl.huslpToRgb);
module['exports']['p']["toHex"] = expandParams(husl.Husl.huslpToHex); |
5acad5c62ec204c5a190bb3f5b2516153b4869b1 | .travis.yml | .travis.yml | node_js:
- "6"
- "7"
- "8"
- "9"
language: node_js
script: "npm run-script test-travis"
after_script: "npm install coveralls@2 && cat ./coverage/lcov.info | coveralls"
before_script:
- curl http://ffmpeg.gusari.org/static/64bit/ffmpeg.static.64bit.latest.tar.gz | sudo tar -C /usr/local/bin/ -xvz
| node_js:
- "6"
- "7"
- "8"
- "9"
language: node_js
script: "npm run-script test-travis"
after_script: "npm install coveralls@2 && cat ./coverage/lcov.info | coveralls"
before_script:
- curl https://johnvansickle.com/ffmpeg/builds/ffmpeg-git-64bit-static.tar.xz | sudo tar -C /usr/local/bin/ -xvJ
| Fix FFMPEG download URL for tests | :green_heart: Fix FFMPEG download URL for tests
| YAML | mit | caffco/get-video-duration,caffco/get-video-duration | yaml | ## Code Before:
node_js:
- "6"
- "7"
- "8"
- "9"
language: node_js
script: "npm run-script test-travis"
after_script: "npm install coveralls@2 && cat ./coverage/lcov.info | coveralls"
before_script:
- curl http://ffmpeg.gusari.org/static/64bit/ffmpeg.static.64bit.latest.tar.gz | sudo tar -C /usr/local/bin/ -xvz
## Instruction:
:green_heart: Fix FFMPEG download URL for tests
## Code After:
node_js:
- "6"
- "7"
- "8"
- "9"
language: node_js
script: "npm run-script test-travis"
after_script: "npm install coveralls@2 && cat ./coverage/lcov.info | coveralls"
before_script:
- curl https://johnvansickle.com/ffmpeg/builds/ffmpeg-git-64bit-static.tar.xz | sudo tar -C /usr/local/bin/ -xvJ
|
52ffc2b264cbacaee56017cd4a67df4511d60392 | celery/managers.py | celery/managers.py | from django.db import models
from celery.registry import tasks
from datetime import datetime, timedelta
__all__ = ["TaskManager", "PeriodicTaskManager"]
class TaskManager(models.Manager):
def get_task(self, task_id):
task, created = self.get_or_create(task_id=task_id)
return task
def is_done(self, task_id):
return self.get_task(task_id).is_done
def get_all_expired(self):
return self.filter(date_done__lt=datetime.now() - timedelta(days=5))
def delete_expired(self):
self.get_all_expired().delete()
def mark_as_done(self, task_id):
task, created = self.get_or_create(task_id=task_id, defaults={
"is_done": True})
if not created:
task.is_done = True
task.save()
class PeriodicTaskManager(models.Manager):
def get_waiting_tasks(self):
periodic_tasks = tasks.get_all_periodic()
waiting = []
for task_name, task in periodic_tasks.items():
task_meta, created = self.get_or_create(name=task_name)
# task_run.every must be a timedelta object.
run_at = task_meta.last_run_at + task.run_every
if datetime.now() > run_at:
waiting.append(task_meta)
return waiting
| from django.db import models
from celery.registry import tasks
from datetime import datetime, timedelta
__all__ = ["TaskManager", "PeriodicTaskManager"]
class TaskManager(models.Manager):
def get_task(self, task_id):
task, created = self.get_or_create(task_id=task_id)
return task
def is_done(self, task_id):
return self.get_task(task_id).is_done
def get_all_expired(self):
return self.filter(date_done__lt=datetime.now() - timedelta(days=5),
is_done=True)
def delete_expired(self):
self.get_all_expired().delete()
def mark_as_done(self, task_id):
task, created = self.get_or_create(task_id=task_id, defaults={
"is_done": True})
if not created:
task.is_done = True
task.save()
class PeriodicTaskManager(models.Manager):
def get_waiting_tasks(self):
periodic_tasks = tasks.get_all_periodic()
waiting = []
for task_name, task in periodic_tasks.items():
task_meta, created = self.get_or_create(name=task_name)
# task_run.every must be a timedelta object.
run_at = task_meta.last_run_at + task.run_every
if datetime.now() > run_at:
waiting.append(task_meta)
return waiting
| Add is_done=True to get_all_expired filter. | Add is_done=True to get_all_expired filter.
| Python | bsd-3-clause | WoLpH/celery,cbrepo/celery,cbrepo/celery,ask/celery,frac/celery,WoLpH/celery,mitsuhiko/celery,ask/celery,mitsuhiko/celery,frac/celery | python | ## Code Before:
from django.db import models
from celery.registry import tasks
from datetime import datetime, timedelta
__all__ = ["TaskManager", "PeriodicTaskManager"]
class TaskManager(models.Manager):
def get_task(self, task_id):
task, created = self.get_or_create(task_id=task_id)
return task
def is_done(self, task_id):
return self.get_task(task_id).is_done
def get_all_expired(self):
return self.filter(date_done__lt=datetime.now() - timedelta(days=5))
def delete_expired(self):
self.get_all_expired().delete()
def mark_as_done(self, task_id):
task, created = self.get_or_create(task_id=task_id, defaults={
"is_done": True})
if not created:
task.is_done = True
task.save()
class PeriodicTaskManager(models.Manager):
def get_waiting_tasks(self):
periodic_tasks = tasks.get_all_periodic()
waiting = []
for task_name, task in periodic_tasks.items():
task_meta, created = self.get_or_create(name=task_name)
# task_run.every must be a timedelta object.
run_at = task_meta.last_run_at + task.run_every
if datetime.now() > run_at:
waiting.append(task_meta)
return waiting
## Instruction:
Add is_done=True to get_all_expired filter.
## Code After:
from django.db import models
from celery.registry import tasks
from datetime import datetime, timedelta
__all__ = ["TaskManager", "PeriodicTaskManager"]
class TaskManager(models.Manager):
def get_task(self, task_id):
task, created = self.get_or_create(task_id=task_id)
return task
def is_done(self, task_id):
return self.get_task(task_id).is_done
def get_all_expired(self):
return self.filter(date_done__lt=datetime.now() - timedelta(days=5),
is_done=True)
def delete_expired(self):
self.get_all_expired().delete()
def mark_as_done(self, task_id):
task, created = self.get_or_create(task_id=task_id, defaults={
"is_done": True})
if not created:
task.is_done = True
task.save()
class PeriodicTaskManager(models.Manager):
def get_waiting_tasks(self):
periodic_tasks = tasks.get_all_periodic()
waiting = []
for task_name, task in periodic_tasks.items():
task_meta, created = self.get_or_create(name=task_name)
# task_run.every must be a timedelta object.
run_at = task_meta.last_run_at + task.run_every
if datetime.now() > run_at:
waiting.append(task_meta)
return waiting
|
7da7354668187167ccc6143ee0f469158c858ed3 | lib/economic/proxies/actions/creditor_contact_proxy/find_by_name.rb | lib/economic/proxies/actions/creditor_contact_proxy/find_by_name.rb | module Economic
module Proxies
module Actions
module CreditorContactProxy
class FindByName
attr_reader :name
def initialize(caller, name)
@caller = caller
@name = name
end
def call
# Get a list of CreditorContactHandles from e-conomic
response = request('FindByName', {
'name' => name
})
# Make sure we always have an array of handles even if the result only contains one
handles = [response[:creditor_contact_handle]].flatten.reject(&:blank?)
# Create partial CreditorContact entities
contacts = handles.collect do |handle|
creditor_contact = build
creditor_contact.partial = true
creditor_contact.persisted = true
creditor_contact.handle = handle
creditor_contact.id = handle[:id]
creditor_contact.number = handle[:number]
creditor_contact
end
if owner.is_a?(Creditor)
# Scope to the owner
contacts.select do |creditor_contact|
creditor_contact.get_data
creditor_contact.creditor.handle == owner.handle
end
else
contacts
end
end
private
def build(*options)
@caller.build(options)
end
def owner
@caller.owner
end
def request(action, data)
@caller.request(action, data)
end
end
end
end
end
end
| module Economic
module Proxies
module Actions
module CreditorContactProxy
class FindByName
attr_reader :name
def initialize(caller, name)
@caller = caller
@name = name
end
def call
# Get a list of CreditorContactHandles from e-conomic
handles = [response[:creditor_contact_handle]].flatten.reject(&:blank?)
contacts = build_partial_contact_entities(handles)
scope_to_owner(contacts)
end
private
def build(*options)
@caller.build(options)
end
def build_partial_contact_entities(handles)
handles.collect do |handle|
creditor_contact = build
creditor_contact.partial = true
creditor_contact.persisted = true
creditor_contact.handle = handle
creditor_contact.id = handle[:id]
creditor_contact.number = handle[:number]
creditor_contact
end
end
def owner
@caller.owner
end
def request(action, data)
@caller.request(action, data)
end
def response
request('FindByName', {'name' => name})
end
def scope_to_owner(contacts)
if owner.is_a?(Creditor)
# Scope to the owner
contacts.select do |creditor_contact|
creditor_contact.get_data
creditor_contact.creditor.handle == owner.handle
end
else
contacts
end
end
end
end
end
end
end
| Clean up FindByName a bit | Clean up FindByName a bit
| Ruby | mit | anderslime/rconomic,kongens-net/rconomic,substancelab/rconomic,lokalebasen/rconomic | ruby | ## Code Before:
module Economic
module Proxies
module Actions
module CreditorContactProxy
class FindByName
attr_reader :name
def initialize(caller, name)
@caller = caller
@name = name
end
def call
# Get a list of CreditorContactHandles from e-conomic
response = request('FindByName', {
'name' => name
})
# Make sure we always have an array of handles even if the result only contains one
handles = [response[:creditor_contact_handle]].flatten.reject(&:blank?)
# Create partial CreditorContact entities
contacts = handles.collect do |handle|
creditor_contact = build
creditor_contact.partial = true
creditor_contact.persisted = true
creditor_contact.handle = handle
creditor_contact.id = handle[:id]
creditor_contact.number = handle[:number]
creditor_contact
end
if owner.is_a?(Creditor)
# Scope to the owner
contacts.select do |creditor_contact|
creditor_contact.get_data
creditor_contact.creditor.handle == owner.handle
end
else
contacts
end
end
private
def build(*options)
@caller.build(options)
end
def owner
@caller.owner
end
def request(action, data)
@caller.request(action, data)
end
end
end
end
end
end
## Instruction:
Clean up FindByName a bit
## Code After:
module Economic
module Proxies
module Actions
module CreditorContactProxy
class FindByName
attr_reader :name
def initialize(caller, name)
@caller = caller
@name = name
end
def call
# Get a list of CreditorContactHandles from e-conomic
handles = [response[:creditor_contact_handle]].flatten.reject(&:blank?)
contacts = build_partial_contact_entities(handles)
scope_to_owner(contacts)
end
private
def build(*options)
@caller.build(options)
end
def build_partial_contact_entities(handles)
handles.collect do |handle|
creditor_contact = build
creditor_contact.partial = true
creditor_contact.persisted = true
creditor_contact.handle = handle
creditor_contact.id = handle[:id]
creditor_contact.number = handle[:number]
creditor_contact
end
end
def owner
@caller.owner
end
def request(action, data)
@caller.request(action, data)
end
def response
request('FindByName', {'name' => name})
end
def scope_to_owner(contacts)
if owner.is_a?(Creditor)
# Scope to the owner
contacts.select do |creditor_contact|
creditor_contact.get_data
creditor_contact.creditor.handle == owner.handle
end
else
contacts
end
end
end
end
end
end
end
|
870723483416ffe81a47facbbc68440735b2dfea | package.json | package.json | {
"name": "heroku-dyno",
"version": "0.1.1",
"description": "Unbaffled heroku dyno management",
"repository": {
"type": "git",
"url": "git://github.com/visionmobile/heroku-dyno.git"
},
"main": "src/Dyno.js",
"scripts": {
"test": "node test/index.js"
},
"keywords": [
"heroku",
"dyno",
"process",
"scale",
"promises",
"simple",
"bluebird"
],
"author": "Dimitrios C. Michalakos <[email protected]>",
"license": "MIT",
"dependencies": {
"bluebird": "^2.9.24",
"heroku-client": "1.9.1",
"lodash": "^3.7.0",
"type-of": "^2.0.1"
},
"devDependencies": {
"chai": "^2.2.0",
"dotenv": "^1.1.0",
"mocha": "^2.2.4"
}
}
| {
"name": "heroku-dyno",
"version": "0.1.1",
"description": "Unbaffled heroku dyno management",
"repository": {
"type": "git",
"url": "git://github.com/visionmobile/heroku-dyno.git"
},
"main": "src/Dyno.js",
"scripts": {
"test": "node test/index.js"
},
"keywords": [
"heroku",
"dyno",
"process",
"scale",
"promises",
"simple",
"bluebird"
],
"author": {
"name": "VisionMobile Ltd",
"email": "[email protected]",
"url": "https://github.com/visionmobile"
},
"maintainers": [
{
"name": "Dimitrios C. Michalakos",
"email": "[email protected]",
"url": "https://github.com/jmike"
}
],
"license": "MIT",
"dependencies": {
"bluebird": "^2.9.24",
"heroku-client": "1.9.1",
"lodash": "^3.7.0",
"type-of": "^2.0.1"
},
"devDependencies": {
"chai": "^2.2.0",
"dotenv": "^1.1.0",
"mocha": "^2.2.4"
}
}
| Change author to visionmobile; add jmike as maintainer | Change author to visionmobile; add jmike as maintainer
| JSON | mit | visionmobile/heroku-dyno | json | ## Code Before:
{
"name": "heroku-dyno",
"version": "0.1.1",
"description": "Unbaffled heroku dyno management",
"repository": {
"type": "git",
"url": "git://github.com/visionmobile/heroku-dyno.git"
},
"main": "src/Dyno.js",
"scripts": {
"test": "node test/index.js"
},
"keywords": [
"heroku",
"dyno",
"process",
"scale",
"promises",
"simple",
"bluebird"
],
"author": "Dimitrios C. Michalakos <[email protected]>",
"license": "MIT",
"dependencies": {
"bluebird": "^2.9.24",
"heroku-client": "1.9.1",
"lodash": "^3.7.0",
"type-of": "^2.0.1"
},
"devDependencies": {
"chai": "^2.2.0",
"dotenv": "^1.1.0",
"mocha": "^2.2.4"
}
}
## Instruction:
Change author to visionmobile; add jmike as maintainer
## Code After:
{
"name": "heroku-dyno",
"version": "0.1.1",
"description": "Unbaffled heroku dyno management",
"repository": {
"type": "git",
"url": "git://github.com/visionmobile/heroku-dyno.git"
},
"main": "src/Dyno.js",
"scripts": {
"test": "node test/index.js"
},
"keywords": [
"heroku",
"dyno",
"process",
"scale",
"promises",
"simple",
"bluebird"
],
"author": {
"name": "VisionMobile Ltd",
"email": "[email protected]",
"url": "https://github.com/visionmobile"
},
"maintainers": [
{
"name": "Dimitrios C. Michalakos",
"email": "[email protected]",
"url": "https://github.com/jmike"
}
],
"license": "MIT",
"dependencies": {
"bluebird": "^2.9.24",
"heroku-client": "1.9.1",
"lodash": "^3.7.0",
"type-of": "^2.0.1"
},
"devDependencies": {
"chai": "^2.2.0",
"dotenv": "^1.1.0",
"mocha": "^2.2.4"
}
}
|
7ece85d1ab3d789ee1f91ce77d1934927339ae16 | react/frontpage/components/PromoBanner.scss | react/frontpage/components/PromoBanner.scss | // Banner
@import '../common';
.promoBanner {
margin-bottom: 20pt;
cursor: pointer;
}
.promoBanner .overlay {
width: 100%;
height: 100%;
top: 0;
left: 0;
position: absolute;
}
.promoBanner:hover .overlay {
background-color: $link-image-active-color;
}
| // Banner
@import '../common';
.promoBanner {
margin-bottom: 20pt;
}
.promoBanner .overlay {
width: 100%;
height: 100%;
top: 0;
left: 0;
position: absolute;
cursor: pointer;
z-index: 1; // So overlay is on top of all other promo banner divs (need this to trigger :hover)
&:hover {
background-color: $link-image-active-color;
}
}
| Fix some hover weirdness in the promo banner | Fix some hover weirdness in the promo banner
| SCSS | agpl-3.0 | uclaradio/uclaradio,uclaradio/uclaradio,uclaradio/uclaradio | scss | ## Code Before:
// Banner
@import '../common';
.promoBanner {
margin-bottom: 20pt;
cursor: pointer;
}
.promoBanner .overlay {
width: 100%;
height: 100%;
top: 0;
left: 0;
position: absolute;
}
.promoBanner:hover .overlay {
background-color: $link-image-active-color;
}
## Instruction:
Fix some hover weirdness in the promo banner
## Code After:
// Banner
@import '../common';
.promoBanner {
margin-bottom: 20pt;
}
.promoBanner .overlay {
width: 100%;
height: 100%;
top: 0;
left: 0;
position: absolute;
cursor: pointer;
z-index: 1; // So overlay is on top of all other promo banner divs (need this to trigger :hover)
&:hover {
background-color: $link-image-active-color;
}
}
|
878c3dc00f31e2d35d6f26b99b3e45be772ba1fd | app/controllers/property_elements_controller.rb | app/controllers/property_elements_controller.rb | class PropertyElementsController < ApplicationController
before_filter :check_for_cancel, :only => [:create]
def index
@user = current_user
@property = Property.find(params[:property_id])
@property_elements = @property.property_elements
if @property_elements.empty?
flash[:error] = "La propiedad no tiene elementos"
redirect_to user_property_path(@user.id, @property.id)
end
end
def new
@property_element = PropertyElement.new
end
# def show
# @property = Property.find(params[:id])
# end
def create
@property_element = PropertyElement.new(property_element_params)
if @property_element.save
flash[:success] = 'Elemento registrado exitosamente.'
redirect_to user_property_property_elements_path(current_user.id, params[:property_id])
else
flash[:error] = @property_element.errors.full_messages.join(',')
render 'new'
end
end
def property_element_params
params.require(:property_element).permit(:name, :description, :image, :element_type_id, :property_id)
end
def check_for_cancel
if(params.size == 5)
flash[:notice] = "Registro cancelado."
redirect_to user_property_property_elements_path(current_user.id, params[:property_id])
end
end
end
| class PropertyElementsController < ApplicationController
before_filter :check_for_cancel, :only => [:create]
def index
@user = current_user
@property = Property.find(params[:property_id])
@property_elements = @property.property_elements
if @property_elements.empty?
flash[:error] = "La propiedad no tiene elementos"
redirect_to new_user_property_property_element_path(@user.id, @property.id)
end
end
def new
@property_element = PropertyElement.new
end
# def show
# @property = Property.find(params[:id])
# end
def create
@property_element = PropertyElement.new(property_element_params)
if @property_element.save
flash[:success] = 'Elemento registrado exitosamente.'
redirect_to user_property_property_elements_path(current_user.id, params[:property_id])
else
flash[:error] = @property_element.errors.full_messages.join(',')
render 'new'
end
end
def property_element_params
params.require(:property_element).permit(:name, :description, :image, :element_type_id, :property_id)
end
def check_for_cancel
if(params.size == 5)
flash[:notice] = "Registro cancelado."
redirect_to user_property_property_elements_path(current_user.id, params[:property_id])
end
end
end
| Fix bug redirecting to new when property has no elements. | Fix bug redirecting to new when property has no elements.
| Ruby | apache-2.0 | gtrujillop/RealEstateWeb,gtrujillop/RealEstateWeb,gtrujillop/RealEstateWeb | ruby | ## Code Before:
class PropertyElementsController < ApplicationController
before_filter :check_for_cancel, :only => [:create]
def index
@user = current_user
@property = Property.find(params[:property_id])
@property_elements = @property.property_elements
if @property_elements.empty?
flash[:error] = "La propiedad no tiene elementos"
redirect_to user_property_path(@user.id, @property.id)
end
end
def new
@property_element = PropertyElement.new
end
# def show
# @property = Property.find(params[:id])
# end
def create
@property_element = PropertyElement.new(property_element_params)
if @property_element.save
flash[:success] = 'Elemento registrado exitosamente.'
redirect_to user_property_property_elements_path(current_user.id, params[:property_id])
else
flash[:error] = @property_element.errors.full_messages.join(',')
render 'new'
end
end
def property_element_params
params.require(:property_element).permit(:name, :description, :image, :element_type_id, :property_id)
end
def check_for_cancel
if(params.size == 5)
flash[:notice] = "Registro cancelado."
redirect_to user_property_property_elements_path(current_user.id, params[:property_id])
end
end
end
## Instruction:
Fix bug redirecting to new when property has no elements.
## Code After:
class PropertyElementsController < ApplicationController
before_filter :check_for_cancel, :only => [:create]
def index
@user = current_user
@property = Property.find(params[:property_id])
@property_elements = @property.property_elements
if @property_elements.empty?
flash[:error] = "La propiedad no tiene elementos"
redirect_to new_user_property_property_element_path(@user.id, @property.id)
end
end
def new
@property_element = PropertyElement.new
end
# def show
# @property = Property.find(params[:id])
# end
def create
@property_element = PropertyElement.new(property_element_params)
if @property_element.save
flash[:success] = 'Elemento registrado exitosamente.'
redirect_to user_property_property_elements_path(current_user.id, params[:property_id])
else
flash[:error] = @property_element.errors.full_messages.join(',')
render 'new'
end
end
def property_element_params
params.require(:property_element).permit(:name, :description, :image, :element_type_id, :property_id)
end
def check_for_cancel
if(params.size == 5)
flash[:notice] = "Registro cancelado."
redirect_to user_property_property_elements_path(current_user.id, params[:property_id])
end
end
end
|
0a48250ec2d7350f11bc864a010954078f6e1588 | vdb-bench-assembly/plugins/vdb-bench-widgets/fileImportControl.html | vdb-bench-assembly/plugins/vdb-bench-widgets/fileImportControl.html | <div>
<!-- Show progress bar on loading -->
<div ng-show="vm.inProgress">
<uib-progressbar class="progress-striped active" value="dynamic" type="info"></uib-progressbar>
</div>
<!-- Import control -->
<div ng-show="! vm.inProgress">
<div class="import-control-title" translate="fileImportControl.selectFileMsg" />
<div class="import-control-button">
<label class="btn btn-primary" translate="shared.BrowseMsg">
<input ng-model="fileToImport" type="file" style="display: none;" file-input-handler="vm.importFile">
</label>
<button class="btn" ng-click="vm.cancel()" ng-show="vm.displayCancel()" translate="shared.Cancel"></button>
</div>
<div class="import-control-response">
<div ng-class="vm.responseStyleClass">{{vm.response}}</div>
</div>
<div class="import-control-error">
{{vm.error}}
</div>
</div>
</div>
| <div>
<!-- Show progress bar on loading -->
<div ng-show="vm.inProgress">
<uib-progressbar class="progress-striped active" value="dynamic" type="info"></uib-progressbar>
</div>
<!-- Import control -->
<div ng-show="! vm.inProgress">
<div class="import-control-title" translate="fileImportControl.selectFileMsg" />
<div class="import-control-button">
<label class="btn btn-primary">
<span translate="shared.BrowseMsg"></span>
<input ng-model="fileToImport" type="file" style="display: none;" file-input-handler="vm.importFile">
</label>
<button class="btn" ng-click="vm.cancel()" ng-show="vm.displayCancel()" translate="shared.Cancel"></button>
</div>
<div class="import-control-response">
<div ng-class="vm.responseStyleClass">{{vm.response}}</div>
</div>
<div class="import-control-error">
{{vm.error}}
</div>
</div>
</div>
| Fix for file import control button | Fix for file import control button
* Seems that translate directive does not transclude the contents of the
element but replaces it with its translated content. This is a problem
for elements that already contain child-elements.
| HTML | apache-2.0 | phantomjinx/vdb-bench,Teiid-Designer/vdb-bench,Teiid-Designer/vdb-bench,mdrillin/vdb-bench,mdrillin/vdb-bench,phantomjinx/vdb-bench | html | ## Code Before:
<div>
<!-- Show progress bar on loading -->
<div ng-show="vm.inProgress">
<uib-progressbar class="progress-striped active" value="dynamic" type="info"></uib-progressbar>
</div>
<!-- Import control -->
<div ng-show="! vm.inProgress">
<div class="import-control-title" translate="fileImportControl.selectFileMsg" />
<div class="import-control-button">
<label class="btn btn-primary" translate="shared.BrowseMsg">
<input ng-model="fileToImport" type="file" style="display: none;" file-input-handler="vm.importFile">
</label>
<button class="btn" ng-click="vm.cancel()" ng-show="vm.displayCancel()" translate="shared.Cancel"></button>
</div>
<div class="import-control-response">
<div ng-class="vm.responseStyleClass">{{vm.response}}</div>
</div>
<div class="import-control-error">
{{vm.error}}
</div>
</div>
</div>
## Instruction:
Fix for file import control button
* Seems that translate directive does not transclude the contents of the
element but replaces it with its translated content. This is a problem
for elements that already contain child-elements.
## Code After:
<div>
<!-- Show progress bar on loading -->
<div ng-show="vm.inProgress">
<uib-progressbar class="progress-striped active" value="dynamic" type="info"></uib-progressbar>
</div>
<!-- Import control -->
<div ng-show="! vm.inProgress">
<div class="import-control-title" translate="fileImportControl.selectFileMsg" />
<div class="import-control-button">
<label class="btn btn-primary">
<span translate="shared.BrowseMsg"></span>
<input ng-model="fileToImport" type="file" style="display: none;" file-input-handler="vm.importFile">
</label>
<button class="btn" ng-click="vm.cancel()" ng-show="vm.displayCancel()" translate="shared.Cancel"></button>
</div>
<div class="import-control-response">
<div ng-class="vm.responseStyleClass">{{vm.response}}</div>
</div>
<div class="import-control-error">
{{vm.error}}
</div>
</div>
</div>
|
5ae47158073bc0f2f43b95c5e5a343e736537f98 | delegatednews/delegatednews-config/src/main/config/dbRepository/oracle/delegatedNews/003/create_table.sql | delegatednews/delegatednews-config/src/main/config/dbRepository/oracle/delegatedNews/003/create_table.sql | create table sc_delegatednews_news (
pubId int not null,
instanceId varchar(50) not null,
status varchar(100) not null,
contributorId varchar(50) not null,
validatorId varchar(50) null,
validationDate timestamp(0) null,
beginDate timestamp(0) null,
endDate timestamp(0) null,
newsOrder int not null DEFAULT (0)
);
| create table sc_delegatednews_news (
pubId int not null,
instanceId varchar(50) not null,
status varchar(100) not null,
contributorId varchar(50) not null,
validatorId varchar(50) null,
validationDate timestamp(0) null,
beginDate timestamp(0) null,
endDate timestamp(0) null,
newsOrder int DEFAULT (0) not null
);
| Fix the creation of the table sc_delegatednews_news for Oracle: the declaration of the column newsOrder was incorrect | Fix the creation of the table sc_delegatednews_news for Oracle: the declaration of the column newsOrder was incorrect
| SQL | agpl-3.0 | ebonnet/Silverpeas-Components,ebonnet/Silverpeas-Components,SilverTeamWork/Silverpeas-Components,SilverYoCha/Silverpeas-Components,CecileBONIN/Silverpeas-Components,CecileBONIN/Silverpeas-Components,SilverYoCha/Silverpeas-Components,CecileBONIN/Silverpeas-Components,Silverpeas/Silverpeas-Components,auroreallibe/Silverpeas-Components,ebonnet/Silverpeas-Components,SilverTeamWork/Silverpeas-Components,auroreallibe/Silverpeas-Components,ebonnet/Silverpeas-Components,CecileBONIN/Silverpeas-Components,mmoqui/Silverpeas-Components,auroreallibe/Silverpeas-Components,Silverpeas/Silverpeas-Components,CecileBONIN/Silverpeas-Components,ebonnet/Silverpeas-Components,CecileBONIN/Silverpeas-Components,mmoqui/Silverpeas-Components,ebonnet/Silverpeas-Components,ebonnet/Silverpeas-Components,Silverpeas/Silverpeas-Components,SilverTeamWork/Silverpeas-Components,mmoqui/Silverpeas-Components,SilverYoCha/Silverpeas-Components | sql | ## Code Before:
create table sc_delegatednews_news (
pubId int not null,
instanceId varchar(50) not null,
status varchar(100) not null,
contributorId varchar(50) not null,
validatorId varchar(50) null,
validationDate timestamp(0) null,
beginDate timestamp(0) null,
endDate timestamp(0) null,
newsOrder int not null DEFAULT (0)
);
## Instruction:
Fix the creation of the table sc_delegatednews_news for Oracle: the declaration of the column newsOrder was incorrect
## Code After:
create table sc_delegatednews_news (
pubId int not null,
instanceId varchar(50) not null,
status varchar(100) not null,
contributorId varchar(50) not null,
validatorId varchar(50) null,
validationDate timestamp(0) null,
beginDate timestamp(0) null,
endDate timestamp(0) null,
newsOrder int DEFAULT (0) not null
);
|
8445a067ae4b8158c108145cc9715b4ae6e5a88c | .travis.yml | .travis.yml | sudo: required
dist: trusty
language: cpp
services:
- docker
branches:
only:
- master
- develop
before_install:
- docker build -t travis-build -f infrastructure/tools/Dockerfile .
script:
# Common style check
- docker run --rm travis-build infrastructure/tools/checkstyle.sh $TRAVIS_COMMIT_RANGE
# Run tests
- docker run --rm travis-build /bin/bash -c "fab test_all"
| sudo: required
dist: trusty
language: cpp
services:
- docker
branches:
only:
- master
- develop
before_install:
- docker build -t travis-build -f infrastructure/tools/Dockerfile .
script:
# Common style check
- docker run --rm travis-build infrastructure/tools/checkstyle.sh $TRAVIS_COMMIT_RANGE
# Test whole project with pylint (can find issues in multiple files)
- docker run --rm travis-build /bin/bash -c "pylint --rcfile=.pylintrc ./pylua"
# Run tests
- docker run --rm travis-build /bin/bash -c "fab test_all"
| Check whole project with pylint on CI | Check whole project with pylint on CI
| YAML | mit | malirod/pylua,malirod/pylua | yaml | ## Code Before:
sudo: required
dist: trusty
language: cpp
services:
- docker
branches:
only:
- master
- develop
before_install:
- docker build -t travis-build -f infrastructure/tools/Dockerfile .
script:
# Common style check
- docker run --rm travis-build infrastructure/tools/checkstyle.sh $TRAVIS_COMMIT_RANGE
# Run tests
- docker run --rm travis-build /bin/bash -c "fab test_all"
## Instruction:
Check whole project with pylint on CI
## Code After:
sudo: required
dist: trusty
language: cpp
services:
- docker
branches:
only:
- master
- develop
before_install:
- docker build -t travis-build -f infrastructure/tools/Dockerfile .
script:
# Common style check
- docker run --rm travis-build infrastructure/tools/checkstyle.sh $TRAVIS_COMMIT_RANGE
# Test whole project with pylint (can find issues in multiple files)
- docker run --rm travis-build /bin/bash -c "pylint --rcfile=.pylintrc ./pylua"
# Run tests
- docker run --rm travis-build /bin/bash -c "fab test_all"
|
d31abe5a6f6fc1045837ebe863e6a0ff1307c2df | requirements.txt | requirements.txt | doctr
ipython
jsonschema
matplotlib
numpydoc
path.py
sphinx
sphinx-gallery
sphinx_rtd_theme
sphinxcontrib-napoleon
| bluesky # for gallery example
doctr
ipython
jsonschema
matplotlib
numpy # for gallery example
numpydoc
path.py
sphinx
sphinx-gallery
sphinx_rtd_theme
sphinxcontrib-napoleon
| Add dependencies for gallery examples. | FIX: Add dependencies for gallery examples.
| Text | bsd-2-clause | NSLS-II/docs,tacaswell/docs,NSLS-II/docs | text | ## Code Before:
doctr
ipython
jsonschema
matplotlib
numpydoc
path.py
sphinx
sphinx-gallery
sphinx_rtd_theme
sphinxcontrib-napoleon
## Instruction:
FIX: Add dependencies for gallery examples.
## Code After:
bluesky # for gallery example
doctr
ipython
jsonschema
matplotlib
numpy # for gallery example
numpydoc
path.py
sphinx
sphinx-gallery
sphinx_rtd_theme
sphinxcontrib-napoleon
|
c7c1c94635bc499b8c33fc4c9b65033ef2c24c89 | source/popup/containers/EntriesPage.js | source/popup/containers/EntriesPage.js | import { connect } from "react-redux";
import EntriesPage from "../components/EntriesPage.js";
import { clearSearchResults } from "../library/messaging.js";
import { searchEntriesForTerm } from "../../shared/library/messaging.js";
export default connect(
(state, ownProps) => ({}),
{
onPrepare: () => () => {
clearSearchResults();
},
onSearchTermChange: searchTerm => () => {
if (searchTerm.trim().length > 0) {
searchEntriesForTerm(searchTerm);
} else {
clearSearchResults();
}
}
}
)(EntriesPage);
| import { connect } from "react-redux";
import EntriesPage from "../components/EntriesPage.js";
import { clearSearchResults } from "../library/messaging.js";
import { searchEntriesForTerm } from "../../shared/library/messaging.js";
export default connect(
(state, ownProps) => ({}),
{
onPrepare: () => () => {},
onSearchTermChange: searchTerm => () => {
if (searchTerm.trim().length > 0) {
searchEntriesForTerm(searchTerm);
} else {
clearSearchResults();
}
}
}
)(EntriesPage);
| Stop clearing search results on popup init | Stop clearing search results on popup init
| JavaScript | mit | buttercup-pw/buttercup-browser-extension,perry-mitchell/buttercup-chrome,buttercup-pw/buttercup-browser-extension,perry-mitchell/buttercup-chrome | javascript | ## Code Before:
import { connect } from "react-redux";
import EntriesPage from "../components/EntriesPage.js";
import { clearSearchResults } from "../library/messaging.js";
import { searchEntriesForTerm } from "../../shared/library/messaging.js";
export default connect(
(state, ownProps) => ({}),
{
onPrepare: () => () => {
clearSearchResults();
},
onSearchTermChange: searchTerm => () => {
if (searchTerm.trim().length > 0) {
searchEntriesForTerm(searchTerm);
} else {
clearSearchResults();
}
}
}
)(EntriesPage);
## Instruction:
Stop clearing search results on popup init
## Code After:
import { connect } from "react-redux";
import EntriesPage from "../components/EntriesPage.js";
import { clearSearchResults } from "../library/messaging.js";
import { searchEntriesForTerm } from "../../shared/library/messaging.js";
export default connect(
(state, ownProps) => ({}),
{
onPrepare: () => () => {},
onSearchTermChange: searchTerm => () => {
if (searchTerm.trim().length > 0) {
searchEntriesForTerm(searchTerm);
} else {
clearSearchResults();
}
}
}
)(EntriesPage);
|
2ed5f3f6567cd7cc541fa58a49b6cf03f07d23e5 | playbooks/roles/jenkins_worker/defaults/main.yml | playbooks/roles/jenkins_worker/defaults/main.yml | ---
jenkins_user: "jenkins"
# the main group for the jenkins user
jenkins_group: "jenkins"
# the groups that the jenkins user must belong to
jenkins_groups: "jenkins,docker"
jenkins_home: /home/jenkins
jenkins_edx_platform_version: master
# System packages
jenkins_debian_pkgs:
- pkg-config
- libffi-dev
- python-dev
- libsqlite3-dev
- libfreetype6-dev
# packer direct download URL
packer_url: "https://releases.hashicorp.com/packer/1.4.4/packer_1.4.4_linux_amd64.zip"
jenkins_node_version: "12"
ansible_distribution_release: "xenial"
jenkins_worker_python_versions:
- 2.7
- 3.5
- 3.6
| ---
jenkins_user: "jenkins"
# the main group for the jenkins user
jenkins_group: "jenkins"
# the groups that the jenkins user must belong to
jenkins_groups: "jenkins,docker"
jenkins_home: /home/jenkins
jenkins_edx_platform_version: master
# System packages
jenkins_debian_pkgs:
- pkg-config
- libffi-dev
- python-dev
- libsqlite3-dev
- libfreetype6-dev
- libpq-dev
# packer direct download URL
packer_url: "https://releases.hashicorp.com/packer/1.4.4/packer_1.4.4_linux_amd64.zip"
jenkins_node_version: "12"
ansible_distribution_release: "xenial"
jenkins_worker_python_versions:
- 2.7
- 3.5
- 3.6
| Add postgres library to handle openedxstats upgrades. | Add postgres library to handle openedxstats upgrades.
| YAML | agpl-3.0 | EDUlib/configuration,arbrandes/edx-configuration,mitodl/configuration,edx/configuration,EDUlib/configuration,mitodl/configuration,arbrandes/edx-configuration,EDUlib/configuration,arbrandes/edx-configuration,mitodl/configuration,appsembler/configuration,edx/configuration,EDUlib/configuration,edx/configuration,stvstnfrd/configuration,EDUlib/configuration,stvstnfrd/configuration,edx/configuration,appsembler/configuration,appsembler/configuration,stvstnfrd/configuration,arbrandes/edx-configuration,mitodl/configuration,appsembler/configuration,arbrandes/edx-configuration,stvstnfrd/configuration,stvstnfrd/configuration | yaml | ## Code Before:
---
jenkins_user: "jenkins"
# the main group for the jenkins user
jenkins_group: "jenkins"
# the groups that the jenkins user must belong to
jenkins_groups: "jenkins,docker"
jenkins_home: /home/jenkins
jenkins_edx_platform_version: master
# System packages
jenkins_debian_pkgs:
- pkg-config
- libffi-dev
- python-dev
- libsqlite3-dev
- libfreetype6-dev
# packer direct download URL
packer_url: "https://releases.hashicorp.com/packer/1.4.4/packer_1.4.4_linux_amd64.zip"
jenkins_node_version: "12"
ansible_distribution_release: "xenial"
jenkins_worker_python_versions:
- 2.7
- 3.5
- 3.6
## Instruction:
Add postgres library to handle openedxstats upgrades.
## Code After:
---
jenkins_user: "jenkins"
# the main group for the jenkins user
jenkins_group: "jenkins"
# the groups that the jenkins user must belong to
jenkins_groups: "jenkins,docker"
jenkins_home: /home/jenkins
jenkins_edx_platform_version: master
# System packages
jenkins_debian_pkgs:
- pkg-config
- libffi-dev
- python-dev
- libsqlite3-dev
- libfreetype6-dev
- libpq-dev
# packer direct download URL
packer_url: "https://releases.hashicorp.com/packer/1.4.4/packer_1.4.4_linux_amd64.zip"
jenkins_node_version: "12"
ansible_distribution_release: "xenial"
jenkins_worker_python_versions:
- 2.7
- 3.5
- 3.6
|
3d22a51134963c3f188edeacb937d2a6f17a29e5 | lib/env.js | lib/env.js | /**
* @file env
* @author treelite([email protected])
*/
/**
* 修复全局的异步函数
* 使之能在异步回调中恢复请求请求上下文
*
* @inner
* @param {Object} app server
*/
function fixGlobalMethods(app) {
var methods = ['setTimeout', 'setInterval', 'setImmediate'];
methods.forEach(function (name) {
var method = global[name];
global[name] = function (callback) {
var args = Array.prototype.slice.call(arguments, 1);
var id = app.stashContext();
args.unshift(function () {
app.revertContext(id);
callback();
});
method.apply(null, args);
};
});
}
/**
* 运行环境设置
*
* @public
* @param {Object} app server
*/
module.exports = function (app) {
// 修复全局异步函数
fixGlobalMethods(app);
};
| /**
* @file env
* @author treelite([email protected])
*/
/**
* 修复异步函数
* 使之能在异步回调中恢复请求请求上下文
*
* @inner
* @param {Object} app server
*/
function fixAsyncMethods(app) {
/**
* 包裹异步函数
*
* @param {Object} obj 模块
* @param {string} name 方法名
*/
function wrap(obj, name) {
var method = obj[name];
obj[name] = function (callback) {
var args = Array.prototype.slice.call(arguments, 1);
var id = app.stashContext();
args.unshift(function () {
app.revertContext(id);
callback();
});
method.apply(null, args);
};
}
var methods = ['setTimeout', 'setInterval', 'setImmediate'];
methods.forEach(function (name) {
wrap(global, name);
});
wrap(process, 'nextTick');
}
/**
* 运行环境设置
*
* @public
* @param {Object} app server
*/
module.exports = function (app) {
// 修复异步函数
fixAsyncMethods(app);
};
| Fix process.nextTick, try to save context | Fix process.nextTick, try to save context
| JavaScript | mit | ecomfe/rebas,ecomfe/rebas,noikiy/rebas | javascript | ## Code Before:
/**
* @file env
* @author treelite([email protected])
*/
/**
* 修复全局的异步函数
* 使之能在异步回调中恢复请求请求上下文
*
* @inner
* @param {Object} app server
*/
function fixGlobalMethods(app) {
var methods = ['setTimeout', 'setInterval', 'setImmediate'];
methods.forEach(function (name) {
var method = global[name];
global[name] = function (callback) {
var args = Array.prototype.slice.call(arguments, 1);
var id = app.stashContext();
args.unshift(function () {
app.revertContext(id);
callback();
});
method.apply(null, args);
};
});
}
/**
* 运行环境设置
*
* @public
* @param {Object} app server
*/
module.exports = function (app) {
// 修复全局异步函数
fixGlobalMethods(app);
};
## Instruction:
Fix process.nextTick, try to save context
## Code After:
/**
* @file env
* @author treelite([email protected])
*/
/**
* 修复异步函数
* 使之能在异步回调中恢复请求请求上下文
*
* @inner
* @param {Object} app server
*/
function fixAsyncMethods(app) {
/**
* 包裹异步函数
*
* @param {Object} obj 模块
* @param {string} name 方法名
*/
function wrap(obj, name) {
var method = obj[name];
obj[name] = function (callback) {
var args = Array.prototype.slice.call(arguments, 1);
var id = app.stashContext();
args.unshift(function () {
app.revertContext(id);
callback();
});
method.apply(null, args);
};
}
var methods = ['setTimeout', 'setInterval', 'setImmediate'];
methods.forEach(function (name) {
wrap(global, name);
});
wrap(process, 'nextTick');
}
/**
* 运行环境设置
*
* @public
* @param {Object} app server
*/
module.exports = function (app) {
// 修复异步函数
fixAsyncMethods(app);
};
|