question
stringlengths
11
28.2k
answer
stringlengths
26
27.7k
tag
stringclasses
130 values
question_id
int64
935
78.4M
score
int64
10
5.49k
I am getting the strange error below in my Jenkins pipeline [Pipeline] withDockerContainer acp-ci-ubuntu-test does not seem to be running inside a container $ docker run -t -d -u 1002:1006 -u ubuntu --net=host -v /var/run/docker.sock:/var/run/docker.sock -v /home/ubuntu/.docker:/home/ubuntu/.docker -w /home/ubuntu/workspace/CD-acp-cassandra -v /home/ubuntu/workspace/CD-acp-cassandra:/home/ubuntu/workspace/CD-acp-cassandra:rw,z -v /home/ubuntu/workspace/CD-acp-cassandra@tmp:/home/ubuntu/workspace/CD-acp-cassandra@tmp:rw,z -e ******** -e ******** -e ******** -e ******** -e ******** -e ******** -e ******** -e ******** -e ******** -e ******** -e ******** -e ******** -e ******** -e ******** -e ******** -e ******** -e ******** -e ******** -e ******** -e ******** -e ******** -e ******** -e ******** -e ******** -e ******** -e ******** -e ******** -e ******** -e ******** -e ******** -e ******** -e ******** quay.io/arubadevops/acp-build:ut-build cat $ docker top 83d04d0a3a3f9785bdde3932f55dee36c079147eb655c1ee9d14f5b542f8fb44 -eo pid,comm [Pipeline] { [Pipeline] sh process apparently never started in /home/ubuntu/workspace/CD-acp-cassandra@tmp/durable-70b242d1 (running Jenkins temporarily with -Dorg.jenkinsci.plugins.durabletask.BourneShellScript.LAUNCH_DIAGNOSTICS=true might make the problem clearer) [Pipeline] } $ docker stop --time=1 83d04d0a3a3f9785bdde3932f55dee36c079147eb655c1ee9d14f5b542f8fb44 $ docker rm -f 83d04d0a3a3f9785bdde3932f55dee36c079147eb655c1ee9d14f5b542f8fb44 [Pipeline] // withDockerContainer The corresponding stage in Jenkins pipeline is stage("Build docker containers & coreupdate packages") { agent { docker { image "quay.io/arubadevops/acp-build:ut-build" label "acp-ci-ubuntu" args "-u ubuntu --net=host -v /var/run/docker.sock:/var/run/docker.sock -v $HOME/.docker:/home/ubuntu/.docker" } } steps { script { try { sh "export CI_BUILD_NUMBER=${currentBuild.number}; cd docker; ./build.sh; cd ../test; ./build.sh;" ciBuildStatus="PASSED" } catch (err) { ciBuildStatus="FAILED" } } } } What could be the reasons why the process is not getting started within the docker container? Any pointers on how to debug further are also helpful.
This error means the Jenkins process is stuck on some command. Some suggestions: Upgrade all of your plugins and re-try. Make sure you've the right number of executors and jobs aren't stuck in the queue. If you're pulling the image (not your local), try adding alwaysPull true (next line to image). When using agent inside stage, remove the outer agent. See: JENKINS-63449. Execute org.jenkinsci.plugins.durabletask.BourneShellScript.LAUNCH_DIAGNOSTICS=true in Jenkins's Script Console to debug. When the process is stuck, SSH to Jenkins VM and run docker ps to see which command is running. Run docker ps -a to see the latest failed runs. In my case it tried to run cat next to custom CMD command set by container (e.g. ansible-playbook cat), which was the invalid command. The cat command is used by design. To change entrypoint, please read JENKINS-51307. If your container is still running, you can login to your Docker container by docker exec -it -u0 $(docker ps -ql) bash and run ps wuax to see what's doing. Try removing some global variables (could be a bug), see: parallel jobs not starting with docker workflow.
Jenkins
58,346,984
36
I am trying to activate a pipeline on any merge request change. This works as long as my pipeline script is in Jenkins UI. Now I outsourced my script on GitLab, and the checkout should happen via the pipeline via scm option. But all I get on build (yes, it triggers) is: java.lang.IllegalArgumentException: Invalid refspec refs/heads/** This happens if I leave the branch specifier empty, this is because I want to listen to any change. If I specify the branch, the build goes through. My refspec: +refs/heads/*:refs/remotes/origin/* +refs/merge-requests/*/head:refs/remotes/origin/merge-requests/*
Most likely this is a Jenkins bug. https://issues.jenkins-ci.org/browse/JENKINS-46588 There seems to a solution anyway: In your project configuration under Pipeline -> SCM -> Branches to build -> "Branch Specifier (blank for 'any'): Do not use blank for any or * or .* or **. Use: */* Another workaround would be to disable Lightweight Checkout. PS: Big thanks to ChrisAnnODell and Omurbek Kadyrbekov for linking the solutions in first place. I'm still a bit puzzled that there is no fix for over 2 years now...
Jenkins
46,684,972
36
I want to get Getting current timestamp in inline pipeline script using pipeline plugin of hudson. For setting up build display name. Inline groovy script used: def jobName = env.JOB_NAME + "_" + new Date() currentBuild.displayName = "$jobName" node { echo "job name $jobName" } Error on console : org.jenkinsci.plugins.scriptsecurity.sandbox.RejectedAccessException: Scripts not permitted to use new java.util.Date
you can also use this, I needed this in ms so: echo "TimeStamp: ${currentBuild.startTimeInMillis}" echo "TimeStamp: ${Util.getTimeSpanString(System.currentTimeMillis())}"
Jenkins
40,261,710
36
I have 10 jenkins job in folder foo. I have created a new sub folder baar in folder foo. How to move the 10 jobs from folder foo to the subfolder baar?
First, you need to install cloudbees folder plugin then you will see Move option in jobs click on it,then option(drop down) will come where you want to move select and move
Jenkins
39,406,546
36
I am triggering a parameterized Jenkins from from outside of jenkins via a http POST request: I have enabled in the job configuration that the job can be triggered from outside and i can really trigger it by sending jenkins a request with a content like this: POST http://myJenkins.com/myJob/buildWithParameters?token=MYTOKEN Parameter: SCREEN_SIZE: 27 Triggering the job creation returns a successfull 201 CREATED http response. My problem is that i dont know the id of the build job that was created. I want to monitor the state of the job. In order to do that i need to know the id. Otherwise, if i just take the latest build of that job, i could take wrong job. Is there a reliable way to get the id of the created job?
Since Jenkins 1.519, enqueuing a build responds with a URL in the Location, pointing you to an item in the build queue: $ nc localhost 8666 POST /jenkins/job/morgRemote/buildWithParameters?jenkins_status=1&jenkins_sleep=20&token=morgRemote HTTP/1.1 Host: localhost:8666 HTTP/1.1 201 Created Location: http://localhost:8666/jenkins/queue/item/39/ Content-Length: 0 Server: Jetty(winstone-2.8) Now if you add api/json (or api/xml and so on) to the end of it (so in this example it would be http://localhost:8666/jenkins/queue/item/39/api/json) then you will get a document that will contain build id for the given job. For json the retrieved object has executable attribute, which in turn has number and url attributes. number is the build id for the given job (35 here) and url is the jenkins build page url. { "actions" : [ { "parameters" : [ { "name" : "jenkins_status", "value" : "1" }, { "name" : "jenkins_sleep", "value" : "20" } ] }, { "causes" : [ { "shortDescription" : "Started by remote host 127.0.0.1", "addr" : "127.0.0.1", "note" : null } ] } ], "blocked" : false, "buildable" : false, "id" : 39, "inQueueSince" : 1423993879845, "params" : "\njenkins_status=1\njenkins_sleep=20", "stuck" : false, "task" : { "name" : "morgRemote", "url" : "http://localhost:8666/jenkins/job/morgRemote/", "color" : "red" }, "url" : "queue/item/39/", "why" : null, "cancelled" : false, "executable" : { "number" : 35, "url" : "http://localhost:8666/jenkins/job/morgRemote/35/" } } be aware of 2 things: inactive items in the build queue are garbage collected after few minutes, so you should retrieve build id ASAP by default it takes few seconds between item is added to the queue until it gets build id. During this time executable and canceled attributes will be missing and why will be not null. You can change this behavior in "Advanced Project Options" of your job config by modifying "Quiet period" setting or in the jenkins global configuration. : ... "url" : "queue/item/39/", "why" : "In the quiet period. Expires in 2.4 sec", "timestamp" : 1423993879845 }
Jenkins
24,507,262
36
I want to deploy with jenkins to the test environment and to the production environment. To do so I need to connect to the server of the wanted environment, something like ssh/scp. I would like to know what the best way is. I found some plugins to do this, like the Jenkins-Deploy-Plug-in or Jenkins Publish over SSH Plugin. The first has lots of issues, which is not really trustworthy to deploy to production and for the second you need to change the global configuration, which is manual work for every deploy. Any ideas how to solve this? Maybe with some scripts or plugins? The only current idea I have is: to connect with jenkins to a server (maybe with the SSH Plugin) and to execute there a script that connects to the wished environment. But that are two connections. Is that really neccessary? I hope for a more straightforward way for this. thanks for any hint.
I suggest the following procedure: one single shell script (stored somewhere on the jenkins server) does everything. Basically, the script does scp of the build artifact and then connects to the server (ssh) and does all the necessary tasks to deploy (setup maintenance page, backup the current app, deploy the new app, ...). On the jenkins server, there are at least 2 jobs: the first one simply does the build (using maven, or any other build script) the second job does the deploy : so this job only runs the shell script. (I suggest one deploy job for each target environment : testing, production, ...) It does not require any "special" jenkins plugin to achieve this "one click deployment". It only requires that the jenkins user has ssh access to the target server. EDIT Here is a sample shell script to illustrate my post #This script will copy the last artifact build by the job "MyApp" to test.myserver.com #and remotely execute the deployment script. #copy the war to the server #(the job "MyApp" is using maven, that's why the war can be found at this location) scp -i <HOME_DIR>/.ssh/id_dsa $HUDSON_HOME/jobs/MyApp_Build/workspace/myapp/target/myapp.war [email protected]:/tmp/ #connect to the server and execute the deployment script ssh -i <HOME_DIR>/.ssh/id_dsa [email protected] #The following is just an example of what a deployment script can be. #of course you must adapt it to your needs and environment "cd <TOMCAT_DIR>; #first copy the current war to a backup directory (additionaly, I have a cron task deleting old undeployed apps) cp -rf myapp-apps/myapp* undeployed/myapp-apps/; #execute a script (stored on the server) to properly stop the app sh bin/myapp.sh stop; #delete current app rm -rf myapp-apps/myapp; rm -rf myapp-apps/myapp.war; #copy the uploaded war in tomcat app directory cp /tmp/myapp.war myapp-apps/; #execute a script (stored on the server) to start the app sh bin/myapp.sh start"
Jenkins
13,976,373
36
I'm running a Jenkins CI server on an OS X machine. The server is running as a standard user 'john', and is started by running launchctl. One of the things this server does is build XCode projects using keys and certificates stored in a keychain 'xcode.keychain': Jenkins (which is running under the user 'john' according to activity monitor) calls these commands from a script when the user presses a button on the web interface. security default-keychain -s /Users/john/Library/Keychains/xcode.keychain security unlock-keychain -p password /Users/john/Library/Keychains/xcode.keychain xcodebuild ... If I happen to be logged into the server as 'john' via the UI, the keychain gets unlocked properly when Jenkins calls those commands. But, if I'm not logged in, xcode.keychain doesn't get unlocked and the build fails. Any ideas?
I had to: Right-click on the private key in my keychain that my build process was trying to use Click "Get Info" Then the "Access Control" tab. You can add specific apps (like "codesign") to the list of apps that are allowed access to that key, or just allow access from all applications. This cleared it up for me. More info in these comments: https://stackoverflow.com/a/12235462/544130 https://stackoverflow.com/a/14761060/544130
Jenkins
6,416,121
36
I need to build my application using Java 11. However the dropdown menu stops at Java 9. What do I do? OpenJDK is acceptable too. I'm on the latest version of Jenkins. Edit: as of now I've downloaded the binaries using wget, extracted them on the machine, and added a JDK JDK_HOME entry via Global Configurations.
I guess you are using the JDK Tool Plugin. Click "Manage Jenkins" > "Global Tool Configuration" > "Add JDK" (near JDK installations) Delete the java.sun.com installer. Just click "Add Installer" below and choose "Extract .zip/.tar.gz" Enter following: Label: openjdk-11 Download URL: https://download.java.net/java/GA/jdk11/13/GPL/openjdk-11.0.1_linux-x64_bin.tar.gz Subdirectory of extracted archive: jdk-11.0.1 (Optional subdirectory of the downloaded and unpacked archive to use as the tool's home directory.) And "Save" the configuration => Use JDK label (openjdk-11) in your build job. The download-link given above appears to be from a time when java 11 hadn't entered LTS; Instead go to Java Platform, Standard Edition 11 Reference Implementations which will provide you the download-link to the most recent release of the... [...] official Reference Implementation for Java SE 11 (JSR 384) [...] based solely upon open-source code available from the JDK 11 Project in the OpenJDK Community. This Reference Implementation applies to both the Final Release of JSR 384 (Sep 2018) and Maintenance Release 1 (Mar 2019). ... which as of this writing is: https://download.java.net/openjdk/jdk11/ri/openjdk-11+28_linux-x64_bin.tar.gz
Jenkins
55,243,120
35
I have a job with as cron: 5 3,21 * * 1-5 This will run my job at 03:05AM and 09:05PM. Now I read it's a best practice to use H. I try: H/5 3,21 * * 1-5 What is the meaning now? Will this schedule a build in a range of 5 minutes or 5 minutes after 3AM and 21PM?
The H will take a numeric hash of the Job name and use this to ensure that different jobs with the same cron settings do not all trigger at the same time. This is a form of Scheduling Jitter H/5 in the first field means Every five minutes starting at some time between 0 and 4 minutes past the hour So H/5 3,21 * * 1-5 is Every five minutes between 03:00 and 03:59 and then between 21:00 and 21:59 on Mon -> Fri but starting at some 'random' time between 03:00 and 03:04 and then the same number of minutes after 21:00 If you want to run once per hour between 03:00-03:59 and 21:00-21:59 on Mon -> Fri, you can use H 3,21 * * 1-5 where H will be replaced by some number between 0-59 depending on job name. The user interface will tell you when the job will last have triggered and will next trigger
Jenkins
47,302,607
35
I want to be able to say something like: git branch: commitHash, credentialsId: credentialsId, url: url The usecase: I'm doing parallel build and test runs on different platforms, and want to ensure each gets the same code. It is C++, and we build on separate platforms as well as building on them. If I do the above, it fails - the underlying code assumes the given branch actually is a branch, or you get something like: [Linux64 Build] > git rev-parse origin/e4b6c976a0a986c348a211579f1e8fd32cf29567^{commit} # timeout=10 [Pipeline] [Linux64 Build] } [Pipeline] [Linux64 Build] // dir [Pipeline] [Linux64 Build] } [Pipeline] [Linux64 Build] // node [Pipeline] [Linux64 Build] } [Linux64 Build] Failed in branch Linux64 Build I've seen variations on this question asked before, although no actual answers - just suggestions like to stash the source code instead, etc. Not really what I'm after. The documentation suggests it ought to be possible to give explicit commit hashes, possibly using branches instead, but I can't work out the syntax and can't find any examples. When I do it, I get the master branch, I think - in our setup, master does not work. So far, the only solution I've found has been to checkout the branch and then explicitly call git to get the commit: git branch: branch, credentialsId: credentialsId, url: url sh 'git checkout ' + commitHash (where branch is the branch I originally got the hash for at the top of the job. It works but is not the neatest. Anybody got a better way?
Use a general scm step checkout([$class: 'GitSCM', branches: [[name: commitHash ]], userRemoteConfigs: [[url: 'http://git-server/user/repository.git']]])
Jenkins
43,611,673
35
In one of my stages I need to copy the contents of two folders after a build is completed and copy to a different directory. I am actually converting a freestyle job to pipeline, and have been using the artifact deployer plugin. Reading around, it looks like stash and unstash commands should help with what I want to achieve. Can someone verify if this is the correct approach below please? stage('Build') { steps { sh ''' gulp set-staging-node-env gulp prepare-staging-files gulp webpack ''' stash includes: '/dist/**/*', name: 'builtSources' stash includes: '/config/**/*', name: 'appConfig' dir('/some-dir') { unstash 'builtSources' unstash 'appConfig' } } } If I change dir in one stage, does that mean all other stages thereafter will try to execute commands from that directory, or do they do back to using the workspace default location? Thanks EDIT I have realised what I actually want to do is to copy built sources to a different node (running a different OS). So in my snippet I have shared, where I am switching directories, that directory is actually to be on a different machine (node) that I have setup. Would I need to wrap the dir() block with a node('my-node-name') block? Im struggling to find examples. Thanks
I hope it is meant to be this: stash includes: 'dist/**/*', name: 'builtSources' stash includes: 'config/**/*', name: 'appConfig' where dist and config are the directories in the workspace path, so it should be a relative path like above. Rest seems alright, only to mention that path "/some-dir" should be writable by jenkins user (user used to run jenkins daemon). And yes it falls back to its then enclosing workspace path (in this case default) when it exits dir block. EDIT So when you stash a path, it is available to be unstashed at any step later in the pipeline. So yes, you could put dir block under a node('<nodename>') block. You could add something like this : stage('Move the Build'){ node('datahouse'){ dir('/opt/jenkins_artifacts'){ unstash 'builtSources' unstash 'appConfig' } } }
Jenkins
43,050,248
35
I'm trying to access a variable from an input step using the declarative pipelines syntax but it seems not to be available via env or params. This is my stage definition: stage('User Input') { steps { input message: 'User input required', ok: 'Release!', parameters: [choice(name: 'RELEASE_SCOPE', choices: 'patch\nminor\nmajor', description: 'What is the release scope?')] echo "env: ${env.RELEASE_SCOPE}" echo "params: ${params.RELEASE_SCOPE}" } } Both echo steps print null. I also tried to access the variable directly but I got the following error: groovy.lang.MissingPropertyException: No such property: RELEASE_SCOPE for class: groovy.lang.Binding at groovy.lang.Binding.getVariable(Binding.java:63) at org.jenkinsci.plugins.scriptsecurity.sandbox.groovy.SandboxInterceptor.onGetProperty(SandboxInterceptor.java:224) What is the correct way to access this choice parameter?
Since you are using declarative pipelines we will need to do some tricks. Normally you save the return value from the input stage, like this def returnValue = input message: 'Need some input', parameters: [string(defaultValue: '', description: '', name: 'Give me a value')] However this is not allowed directly in declarative pipeline steps. Instead, what you need to do is wrap the input step in a script step and then propagate the value into approprierte place (env seems to work good, beware that the variable is exposed to the rest of the pipeline though). pipeline { agent any stages { stage("foo") { steps { script { env.RELEASE_SCOPE = input message: 'User input required', ok: 'Release!', parameters: [choice(name: 'RELEASE_SCOPE', choices: 'patch\nminor\nmajor', description: 'What is the release scope?')] } echo "${env.RELEASE_SCOPE}" } } } } Note that if you have multiple parameters in the input step, then input will return a map and you need to use map references to get the entry that you want. From the snippet generator in Jenkins: If just one parameter is listed, its value will become the value of the input step. If multiple parameters are listed, the return value will be a map keyed by the parameter names. If parameters are not requested, the step returns nothing if approved.
Jenkins
42,501,553
35
I have a file pipeline.gdsl that contains the Syntax for my Jenkins Pipeline DSL. Following this blog post I put the file into the /src folder of my Java project. When I now edit my Jenkinsfile (residing in the root folder of my project), I don't get any code completion / syntax explanation as I would expect. My project is a Java / Gradle project and I cannot make a Groovy project out of it. Is there some other way to make IntelliJ aware of the .gdsl file and provide code completion?
The problem was, that /src was not marked as a source root folder in my project. Creating a folder /src/main/groovy, putting the file in there and marking it as a sources root (right click on the folder -> Mark directory as -> Sources Root) did the trick.
Jenkins
41,062,514
35
I am attempting to write a pipeline script to use with Jenkins 2.0 to replicate our existing build. This original build used the envInject plugin to read a Java properties file, but I can't see how to do this from the pipeline Groovy script. I have Googled and found the following, but it doesn't work (FileNotFoundException): Properties props = new Properties() File propsFile = new File('./Builder/project.properties') props.load(propsFile.newDataInputStream()) Thanks!
I just fought with this yesterday and today. I wish the availability of this was easier to find. Grab the 'Pipeline Utility Steps' plugin. Use the readProperties step. def props = readProperties file: 'dir/my.properties' One word of warning - what I expected to be booleans in the properties files were treated as strings.
Jenkins
39,619,093
35
I am currently testing the pipeline approach of Jenkins 2.0 to see if it works for the build environment I am using. First about the environment itself. It currently consists of multiple SCM repositories. Each repository contains multiple branches, for the different stages of the development and each branch is build with multiple configurations. Not all configurations apply to every repository. Currently every repository/branch is setup as a Matrix Project for the different configurations. Each project exposes it's building results as a artifact and these artifacts are used in the downstream projects. The different repositories depend on each other, so a successful build on a upstream job triggers some specific down stream jobs. Currently all that works, but the amount of work required to setup a new branch or to tweak the building process is a lot, since many different projects need to be altered by hand. Now I wanted to give the new pipelines a try. My idea was to create multi-branch pipeline projects and place a Jenkinsfile inside the repository containing the instructions for the build. The main problem is getting the builds to trigger each other, because basically a build in a specific upstream branch, needs to trigger a downstream branch. How ever the information what downstream branches need to be triggered is not known to the upstream project. Each downstream project fetches the artifacts from some upstream branches and the ideal solution would be if the downstream build would be triggered in case the upstream build that is the source for the artifact finishes it's build. The problem is only the downstream projects really know what artifacts they require. The branch names are unlikely to match in most cases and that makes triggering the builds from the upstream project very difficult. Currently this is solved using the ReverseBuildTrigger. But this thing stops working as soon as it gets anywhere close to a pipeline. I am really at a loss how to get this working. Is there any way to get something like the ReverseBuildTrigger working inside pipeline scripts? Also triggering the entire downstream build for all branches in case a single branch upstream is changed is not a option. This would create far too many equal builds.
If you're using a declarative multi-branch pipeline, you can use: triggers { upstream(upstreamProjects: "some_project/some_branch", threshold: hudson.model.Result.SUCCESS) } If you wish for branch matching to occur across dependencies you can use: triggers { upstream(upstreamProjects: "some_project/" + env.BRANCH_NAME.replaceAll("/", "%2F"), threshold: hudson.model.Result.SUCCESS) } Note: if you use "organizations" in github, then you must add your organization name as a prefix before the project name so it looks like this: triggers { upstream(upstreamProjects: 'some_organization/some_project/some_branch', threshold: hudson.model.Result.SUCCESS) } Note on 2 useless logs: after adding "triggers{upstream(...)" section, in case your "upstreamProjects" is not good (for example you forgot to add organization or did typo in the upstream project name) you will see Jenkins successfully rebuilding your 1 project, but you will not see anything trigger-related in either of 2 logs: Jenkins system log: [jenkins-fqdn]/manage/log/all Job-related pipeline log: blue ocean -> job -> Artifacts -> pipeline.org available by URL: [jenkins-fqdn]/blue/rest/organizations/jenkins/pipelines/some_organization/pipelines/some_project/branches/master/runs/4/log/?start=0 Only after your "triggers" section is good/working you will see at the end of you upstream project pipeline.log line like this: [Pipeline] End of Pipeline Triggering a new build of some_organization » some_project » master #4 GitHub has been notified of this commit’s build result Finished: SUCCESS And the corresponding downstream project will have pipeline.log 1st line: Started by upstream project "some_organization/some_project/some_branch" build number 4 Instead of what you saw before when the build was triggered by github webhooks: Push event to branch some_branch
Jenkins
36,825,103
35
I have no idea why after Jenkins is updated to version 1.591 (Ubuntu Server 12.04), the originally correctly set up reverse proxy now becomes broken. My current setting is exactly the same as said in Jenkins wiki: ProxyPass /jenkins http://localhost:8081/jenkins nocanon ProxyPassReverse /jenkins http://localhost:8081/jenkins ProxyPreserveHost On ProxyRequests Off AllowEncodedSlashes NoDecode <Proxy http://localhost:8081/jenkins*> Order deny,allow Allow from all </Proxy> also --prefix=/jenkins has been added into /etc/default/jenkins file Is that a bug in Jenkins?
I was faced with this issue with Jenkins as a Windows Service Package. According to their wiki: Make sure the Jenkins URL configured in the System Configuration matches the URL you're using to access Jenkins. To reach the System Configuration: Go to your Jenkins page Click Manage Jenkins Click Configure System Scroll to Jenkins Location and find Jenkins URL. Ensure that port value matches with the port value set in the <arguments> section of the jenkins.xml file located in the Jenkins folder on your machine.
Jenkins
27,161,854
35
It seems like this should be easy to integrate CMake+CTest in jenkins. The cmakebuilder plugin is extremely easy to configure (just set the source tree and the build tree, done!). However I fail to understand how to call the CTest steps. According to the main xUnit page, since version 1.58 the XML output from CTest is supported, see bug report. That's about all the documentation I could find. When I search on google or on stackoverflow, I can only find very old documentation requiring manual steps. I would like to know how to setup a recent jenkins (1.532.1) with xUnit (1.81). Should I create a 'Add build-step' ? Should I create a 'post-build action' ? What do I need to fill in to get CTest to run and to produce proper XML files, so that jenkins can integrate them ?
Here is a small example that demonstrates how to have xUnit pick up a CTest generated XML test result file. The example consists of a single C++ source file main.cpp #include <cstdlib> int main() { std::exit(-1); } and an accompanying CMakeLists.txt: cmake_minimum_required(VERSION 2.8) project(JenkinsExample) enable_testing() add_executable(main main.cpp) add_test(test1 main) add_test(test2 main) set_tests_properties(test2 PROPERTIES WILL_FAIL ON) The CMake list file generates an executable main and runs this executable from two CMake tests, where for demonstration purposes one will always fail and the other one will always succeed. Using Jenkins, set up a new job and a add new cmakebuilder build step. Point the CMake source directory to the folder that contains the example project on your local machine. The CMake build directory should be set to build. This will make Jenkins create a CMake build folder in the job's workspace directory. It's a good idea to set the Clean Build flag to make the job always start with a clean state. Then, assuming you are running Unix, add an Execute shell job step and enter the following shell script in the Command box: cd build /usr/local/bin/ctest --no-compress-output -T Test || /usr/bin/true Running ctest with the option -T Test will make CTest generate an XML output file in a sub-folder Testing inside the build folder, which can be picked up by the xUnit plug-in in a post-build action then. The || /usr/bin/true is necessary to prevent Jenkins from aborting the build prematurely (without running the xUnit plug-in) if some tests fail. If you are using Windows, set up a similar Execute Windows batch command job step instead: cd build "C:\Program Files (x86)\CMake 2.8\bin\ctest.exe" --no-compress-output -T Test || verify > NUL Finally the xUnit plug-in must be configured in the following way: Add a Publish xUnit test result report post-build action and then use the plug-in's Add button to create a CTest-Version test result report. In the CTest-Version (default) Pattern enter the file pattern build/Testing/**/Test.xml.
Jenkins
21,633,716
35
I have installed Jenkins by deploying its WAR file to Tomcat. On typing http://localhost:8080/jenkins In browser, jenkins home page is opening which means jenkins is successfully installed. I configured system settings, gave jdk and maven path and save them. Then to install plugins, I clicked on Jenkins->Manage plugins and clicked on Available tab but could not find any plugins. I tried three solutions: Configured proxy for Jenkins by going to Jenkins->Manage Plugins->Advanced(did not find plugins) Restarted server, refreshed browser and went to Jenkins->Manage plugins->Available (still did not find any plugins). So, I read somewhere that we have update plugins forcefully if they are not updated automatically. So, went to Jenkins->Manage Plugins->Advanced and clicked the tab 'Check now' (Still did not find any plugins on clicking on Available tab). Finally I read somewhere that if we add the pluginGroup 'org.jvnet.hudson.tools' to settings.xml file of maven, problem may be resolved. So, added the corresponding code to settings.xml: Then I tried again but still could not find any plugins in Jenkins->Manage plugins->Available If any other solution is there which can resolve this problem please let me know.
Go to: Manage Jenkins → Manage Plugins → Advanced, then click Check now in the bottom right-hand corner. When you go back to Available tab all plugins should be listed.
Jenkins
16,213,982
35
How can I change the location where jenkins store temp data in its slaves. Currently, it shuts down the connection to my slaves because it complains about the following Disk space is too low. Only 0.119GB left on /tmp. I want to move the tmpdir location to something like /var/tmp/ instead of /tmp. How can I do that?
Just add "-Djava.io.tmpdir=/path/to/tmp" to the java command line options (you don't need any extra service wrapper). Depending on your installation there might be an existing startup script and/or config file this can go into. On my fedora system, I can add the option to the /etc/sysconfig/jenkins file: ## Type: string ## Default: "-Djava.awt.headless=true" ## ServiceRestart: jenkins # # Options to pass to java when running Jenkins. # JENKINS_JAVA_OPTIONS="-Djava.awt.headless=true -Djava.io.tmpdir=/var/tmp"
Jenkins
15,675,783
35
When I run my selenium test (mvn test) from jenkins (windows) I see only the console output. I don't see the real browsers getting opened . How can I configure jenkins so that I can see the browsers running the test?
I had the same problem, i got the solution after many attempts. This solution works ONLY on windows XP If you are using jenkins as a windows service you need to do the following : 1) In windows service select the service of jenkins 2) Open properties window of the service -> Logon-> enable the checkbox "Allow service to interact with desktop" After then you should reboot the service jenkins Hope this help you :) UPDATE: Actually, I'm working on a an automation tool using Selenium on Windows 10, I've installed Jenkins ver. 2.207 as windows application (EXE file), it's running as windows service and ALL drivers (Chrome, FireFox, IE) are visible during test executions WITHOUT performing a mere configuration on the System or Jenkins
Jenkins
9,618,774
35
I am trying to create a bash script for setting up Jenkins. Is there any way to update a plugin list from the Jenkins terminal? At first setup there is no plugin available on the list i.e.: java -jar jenkins-cli.jar -s `http://localhost:8080` install-plugin dry won't work
A simple but working way is first to list all installed plugins, look for updates and install them. java -jar /root/jenkins-cli.jar -s http://127.0.0.1:8080/ list-plugins Each plugin which has an update available, has the new version in brackets at the end. So you can grep for those: java -jar /root/jenkins-cli.jar -s http://127.0.0.1:8080/ list-plugins | grep -e ')$' | awk '{ print $1 }' If you call install-plugin with the plugin name, it is automatically upgraded to the latest version. Finally you have to restart jenkins. Putting it all together (can be placed in a shell script): UPDATE_LIST=$( java -jar /root/jenkins-cli.jar -s http://127.0.0.1:8080/ list-plugins | grep -e ')$' | awk '{ print $1 }' ); if [ ! -z "${UPDATE_LIST}" ]; then echo Updating Jenkins Plugins: ${UPDATE_LIST}; java -jar /root/jenkins-cli.jar -s http://127.0.0.1:8080/ install-plugin ${UPDATE_LIST}; java -jar /root/jenkins-cli.jar -s http://127.0.0.1:8080/ safe-restart; fi
Jenkins
7,709,993
35
I have set up Jenkins, but I would like to find out what files were added/changed between the current build and the previous build. I'd like to run some long running tests depending on whether or not certain parts of the source tree were changed. Having scoured the Internet I can find no mention of this ability within Hudson/Jenkins though suggestions were made to use SVN post-commit hooks. Maybe it's so simple that everyone (except me) knows how to do it! Is this possible?
I have done it the following way. I am not sure if that is the right way, but it seems to be working. You need to get the Jenkins Groovy plugin installed and do the following script. import hudson.model.*; import hudson.util.*; import hudson.scm.*; import hudson.plugins.accurev.* def thr = Thread.currentThread(); def build = thr?.executable; def changeSet= build.getChangeSet(); changeSet.getItems(); ChangeSet.getItems() gives you the changes. Since I use accurev, I did List<AccurevTransaction> accurevTransList = changeSet.getItems();. Here, the modified list contains duplicate files/names if it has been committed more than once during the current build window.
Jenkins
6,260,383
34
I am using the Pipeline plugin in Jenkins by Clouldbees (the name was Workflow plugin before), I am trying to get the user name in the Groovy script but I am not able to achieve it. stage 'checkout svn' node('master') { // Get the user name logged in Jenkins }
Did you try installing the Build User Vars plugin? If so, you should be able to run node { wrap([$class: 'BuildUser']) { def user = env.BUILD_USER_ID } } or similar.
Jenkins
35,902,664
34
Maybe a fool question, I installed jenkins on windows by default, have set no user/password, it worked at first, no need to login. But when launch the 8080 webpage now, it hangs the login page, I've tried some normal user/password combinations, none could pass. Also searched the resolution on website, only find some about linux, no about windows, so need help. jenkins login page
You can try to re-set your Jenkins security: Stop the Jenkins service Open the config.xml with a text editor (i.e notepad++), maybe be in C:\jenkins\config.xml (could backup it also). Find this <useSecurity>true</useSecurity> and change it to <useSecurity>false</useSecurity> Start Jenkins service You might create an admin user and enable security again. Note: On more recent Jenkins versions running on Windows the config.xml file is found here: C:\Windows\System32\config\systemprofile\AppData\Local\Jenkins\.jenkins\
Jenkins
39,340,322
34
I'm trying to mask a password in my Jenkins build. I have been trying the mask-passwords plugin. However, this doesn't seem to work with my Jenkins pipeline script, because if I define the password PASSWD1 and then I use it in the script like this ${PASSWD1}, I am getting: No such DSL method '$' found among steps [addToClasspath, ansiColor, ansiblePlaybook, ....] If I use env.PASSWD1, then its value will be resolved to null. So how should I mask a password in a Jenkins pipeline script?
The simplest way would be to use the Credentials Plugin. There you can define different types of credential, whether it's a single password ("secret text"), or a file, or a username/password combination. Plus other plugins can contribute other types of credentials. When you create a credential (via the Credentials link on the main Jenkins page), make sure you set an "ID". In the example below, I've called it my-pass. If you don't set it, it will still work, Jenkins will allocate an opaque UUID for you instead. In any case, you can easily generate the required syntax with the snippet generator. withCredentials([string(credentialsId: 'my-pass', variable: 'PW1')]) { echo "My password is '${PW1}'!" } This will make the password available in the given variable only within this block. If you attempt to print the password, like I do here, it will be masked.
Jenkins
42,371,909
34
I have a fresh install of Jenkins as a service on my Linux machine. When Jenkins installs, it creates a 'jenkins' user, but I can't seem to find the default password for it anywhere. I'm trying to secure my system, so if the default password is '123' or something insecure that I just haven't thought of yet, that's a problem. Thanks!
I don't believe it has any password. You should be able to do: sudo passwd jenkins This will prompt for you to set a password. Alternatively you could create the jenkins user prior to installing, and it would leverage that one.
Jenkins
25,041,125
34
I have written a simple script via PowerShell to gather some files and zip them into one folder, lets call it Script.ps1. I want to make the script run every time Jenkins does a new build, however I also would like the name of the zip file to be the BUILD_NUMBER. How can I create a variable in PowerShell that is Jenkins's current build number? As of the moment I am calling the Script.ps1 in the execute shell section of configuration.
I'm not familiar with Jenkins, but I believe BUILD_NUMBER is an environment variable. To access it in PowerShell, use $env:BUILD_NUMBER E.g. If using 7-zip 7z.exe a "$env:BUILD_NUMBER.zip" C:\Build\Path
Jenkins
24,291,827
34
I've been researching this for a good few hours now, but I've only found pieces of the big picture. Everywhere they are assuming that the reader already has a part of the system set up. I think it will be useful to have a big picture description of the parts needed to put the whole thing together. They all say "use your maven selenium tests" and so on and so forth. EDIT: After some research I found out I need to install Maven in Jenkins and on my computer, install a maven plugin for Eclipse, and create/convert my projects as Maven projects. How do I transfer my Maven projects in Jenkins? Do I export to .jar, or do I move the whole folder on the server? How do I connect the whole thing together with xvfb? So here is what I know so far Install Jenkins (we already have that on our server) Install plugins for Jenkins (which ones?) Install xvfb so tests are run in a headless browser (how do I specify that in the Java written test?) Install Maven on computer, jenkins and eclipse, use maven projects. Which part of my project folder from the eclipse workplace should I upload on the server and where? I have a testng.xml file and some classes (which are the acutal tests) How do I tell Jenkins to automatically run the Selenium Webdriver tests after deploy, and which file do I point to? How to get reports - through TestNg or through some Jenkins feature?
Responses, following your list: Q1. Install Jenkins (we already have that on our server) A1. None needed. Q2. Install plugins for Jenkins (which ones?) A2. As far as I remember no specific plugin is required just for this purpose. Jenkins should be able to run maven or ant job, it's out of the box. Q3. Install xvfb so tests are run in a headless browser (how do I specify that in the Java written test?) A3. In your Java tests you will be specifying the host where the browser should be launched (more technically, the host that runs selenium server). It's normally 'localhost', but for this case it will be different (it is generally not a good idea to run jenkins and selenium on the same box). So, in your java code you indicate that host with xvfb AND with selenium grid (that listens to port 4444 by default). It is also considered some good practice to factor this information out of the code (property files and, further, variables in the pom file, or provided by jenkins). Q4. Install Maven on computer, jenkins and eclipse, use maven projects. A4. Maven should be installed on jenkins host (and your local machine, the one you use to develop tests). Q5. Which part of my project folder from the eclipse workplace should I upload on the server and where? I have a testng.xml file and some classes (which are the acutal tests) A5. Your code is placed under version control (right?), so you point jenkins to fetch your project (then compile code, compile tests, run tests...). The answer is "at least all code that is needed to compile your tests and run them". Jenkins builds your project from source and test execution is just a phase of this process. Q6. How do I tell Jenkins to automatically run the Selenium Webdriver tests after deploy, and which file do I point to? A6. use 'integration-test' phase served by surefire plugin. Q7. How to get reports - through TestNg or through some Jenkins feature? A7. Jenkins will display (and distribute, if set up this way) the reports generated by testng.
Jenkins
17,719,385
34
I have a Maven job in Jenkins. Before the actual build step I have an "Execute shell" pre-build step. In that shell I set a variable: REVISION=$(cat .build_revision) I would like to use that variable in the Maven build job in "Goals and options": clean install -Drevision=${REVISION} But that does not work! The "Drevision" is set to "${REVISION}" not the actual value of ${REVISION}. Output: Executing Maven: -B -f /home/gerrit/.jenkins/jobs/<job_name>/workspace/pom.xml clean install -Drevision=${REVISION} It works with Jenkins environment variables: clean install -Dbuild=${BUILD_NUMBER} It sets "Dbuild" to the actual build number. Output: Executing Maven: -B -f /home/gerrit/.jenkins/jobs/<job_name>/workspace/pom.xml clean install -Dbuild=54 My question: How to use a shell variable in Maven "Goals and options"?? EDIT: I tried using Jenkins EnvInject Plugin to "Inject environment variables" after the pre-build shell, and my variable is now accessible by e.g. post-build shells, but it is still not available in Maven "Goals and options". Then it is possible to set "Inject environment variables to the build process" using the EnvInject Plugin, which actually makes those variables available in Maven "Goals and options", but those are set right after SCM checkout, i.e. before pre-build steps, and do not support expression evaluations.
You're on the right track here, but missed a third feature of the EnvInject-Plugin: The "Inject environment variables" build step that can inject variables into following build steps based on the result of a script or properties. We're using the EnvInject plugin just like that; A script sets up a resource and communicates its parameters using properties that are then propagated by the plugin as environment variables. i.e. setting up a temporary database for the build:
Jenkins
16,332,659
34
I am new to Jenkins, I am getting following error while cloning repository from GitHub. I tried to search all relevant issues here but could find exact stacktstrace with answers. I am trying to clone repository which requires username and password, I am providing SSH:// repository-path in job configuration settings for my job. I have not done any .ssh related settings yet as this seems to be different problem than security issue with GIT repository. I couldn't figure out what comment Jenkins is trying to execute, which file/directory it is not finding clueless. Here is my exact stack trace from Jenkins job: Caused by: java.io.IOException: Cannot run program: Error trying to determine the git version: Error performing command: --version Assuming 1.6 ERROR: Error cloning remote repo 'myRE' : Could not clone [email protected]:myORG/RVL.myProj.git hudson.plugins.git.GitException: Could not clone [email protected]:myORG/RVL.myProj.git at hudson.plugins.git.GitAPI.clone(GitAPI.java:268) at hudson.plugins.git.GitSCM$2.invoke(GitSCM.java:1122) at hudson.plugins.git.GitSCM$2.invoke(GitSCM.java:1064) at hudson.FilePath.act(FilePath.java:842) at hudson.FilePath.act(FilePath.java:824) at hudson.plugins.git.GitSCM.checkout(GitSCM.java:1064) at hudson.model.AbstractProject.checkout(AbstractProject.java:1256) at hudson.model.AbstractBuild$AbstractBuildExecution.defaultCheckout(AbstractBuild.java:589) at jenkins.scm.SCMCheckoutStrategy.checkout(SCMCheckoutStrategy.java:88) at hudson.model.AbstractBuild$AbstractBuildExecution.run(AbstractBuild.java:494) at hudson.model.Run.execute(Run.java:1502) at hudson.maven.MavenModuleSetBuild.run(MavenModuleSetBuild.java:477) at hudson.model.ResourceController.execute(ResourceController.java:88) at hudson.model.Executor.run(Executor.java:236) Caused by: hudson.plugins.git.GitException: Error performing command: clone -o RVL.myProj [email protected]:myORG/RVL.myProj.git /var/lib/jenkins/jobs/myProj/workspace at hudson.plugins.git.GitAPI.launchCommandIn(GitAPI.java:862) at hudson.plugins.git.GitAPI.access$000(GitAPI.java:40) at hudson.plugins.git.GitAPI$1.invoke(GitAPI.java:264) at hudson.plugins.git.GitAPI$1.invoke(GitAPI.java:244) at hudson.FilePath.act(FilePath.java:842) at hudson.FilePath.act(FilePath.java:824) at hudson.plugins.git.GitAPI.clone(GitAPI.java:244) ... 13 more Caused by: java.io.IOException: Cannot run program "": error=2, No such file or directory at java.lang.ProcessBuilder.start(Unknown Source) at hudson.Proc$LocalProc.(Proc.java:244) at hudson.Proc$LocalProc.(Proc.java:216) at hudson.Launcher$LocalLauncher.launch(Launcher.java:709) at hudson.Launcher$ProcStarter.start(Launcher.java:338) at hudson.Launcher$ProcStarter.join(Launcher.java:345) at hudson.plugins.git.GitAPI.launchCommandIn(GitAPI.java:843) ... 19 more Caused by: java.io.IOException: error=2, No such file or directory at java.lang.UNIXProcess.forkAndExec(Native Method) at java.lang.UNIXProcess.(Unknown Source) at java.lang.Pro
I encountered and fixed the same problem :) There are two way to configure the path of git: On Jenkins Master a. Enter Jenkins System Configure (Jenkins -> Manage Jenkins -> Configure System ) b. Find the Git item and Configure the git installation (specify the git path on Jenkins Master) On Jenkins Slave a. Enter Jenkins Slave's Configure b. Check the "Tool Locations" Checkbox and specify the path of git on the Jenkins Slave. In my situation, I don't have the privilege to access the Jenkins Master. So I install the git on the Jenkins Slave and add it to Jenkins Slave's Configure.
Jenkins
12,202,078
34
In Jenkins is there a plugin for parameterized builds to make the parameters required? The fields under the standard "This build is parameterized" option do not seem to provide that. Clarification: by "required" I mean that the build will not execute until the field is populated with a value. This would obviously preclude automated triggers.
The accepted answer is no longer valid. There was a plugin that did that but is no longer maintained. There's an open bug to support it. In the mean time what you can do is check if your parameter is present and if not throw an error like: if (!params.SomeParam) { error("Build failed because of this and that..") }
Jenkins
10,742,401
34
I have two Jenkins projects that share a database. They must not be run simultaneously. Strictly speaking, there is no particular dependency between them beyond non concurrency, but at the moment I partially manage this constraint by running one "downstream" of the other. This works most of the time, but not always. If a source control change happens while the second is running, the first will start up again, and they'll be running concurrently and probably both fail miserably. This is similar, but not identical, to How to prevent certain Jenkins jobs from running simultaneously? The difference is that I don't have a "number of threads" problem -- I'm already only running at most one thread of any given project at any one time, even in the case where two (different-project) builds stomp each other. This seems to rule out all the several suggestions in that thread.
The Locks and Latches plugin should resolve your problem. Create a lock and have both jobs use the same lock. That will prevent the jobs from running concurrently. Install the plugin in "Manage Jenkins: Manage Plugins." Define (provide a name for) your lock(s) in "Manage Jenkins: Configure System." For each job you want to participate in the exclusion, in ": Configure: Build Environment," check "Locks", and pick your lock name from the drop list.
Jenkins
10,115,759
34
I currently set up a Jenkins Multibranch Pipeline job that is based on a Git repository hosted on our GitLab server. Jenkins can read the branches in the repository and creates a job for every branch in the repository. But I can't figure out how to trigger the jobs with webhooks in GitLab. My questions are: How can I trigger the creation of a new branch job in Jenkins from our GitLab server? I can't see a webhook for a new branch being pushed. How do I trigger the actual build job for a single branch? I can only add a webhook for push events but then I would have to add the branch name which I don't know how to do. How can I make sure that GitLab always triggers the "creation of the branch job" before a push to a branch triggers the build job itself. What I tried so far is triggering the multi-branch job, but this has no effect and following this post does not work at all.
You need to install the GitLab Plugin on Jenkins. This will add a /project endpoint on Jenkins. (See it in Jenkins => Manage Jenkins => Configure System => GitLab) Now add a webhook to your GitLab project => Settings => Integrations. (or in older GitLab versions: GitLab project => Wheel icon => Integrations, it seems you need to be owner of the project in this case) Set the URL to http://*yourjenkins.com*/**project**(/*foldername*)?/*yourprojectname* then click "Add webhook". When you click "Test" on your webhook it should trigger your Jenkins pipeline build (you should get a 200 HTTP response). It works without authentication in the GitLab plugin, configuration with authentication are welcome.
Jenkins
40,979,405
33
I recently ran into something of a puzzler while working on some Jenkins builds with a coworker. He's been using params.VARIABLE and env.VARIABLE interchangably and having no issues with it. Meanwhile, I started getting null object errors on one of his calls to a parameter object through the environment on this line of code: if(!deploy_environments.contains(env.ENVIRONMENT_NAME.trim()) || params.INVOKE_PARAMETERS ) { ENVIRONMENT_NAME here is a parameter. I started getting this error: java.lang.NullPointerException: Cannot invoke method trim() on null object This build is executing as a child of another build. The ENVIRONMENT_NAME parameter is passed down to the child from that parent build. He was not seeing this error at all on a different Jenkins master. When I changed the reference above from env.ENVIRONMENT_NAME to params.ENVIRONMENT_NAME the issue went away. I could find no reference to params == env in the Jenkins documentation, so I created a build to try to clarify their relationship. pipeline { agent { label 'jenkins-ecs-slave' } environment { ENV_VARIABLE = 'Environment' } parameters { string(description: 'Parameter', name: 'PARAMETER_VARIABLE', defaultValue: 'Parameter') } stages { stage('Output Parameters'){ steps { script { echo "Environment: ${env.ENV_VARIABLE}" echo "Parameter: ${params.PARAMETER_VARIABLE}" echo "Environment from params: ${params.ENV_VARIABLE}" echo "Parameter from Env: ${env.PARAMETER_VARIABLE}" echo "Inspecific reference ENV_VARIABLE: $ENV_VARIABLE" echo "Inspecific reference PARAMETER_VARIABLE: $PARAMETER_VARIABLE" sh 'echo "Shell environment: $ENV_VARIABLE"' sh 'echo "Shell parameter: $PARAMETER_VARIABLE"' } } } } } The first time I ran this on my Jenkins master, it only included the first four lines (echo env.ENV, echo param.PARAM, echo env.PARAM, echo param.ENV) it succeeded with the following output: [Pipeline] { [Pipeline] withEnv [Pipeline] { [Pipeline] stage [Pipeline] { (Output Parameters) [Pipeline] script [Pipeline] { [Pipeline] echo Environment: Environment [Pipeline] echo Parameter: Parameter [Pipeline] echo Environment from params: null [Pipeline] echo Parameter from Env: null [Pipeline] } [Pipeline] // script [Pipeline] } [Pipeline] // stage [Pipeline] } [Pipeline] // withEnv [Pipeline] } [Pipeline] // node [Pipeline] End of Pipeline Finished: SUCCESS And I thought, "Aha!" Problem solved. They're not the same. However, that box promptly froze up on me afterwards and refused to queue anymore builds. I haven't finished debugging it, but it's not out of line to wonder if that master is just messed up. So I went and ran it on a third Jenkins master we have hanging around. It's at this point I added the additional lines you see in the script above to further clarify. The first time I ran this script on that box, it failed on the "Inspecific reference to $PARAMETER_VARIABLE line" with the following output: [Pipeline] { [Pipeline] withEnv [Pipeline] { [Pipeline] stage [Pipeline] { (Output Parameters) [Pipeline] script [Pipeline] { [Pipeline] echo Environment: Environment [Pipeline] echo Parameter: Parameter [Pipeline] echo Environment from params: null [Pipeline] echo Parameter from Env: null [Pipeline] echo Inspecific reference ENV_VARIABLE: Environment [Pipeline] } [Pipeline] // script [Pipeline] } [Pipeline] // stage [Pipeline] } [Pipeline] // withEnv [Pipeline] } [Pipeline] // node [Pipeline] End of Pipeline groovy.lang.MissingPropertyException: No such property: PARAMETER_VARIABLE for class: groovy.lang.Binding Okay, so far so good. This makes sense. They aren't the same. You can reference Environment variables in echos and shells with out specifically referencing the environment object, but can't do the same with parameters. Consistent, reasonable, I'm good with this. So then I removed the two lines doing the "inspecific reference" and the script succeeded with the following output: [Pipeline] { [Pipeline] withEnv [Pipeline] { [Pipeline] stage [Pipeline] { (Output Parameters) [Pipeline] script [Pipeline] { [Pipeline] echo Environment: Environment [Pipeline] echo Parameter: Parameter [Pipeline] echo Environment from params: null [Pipeline] echo Parameter from Env: Parameter [Pipeline] sh [Environment Testing] Running shell script + echo 'Shell environment: Environment' Shell environment: Environment [Pipeline] sh [Environment Testing] Running shell script + echo 'Shell parameter: Parameter' Shell parameter: Parameter [Pipeline] } [Pipeline] // script [Pipeline] } [Pipeline] // stage [Pipeline] } [Pipeline] // withEnv [Pipeline] } [Pipeline] // node [Pipeline] End of Pipeline Finished: SUCCESS And now I'm completely confuddled. What the hell? I ran it a couple of times just to be sure, and got the same successful output as above, consistently. Granted, none of the previous builds that showed env.PARAM as null had really succeeded in a clean environment (the one that succeeded was in an environment the promptly imploded on me afterwards). So maybe if there's an error in the Jenkins pipeline, it short circuits the loading of parameters into the environment or something? I tried adding echo "$I_SHOULD_FAIL" to the script to force an error in an attempt to reproduce what was I seeing. No dice: [Pipeline] { [Pipeline] withEnv [Pipeline] { [Pipeline] stage [Pipeline] { (Output Parameters) [Pipeline] script [Pipeline] { [Pipeline] echo Environment: Environment [Pipeline] echo Parameter: Parameter [Pipeline] echo Environment from params: null [Pipeline] echo Parameter from Env: Parameter [Pipeline] sh [Environment Testing] Running shell script + echo 'Shell environment: Environment' Shell environment: Environment [Pipeline] sh [Environment Testing] Running shell script + echo 'Shell parameter: Parameter' Shell parameter: Parameter [Pipeline] } [Pipeline] // script [Pipeline] } [Pipeline] // stage [Pipeline] } [Pipeline] // withEnv [Pipeline] } [Pipeline] // node [Pipeline] End of Pipeline groovy.lang.MissingPropertyException: No such property: I_SHOULD_FAIL for class: groovy.lang.Binding So what's going on here? What's the relationship between environment and parameters in Jenkins pipelines? What is that relationship supposed to be and why does it seem to be inconsistent?
Basically this works as follows env contains all environment variables. Jenkins pipeline automatically creates a global variable for each environment variable params contains all build parameters. Jenkins also automatically creates an environment variable for each build parameter (and as a consequence of the second point a global variable). Environment variables can be overridden or unset, but params is an immutable Map and cannot be changed. Best practice is to always use params when you need to get a build parameter. See Global Variable Reference for more details regarding the variables.
Jenkins
50,398,334
33
I would like to use Pipeline to keep track of my Jenkin Jobs within my SCM. (Source control manager). Is there a way I can take my existing jobs and export them to a valid Jenkinsfile which can be read by Pipeline? The main plugins I'm using which I would need to be exported are Github Pull Request Builder, Test result reporters, code coverage reporters, as well as slack notification post-build tasks. My main question is how to export my Jenkins settings into a Jenkinsfile as mentioned in the above link so that I don't have to write them all by hand.
Turns out the short answer is that you can't. You need to look up each plugin you use and see if it has a syntax or support for Jenkinsfile and Pipelines.
Jenkins
41,224,533
33
I am trying to setup a project that uses the shiny new Jenkins pipelines, more specifically a multibranch project. I have a Jenkinsfile created in a test branch as below: node { stage 'Preparing VirtualEnv' if (!fileExists('.env')){ echo 'Creating virtualenv ...' sh 'virtualenv --no-site-packages .env' } sh '. .env/bin/activate' sh 'ls -all' if (fileExists('requirements/preinstall.txt')){ sh 'pip install -r requirements/preinstall.txt' } sh 'pip install -r requirements/test.txt' stage 'Unittests' sh './manage.py test --noinput' } It's worth noting that preinstall.txt will update pip. I am getting error as below: OSError: [Errno 13] Permission denied: '/usr/local/lib/python2.7/dist-packages/pip' Looks like it's trying to update pip in global env instead of inside virtualenv, and looks like each sh step is on its own context, how do I make them to execute within the same context?
What you are trying to do will not work. Every time you call the sh command, jenkins will create a new shell. This means that if you use .env/bin/activate in a sh it will be only sourced in that shell session. The result is that in a new sh command you have to source the file again (if you take a closer look at the console output you will see that Jenkins will actually create temporary shell files each time you run the command. So you should either source the .env/bin/activate file at the beginning of each shell command (you can use triple quotes for multiline strings), like so if (fileExists('requirements/preinstall.txt')) { sh """ . .env/bin/activate pip install -r requirements/preinstall.txt """ } ... sh """ . .env/bin/activate pip install -r requirements/test.txt """ } stage("Unittests") { sh """ . .env/bin/activate ./manage.py test --noinput """ } or run it all in one shell sh """ . .env/bin/activate if [[ -f requirements/preinstall.txt ]]; then pip install -r requirements/preinstall.txt fi pip install -r requirements/test.txt ./manage.py test --noinput """
Jenkins
40,836,570
33
I got very strange behavior that has never happened before, when I try to configure the GitHub server in Jenkins general configuration to set up webhooks auto. The drop down menu doesn't display my registered credentials. I was always be able to do that, but suddenly I don't know what's happening. I tried to uninstall the plugin, restarting Jenkins, kill the Jenkins Docker container and configure all the stuff again ... still I got the same issue. There is no other option other than none, and when I add new credentials I still get None as the only option...
The issue is that the GitHub plugin only accepts plain text credentials. The GitHub access token can be created manually, or automatically via the Advanced... options as described here. In case you already have an access token in GitHub (you'll get an error in Jenkins), you can remove it in Github. Then you can let Jenkins generate the token and select it in the Credentials menu.
Jenkins
36,500,729
33
I have a React app that has Jest tests. I'm configuring Jest in my package.json: … "jest": { "setupEnvScriptFile": "./test/jestenv.js", "setupTestFrameworkScriptFile": "./test/setup-jasmine-env.js", "testRunner": "node_modules/jest-cli/src/testRunners/jasmine/jasmine2.js", "unmockedModulePathPatterns": [ "./node_modules/q", "./node_modules/react" ] }, … The setup-jasmine-env.js looks like this: var jasmineReporters = require('jasmine-reporters'); jasmine.VERBOSE = true; jasmine.getEnv().addReporter( new jasmineReporters.JUnitXmlReporter({ consolidateAll: true, savePath: "output/", filePrefix: "test-results" }) ); It took a bit of working to get that jasmine env setup correctly, but I"m not seeing anything in the output directory (indeed, it isn't created and creating it myself doesn't help). I suspect that my alterations to the jasmine var aren't the same one that Jest is using, but I can't figure out how to hook them together.
If you use a more recent version of jest (I'm looking at 16.0.2), you don't need to specify the testrunner because jasmine is the default. You also don't need the unmockedModulePathPatterns section of the jest config. I.e. you just need to include the following devDependencies in your package.json: "jasmine-reporters": "^2.2.0", "jest": "^16.0.2", "jest-cli": "^16.0.2" And add this jest config to your package.json (note: you no longer need the unmockedModulePathPatterns section): "jest": { "setupTestFrameworkScriptFile": "./setup-jasmine-env.js" } And then use Drew's setup-jasmine-env.js from the question.
Jenkins
34,427,553
33
I want to use Java 11 syntax in my unit tests, but my 'main' code needs to be compiled for Java 8 since my production environment only has JDK 8 installed. Is there a way of doing this with the maven-compiler-plugin? My Jenkins server has Java 11 installed. I will accept the risk that I can accidental use Java 11 specific functionality in my production code.
In Maven compile and testCompile goals are different. And Maven even has parameters for testCompile: testTarget and testSource. So: <plugin> <groupId>org.apache.maven.plugins</groupId> <artifactId>maven-compiler-plugin</artifactId> <version>3.0</version> <configuration> <source>1.7</source> <target>1.7</target> <testSource>1.8</testSource> <testTarget>1.8</testTarget> </configuration> </plugin>
Jenkins
24,323,176
33
I've seen similar posts to this on SO, but not quite exactly what I am trying to do (or at least no full examples of a command to run). I am trying to remotely trigger a parameterized build on Jenkins using curl. I have 'Prevent Cross Site Request Forgery' enabled so I also need to pass a valid crumb. The script I have is below: #!/bin/bash json="{\"parameter\": [{ \"P1\": \"param1\", \"P2\": \"param2\", \"P3\": \"param3\" }]}" crumb=`curl "http://SERVER/crumbIssuer/api/xml?xpath=concat(//crumbRequestField,%22:%22,//crumb)"` curl -v -H $crumb -X POST http://SERVER/job/JOB_NAME/buildWithParameters -d token=runme --data-urlencode json="$json" I've also tried modifying the URL I'm passing to curl to either: USERNAME:APITOKEN@SERVER and USERNAME:PASSWORD@SERVER Output from curl is: * About to connect() to SERVER port 8080 (#0) * Trying SERVER... connected * Connected to SERVER (SERVER) port 8080 (#0) * Server auth using Basic with user 'USERNAME' > POST /job/JOB_NAME/buildWithParameters HTTP/1.1 > Authorization: Basic bjAwNjY5MjI6YWxLaW5kaTg= > User-Agent: curl/7.19.7 (x86_64-redhat-linux-gnu) libcurl/7.19.7 NSS/3.13.1.0 zlib/1.2.3 libidn/1.18 libssh2/1.2.2 > Host: SERVER:8080 > Accept: */* > .crumb:776eb589e8b930d9f06cfc2df885314c > Content-Length: 168 > Content-Type: application/x-www-form-urlencoded > < HTTP/1.1 403 No valid crumb was included in the request < Content-Type: text/html;charset=ISO-8859-1 < Cache-Control: must-revalidate,no-cache,no-store < Content-Length: 1469 < Server: Jetty(8.y.z-SNAPSHOT) < So it looks like I'm not passing the crumb properly, but I'm not sure what the correct format of the command should be.
What worked for me: SERVER=http://localhost:8080 CRUMB=$(curl --user $USER:$APITOKEN \ $SERVER/crumbIssuer/api/xml?xpath=concat\(//crumbRequestField,%22:%22,//crumb\)) curl --user $USER:$APITOKEN -H "$CRUMB" -d "script=$GROOVYSCRIPT" $SERVER/script
Jenkins
23,497,819
33
I am trying to get the git short hash in a variable. I tried to set GIT_COMMIT_SHORT variable to run 'git rev-parse --short HEAD' but it didn't work. I need this variable to pass to ant build script so the package name include this short hash. I am running Jenkins on windows 2008 server. Thanks
Probably the simplest way to achieve the result you want would be to use the GIT_REVISION token makro, like this: ${GIT_REVISION,length=6} Have a look at https://wiki.jenkins-ci.org/display/JENKINS/Token+Macro+Plugin for more details. Hope this helps, Jan
Jenkins
16,943,665
33
We are using maven. I want to set up infrastructure, so that automatically built artifacts would go to Nexus repository. And then they could be used by developers. I have already set up Jenkins with 1 job for our project. And I set up Nexus to on the same server. On developers' PCs I copied default maven setting to C:\Users{user}.m2\settings.xml adding this section. References: Configuring Maven to Use a Single Nexus Maven Settings Reference <mirror> <!--This sends everything else to /public --> <id>nexus</id> <mirrorOf>*</mirrorOf> <url>http://myserver:8081/nexus/content/groups/public</url> </mirror> (I just follow Repository Management with Nexus book) What are my next steps should be? Should Jenkins job have mvn install? How to create Nexus repository for company artifacts?
To deploy artifacts to Nexus, you'll need to include a distributionManagement section in your pom. Nexus ships with specific repositories already set up for both snapshots and releases. You should give the correct path to each of those so that maven will deploy snapshot and release artifacts to the correct repos. Then any time you deploy artifacts--typically with mvn deploy or using the maven release plugin, the artifacts will be deployed there. Nexus has write authentication on by default, so you'll need to make sure to add a server section with the correct credentials to the settings.xml of anyone who will be deploying artifacts. Jenkins can be treated pretty much like any other user. If you have it do a deploy as its build, then every build will deploy to Nexus. There's also a post-build action for deploying artifacts in case you want it to happen later in the Jenkins job.
Jenkins
6,950,346
33
I'm trying to set up hudson with git according to this article, but I still get git errors during build: FATAL: Could not apply tag-PROJECTNAME-ID ... Caused by: hudson.plugins.git.GitException: Command returned status code 128: *** Please tell me who you are. running: git config --global user.name shows valid data, .gitconfig is accessible. How to correct those errors?
After installing the git plugin you can configure git name and email in Jenkins "Configure System" page...
Jenkins
2,671,296
33
I have a virtual machine hosting Oracle Linux where I've installed Docker and created containers using a docker-compose file. I placed the jenkins volume under a shared folder but when starting the docker-compose up I got the following error for Jenkins : jenkins | touch: cannot touch ‘/var/jenkins_home/copy_reference_file.log’: Permission denied jenkins | Can not write to /var/jenkins_home/copy_reference_file.log. Wrong volume permissions? jenkins exited with code 1 Here's the volumes declaration volumes: - "/media/sf_devops-workspaces/dev-tools/continuous-integration/jenkins:/var/jenkins_home"
The easy fix it to use the -u parameter. Keep in mind this will run as a root user (uid=0) docker run -u 0 -d -p 8080:8080 -p 50000:50000 -v /data/jenkins:/var/jenkins_home jenkins/jenkins:lts
Jenkins
44,065,827
32
In my Jenkins pipelines I generally use post declarative function to send me an email incase the pipeline has failed. A simple syntax of the post function is as under: post { failure { mail to: '[email protected]', subject: "Failed Pipeline: ${currentBuild.fullDisplayName}", body: "Something is wrong with ${env.BUILD_URL}" } } In the above email, I also want to mention which stage (lets say the pipeline has 5 to 6 stages) of the pipeline has failed. How can I do that? Any help is much appreciated. An extension to the above requirement will be to provide the user with the actual error log (of the stage that has failed) also as a part of the failure notification email. Idea is, when a user receives a failure notification from jenkins, he should know which stage of the pipeline has failed along with the error log. Thanks in advance.
There is a variable called env.STAGE_NAME which you can use. However, in your case you will probably need to store the stage name in a different variable, because when you get the env.STAGE_NAME in a post stage the result will be Declarative: Post Actions. Instead, you will need to store the stage name in a variable in all stages. So once a stage fails - Jenkins will not continue with the next stages and therefore you will have the "failed" stage name. Here's an example: def FAILED_STAGE pipeline { agent { label "master" } stages { stage("Stage 1") { steps { script { FAILED_STAGE=env.STAGE_NAME echo "stage 1" } } } stage("Stage 2") { steps { script { FAILED_STAGE=env.STAGE_NAME echo "stage 2" error "failed for some reason." } } } stage("Stage 3") { steps { script { FAILED_STAGE=env.STAGE_NAME echo "stage 3" } } } } post { failure { echo "Failed stage name: ${FAILED_STAGE}" } } } There might be a better way to do it, but I haven't found it so far. Regarding the logs - As of JENKINS-40526 you could possibly use the API and get the log file from there, but I am not sure you can get the parameters you need from within the pipeline. The other solution would be to use emailext and email the entire build log file: emailext attachLog: true, body: '', compressLog: true, subject: 'Build failed.', to: '[email protected]'
Jenkins
50,411,381
32
I'm developing Jenkins pipelines as Groovy scripts (scripted pipelines, not declarative), and having a real hard time. Jenkins is always very generic regarding syntax/semantic errors, outputting stacks like below: groovy.lang.MissingPropertyException: No such property: caughtError for class: groovy.lang.Binding at groovy.lang.Binding.getVariable(Binding.java:63) at So I have to figure where the error is completely by myself, inspecting line per line of code. Is there a better way to debug it? What you guys use to do?
I have seen this post, http://notes.asaleh.net/posts/debugging-jenkins-pipeline/ Which describe how to debug a groovy script for jenkins pipeline. it's clearly describe the steps how to do it.
Jenkins
47,993,538
32
env.JOB_NAME Is the pipeline name suffixed with the branch name. So env.JOB_NAME will be <jenkins_pipeline_name>_<my_branch> How can I just get the pipeline name and store it in a var in the environment{} block at the top of my jenkinsfile to use through the file? I don't want to resort to scripted pipeline just the declarative.
@red888 pointed out the following answer that worked like magic for me. I am pointing it out in an actual answer because I almost missed it: env.JOB_BASE_NAME Credit to @red888 in the comment above. Send upvotes his/her way.
Jenkins
45,746,902
32
I have to create this JSON file in Groovy. I have try many things (JsonOutput.toJson() / JsonSlurper.parseText()) unsuccessfully. { "attachments":[ { "fallback":"New open task [Urgent]: <http://url_to_task|Test out Slack message attachments>", "pretext":"New open task [Urgent]: <http://url_to_task|Test out Slack message attachments>", "color":"#D00000", "fields":[ { "title":"Notes", "value":"This is much easier than I thought it would be.", "short":false } ] } ] } This is for posting a Jenkins build message to Slack.
JSON is a format that uses human-readable text to transmit data objects consisting of attribute–value pairs and array data types. So, in general json is a formatted text. In groovy json object is just a sequence of maps/arrays. parsing json using JsonSlurperClassic //use JsonSlurperClassic because it produces HashMap that could be serialized by pipeline import groovy.json.JsonSlurperClassic node{ def json = readFile(file:'message2.json') def data = new JsonSlurperClassic().parseText(json) echo "color: ${data.attachments[0].color}" } parsing json using pipeline node{ def data = readJSON file:'message2.json' echo "color: ${data.attachments[0].color}" } building json from code and write it to file import groovy.json.JsonOutput node{ //to create json declare a sequence of maps/arrays in groovy //here is the data according to your sample def data = [ attachments:[ [ fallback: "New open task [Urgent]: <http://url_to_task|Test out Slack message attachments>", pretext : "New open task [Urgent]: <http://url_to_task|Test out Slack message attachments>", color : "#D00000", fields :[ [ title: "Notes", value: "This is much easier than I thought it would be.", short: false ] ] ] ] ] //two alternatives to write //native pipeline step: writeJSON(file: 'message1.json', json: data) //but if writeJSON not supported by your version: //convert maps/arrays to json formatted string def json = JsonOutput.toJson(data) //if you need pretty print (multiline) json json = JsonOutput.prettyPrint(json) //put string into the file: writeFile(file:'message2.json', text: json) }
Jenkins
44,707,265
32
My Jenkins is not run in Docker container, just tradional install to VPS. I got the following error when executing a simple test project. I am using Ubuntu 14, java 7, and stable Jenkins. I tried all methods I can find on google, but can't get it work. I am trying to execute this shell docker build --pull=true -t nick/hello-jenkins:$GIT_COMMIT . After code change. Here is error: Got permission denied while trying to connect to the Docker daemon socket at unix: .... Started by user nicolas xu Building in workspace /var/lib/jenkins/workspace/hello-Jenkins > git rev-parse --is-inside-work-tree # timeout=10 Fetching changes from the remote Git repository > git config remote.origin.url https://github.com/nicolasxu/hello-nick-jenkins.git # timeout=10 Fetching upstream changes from https://github.com/nicolasxu/hello-nick-jenkins.git > git --version # timeout=10 > git fetch --tags --progress https://github.com/nicolasxu/hello-nick-jenkins.git +refs/heads/*:refs/remotes/origin/* > git rev-parse refs/remotes/origin/master^{commit} # timeout=10 > git rev-parse refs/remotes/origin/origin/master^{commit} # timeout=10 Checking out Revision d94ae21a8a2cf58ffc790dcad15bd851fb17df5a (refs/remotes/origin/master) > git config core.sparsecheckout # timeout=10 > git checkout -f d94ae21a8a2cf58ffc790dcad15bd851fb17df5a > git rev-list d94ae21a8a2cf58ffc790dcad15bd851fb17df5a # timeout=10 [hello-Jenkins] $ /bin/sh -xe /tmp/hudson5076309502904684976.sh + docker build --pull=true -t nick/hello-jenkins:d94ae21a8a2cf58ffc790dcad15bd851fb17df5a . Got permission denied while trying to connect to the Docker daemon socket at unix:///var/run/docker.sock: Post http://%2Fvar%2Frun%2Fdocker.sock/v1.27/build?buildargs=%7B%7D&cachefrom=%5B%5D&cgroupparent=&cpuperiod=0&cpuquota=0&cpusetcpus=&cpusetmems=&cpushares=0&dockerfile=Dockerfile&labels=%7B%7D&memory=0&memswap=0&networkmode=default&pull=1&rm=1&shmsize=0&t=nick%2Fhello-jenkins%3Ad94ae21a8a2cf58ffc790dcad15bd851fb17df5a&ulimits=null: dial unix /var/run/docker.sock: connect: permission denied Build step 'Execute shell' marked build as failure Finished: FAILURE I can run 'docker' in console as root no problem, why jenkins can't try a shell command which runs 'docker'? What is going on? Totally confused.......
In your VPS server terminal, do this to add your jenkins user to the docker group: sudo usermod -aG docker jenkins Then restart your jenkins server to refresh the group. Take into account any security issue that this could produce: Warning: The docker group grants privileges equivalent to the root user. For details on how this impacts security in your system, see Docker Daemon Attack Surface. Refer to the docs Edit (mentioned by @iger): Just make sure to restart the Jenkins from command-line (i.e. sudo service jenkins restart), but not through the rest endpoint (http:///restart)
Jenkins
44,444,099
32
Tried with the configure option, not able to find the tools configuration option and the git executable section. Seems like it occurs after a successful build only. Please help. Here's the output I receive after building the project on the console output section: Building in workspace C:\Users\Anishas\.jenkins\workspace\Sample123 Cloning the remote Git repository Cloning repository https://github.com/AnishaSalunkhe/HelloWorld.git > C:\Users\Anishas\git init C:\Users\Anishas\.jenkins\workspace\Sample123 # timeout=10 ERROR: Error cloning remote repo 'origin' hudson.plugins.git.GitException: Could not init C:\Users\Anishas\.jenkins\workspace\Sample123 at org.jenkinsci.plugins.gitclient.CliGitAPIImpl$5.execute(CliGitAPIImpl.java:656) at org.jenkinsci.plugins.gitclient.CliGitAPIImpl$2.execute(CliGitAPIImpl.java:463) at hudson.plugins.git.GitSCM.retrieveChanges(GitSCM.java:1057) at hudson.plugins.git.GitSCM.checkout(GitSCM.java:1097) at hudson.scm.SCM.checkout(SCM.java:485) at hudson.model.AbstractProject.checkout(AbstractProject.java:1269) at hudson.model.AbstractBuild$AbstractBuildExecution.defaultCheckout(AbstractBuild.java:607) at jenkins.scm.SCMCheckoutStrategy.checkout(SCMCheckoutStrategy.java:86) at hudson.model.AbstractBuild$AbstractBuildExecution.run(AbstractBuild.java:529) at hudson.model.Run.execute(Run.java:1738) at hudson.model.FreeStyleBuild.run(FreeStyleBuild.java:43) at hudson.model.ResourceController.execute(ResourceController.java:98) at hudson.model.Executor.run(Executor.java:410) Caused by: hudson.plugins.git.GitException: Error performing command: C:\Users\Anishas\git init C:\Users\Anishas\.jenkins\workspace\Sample123 at org.jenkinsci.plugins.gitclient.CliGitAPIImpl.launchCommandIn(CliGitAPIImpl.java:1726) at org.jenkinsci.plugins.gitclient.CliGitAPIImpl.launchCommandIn(CliGitAPIImpl.java:1695) at org.jenkinsci.plugins.gitclient.CliGitAPIImpl.launchCommandIn(CliGitAPIImpl.java:1691) at org.jenkinsci.plugins.gitclient.CliGitAPIImpl.launchCommand(CliGitAPIImpl.java:1321) at org.jenkinsci.plugins.gitclient.CliGitAPIImpl$5.execute(CliGitAPIImpl.java:654) ... 12 more Caused by: java.io.IOException: Cannot run program "C:\Users\Anishas\git" (in directory "C:\Users\Anishas\.jenkins\workspace\Sample123"): CreateProcess error=5, Access is denied at java.lang.ProcessBuilder.start(Unknown Source) at hudson.Proc$LocalProc.<init>(Proc.java:240) at hudson.Proc$LocalProc.<init>(Proc.java:212) at hudson.Launcher$LocalLauncher.launch(Launcher.java:815) at hudson.Launcher$ProcStarter.start(Launcher.java:381) at org.jenkinsci.plugins.gitclient.CliGitAPIImpl.launchCommandIn(CliGitAPIImpl.java:1715) ... 16 more Caused by: java.io.IOException: CreateProcess error=5, Access is denied at java.lang.ProcessImpl.create(Native Method) at java.lang.ProcessImpl.<init>(Unknown Source) at java.lang.ProcessImpl.start(Unknown Source) ... 22 more ERROR: null Finished: FAILURE
This wasted so much time on my Jenkins Windows slave. I knew git was in the path because I executed "where git" in the build job's batch command. where git C:\Program Files (x86)\Git\cmd\git.exe Apparently the Jenkins Git Plugin executes ** before ** the environment is inherited. SET YOUR SLAVE's PATH to Git ( Just DO IT !! ) 1) Go to your Windows slave configuration Manage Jenkins > Manage Nodes 2) Select your slave configuration 3) Check Tool Locations under Node Properties 4) Enter complete path to git executable including git.exe [x] Tool Locations Name: (GIT) git Home: C:\Program Files (x86)\Git\cmd\git.exe See screenshot:
Jenkins
37,155,321
32
It looks like the GitHubPullRequestBuilder is not compatible with Jenkins v2.0 pipeline jobs. How do you configure a pipeline job to be triggered from a GitHub pull request event? The documentation on this topic is sparse and I cannot find any examples of this. Or is it better to create a web-hook in GitHub to trigger the pipeline job on the PR event?
I had similar issue. Here’s what worked for me Pre-req Jenkins ver. 2+ (I was using Jenkins 2.60) Github (or Githhub enterprise) account Your github and Jenkins must be able to talk to each other. On Github create a github Personal Access Token (PAT) with relevant rights. For your repo, create a webhook with URL as YourJenkinsURL/github-webhook/ Choose ‘Let me select individual events’ and check ‘Pull Request’ Add a Jenkinsfile to the root folder of your repo. For testing purpose you could put content as a basic hello world like below pipeline { agent any stages { stage('Test') { steps { echo 'Hello World ...' } } } } On Jenkins Install GitHub Pull Request Builder plugin. (You also need “Github” plugin but that should normally be installed as part of Jenkins ver 2+) Jenkins – Credentials Add github Personal Access Token (PAT) as a ‘secret text’ credential. Add github username-password as ‘username-password’ credential. Manage Jenkins – Configure System Github – Github Servers : This is part of the Github plugin. Add a github server. ‘API URL’ It will default to https://api.github.com. If you are using enterprise github, replace with enterprise github url followed by /api/v3. For credential select the PAT option. Test the connection. ‘Manage Hooks’ is checked. GitHub Pull Request Builder : for ‘GitHub Server API URL’ use same url as specified in Github Server section. Leave ‘Shared Secret’ blank. For credentials use ‘username-password’ credential. Test credentials to ensure its working. In my settings, ‘Auto-manage webhooks’ was checked. Pipeline Job Create a new item using ‘Pipeline’ option. Note: This is the vanilla Pipeline job, not Multibranch Pipeline. General Section: Check ‘Github Project’ – Project URL : Enter your github repo url Build Triggers: Check ‘GitHub Pull Request Builder’ For ‘GitHub API credentials’ select option you set for GitHub pull request builder in ‘Manage Jenkins – Configure System’ screen For admin list: add your username Check Use github hooks for build triggering Pipeline: Select ‘Pipeline Script from SCM’. Note this assumes that the root folder of your repo will contain a ‘Jenkinsfile’ SCM: Select ‘Git’ Repositories – enter repo detail. For credentials use ‘username-password’ based credentials. Click Advanced and add refspec as +refs/pull/*:refs/remotes/origin/pr/* Branch – should be ${sha1} Script Path: defaulted to Jenkinsfile, leave as is. Lightweight Checkout - Uncheck this (https://github.com/jenkinsci/ghprb-plugin/issues/507) That’s it. You are all set. Creating a PR on master branch of your repo should now trigger your Jenkins Pipeline job Some observations Redelivering the webhook payload of a PR from github does not trigger the pipeline but opening a new PR or even re-opening a closed PR on github, triggers the pipeline job In Pipeline Job Configuration, if you choose “Pipeline Script” and paste your pipeline script in there, the job doesn't trigger !!!
Jenkins
36,850,485
32
I have found a way to access the credentials store in Jenkins: def getPassword = { username -> def creds = com.cloudbees.plugins.credentials.CredentialsProvider.lookupCredentials( com.cloudbees.plugins.credentials.common.StandardUsernamePasswordCredentials.class, jenkins.model.Jenkins.instance ) def c = creds.findResult { it.username == username ? it : null } if ( c ) { println "found credential ${c.id} for username ${c.username}" def credentials_store = jenkins.model.Jenkins.instance.getExtensionList( 'com.cloudbees.plugins.credentials.SystemCredentialsProvider' )[0].getStore() println "result: " + credentials_store } else { println "could not find credential for ${username}" } } getPassword("XYZ") But now i would like to get the password for the appropriate user which i can't do... I always get unknown method etc. if i try to access passord etc. The reason for doing this is to use this user/password to call git and extract information from repository.. I always get something like this: result: com.cloudbees.plugins.credentials.SystemCredentialsProvider$StoreImpl@1639eab2 Update After experimenting more (and the hint of Jeanne Boyarsky) with it i found that i was thinking to compilcated. The following already gives me the password for the user: def getUserPassword = { username -> def creds = com.cloudbees.plugins.credentials.CredentialsProvider.lookupCredentials( com.cloudbees.plugins.credentials.common.StandardUsernamePasswordCredentials.class, jenkins.model.Jenkins.instance ) def c = creds.findResult { it.username == username ? it : null } if ( c ) { return c.password } else { println "could not find credential for ${username}" } } Furthermore by using the following snippet you can iterate over the whole credentials store: def credentials_store = jenkins.model.Jenkins.instance.getExtensionList( 'com.cloudbees.plugins.credentials.SystemCredentialsProvider' ) println "credentials_store: ${credentials_store}" println " Description: ${credentials_store.description}" println " Target: ${credentials_store.target}" credentials_store.each { println "credentials_store.each: ${it}" } credentials_store[0].credentials.each { it -> println "credentials: -> ${it}" if (it instanceof com.cloudbees.plugins.credentials.impl.UsernamePasswordCredentialsImpl) { println "XXX: username: ${it.username} password: ${it.password} description: ${it.description}" } } And you will get an output like this: [(master)]: credentials_store: [com.cloudbees.plugins.credentials.SystemCredentialsProvider@5a2822be] Description: [The descriptions...] Target: [com.cloudbees.plugins.credentials.SystemCredentialsProvider@5a2822be] credentials_store.each: com.cloudbees.plugins.credentials.SystemCredentialsProvider@5a2822be credentials: -> com.cloudbees.jenkins.plugins.sshcredentials.impl.BasicSSHUserPrivateKey@38357ca1 credentials: -> com.cloudbees.jenkins.plugins.sshcredentials.impl.BasicSSHUserPrivateKey@47cf7703 credentials: -> com.cloudbees.plugins.credentials.impl.UsernamePasswordCredentialsImpl@739abac5 XXX: username: User1 password: Password description: The description of the user. credentials: -> com.cloudbees.plugins.credentials.impl.UsernamePasswordCredentialsImpl@884a53e6 XXX: username: User2 password: Password1 description: The description of the user1. Result: [com.cloudbees.jenkins.plugins.sshcredentials.impl.BasicSSHUserPrivateKey@38357ca1, com.cloudbees.jenkins.plugins.sshcredentials.impl.BasicSSHUserPrivateKey@47cf7703, com.cloudbees.plugins.credentials.impl.UsernamePasswordCredentialsImpl@739abac5, com.cloudbees.plugins.credentials.impl.UsernamePasswordCredentialsImpl@884a53e6] So by using the appropriate class in the instanceof clause you can select what you need.
This works. It gets the credentials rather than the store. I didn't write any error handling so it blows up if you don't have a credentials object set up (or probably if you have two). That part is easy to add though. The tricky part is getting the right APIs! def getPassword = { username -> def creds = com.cloudbees.plugins.credentials.CredentialsProvider.lookupCredentials( com.cloudbees.plugins.credentials.common.StandardUsernamePasswordCredentials.class, jenkins.model.Jenkins.instance ) def c = creds.findResult { it.username == username ? it : null } if ( c ) { println "found credential ${c.id} for username ${c.username}" def systemCredentialsProvider = jenkins.model.Jenkins.instance.getExtensionList( 'com.cloudbees.plugins.credentials.SystemCredentialsProvider' ).first() def password = systemCredentialsProvider.credentials.first().password println password } else { println "could not find credential for ${username}" } } getPassword("jeanne")
Jenkins
35,205,665
32
To free up space on C:, I would like to move my Jenkins data files (specifically the \jobs directory) from the default installation directory C:\Program Files (x86)\Jenkins to F:\Jenkins\home. I think what I need to do is set the JENKINS_HOME environment variable to F:\Jenkins\home. But no matter what I try, the JENKINS_HOME environment variable is always set to the location of jenkins.exe. Related: How to change Jenkins default folder on Windows? JIRA issue JENKINS-13530 JENKINS_HOME ignored on Bundled Windows EXE was closed as "not an issue" Here is what I've tried so far: Moved jenkins data to F:\Jenkins\home Stop the running jenkins service Uninstall the jenkins service with jenkins.exe uninstall Uninstall jenkins Delete %HOMEPATH%\.jenkins directory Delete old jenkins install directory Download latest MSI installer v1.597 Installed to C:\Program Files (x86)\Jenkins2 (renamed to ensure there are no stale values in the registry or config files) Set system-level environment variable JENKINS_HOME to F:\Jenkins\home Set user-level environment variable JENKINS_HOME to F:\Jenkins\home Modified jenkins.xml to use <env name="JENKINS_HOME" value="F:\Jenkins\home"/> Started the Jenkins service At this point, when I look at the system configuration, JENKINS_HOME is set to C:\Program Files (x86)\Jenkins2. So it seems it must always be set to the location of jenkins.exe. Maybe I've answered my own question. I'd like to have the program and data separate, if possible. Do I have to install jenkins to my F:\ drive? Or, is there a way to simply split off the jobs directory and leave everything else on C:? Thanks! EDIT : I did not have to move JENKINS_HOME, but instead was able to configure the workspace and builds directories, which moved all the heavy disk usage over to F:. The settings I chose were: Workspace Root Directory = F:/Jenkins/workspace/${ITEM_FULLNAME} Build Record Root Directory = F:/Jenkins/jobs/${ITEM_FULL_NAME}/builds I manually migrated these directories so they would not have to be recreated. During this process I did lose my build history, but I'm okay with that for now.
Pre Jenkins 2.121 JENKINS_HOME is where Jenkins is installed which is not what you want to change. After you start up Jenkins, go to: Manage Jenkins System Configuration Click the first "advanced" button This gives you text fields where you can change the directory for the workspace and builds directories. Those are the two directories that use a good bit of disk space. Note that it will not move history. If you want to move the existing workspaces/etc, you'll need to manually copy them over. Post 2.121 You have to set properties (not through the UI). The system property to use is jenkins.model.Jenkins.buildsDir. https://jenkins.io/doc/upgrade-guide/2.121/#ui-option-for-custom-builds-and-workspace-directories-on-the-master-has-been-removed https://wiki.jenkins.io/display/JENKINS/Features+controlled+by+system+properties
Jenkins
28,034,663
32
Is there a easy way to get a list of all node labels in Jenkins? I can see which labels are set on each node (.../computer/) and which nodes have the same label (.../label/). But similar to listing all nodes on .../computer/ there is no listing of all the labels on .../label/ The approach with python and jenkinsapi or similar seem a bit too advanced for a listing that probably already is available in Jenkins (but not visible?)
Haven't installed/tried it myself, but the "label linked jobs" jenkins plugin has a label dashboard as one of its features.. it sounds like this is what you're looking for
Jenkins
27,384,481
32
I'm trying to restart the Jenkins service using Ansible: - name: Restart Jenkins to make the plugin data available service: name=jenkins state=restarted - name: Wait for Jenkins to restart wait_for: host=localhost port=8080 delay=20 timeout=300 - name: Install Jenkins plugins command: java -jar {{ jenkins_cli_jar }} -s {{ jenkins_dashboard_url }} install-plugin {{ item }} creates=/var/lib/jenkins/plugins/{{ item }}.jpi with_items: jenkins_plugins But on the first run, the third task throws lots of Java errors including this: Suppressed: java.io.IOException: Server returned HTTP response code: 503 for URL, which makes me think the web server (handled entirely by Jenkins) wasn't ready. Sometimes when I go to the Jenkins dashboard using my browser it says that Jenkins isn't ready and that it will reload when it is, and it does, it works fine. But I'm not sure if accessing the page is what starts the server, or what. So I guess what I need is to curl many times until the http code is 200? Is there any other way? Either way, how do I do that? How do you normally restart Jenkins?
Using the URI module http://docs.ansible.com/ansible/uri_module.html - name: "wait for ABC to come up" uri: url: "http://127.0.0.1:8080/ABC" status_code: 200 register: result until: result.status == 200 retries: 60 delay: 1
Jenkins
23,919,744
32
I've set up a build on Jenkins for a Maven project, and I would like to build it without running any of the tests. I've tried entering "clean install -DskipTests" in the goals field, like this: But it doesn't work. What am I doing incorrectly? Note: I want to skip the tests without touching the pom. I have a separate build that DOES run the tests.
The problem is that I omitted =true. I was able to build without running tests by entering: clean install -DskipTests=true
Jenkins
22,513,839
32
I just joined a company that uses batch files to build a C++ project. The batch does all sorts of things (updates svn, which is now done by jenkins), creates build folders, deletes unnecessary files after building, copies library files to the build folder, etc. My problem is Jenkins always considers the build successful, even when it´s not. The .bat file creates a file called errormake.txt when something goes wrong. How do I make jenkins read that and mark the build as a failure? Also, is there any way I can find out the build folder Jenkins created from inside the .bat file (maybe send a variable when I call the batch file)? This is the single line I'm currently using to call the .bat file: call "C:\Source\BuildVersion\AppName\build_version.bat" Edit: Also, this project is split up into several SVN repositories. %SVN_REVISION% is blank. How can I get the correct %SVN_REVISION% from the first repository (the main one)?
To answer each of your questions - Jenkins always return "SUCCESS", even when the Job actually failed: Jenkins sets the status of the Job, based on the return-code of the last command that ran in each "Execute windows batch command" block. If your last command is copy some.log D:, Jenkins thinks everything is OK (If the 'copy' command went fine...) Use EXIT xx or EXIT /B xx, depending on your OS, where 'xx' is some integer greater than zero. How do I make Jenkins mark the build as a failure, based on a log-file: Use the Text-finder Plugin, as mentioned by sdmythos_gr . Is there any way I can find out the build folder Jenkins created: There are a few parameters that are available as environment-variables for each script that runs under Jenkins - see here for the list of those parameters: Jenkins Environment Variables. To your question: %WORKSPACE% - The absolute path of the workspace %BUILD_NUMBER% - The current build number, such as "153" %BUILD_ID% - The current build id, such as "2005-08-22_23-59-59" (YYYY-MM-DD_hh-mm-ss) How can I get the correct %SVN_REVISION% from the first repository: This answer is from the same link: %SVN_REVISION% - For Subversion-based projects, this variable contains the revision number of the module. If you have more than one module specified, this won't be set. Hope that helps
Jenkins
13,972,636
32
I have a GitHub repo that's big and contains several independently build-able bits. If I configure Jenkins with a job (or two) for each of these, I end up with having to pull gigabytes of data multiple times (one clone of the repo for each job). This takes both diskspace and bandwidth. What I'd like to do is have "Refresh local repo" job that clones github once, then configure each of the jobs to clone themselves from that repo, and build. Then by setting up the sub-jobs as dependent builds, I can run "Refresh local repo", have it pull all the latest stuff from GitHub, then have each of the builds run. So far I've got the "Refresh local repo" working - it clones successfully, and if I go to the workspace, I see that it has the HEAD commit of origin/master. The problem is the other jobs - these don't seem to be picking up updates. Here's how I've got one of them configured: Git Repository URL file:////Users/malcolmbox/.jenkins/jobs/Refresh Local repo/workspace Branches to build master Instead of this updating to the latest commit, it's stuck several days in the past. How can I get it to pull the tip and do the right thing? To clarify: the .../Refresh Local repo/workspace has commit 6b20268389064590147d5c73d2b6aceb6ba5fe70 submitted 28/3 The dependent build, after running a build (so presumably doing a git clone/pull step) is checked out to 79a25992cc192376522bcb634ee0f7eb3033fc7e submitted 26/3 - so it's a couple of days behind.
If you open the job configuration and click on the Advanced button of the git SCM configuration, you will see a place to specify "Path of the reference repo to use during clone (optional)". If you have a local clone of your repository, add the path to the reference repo field. Git will then use the local clone and share most of the git objects on the disk and pulling from github only what is missing from the local clone resulting in lightning fast clones and saved disk space. Or is this exactly how you have configured your job and it is not picking up latest commits? If that is so, please provide more details. Consider publishing your job configuration.
Jenkins
9,914,664
32
I'm trying to follow the directions here: https://wiki.jenkins-ci.org/display/JENKINS/Running+Jenkins+behind+Apache to set up my Jenkins server to appear at http://myhost/jenkins. It works, but the Jenkins website thinks http://myhost/ is the jenkins/ root. I believe this problem is caused by the first warning flag on that web page, i.e. that my context path is not set correctly. However, I can't figure out where to set the context path. The instructions for ubuntu and windows are clear enough, but on Mac OS X 10.6, there is no jenkins.xml file, no /etc/default/jenkins file, and nothing of relevance I can see in ~/.jenkins/config.xml. So, what am I missing? Where can I tell jenkins that its root is in /jenkins/ instead of /?
Paraphrasing from the document you mentioned; You need to specify the context/prefix of the Jenkins instance, this can be done by modifying the Jenkins configuration as follows; Either, set the context path by modifying the jenkins.xml configuration file and adding --prefix=/jenkins (or similar) to the entry. Or Set the context path when using by adding --prefix=/jenkins to JENKINS_ARGS in /etc/default/jenkins (Ubuntu) or in an appropriate startup file. So, how to find these things... The Jenkins.xml file should be in the $JENKINS_HOME directory, I'm not sure if Mac OS has the "updatedb" and "locate " commands, but you could try doing updatedb && locate jenkins.xml Also, have a look in the startup scripts; /etc/init.d if installed from a package, or add the JENKINS_ARGS to the environment properties for the User running Jenkins (append to ~user/.profile) or the arguments for the container running Jenkins. Be aware that if your Jenkins installation (without the prefix argument) was running under: http://myserver:8080/ => 200 Jenkins is here adding --prefix=/ci/dashboard in the arguments will produce this behaviour: http://myserver:8080/ => 404 http://myserver:8080/ci/dashboard => 200 Jenkins is now here
Jenkins
9,089,566
32
Is there any option to install jenkins plugins from command line ? I found a command for this after a bit google search : java -jar /var/lib/jenkins/jenkins.war -s http://127.0.0.1:8080/ install-plugin ${Plugin_Name} But it's not working.
As per the Jenkins command line interface documentation, you need to use the client JAR file (not the server WAR file you're using), which you can obtain directly from Jenkins, e.g. via the links on http://localhost:8080/cli Then you can run the command using this JAR: java -jar jenkins-cli.jar -s http://127.0.0.1:8080/ install-plugin <name> This will download install the plugin you want, along with any of its dependencies.
Jenkins
34,761,047
31
In a declarative pipeline, I can specify the parameter that the pipeline expects right in the pipeline script like so: pipeline { parameters([ string(name: 'DEPLOY_ENV', defaultValue: 'TESTING' ) ]) } is it possible do to in a scripted pipline? I know I can do this : BUT, IS IT POSSIBLE TO DO THIS: node{ parameters([ string(name: 'DEPLOY_ENV', defaultValue: 'TESTING' ) ]) }
I found a solution by experimentation so want to share it: properties( [ parameters([ string(defaultValue: '/data', name: 'Directory'), string(defaultValue: 'Dev', name: 'DEPLOY_ENV') ]) ] ) node { // params.DEPLOY_ENV ... }
Jenkins
53,747,772
31
I'm trying to trigger a downstream job from my current job like so pipeline { stages { stage('foo') { steps{ build job: 'my-job', propagate: true, wait: true } } } } The purpose is to wait on the job result and fail or succeed according to that result. Jenkins is always failing with the message Waiting for non-job items is not supported . The job mentioned above does not have any parameters and is defined like the rest of my jobs, using multibranch pipeline plugin. All i can think of is that this type of jenkins item is not supported as a build step input, but that seems counterintuitive and would prove to be a blocker to me. Can anyone confirm if this is indeed the case? If so, can anyone suggest any workarounds? Thank you
I actually managed to fix this by paying more attention to the definition of the build step. Since all my downstream jobs are defined as multibranch pipeline jobs, their structure is folder-like, with each item in the folder representing a separate job. Thus the correct way to call the downstream jobs was not build job: 'my-job', propagate: true, wait: true, but rather build job: "my-job/my-branch-name", propagate: true, wait: true. Also, unrelated to the question but related to the issue at hand, make sure you always have at least one more executor free on the jenkins machine, since the wait on syntax will consume one thread for the waiting job and one for the job being waited on, and you can easily find yourself in a resource-starvation type situation. Hope this helps
Jenkins
46,471,467
31
Is there a way to trigger a Jenkins job to run every hour using the Jenkinsfile scripted pipeline syntax? I have seen examples using the declarative syntax, but none using the pipeline syntax. Declarative Syntax Example pipeline { agent any triggers { cron '@daily' } ... }
You could use this snippet for Scripted pipeline syntax: properties( [ ... , // other properties that you have pipelineTriggers([cron('0 * * * *')]), ] ) Reference for properties is here. You can search for "pipelineTriggers" string and find out that triggers for build can be for example artifactory or something else from this list (extracted 2019-03-23 from linked doc page): $class: 'ArtifactoryTrigger' $class: 'AssemblaBuildTrigger' bitBucketTrigger bitbucketPush $class: 'BuildResultTrigger' $class: 'CIBuildTrigger' $class: 'CodingPushTrigger' $class: 'CronFolderTrigger' $class: 'DeployDbTrigger' $class: 'DockerHubTrigger' $class: 'DosTrigger' $class: 'ElOyente' $class: 'FanInReverseBuildTrigger' $class: 'FeatureBranchAwareTrigger' $class: 'FilesFoundTrigger' $class: 'FogbugzStatePoller' $class: 'FolderContentTrigger' GenericTrigger gerrit $class: 'GhprbTrigger' $class: 'GitBucketPushTrigger' githubBranches githubPullRequests githubPush gitee $class: 'GogsTrigger' issueCommentTrigger $class: 'IvyTrigger' $class: 'JiraChangelogTrigger' $class: 'JiraCommentTrigger' $class: 'KanboardQueryTrigger' $class: 'MailCommandTrigger' $class: 'MavenDependencyUpdateTrigger' $class: 'NugetTrigger' p4Trigger $class: 'PeriodicFolderTrigger' $class: 'PollMailboxTrigger' $class: 'PullRequestBuildTrigger' $class: 'QuayIoTrigger' $class: 'RemoteBuildTrigger' upstream $class: 'RundeckTrigger' <code>scm</code> $class: 'SelfieTrigger' $class: 'SpoonTrigger' $class: 'SqsBuildTrigger' $class: 'TeamPRPushTrigger' $class: 'TeamPushTrigger' cron $class: 'URLTrigger' snapshotDependencies $class: 'io.relution.jenkins.awssqs.SQSTrigger' $class: 'io.relution.jenkins.scmsqs.SQSTrigger' $class: 'org.cloudbees.literate.jenkins.promotions.PromotionTrigger' $class: 'org.jenkinsci.plugins.deploy.weblogic.trigger.DeploymentTrigger' $class: 'org.jenkinsci.plugins.deployment.DeploymentTrigger' More info about scripted way here (sample from another question). Documentation that covers declarative pipeline is here.
Jenkins
44,113,834
31
I have an external tool that should be called as build-step in one of my jenkins jobs. Unfortunately, this tool has some issues with quoting commands to avoid problems with whitespaces in the path that is called from. Jenkins is installed in C:\Program Files (x86)\Jenkins. Hence I'm having trouble with jenkins calling the external tool. What I tried is to set "Workspace Root Directory" in Jenkins->configuration to C:\jenkins_workspace in order to avoid any whitespaces. This works for Freestyle Projects but my Multibranch Pipeline Project is still checked out and built under C:\Program Files (x86)\Jenkins\workspace. One solution would be to move the whole jenkins installation to e.g. C:\jenkins. This I would like to avoid. Is there a proper way to just tell Jenkins Pipeline jobs to use the "Workspace Root Directory" as well? Thanks for any help
the ws instruction sets the workspace for the commands inside it. for declarative pipelines, it's like this: ws("C:\jenkins") { echo "awesome commands here instead of echo" } You can also call a script to build the customWorkspace to use: # if the current branch is master, this helpfully sets your workspace to /tmp/ma partOfBranch = sh(returnStdout: true, script: 'echo $BRANCH_NAME | sed -e "s/ster//g"') path = "/tmp/${partOfBranch}" sh "mkdir ${path}" ws(path) { sh "pwd" } you can also set it globally by using the agent block (generally at the top of the pipeline block), by applying it to a node at that level: pipeline { agent { node { label 'my-defined-label' customWorkspace '/some/other/path' } } stages { stage('Example Build') { steps { sh 'mvn -B clean verify' } } } } Another node instruction later on might override it. Search for customWorkspace at https://jenkins.io/doc/book/pipeline/syntax/. You can also it use it with the docker and dockerfile instructions.
Jenkins
43,627,358
31
The Extended Choice Parameter plugin is great and I use it in jobs configured via the UI https://wiki.jenkins-ci.org/display/JENKINS/Extended+Choice+Parameter+plugin However, I'm struggling to get it working in a Jenkinsfile style pipeline script. It would appear that the Extended Choice Parameter plugin isn't yet fully compatible with Pipeline scripts since Jenkins pipeline-syntax generator creates the following snippet: parameters([<object of type com.cwctravel.hudson.plugins.extended_choice_parameter.ExtendedChoiceParameterDefinition>]) If I create the parameters manually I get the same behavior as mentioned in https://issues.jenkins-ci.org/browse/JENKINS-32188 org.kohsuke.stapler.NoStaplerConstructorException: There's no @DataBoundConstructor on any constructor of class Does anyone know of any workarounds that can get around the issue of ExtendedChoiceParameterDefinition not using @DataBoundConstructor? Jenkins 2.19.2 Extended Choice Parameter plugin 0.75
Since April's 2nd, 2019 it's now possible because of this commit: https://github.com/jenkinsci/extended-choice-parameter-plugin/pull/25 You can use it like this for instance: properties([ parameters([ extendedChoice( name: 'PROJECT', defaultValue: '', description: 'Sélectionnez le projet à construire.', type: 'PT_SINGLE_SELECT', groovyScript: valueKeysScript, descriptionGroovyScript: valueNamesScript ) ]) ]) If you want to know every possible parameter you have to refer to the source code. If you want to know every possible value for the "type" key, have a look at the PT_* constants.
Jenkins
42,392,247
31
I wish to change the time zone of the Jenkins. I have changed the time zone of the Jenkins installed server, but the Jenkins UI shows the different time. I need to set the PST time for Jenkins UI. How can I do it?
On Jenkins2 you can set the timezone at runtime via the Groovy Console. Just open "Manage Jenkins >> Script Console" and type System.setProperty('org.apache.commons.jelly.tags.fmt.timeZone', 'America/Los_Angeles') for example. Particularly helpful if you have no chance to change the startup variables but have admin rights on the instance. (often found in containerized setups). Only downside: Setting is gone on restart.
Jenkins
42,202,070
31
I am considering to use Jenkins pipeline script recently, one question is that I don't figure out a smart to way to create internal reusable utils code, imagine, I have a common function helloworld which will be used by lots of pipeline jobs, so I hope to create a utils.jar can injected it into the job classpath. I notice Jenkins have a similar concept with the global library, but my concern regarding this plugin: Since it is a plugin, so we need to install/upgrade it through jenkins plugin manager, then it may require reboot to apply the change, this is not what I want to see since utils may change, add always, we hope it could be available immediately. Secondly, it is official jenkins shared lib, I dont want to (Or they will not apply us) put private code into jenkins repo. Any good idea?
The Shared Libraries (docs) allows you to make your code accessible to all your pipeline scripts. You don't have to build a plugin for that and you don't have to restart Jenkins. E.g. this is my library and this a Jenkinsfile that calls this common function. EDIT (Feb 2017): The library can be accessed through Jenkins' internal Git server, or deployed through other means (e.g. via Chef) to the workflow-lib/ directory within the jenkins user's home directory. (still possible, but very unhandy). The global library can be configured through the following means: an @Library('github.com/...') annotation in the Jenkinsfile pointing to the URL of the shared library repo. configured on the folder level of Jenkins jobs. configured in Jenkins configuration as global library, with the advantage that the code is trusted, i.e., not subject to script security. A mix of the first and last method would be a not explicitly loaded shared library that is then requested only using its name in the Jenkinsfile: @Library('mysharedlib').
Jenkins
38,695,237
31
Our build server runs Jenkins 1.502 with Subversion plugin upgraded to version 1.45. This plugin uses svnkit-1.7.6-jenkins-1.jar. Also we have SVN client 1.7.8 installed. Jenkins successfully checks out source code from SVN repository. But when I go to workspace directory and try to run some svn command manually, it fails: # cd /var/lib/jenkins/jobs/myproject/workspace/ # svnversion svn: E155036: Working copy '/var/lib/jenkins/jobs/myproject/workspace' is too old (format 8, created by Subversion 1.4) The error message indicates that working copy was created by SVN 1.4, but version of svnkit is 1.7.6. How could it be? I searched entire file system, there are no any other svnkit-*.jar files.
There is an option in jenkins to tell svn which working copy format to use(manage jenkins > configure system), look for a 'Subversion Workspace Version' pulldown - it's likely set to 1.4. change it to the latest version in the list.
Jenkins
15,107,857
31
I'm attempting to build an ASP.NET vNext project in TeamCity. When it tries to build, I get the following error: C:\...\MyApp.kproj(7, 3): error MSB4019: The imported project "C:\Program Files (x86)\MSBuild\Microsoft\VisualStudio\v12.0\AspNet\Microsoft.Web.AspNet.Props" was not found. Confirm that the path in the <Import> declaration is correct, and that the file exists on disk. The file it's looking for is actually located at C:\Program Files (x86)\MSBuild\Microsoft\VisualStudio\v14.0\AspNet\Microsoft.Web.AspNet.Props I'm assuming that I need to get TeamCity to use the version of msbuild that ships with Visual Studio 2015. Is this even possible?
Edit: As of TeamCity 9.x, all works out of the box, but for earlier versions, the below is a solution. The project import problem should be solved by setting a env.VSToolsPath environment property to C:\Program Files (x86)\MSBuild\Microsoft\VisualStudio\v14.0. However, you will not be able to build using the TeamCity included MSBuild runner. But using a command-line runner is very simple. I extracted a meta-runner like this. It has almost the same functionality as the included TeamCity MSBuild 2013 runner. If you need more configurability, just add more parameters. My meta-runner ended up looking like this: <?xml version="1.0" encoding="UTF-8"?> <meta-runner name="MSBuild 2015"> <description>MSBuild 2015 command line runner</description> <settings> <parameters> <param name="solutionFile" /> <param name="target" value="Build" /> </parameters> <build-runners> <runner name="MSBuild 2015" type="simpleRunner"> <parameters> <param name="command.executable" value="C:\Program Files (x86)\MSBuild\14.0\Bin\MSBuild.exe" /> <param name="command.parameters" value="/v:m /m /t:%target% %solutionFile%" /> <param name="teamcity.step.mode" value="default" /> </parameters> </runner> </build-runners> <requirements /> </settings> </meta-runner> Note: TeamCity 9.1, due for Q2 2015 is expected to build VS2015 projects natively.
TeamCity
27,095,531
13
A little background. In my environment we have a large number small .NET solutions each in their own Subversion repositories (500+). We not a TFS shop and are currently evaluating moving our home grown CI process to TeamCity. Instead of having these 500+ repos polling our Subversion server every 5-10 minutes or so I'd like to kick off a Project build via a post-commit-hook REST http call (as our current solution does). I would then want TeamCity to update from SVN and commence the build process. Is this possible? I see TeamCity has a REST API, just that the documentation is sparse. I'm not sure how this example ties to anything I've got configured. What is bt7? How does it tie to the projects I've configured? http://buildserver:8111/httpAuth/action.html?add2Queue=bt7
bt7 is a build type identifier. Each build configuration has one. You can get the full list using the rest api as follows http://buildserver:8111/httpAuth/app/rest/buildTypes You can also see the build type in the url if you click any of the build configurations on your team city page. You will see a url parameter such as buildTypeId=bt7
TeamCity
9,436,792
13
I am trying to setup a build trigger for TeamCity using Mercurial as the VCS. Right now the trigger looks like: +:/** This trigger get fired when changesets are committed. However, I have TeamCity setup to tag each build in the VCS. The tagging process is firing the above build trigger so the build gets caught in a loop. Can anyone suggest a VCS build trigger that will filter out the tagging process?
Adding the trigger pattern: -:/.hgtags filters out the .hgtags file from the build trigger. This is the file that gets modified when the source is tagged by TeamCity. When this file is excluded tagging operations will not fire the build trigger.
TeamCity
1,478,297
13
I have a rather strange problem with TeamCity. I have a TeamCity installation, with local and remote build agents. The TeamCity server is hidden behind IIS with Application Request Routing (ARR), to enable SSL, etc. I have a feeling this might be part of the problem, but I am not sure. Another reason to suspect IIS being part of the problem is, I tried to host TeamCity on an Azure Web App, and got exactly the same behaviour. The trouble is, after building, when the build agents try to publish the artifacts to the server, I get a 404 back from the TeamCity server. TeamCity thinks it is a recoverable error (see log), and keeps trying again some times. Eventually, the publishing fails. If I configure the local agents to access TeamCity via http://localhost, everything works smooth. But, when accessing via the public address (which is served via IIS), I get 404s. The 404 content looks like a standard IIS 404 page. I have tries setting agent logging verbosity to DEBUG, but it still doesn't output the actual URL it is trying to call. Does anyone have any clues on how to troubleshoot this? Getting the TeamCity agent to output the URL for which it gets the 404 would be a good start. [Publishing artifacts] Publishing 1 file [F:/tc/ba3/temp/buildTmp/out/_PublishedWebSites/**/* => dist.zip] using [WebPublisher] [15:34:15][Publishing artifacts] Publishing 1 file [F:/tc/ba3/temp/buildTmp/out/_PublishedWebSites/**/* => dist.zip] using [ArtifactsCachePublisher] [15:35:10] [Publishing artifacts] Recoverable problem publishing artifacts (will retry): <!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd">; <html xmlns="http://www.w3.org/1999/xhtml">; <head> <meta http-equiv="Content-Type" content="text/html; charset=iso-8859-1"/> <title>404 - File or directory not found.</title> <style type="text/css"> <!-- body{margin:0;font-size:.7em;font-family:Verdana, Arial, Helvetica, sans-serif;background:#EEEEEE;} fieldset{padding:0 15px 10px 15px;} h1{font-size:2.4em;margin:0;color:#FFF;} h2{font-size:1.7em;margin:0;color:#CC0000;} h3{font-size:1.2em;margin:10px 0 0 0;color:#000000;} #header{width:96%;margin:0 0 0 0;padding:6px 2% 6px 2%;font-family:"trebuchet MS", Verdana, sans-serif;color:#FFF; background-color:#555555;} #content{margin:0 0 0 2%;position:relative;} .content-container{background:#FFF;width:96%;margin-top:8px;padding:10px;position:relative;} --> </style> </head> <body> <div id="header"><h1>Server Error</h1></div> <div id="content"> <div class="content-container"><fieldset> <h2>404 - File or directory not found.</h2> <h3>The resource you are looking for might have been removed, had its name changed, or is temporarily unavailable.</h3> </fieldset></div> </div> </body> </html>
EDIT: Found this documented on the TeamCity pages as well: https://confluence.jetbrains.com/display/TCD9/Known+Issues#KnownIssues-FailuretopublishartifactstoserverbehindIISreverseproxy Failed Request Tracing (as Terri Rougeou Donahue mentioned) was the tool to help me. I had two errors. Firstly, the StaticFileHandler was not turned off. So, when trying to POST to the /httpAuth/artefactUpload.html URL, the StaticFileHandler tried to handle the request before ARR could handle it. When I turned off StaticFileHandler, the RequestFiltering module kicked in, and returned an error code of 404.13, which is "Content Length Too Large". After a bit of googling, I found this, http://www.iis.net/configreference/system.webserver/security/requestfiltering/requestlimits, describing the parameter maxAllowedContentLength, and says "The default value is 30000000, which is approximately 28.6MB." The solution was: Turn off StaticFileHandler for the web site (Handler mappings) Edit the properties of "Request filtering" settings on the Web site, set "Maximum allowed content length" to something sensible. I added a 0 (makes it approximately 286MB, as artifacts can get quite large).
TeamCity
31,811,110
12
Using TeamCity in combiniation with git. Currently, TeamCity is set up with "master" as the default branch. Typically, development takes place on another branch (e.g. "dev") - TeamCity is set to watch for changes on "dev" and build automatically. If DEADBEEF-SOME-SHA has been built & tagged by TeamCity as build 1.2.3.4 on "dev" and we fast-forward merge that git SHA1 to "master", TeamCity still performs a build - so we end up with DEADBEEF-SOME-SHA being tagged as both 1.2.3.4 and 1.2.3.5. As I understand it, making "dev" the default branch would prevent this. Is there another way to prevent TeamCity performing a build if a build has already succeeded for that same SHA1? Note if we push directly to master (and that SHA1 doesn't exist on other branches / hasn't been built) I'd still like to see a build. I'd like to achieve this entirely in TeamCity if possible - no additional scripts/writing of files etc etc.
You can query builds for a particular SHA1... but you have to know your previous buildID for that. So what I would do is: write in a dedicated folder (accessible by all agents) the sha1 built at the end of each job only triggers a new job if that sha1 file is not already present.
TeamCity
46,826,665
12
I am trying to build an ASP.NET Core 2.0 application for .NET Framework 4.6.2 (not .NET Core) with TeamCity on Windows Server 2012R2. The following components are installed in the server: Microsoft .Net Core SDK - 2.0.0.0 Microsoft .Net Framework (4.5.2, 4.6, 4.6.2) Microsoft Build Tools (2013, 2015, 2017) Windows SDK 10.0 ... TeamCity uses nuget version 4.1 and the solution file to restore dependent packages (see full log below for details). Error: [22:57:51] [22:57:51]Errors in C:\path\to\project\Server\Server.csproj [22:57:51] Package Microsoft.AspNetCore 2.0.0 is not compatible with net462 (.NETFramework,Version=v4.6.2). Package Microsoft.AspNetCore 2.0.0 supports: netstandard2.0 (.NETStandard,Version=v2.0) [22:57:51] Package Microsoft.AspNetCore.Mvc 2.0.0 is not compatible with net462 (.NETFramework,Version=v4.6.2). Package Microsoft.AspNetCore.Mvc 2.0.0 supports: netstandard2.0 (.NETStandard,Version=v2.0) [22:57:51] One or more packages are incompatible with .NETFramework,Version=v4.6.2. [22:57:51] Package Microsoft.AspNetCore 2.0.0 is not compatible with net462 (.NETFramework,Version=v4.6.2) / win7-x86. Package Microsoft.AspNetCore 2.0.0 supports: netstandard2.0 (.NETStandard,Version=v2.0) [22:57:51] Package Microsoft.AspNetCore.Mvc 2.0.0 is not compatible with net462 (.NETFramework,Version=v4.6.2) / win7-x86. Package Microsoft.AspNetCore.Mvc 2.0.0 supports: netstandard2.0 (.NETStandard,Version=v2.0) [22:57:51] One or more packages are incompatible with .NETFramework,Version=v4.6.2 (win7-x86). [22:57:51] [22:57:51]Errors in C:\path\to\project\Server\Server.csproj [22:57:51] Package Microsoft.Extensions.FileProviders.Physical 2.0.0 is not compatible with netcoreapp2.0 (.NETCoreApp,Version=v2.0). Package Microsoft.Extensions.FileProviders.Physical 2.0.0 supports: netstandard2.0 (.NETStandard,Version=v2.0) [22:57:51] Package Microsoft.VisualStudio.Web.CodeGeneration.Contracts 2.0.0 is not compatible with netcoreapp2.0 (.NETCoreApp,Version=v2.0). Package Microsoft.VisualStudio.Web.CodeGeneration.Contracts 2.0.0 supports: netstandard2.0 (.NETStandard,Version=v2.0) [22:57:51] One or more packages are incompatible with .NETCoreApp,Version=v2.0. [22:57:51]Process exited with code 1 Question: I thought that .NET Framework 4.6.2 implements .NET Standard 2.0, therefore I don't understand why the packages are incompatible. On my local machine (with Visual Studio 2017 Update 3 (15.3)) nuget restore works fine. Any ideas how to solve that error / how to further analyze the problem? Project File (containing the nuget packages): <Project Sdk="Microsoft.NET.Sdk.Web"> <PropertyGroup> <TargetFramework>net462</TargetFramework> </PropertyGroup> <ItemGroup> <Folder Include="wwwroot\" /> </ItemGroup> <ItemGroup> <PackageReference Include="Microsoft.AspNetCore" Version="2.0.0" /> <PackageReference Include="Microsoft.AspNetCore.Mvc" Version="2.0.0" /> <PackageReference Include="Newtonsoft.Json" Version="10.0.3" /> </ItemGroup> <ItemGroup> <DotNetCliToolReference Include="Microsoft.VisualStudio.Web.CodeGeneration.Tools" Version="2.0.0" /> </ItemGroup> <ItemGroup> <ProjectReference Include="..\Server.Common\Server.Common.csproj" /> </ItemGroup> </Project> Full Log Output [Step 1/4] restore: Restoring NuGet packages for Server.sln (24s) [22:57:27][restore] NuGet command: C:\path\to\buildagent\tools\NuGet.CommandLine.4.1.0\tools\NuGet.exe restore C:\path\to\project\Server.sln [22:57:27][restore] Starting: C:\path\to\teamcity\temp\agentTmp\custom_script710236021428854.cmd [22:57:27][restore] in directory: C:\path\to\project [22:57:28][restore] MSBuild auto-detection: using msbuild version '15.3.409.57025' from 'C:\Program Files (x86)\Microsoft Visual Studio\2017\BuildTools\MSBuild\15.0\bin'. [22:57:31][restore] Alle in "packages.config" aufgef�hrten Pakete sind bereits installiert. [22:57:31][restore] Restoring packages for C:\path\to\project\Server\Server.csproj... [22:57:31][restore] Restoring packages for C:\path\to\project\Server\Server.csproj... [22:57:32][restore] GET https://api.nuget.org/v3-flatcontainer/microsoft.visualstudio.web.codegeneration.tools/index.json [22:57:32][restore] GET https://api.nuget.org/v3-flatcontainer/microsoft.aspnetcore.mvc/index.json [22:57:32][restore] GET https://api.nuget.org/v3-flatcontainer/microsoft.aspnetcore/index.json [...] [22:57:41][restore] OK https://api.nuget.org/v3-flatcontainer/microsoft.netcore.dotnetapphost/index.json 475ms [22:57:41][restore] GET https://api.nuget.org/v3-flatcontainer/microsoft.netcore.dotnetapphost/2.0.0/microsoft.netcore.dotnetapphost.2.0.0.nupkg [22:57:41][restore] OK https://api.nuget.org/v3-flatcontainer/microsoft.netcore.dotnetapphost/2.0.0/microsoft.netcore.dotnetapphost.2.0.0.nupkg 453ms [22:57:42][restore] Installing System.Xml.XmlSerializer 4.0.11. [22:57:42][restore] Installing System.Threading.Overlapped 4.0.1. [22:57:42][restore] Installing System.Security.Principal 4.0.1. [22:57:42][restore] Installing System.Dynamic.Runtime 4.0.11. [22:57:42][restore] Installing System.Private.DataContractSerialization 4.1.1. [22:57:42][restore] Installing Microsoft.Win32.Registry 4.0.0. [...] [22:57:48][restore] Installing System.Diagnostics.Contracts 4.0.1. [22:57:48][restore] Installing System.Threading.Tasks.Dataflow 4.6.0. [22:57:48][restore] Installing System.IO.Pipes 4.0.0. [22:57:51][restore] Package Microsoft.Extensions.FileProviders.Physical 2.0.0 is not compatible with netcoreapp2.0 (.NETCoreApp,Version=v2.0). Package Microsoft.Extensions.FileProviders.Physical 2.0.0 supports: netstandard2.0 (.NETStandard,Version=v2.0) [22:57:51][restore] Package Microsoft.VisualStudio.Web.CodeGeneration.Contracts 2.0.0 is not compatible with netcoreapp2.0 (.NETCoreApp,Version=v2.0). Package Microsoft.VisualStudio.Web.CodeGeneration.Contracts 2.0.0 supports: netstandard2.0 (.NETStandard,Version=v2.0) [22:57:51][restore] One or more packages are incompatible with .NETCoreApp,Version=v2.0. [22:57:51][restore] Committing restore... [22:57:51][restore] Restore failed in 19,33 sec for C:\path\to\project\Server\Server.csproj. [22:57:51][restore] [22:57:51][restore] Errors in C:\path\to\project\Server\Server.csproj [22:57:51][restore] Package Microsoft.AspNetCore 2.0.0 is not compatible with net462 (.NETFramework,Version=v4.6.2). Package Microsoft.AspNetCore 2.0.0 supports: netstandard2.0 (.NETStandard,Version=v2.0) [22:57:51][restore] Package Microsoft.AspNetCore.Mvc 2.0.0 is not compatible with net462 (.NETFramework,Version=v4.6.2). Package Microsoft.AspNetCore.Mvc 2.0.0 supports: netstandard2.0 (.NETStandard,Version=v2.0) [22:57:51][restore] One or more packages are incompatible with .NETFramework,Version=v4.6.2. [22:57:51][restore] Package Microsoft.AspNetCore 2.0.0 is not compatible with net462 (.NETFramework,Version=v4.6.2) / win7-x86. Package Microsoft.AspNetCore 2.0.0 supports: netstandard2.0 (.NETStandard,Version=v2.0) [22:57:51][restore] Package Microsoft.AspNetCore.Mvc 2.0.0 is not compatible with net462 (.NETFramework,Version=v4.6.2) / win7-x86. Package Microsoft.AspNetCore.Mvc 2.0.0 supports: netstandard2.0 (.NETStandard,Version=v2.0) [22:57:51][restore] [22:57:51][restore] One or more packages are incompatible with .NETFramework,Version=v4.6.2 (win7-x86). [22:57:51][restore] NuGet Config files used: [22:57:51][restore] [22:57:51][restore] C:\Windows\system32\config\systemprofile\AppData\Roaming\NuGet\NuGet.Config [22:57:51][restore] Errors in C:\path\to\project\Server\Server.csproj [22:57:51][restore] [22:57:51][restore] Package Microsoft.Extensions.FileProviders.Physical 2.0.0 is not compatible with netcoreapp2.0 (.NETCoreApp,Version=v2.0). Package Microsoft.Extensions.FileProviders.Physical 2.0.0 supports: netstandard2.0 (.NETStandard,Version=v2.0) [22:57:51][restore] Feeds used: [22:57:51][restore] https://api.nuget.org/v3/index.json [22:57:51][restore] Package Microsoft.VisualStudio.Web.CodeGeneration.Contracts 2.0.0 is not compatible with netcoreapp2.0 (.NETCoreApp,Version=v2.0). Package Microsoft.VisualStudio.Web.CodeGeneration.Contracts 2.0.0 supports: netstandard2.0 (.NETStandard,Version=v2.0) [22:57:51][restore] One or more packages are incompatible with .NETCoreApp,Version=v2.0. [22:57:51][restore] [22:57:51][restore] Installed: [22:57:51][restore] 151 package(s) to C:\path\to\project\Server\Server.csproj [22:57:51][restore] Process exited with code 1 [22:57:51][restore] Process exited with code 1 [22:57:51][Step 1/4] Step Nuget Restore (NuGet Installer) failed
Using Nuget Version 4.3 fixed it :).
TeamCity
45,723,797
12
I want to inhibit the building of certain projects within a solution from building (within a TeamCity Build Configuration in order to optimize the speed of my Commit Build feedback if you must know). I'm aware of the Solution Configurations mechanism but don't want to have to force lots of .sln files to end up with every permutation of things I want to be able to switch off. I have Convention based rule where I want to say "If I'm doing the Commit Build, I dont want to do the final installer packaging". (And I don't want to break it out into a separate solution). I'd prefer not to use a solution involving find and replace in the .sln file or in a .proj file created via [MsBuildEmitSolution][1]. I'm aware of questions here which cover the out of the box solution and this slightly related question. I see MSBuild /v:diag is saying: 2>Target "Build" in file "Z.sln.metaproj" from project "Z.sln" (entry point): Using "MSBuild" task from assembly "Microsoft.Build.Tasks.v4.0, Version=4.0.0.0, Culture=neutral, PublicKeyToken=b03f5f7f11d50a3a". Task "MSBuild" Global Properties: BuildingSolutionFile=true CurrentSolutionConfigurationContents=<SolutionConfiguration> <ProjectConfiguration Project="{C83D035D-169B-4023-9BEE-1790C9FE22AB}" AbsolutePath="X.csproj" BuildProjectInSolution="True">Debug|AnyCPU</ProjectConfiguration> <ProjectConfiguration Project="{15E7887D-F1DB-4D85-8454-E4EF5CBDE6D5}" AbsolutePath="Y.csproj" BuildProjectInSolution="True">Debug|AnyCPU</ProjectConfiguration> </SolutionConfiguration> So the question is: Is there a neat way of me getting to do an XPath replace or similar to have the effect of changing BuildProjectInSolution="True" to BuildProjectInSolution="False" for Project Y above Failing that, is there a relatively simple edit I can do within a .ccproj (An Azure 1.4 Package) or a .csproj (a general project) file to cause the effects (including triggering of dependent projects) of the project being enabled within a commandline msbuild Z.sln solution build to be nullified?
You could always pass the particular projects you want to build as parameters to the MSBuild. The MSBuild command line would look like this: MSBuild /t:<Project Name>:Rebuild;<Another Project Name>:Rebuild In TeamCity, you would put <Project Name>:<Target Action> in the target field in the MSBuild runner.
TeamCity
5,977,444
12
I'm setting up TeamCity (migrating from CruiseControl.NET) and I'm struggling to get it to perform incremental builds through MSBuild. I've got a small .proj file which contains a basic build script to invoke a build of my solution with some parameters fed in from TeamCity. When I invoke the script manually, MSBuild's Incremental Build features kick in and skip the build entirely on subsequent runs. When calling this script via Team City, the build log shows the output of a clean compile every time. I've observed the working directory during builds and can see the output from the previous build hasn't gone anywhere. I also manually called the build script from that directory by remoting onto the server and running MSBuild from the command-prompt. Running it this way triggers the expected incremental builds after the first invocation. Even when starting the build from the dashboard with no changes made, a complete rebuild occurs. I can't pinpoint the cause, but something appears to be giving MSBuild the impression that it's getting new changes and causing it to perform a rebuild on every run. I can't see much in the TeamCity documentation that would explain this - my expectation is that if there are no changes in the source control system, it would not update the working folder. Is TeamCity passing some parameter to the build process which triggers a rebuild? Can I view these parameters? Having examined a detail MSBuild log (/v:d command-line switch), the reason a complete rebuild is occurring is due to the file .NETFramework,Version=v4.0.AssemblyAttributes.cs being updated in the <Agent>\temp\buildTmp directory on every build. This file is normally found at %TMP%\.NETFramework,Version=v4.0.AssemblyAttributes.cs; TeamCity is changing the local temp directory environment variable to reference the agent's temp folder. Unfortunately, this file is created by the Microsoft.Common.targets part of the build process when absent. Deletion of the "temp" file before every build causes it to be created every build and is dynamically referenced in the build of every project file. I need to find a way to prevent this file from being re-created on every build.
A workaround for this problem is to customize the MSBuild process to set the path at which the "Target Framework Moniker Assembly Attributes" file (the proper name for the file mentioned in the question) will be created. The TargetFrameworkMonikerAssemblyAttributesPath property is defined in Microsoft.Common.targets determines where the file should be created. By overriding this property, the location can be changed to use a different location. Here's a script that can be used to achieve a suitable replacement: <Project xmlns="http://schemas.microsoft.com/developer/msbuild/2003"> <PropertyGroup> <PrepareForBuildDependsOn> $(PrepareForBuildDependsOn); _SetTargetFrameworkMonikerAssemblyAttributesPath </PrepareForBuildDependsOn> </PropertyGroup> <Target Name="_SetTargetFrameworkMonikerAssemblyAttributesPath" Condition="'$(TEAMCITY_VERSION)' != ''"> <PropertyGroup> <TargetFrameworkMonikerAssemblyAttributesDir Condition="'$(TargetFrameworkMonikerAssemblyAttributesDir)' == ''"> $([MSBuild]::GetRegistryValue("HKEY_CURRENT_USER\Environment", "TMP")) </TargetFrameworkMonikerAssemblyAttributesDir> <TargetFrameworkMonikerAssemblyAttributesDir Condition="'$(TargetFrameworkMonikerAssemblyAttributesDir)' == ''"> $([MSBuild]::GetRegistryValue("HKEY_CURRENT_USER\Environment", "TEMP")) </TargetFrameworkMonikerAssemblyAttributesDir> <TargetFrameworkMonikerAssemblyAttributesDir Condition="'$(TargetFrameworkMonikerAssemblyAttributesDir)' == ''"> $(USERPROFILE) </TargetFrameworkMonikerAssemblyAttributesDir> <TargetFrameworkMonikerAssemblyAttributesDir Condition="'$(TargetFrameworkMonikerAssemblyAttributesDir)' == ''"> $([System.IO.Path]::Combine('$(WINDIR)', 'Temp')) </TargetFrameworkMonikerAssemblyAttributesDir> <TargetFrameworkMonikerAssemblyAttributesPath> $([System.IO.Path]::Combine('$(TargetFrameworkMonikerAssemblyAttributesDir)','$(TargetFrameworkMoniker).AssemblyAttributes$(DefaultLanguageSourceExtension)')) </TargetFrameworkMonikerAssemblyAttributesPath> </PropertyGroup> <Message Text="Target Framework Moniker Assembly Attributes path is &quot;$(TargetFrameworkMonikerAssemblyAttributesPath)&quot;" Importance="low" /> </Target> The target is only executed when TEAMCITY_VERSION is specified as a property, which should be when the build is being executed by the TeamCity agent. NOTE: The child elements of the PropertyGroup should each be on a single line. They have been spread over multiple lines to increase readability here, but the additional line-breaks cause the script to fail. When the target runs, it tries to build a suitable path based on the user's environment variables as defined in the registry, first looking for TMP and TEMP, before falling back to the user's profile folder and finally the C:\Windows\Temp directory. This matches the order documented by System.Path.GetTempPath(), and should result in behaviour matching MSBuild execution outside of TeamCity. This should be saved as a .targets file somewhere on the system and imported to the .csproj file of projects being built by the TeamCity server, using an <Import> element. I added the script under my MSBuild extensions directory (C:\Program Files\MSBuild\) and referenced it by adding the following import element: <Import Project="$(MSBuildExtensionsPath)\TeamCity\TeamCity.Incremental.targets" /> The location/ordering of Import elements doesn't matter, but I suggest including it after the <Import Project="$(MSBuildBinPath)\Microsoft.CSharp.targets" /> which should appear in every .csproj file.
TeamCity
11,888,275
12
I am using Approval Tests. On my dev machine I am happy with DiffReporter that starts TortoiseDiff when my test results differ from approved: [UseReporter(typeof (DiffReporter))] public class MyApprovalTests { ... } However when the same tests are running on Teamcity and results are different tests fail with the following error: System.Exception : Unable to launch: tortoisemerge.exe with arguments ... Error Message: The system cannot find the file specified ---- System.ComponentModel.Win32Exception : The system cannot find the file specified Obviously it cannot find tortoisemerge.exe and that is fine because it is not installed on build agent. But what if it gets installed? Then for each fail another instance of tortoisemerge.exe will start and nobody will close it. Eventually tons of tortoisemerge.exe instances will kill our servers :) So the question is -- how tests should be decorated to run Tortoise Diff on local machine and just report errors on build server? I am aware of #IF DEBUG [UseReporter(typeof (DiffReporter))] but would prefer another solution if possible.
There are a couple of solutions to the question of Reporters and CI. I will list them all, then point to a better solution, which is not quite enabled yet. Use the AppConfigReporter. This allows you to set the reporter in your AppConfig, and you can use the QuietReporter for CI. There is a video here, along with many other reporters. The AppConfigReporter appears at 6:00. This has the advantage of separate configs, and you can decorate at the assembly level, but has the disadvantage of if you override at the class/method level, you still have the issue. Create your own (2) reporters. It is worth noting that if you use a reporter, it will get called, regardless as to if it is working in the environment. IEnvironmentAwareReporter allows for composite reporters, but will not prevent a direct call to the reporter. Most likely you will need 2 reporters, one which does nothing (like a quiet reporter) but only works on your CI server, or when called by TeamCity. Will call it the TeamCity Reporter. And One, which is a multiReporter which Calls teamCity if it is working, otherwise defers to . Use a FrontLoadedReporter (not quite ready). This is how ApprovalTests currently uses NCrunch. It does the above method in front of whatever is loaded in your UseReporter attribute. I have been meaning to add an assembly level attribute for configuring this, but haven't yet (sorry) I will try to add this very soon. Hope this helps. Llewellyn
TeamCity
9,939,209
12
I'm trying to configure TeamCity to build the project located on the Visual Studio Team Services with Git as VCS. The project contains spaces in the URL, so it looks like: https://mysrv.visualstudio.com/DefaultCollection/_git/some%20project Clone from Visual Studio 2013 works fine, from command line too. When I'm configuring VCS Root in TeamCity and press the Test Connection button it says that connection established, but when I'm trying to run the build TeamCityt reports that there is no compatible agents, and on the Compatible Agents tab of the build I see the message: Implicit requirements: 20project defined in VCS Root: Git VS MySrv Is it possible to fix this issue? P.S. I tried to rename repository on VS Team Services, but it adds to the Url collection name with the spaces :(
Do you still get this behaviour, if you try to use unescaped url (without %20 replacing space)? Another option is to escape '%' sign itself with another '%' - so escaped url of your repository will look like this https://mysrv.visualstudio.com/DefaultCollection/_git/some%%20project
TeamCity
23,091,358
12
We are well into our deployment of continuous integration environment using TeamCity. As we work through the CI process and move toward continuous deployment, we have run into a problem with how we manage production passwords. For other changes in the config, we use the Web.Config transform. However, I don't really want to bake the production password in a build profile. Before CI/CD, we would take the Web.config, use aspnet_regiis to decrypt the connection strings, change the password, then re-encrypt. Obviously, this is error prone and not at all in the spirit of CI/CD. I've had several other thoughts that were basically all about using something in the deploy script to re-write and then encrypt the connection strings section of the file, but it seems like this must be a common problem and that there must be some generally accepted solution. But so far, I can't find it. Is there a "right way"? Thanks!
One possible solution, available since TeamCity 7.0, is to use typed parameters. You can define a parameter in TeamCity of type password, and pass it somehow to your build script (either as environment variable or as your build script property). TeamCity stores values of such parameters in its own configuration files and in database in scrambled form. If password appears in build log or on build parameters page, it will be replaced with ***.
TeamCity
9,470,703
12
I have a TeamCity build configuration A and B, where B is dependent on A. I need to pass a parameter from B to A when B is triggered. This is related to question: Override dependencies properties by parameters value in TeamCity 9 and the teamcity documentation here I need to find WHERE/HOW to use this reverse.dep to set the parameter in the dependent build? In the Project Configuration Parameters section, I can add Configuration/Environment/build parameters, but they take a Name/Value pair. So, pardon my ignorance here, but am not able to make out where to specify this reverse logic. Thanks
Found it! We just need to add a new Configuration Parameter in B with name as reverse.dep.<btId>.paramName and its value as the intended value that needs to be passed. Imp: As noted in the TeamCity documentation - As the parameter's values should be known at that stage, they can only be defined either as build configuration parameters or in the custom build dialog.
TeamCity
37,857,187
12
Since GitLab 7.6, or thereabouts, there is a new option to use TeamCity directly from GitLab projects. In the setup there is this message: The build configuration in Teamcity must use the build format number %build.vcs.number% you will also want to configure monitoring of all branches so merge requests build, that setting is in the vsc root advanced settings. I'm not sure how this works. Lets say I have a repository Foo. I have setup a build on TeamCity to listen to Foo with branch specification: +:refs/pull/*/merge I then fork Foo in gitlab as FooFork, make a change, then request a merge FooFork -> Foo. But nothing happens to test this merge, which is what I was expecting GitLab to do. If I accept the merge than the build server jumps into action (immediately) and builds twice (master and /ref/master). I've also set the build configuration to use exactly: %build.vcs.number% as the build number as prescribed, but gitlab doesn't seem to give me any information about the build result. So I'm a bit confused really as to what exactly this GitLab -> TeamCity integration is supposed to do and whether I'm doing wrong. I'm currently running GitLab 7.9 and TeamCity 8.1.4 Update: Seems this use case was not supported prior to version 8 - https://github.com/gitlabhq/gitlabhq/issues/7240
I'm running GitLab 8.0.2 and TeamCity 9.1.1 and am able to run CI builds on branches and merge requests. I trigger CI builds for specific branches by setting a VCS trigger together with the branch specification +:refs/heads/(xyz*) where xyz is the string for our ticket system prefix since all active branches need to be named after an entry in our issue tracker. I trigger builds for merge requests via the branch specification +:refs/(merge-requests/*) Everything works as as expected and lets us know the status of all feature / bug branches and merge requests automatically. Thanks to Rob's comment linking to the GitLab 8 release notes entry on the merge request spec.
TeamCity
29,282,548
12
I've had to revert to a previous commit in my master branch in git which has meant I've had to force push the changes up to Teamcity. It's seems as though Teamcity has got into a bind and it thinks that any newly triggered builds are actually building an older version of the project (it's correct, I reverted from Build Number 750 to 747) When running the build it displays this: When I look in the history it looks like this (all builds after I reverted are grey) Is there anything I can do to make Teamcity think I am building the latest? Maybe clearing logs or something similar?
You could always delete the builds for the reverted commits that no longer exist. To do this go to the build details page then click "Actions" > "Remove".
TeamCity
25,887,582
12
I came across an interesting issue. I want to build nuget packages with Teamcity. I did set up the configuration which is really straight forward (Good job JetBrains!) However I am not able to run it on one of our build agents. The agent does pass the agent requirements for the configuration, but next to it's name the following is shown: not allowed to run this configuration My question: Why? On the agent configuration parameters page I have Nuget as possible configurations: teamcity.tool.NuGet.CommandLine.2.8.2.nupkg D:\BuildAgent\tools\NuGet.CommandLine.2.8.2.nupkg teamcity.tool.NuGet.CommandLine.DEFAULT.nupkg D:\BuildAgent\tools\NuGet.CommandLine.DEFAULT.nupkg We have a second build agent which is able to run Nuget Packager configuration. The main difference between this two machines is that one is a Windows 8, version 6.2 machine (not allowed to run this configuration) and the other one is a Windows 7, version 6.1 (allowed run this configuration) We are running: TeamCity Enterprise 8.0.5 (build 27692) Any hints and help will be greatly appreciated! Thank you!
Most likely the agent is configured to run only explicitly assigned configurations. Plesase, check the Agents -> -> "Compatible configurations" tab. There is a combo box with options "Run all compatible" / "Run assined .. ". Make sure "Run all compatible" is selected
TeamCity
25,030,483
12
Is there a way to push up a commit containing a param in the commit message such as "--nobuild" which would disable building the project in TeamCity?
Yes, you should change your build trigger. There are trigger rules, and you can add new rule -:comment=--nobuild:** More info: http://confluence.jetbrains.com/display/TCD8/Configuring+VCS+Triggers
TeamCity
23,360,619
12
As the first step in a build configuration I am trying to dynamically change a parameter and use it in the subsequent steps. Reading online, it seems that the way to do this is to call ##teamcity[setParameter. But this doesn't seem to be working. It doesn't even change the value in the same step. For example, I have created a test parameter and set it's default value to '1'. Inside a powershell script, I tried to change it to 2, as shown below. But the output remains unchanged as can be seen below I am currently using TeamCity 8.0.3 (build 27540). What am I doing wrong?
EDIT: I think the problem might be the command you're using to set the parameter. Try: Write-Host "##teamcity[setParameter name='TestParameter' value='2']" -- We've experienced the same behavior. The key here is 'subsequent steps.' You must modify the parameter in a separate build step that is run before the step in which you want to use the new parameter. It's my understanding that all parameters in a build step are evaluated immediately before the execution of that step. The tokens will be replaced with the unmodified values of those parameters. Thus, what actually gets executed by the build agent is: Write-Host "TestParameter value is 1" Write-Host "##teamcity[setParameter name='TestParameter' value='2']" Write-Host "TestParameter value is 1"
TeamCity
22,141,259
12
I'm running into an issue with unit tests on our Team City (8.0.4) build server - the code builds & runs all tests locally via Resharper and nCrunch. But when running on the server I get the following error, even though the Unity assembly exists in the same directory as the unit test assembly, and is referenced in the unit test assembly. SetUp method failed. SetUp : System.IO.FileNotFoundException : Could not load file or assembly 'Microsoft.Practices.Unity, Version=2.0.414.0, Culture=neutral, PublicKeyToken=31bf3856ad364e35' or one of its dependencies. The system cannot find the file specified. at XXXX.Unity.UnityContainerAdapter..ctor() at XXXX.GraphExtensionsTests..ctor() in c:\TeamCityV7\Agent-1\work\f02f7e27c0bedfa2\XXXX\Graph.Tests\Extensions\GraphExtensionsTests.cs:line 44 I've confirmed the copy of Microsoft.Practices.Unity is the correct version. I've also confirmed the assemblies are built using the full version of the framework - not using client profile. Any ideas why Team City might be failing?
Check the pattern you're using to locate your test assemblies. I had a similar problem with another library and it turns out the pattern was finding the test assembly under bin\Release and obj\Release; the obj folder doesn't contain all the assemblies referenced by the project and is really just a scratch folder for the compiler.
TeamCity
21,164,646
12
When trying to deploy my site using TeamCity and Web Deploy I get this error: error MSB4057: The target "MsDeployPublish" does not exist in the project. Is there something I have to install on a build server? It's a clean Windows Server 2012 with Web Deploy 3.5 installed.
Or you can use this NuGet package with portable version of the targets: https://www.nuget.org/packages/MSBuild.Microsoft.VisualStudio.Web.targets and modify your csproj file to include it like this: <Import Project="..\packages\MSBuild.Microsoft.VisualStudio.Web.targets.12.0.1\tools\VSToolsPath\WebApplications\Microsoft.WebApplication.targets" />
TeamCity
19,295,854
12
I created build step with type "MSBuild", set Target to "Clean;Build;Publish", added command line parameters to /p:Configuration=Release;PublishDir=M:\MyPackage after running configuration I got "success" status but M:\MyPackage folder is empty. I need just revive deployment package files in directory on same computer but do not deploy to server or somewhere else
I've solved this problem by creating "Visual Studio" build step and add next build parameters /p:Configuration=QA /p:DeployOnBuild=true /p:PublishDir=M:\MyPackage It still do not copy deployment package to MyPackage folder, but it is available in "obj" directory of project sources and this is enough for me.
TeamCity
18,993,874
12
I've just installed Teamcity 8.0.3 on a fresh Windows Server 2012 machine. Installation was successful, and I'm trying to configure an agent in order to fetch a project stored in a git server. This server uses a ssh key. I've added it to my agent, but when it tries to retrieve the project this error appears. Failed for the root 'rtogit' #1: List remote refs failed: com.jcraft.jsch.JSchException: The cipher 'aes256-cbc' is required, but it is not available. I've seen, for example here that I must change my policy, but I'm not a java expert and I don't know what I must do. Can someone help me please? Edit: I forgot to say that I've also installed GitExtensions 2.46 complete.
I had this problem and found out my private key file was in the wrong format. I'm not sure if you used PuTTYgen to generate the key but if so try "Export OpenSSH key" from the Conversions menu and use that file instead.
TeamCity
18,813,237
12
I have a command line 'custom script' build step involving robocopy. Unfortunately when robocopy is successful, it returns exit code 1 instead of the more common exit code 0. When it does this, it fails my teamcity build configuration. How can I tell teamcity to fail the build if the exit code != 1 for this build step only? Can this be done? What about editing project-config.xml somehow?
There's two ways: In that Build Configuration, go to the Build Failure Conditions step. Look for the heading Fail build if: . The first checkbox is "build process exit code is not zero". Make sure that sucker isn't checked. When you run robocopy, check the result of the call to robocopy. You can explicitly exit 0 from inside the script if robocopy works, or do something else. This is necessary if you need to fail the build upon other conditions (e.g., first exit 1 if the source folder doesn't exist, then run robocopy, then send a message if robocopy is successful).
TeamCity
14,477,592
12
I'm trying to integrate the sonar analysis into by TeamCity build process. I have a NUnit build step which runs my unit tests and then runs dotCover for the coverage. My next step is the sonar-runner. The configuration that currently exists is; gallio.mode=dotCover, sonar.gallio.mode=reuseReport but I also need sonar.gallio.reports.path. Does anybody know the path to the dotCover report generated in the the previous step?
Spent some amount of time on the same issue, but with newer Sonar c# plugin (v.2.3) - Gallio support has been dropped, but the report is still required. To answer the question directly, TeamCity puts dotcover snapshot file into a temp folder with a name like coverage_dotcover27574681205420364801.data (where digits are random). So The procedure is: Create a PowerShell Build step in Team City after the step with test and coverage you may use Command line if you prefer Get the full dotCover snapshot name in temp folder Run dotCover to produce a HTML report from a snapshot Note - Sonar (c# plugin v 2.3) supports only dotCover HTML reports Pass the produced HTML report to sonar PowerShell script: $snapshot = Get-ChildItem "%system.teamcity.build.tempDir%" ` -Filter coverage_dotcover*.data ` | select -ExpandProperty FullName -First 1 %teamcity.dotCover.home%\dotCover.exe report ` /ReportType=HTML /Source="$snapshot" ` /Output="%sonar.coverageReport%" Now you can specify your report in sonnar runner as sonar.cs.dotcover.reportsPaths='%sonar.coverageReport%' Where %sonar.coverageReport% is a defined property in a TeamCity
TeamCity
13,170,780
12
I have two svn VCS roots (ProjectX, ProjectY). Correct build path should be: ParrentFolder\ProjectX (svn://svn_server1/ProjectX) ParrentFolder\ProjectY (svn://svn_server2/Folder1/ProjectY) How to configure shared ParrentFolder for both projects? I looked into Checkout directory parameter but its seams there is no system variable which can create folder by project name. Thanks a lot
So if anyone have similar issue you need todo next: Configure checkout rule for the first project +:.=>ProjectX Configure checkout rule for the second project +:.=>ProjectY Configure correct build paths /ProjectX/ProjectX.sln, /ProjectY/ProjectY.sln
TeamCity
13,117,953
12
We use the build in coverage application in TeamCity 6 (about to upgrade to 7.1) If we wish to see the code coverage (or other metrics) of a particular build it is fine as we can navigate to that build, but it would be great if we could pluck out a few interesting metrics from all/some of the current projects/build configurations and display them all together. For convenience I would expect the new display to be accessible from within TeamCity itself, however if there are solutions that require a separate solution we could look at them.
If you want to compare a set of common metrics (e.g. code coverage) across different projects and over time then SonarQube is probably what you want. You can integrate it with TeamCity by adding a sonar-project.properties file to each project and calling sonar-runner from a command line build step.
TeamCity
12,844,190
12
We use TeamCity and GitHub Enterprise. We use an open-source-esque workflow with git: there's a mainline repository for each component, and when people want to make changes, they fork mainline to their own account (so there might be many forks) create a branch in their fork implement change bring up to date with mainline/master for changes that have happened in the meantime submit a pull request of fork/feature-branch -> mainline/master We're very happy with this workflow; it forces a code-review (well, at least a manual step, which hopefully involves actually reading the code and running its tests) before mainline sees any changes, which historically has been a problem. We'd like to use the GH Status API (blog post, API doc) to turn the merge button non-green if the author is the person looking at the pull-request, but that's for later. We have TeamCity 7.1 set up to watch the mainline repositories and build when changes are seen. However, the way it's currently set up, CI only builds when it sees changes to mainline/master. How should we configure our VCS roots in TeamCity such that we can have the same workflow, but CI will trigger a build based on branches in forks of the mainline repo? Preferably without our having to register every fork individually? I've read TeamCity 7.1's feature-branch documentation (blog post, release notes, documentation), but I don't see how to apply it to our model of arbitrary-number-of-forks as opposed to everyone-commits-to-mainline-in-feature-branches.
You can monitor pull-requests by teamcity: http://blog.jetbrains.com/teamcity/2013/02/automatically-building-pull-requests-from-github-with-teamcity/
TeamCity
12,494,759
12
I have looked into both. Would like your suggestions as to which one is better for automated web deployment on multiple servers.
I think you should definitely give TeamCity and Octopus a try. We use TeamCity to create Octopus (NuGet) packages and the Octo tool to automatically trigger deployment to a test environment after each succesfull build. After that we use the Octopus portal to promote deployments to other environments. We use the following Octo command line to trigger deployments from TeamCity: Octo.exe create-release --apiKey=YourOctopusAPIKey --server=http://YourOctopusServer:9015/api --project=YourOctopusProjectName --deployto=YourOctopusEnvironment The Octo create-release step needs to be in a separate TeamCity project, otherwise the NuGet won't be updated with the resulting package from the build.
TeamCity
11,411,436
12
Is it possible to deploy a VS 2010 database project using TeamCity? I am building my whole solution, and deploying a website to my server, this all works fine. The final step I want to trigger is the deploy of the database project which generates a sql script and deploys it. I have the "Create a deployment script (.sql) and deploy to the database" option selected as a deploy action, my Configuration target is set to build and deploy the database project, but I just can't figure how to get TeamCity and MSBuild to trigger it.
Visual Studio must be installed for this to work. For the original SQL Server 2005/2008 Database Project types: Create a build step of runner type Visual Studio to build the solution. Create a build step of runner type Command Line. Set Command Executable to C:\Program Files\Microsoft Visual Studio 10.0\VSTSDB\Deploy\VSDBCMD.exe. Set Command Parameters to /a:Deploy /dd:+ /manifest:%system.teamcity.build.checkoutDir%\<PROJECT PATH>\sql\debug\<PROJECT NAME>.deploymanifest. (See here for VSDBCMD.exe parameters). For the SQL Server Database Project provided by SQL Server Data Tools or Visual Studio 2012/2013: Create a build step of runner type Visual Studio to build the solution. Create a build step of runner type Command Line. Set Command Executable to C:\Program Files\Microsoft Visual Studio 10.0\Microsoft SQL Server Data Tools\sqlpackage.exe. See here for sqlpackage.exe parameters. Here's an example with the deployment settings stored in an XML file (created via the Publish option): /Action:Publish /SourceFile:%system.teamcity.build.checkoutDir%\<PROJECT PATH>\bin\Debug\<PROJECT NAME>.dacpac /Profile:%system.teamcity.build.checkoutDir%\<PATH TO PROJECT>\PublishSettings.xml.
TeamCity
11,291,250
12
I have a TeamCity agent configured to build my XCode projects and I use github. I would like to automatically include in my release notes the descriptions from all pending commits in TeamCity. How can I fetch them from github and store them in teamcity? Once I put them in a teamcity variable I can easily add them to my build script.
THis is how I ended up doing this using a bash script: #!/bin/bash curl -o lastBuild.tmp "http://localhost:8111/app/rest/buildTypes/id:bt2/builds/status:SUCCESS" --user rest:rest last_commit=`xpath lastBuild.tmp '/build/revisions/revision/@version'| awk -F"\"" '{print $2}'` echo "##Last commit = $last_commit" # prepare build notes NOTES=`git log --pretty=format:"- %s" $last_commit..origin/master` echo "this is it:$NOTES" Some explanations: Use curl to fetch the last successful build from your build configuration. In my sample this is bt2, make sure to replace it with yours Use XPath/AWK to parse the XML response and get the last git version Use git log to get all changes form last build and format them anyway you want. I wanted to just get the commit descriptions.
TeamCity
10,794,300
12
Is it possible to format powershell output so that it renders as a collapsible section in the TeamCity build log, Tree view? So for example, my build step uses a powershell runner, and issues a write-host " ################# deployment manifest ############################" ls -r -i *.* | %{ $_.FullName } which outputs this: [15:28:13] ################# deployment manifest ############################ [15:28:13]\\10.10.10.49\d$\sites\account.foo.net\v32\Bin [15:28:13]\\10.10.10.49\d$\sites\account.foo.net\v32\contact [15:28:13]\\10.10.10.49\d$\sites\account.foo.net\v32\Content [15:28:13]\\10.10.10.49\d$\sites\account.foo.net\v32\controls [15:28:13]\\10.10.10.49\d$\sites\account.foo.net\v32\error I'd like that chunk of the log to be collapsible in the Tree View.
Yes we do this with our powershell scripts, you need to get your build script to update Teamcity with the build status. More specifically you need to report the build progress which will tell Teamcity when the start and the end of a block of work occurs. After the build has finished Teamcity will use this information to create nodes on the tree view of the log. In powershell do the following: write-host "##teamcity[progressStart '<message>']" do work write-host "##teamcity[progressFinish '<message>']" Note You need to make sure that the message is the same in the start and finish message, blocks can be nested. You can also use the block message instead. I don't know exactly what the difference is but you appear to get the same results: write-host "##teamcity[blockOpened name='<blockName>']" do work write-host "##teamcity[blockClosed name='<blockName>']"
TeamCity
10,357,525
12
In our project, deployment is always a pain, mostly because of the mistakes done by the release management team. Either they screw up the configuration or get the wrong version installed somehow. We use teamcity as our CI server, and it produces the artifacts as zip files(dll's and exe) which is usually passed on to the release team. My question is, is there a way to automate the whole deployment process? Is there a commercial tool, which supports this? We will want to do the following: Update the config files with environment specific values. Install windows services to the server. Upload the UI(WPF) bundle to the centralized location(which is pulled down by another application, sort of a launcher). Change the DB connection strings. Do all the above for various environments(like int,uat and prod) DB deployment since is a separate beast as such, need not be covered in this. Any best practices, tools or solutions will be highly helpful. Thanks, -Mike
I have used TeamCity for some fairly large projects and I have automated every aspect of deployments apart from the database. The main steps I use for each project are: Get a TeamCity agent installed on the production server Have the build get everything out of source control (you do have everything in source control right?). Have a build step that builds and publishes your solution. This can be achieved by adding the following command line argument to your MSBuild call: /p:Configuration=[Your Config];DeployOnBuild=True;PackageAsSingleFile=False Your published files (and tranformed config files) will be written to the following directory: [Your Project Directory]\obj\[Your Config]\Package\PackageTmp Using a scripting language (in my case Powershell) to copy the published artifacts to your deployment directory and make environment specific changes you mentioned. E.g. extracting archives, copying files, starting/stopping websites etc.. Run any automated testing (e.g. nUnit, Selenium etc...) I find the best strategy is to have a .Net post-build event that invokes an appropriate powershell script passing in relevant details like the solution path and configuration name (alternatively, I have also had TeamCity pass the environment name to the Powershell script) so that it knows what it needs to do (e.g. Staging, Production etc...). You should find that a scripting language like Powershell can do everything that a person can do (and about 100x faster and 100% reliably). There is so much content on Powershell out there that you can just google anything you need to do in Powershell and you will get an example. E.g. "powershell deploy WPF", "powershell upload FTP" etc... In a previous job I needed to deploy windows services remotely and I found that with enough research, I was able to get the MSI for the service to uninstall the existing service and install the new one completely silently (i.e. no dialogs). This will help a lot in your quest for automation. I can elaborate on this if you would like. Below is an example of a Powershell post build script I generally use: Note how I use some default parameter values so that I can execute the script directly from my Powershell editor to simulate and test different configurations on my local machine. param( [string]$configurationName="Debug", [string]$sourceDirectory="C:\SVN\<Your local solution location>") Set-StrictMode -v latest $ErrorActionPreference = "Stop" # Load required functions $private:scriptFolder = & { (Split-Path $MyInvocation.ScriptName -Parent) } . (Join-Path $scriptFolder DebugBuild.ps1) . (Join-Path $scriptFolder StagingBuild.ps1) . (Join-Path $scriptFolder ProductionBuild.ps1) . (Join-Path $scriptFolder CommonBuildFunctions.ps1) #Execute appropriate build switch ($configurationName) { "Debug" { RunDebugBuild $sourceDirectory } "Staging" { RunStagingBuild $sourceDirectory } "Production" { RunReleaseBuild $sourceDirectory } } To execute a publish on development machines, I setup a VS publish profile for the solution that is committed to SVN so the other developers can use it. This profile publishes directly to the local deployment directory.
TeamCity
8,902,468
12
I am on Windows and trying to run multiple (currently two) instances of TeamCity on the same server. I chose not to install the Windows services and instead run the server via runAll.bat start command. When I ran the installer I chose different ports, names and paths for each one. The first server starts successfully, but the second one does not - I see the tomcat window showing up, but it's gone after a while, though the build agent window stays. Here's some output: c:\Tools\TeamCity\bin>runAll.bat start starting TeamCity server and agent... Using CATALINA_BASE: "c:\Tools\TeamCity\bin\.." Using CATALINA_HOME: "c:\Tools\TeamCity\bin\.." Using CATALINA_TMPDIR: "c:\Tools\TeamCity\bin\..\temp" Using JRE_HOME: "c:\Tools\TeamCity\bin\..\jre" Using CLASSPATH: "c:\Tools\TeamCity\bin\..\bin\bootstrap.jar" Starting TeamCity Build Agent Launcher... Agent home directory is C:\Tools\TeamCity\buildAgent Current JRE version is 1.6 c:\Tools\TeamCity\bin>cd c:\Tools\TeamCity2\bin c:\Tools\TeamCity2\bin>runAll.bat start starting TeamCity server and agent... Using CATALINA_BASE: "c:\Tools\TeamCity2\bin\.." Using CATALINA_HOME: "c:\Tools\TeamCity2\bin\.." Using CATALINA_TMPDIR: "c:\Tools\TeamCity2\bin\..\temp" Using JRE_HOME: "c:\Tools\TeamCity2\bin\..\jre" Using CLASSPATH: "c:\Tools\TeamCity2\bin\..\bin\bootstrap.jar" Starting TeamCity Build Agent Launcher... Agent home directory is C:\Tools\TeamCity2\buildAgent Current JRE version is 1.6
You need to edit conf\server.xml for the second Tomcat/TeamCity instance so that it uses different ports or binds to different network interfaces, changing the following ports should be enough: Server port="8005" Connector port="8080" Two servers cannot share the same database, so you must configure them to use different locations via TEAMCITY_DATA_PATH environment variable. If you also want to run multiple agents on the same machine, refer to my answer here.
TeamCity
6,267,737
12