Currently We are using Bitbucket repository for SQL scripts version control and manually deploying the changes to database.
For now we are going to automate the deployment. So Can any one suggest how we can automate the sql scripts from bitbucket repository.
Related
I have an MVC 5 application we're moving from on-premise to the Azure cloud. Currently, we have several publish profiles, one per environment, which we determine using a powershell script. One of our goals is to make the building scripts and infrastructure as simple as possible, so I was wondering if I could make it so that using only my appveyor.yml file I could set the publish profile to be used, so
Is there a way to set the publish profile from the appveyor.yml file?
If not what are my choices?
You can run your PowerShell script as part of desired build step in pipeline. It is possible can run commands right from YAML file or UI or check-in your PowerShell script into repository and run .ps1 file. You might consider using secure variables to avoid checking in things like connection strings into repo in clear text.
However this custom script/profiles approach will not allow you to use built-in WAP artifacts packaging and you will be also needed to use custom script instead of automatic MSBuild mode. Which is OK, but a little bit more scripting. Also you will be needed to publish artifacts so it will be available for deployment.
Maybe easier option is to let AppVeyor do all build and WAP artifacts packaging/publishing automatically, and then use built-in Web Deployment with Web Deploy parametrization instead of multiple publishing profiles.
But if you decide to go with custom scripts, and multiple publishing profiles, you still can use use built-in Web Deployment with artifacts created by your scripts.
I have been trying to find out how should I execute the Selenium Test (Java) using gitlab CI
I have created an automation framework and I am able to run the maven project via jenkins
I wanted to run the same maven project with the help of gitlab ci runner
My Code will be available on git and just need to trigger the execution as a when developer checks in the code
Please help me out with this setup, I have been trying to find out the solution but couldn't figure out any
I suggest you to read about jeknins and gitlab hooks here: https://docs.gitlab.com/ee/integration/jenkins.html
, In general, these hooks "follows" any push you perform to you gitlab repository, and run the desired build on them, including pulling the latest code.
I've spent countless hours trying to understand how to do this correctly.
I have a nodejs application with angular front-end which are both contained in the same project.
I would simply like to have Gitlab CE CI build the project and then copy the resulting dist folder and package.json file to the production server and restart
I have a shared gitlab runner setup and was able to successfully configure the ssh runner.
Using the Gitlab runner ssh I was able to copy the entire project to the production server but cannot get it to build (plus I really don't want to have all files on the server, just the production required files.
what am I missing. do you use a docker runner with a node image to build the project and they scp files to the production server?
Any guidance would greatly be appreciated.
Yes, I think you already state the solution.
I would recommend looking into Job Artifacts to build your application in a Docker-based build job and upload the production files as an artifact.
The next job could then deploy (by scp or ssh+wget/curl) the artifact files to your production server and restart the webserver.
I'm trying to reconfigure our TeamCity build server (running on WIN 2008) and switch from SVN to Mercurial repository.
I use the following link as Configuring Mercurial (Hg) with TeamCity over SSH at RepositoryHosting.
I spent several days and still can't get it work.
I can reach the Mercurial repository from TeamCity (via Test Connect) and I can pull code from Mercurial via TrotoiseHG, but the build fails every time because TeamCity cannot pull source code from the server.
Does anybody had the same issue?
Have you tried using Agent-Side checkout instead of the default? Otherwise, TeamCity tries to "intelligently" detect changes and distribute them to build nodes. Failure can occur whenever something happens in the VCS that TeamCity doesn't understand how to replicate, such as using subrepos.
I have the following setup:
Development server for continuous build with Hudson, Maven and Artifactory up and running.
I am able to build an artifact by Maven, and let Hudson deploy it to Artifactory.
However, I would like to do the following:
Select an artifact in Artifactory on the local build server.
Deploy that artifact to a remote repository without doing a new build.
The reason I want to do that is, that the first server is a development server where new artifacts are changed at will. The other server / repository is the public visible server for all artefacts of the company.
I know that I am able to create a second build job, but that would be a duplication, and Hudson is only able to manage one Artifactory repository at one time. I could deploy that artifact by hand to a different URL, but this is error prune and I need a login on the build server and have to use the shell to do that.
Has anyone an idea how to finally reach the goal?
If I understood your question correctly it looks like a good scenario for User Plugin. It should be straightforward, you select the needed artifacts using the searches object, and then execute REST PUT call to put them on the remote server.
The Maven Stage plugin seems to support this case: http://maven.apache.org/plugins/maven-stage-plugin/usage.html
but I'm not sure if it works correctly. I'm working on it currently, so if it doesn't work perhapse I will find time to correct it.