conversations
listlengths
2
3
source
stringclasses
1 value
stack
stringclasses
1 value
[ { "from": "system", "value": "Add a gitter chat badge\n" }, { "from": "human", "value": "= JSR 377 - Desktop|Embedded Application API\n\nContains the specification and API sources for link:https://jcp.org/en/jsr/detail?id=377[JSR-377].\n\n== Additional resources\nlink:http://jsr377-api.40747.n7.nabble.com[Forum / Mailing List]\n" }, { "from": "gpt", "value": "= JSR 377 - Desktop|Embedded Application API\n\nContains the specification and API sources for link:https://jcp.org/en/jsr/detail?id=377[JSR-377].\n\nimage:https://badges.gitter.im/Join%20Chat.svg[Gitter, link=\"https://gitter.im/jsr377/jsr377-api?utm_source=badge&utm_medium=badge&utm_campaign=pr-badge\"]\n\n== Additional resources\nlink:http://jsr377-api.40747.n7.nabble.com[Forum / Mailing List]\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Fix close tag on table\n" }, { "from": "human", "value": "= Badging Your App\n\nYou can set the badge count on your app's icon using:\n[source,csharp]\n----\nbool SetBadgeCount(int count)\n----\n\n.Parameters\n[cols=\"1,2a\"]\n|===\n|Parameter |Description\n|count |The number to assign as a badge to the application's icon\n|===\n\nThis function will return `true` if it was able to set the badge count for your\napp's icon, or `false` otherwise.\n\nOn Android, this functionality uses the https://github.com/leolin310148/ShortcutBadger[ShortcutBadger library by Leo Sin].\n\n== Known Issues\n\n[cols=\"1,2a\"]\n|===\n|Date |Issue\n\n|October 11, 2018 |Google Nexus and Pixel devices are not supported, because the\nlauncher used does not support badging.\n\nSamsung devices running Android 8+ are not supported (https://github.com/leolin310148/ShortcutBadger/pull/268[GitHub Issue]).\n\n[cols=\"1,2a\"]\n!===\n!Since Version !Status\n\n!All Versions !Can't Fix\n!===\n" }, { "from": "gpt", "value": "= Badging Your App\n\nYou can set the badge count on your app's icon using:\n[source,csharp]\n----\nbool SetBadgeCount(int count)\n----\n\n.Parameters\n[cols=\"1,2a\"]\n|===\n|Parameter |Description\n|count |The number to assign as a badge to the application's icon\n|===\n\nThis function will return `true` if it was able to set the badge count for your\napp's icon, or `false` otherwise.\n\nOn Android, this functionality uses the https://github.com/leolin310148/ShortcutBadger[ShortcutBadger library by Leo Sin].\n\n== Known Issues\n\n[cols=\"1,2a\"]\n|===\n|Date |Issue\n\n|October 11, 2018 |Google Nexus and Pixel devices are not supported, because the\nlauncher used does not support badging.\n\nSamsung devices running Android 8+ are not supported (https://github.com/leolin310148/ShortcutBadger/pull/268[GitHub Issue]).\n\n[cols=\"1,2a\"]\n!===\n!Since Version !Status\n\n!All Versions !Can't Fix\n!===\n|===\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "OSSMDOC-363: Add admonition to overview of Installing the Operators topic\n" }, { "from": "human", "value": ":_content-type: ASSEMBLY\n[id=\"installing-ossm\"]\n= Installing the Operators\ninclude::modules/ossm-document-attributes.adoc[]\n:context: installing-ossm\n\ntoc::[]\n\nTo install {ProductName}, first install the required Operators on {product-title} and then create a `ServiceMeshControlPlane` resource to deploy the control plane.\n\n.Prerequisites\n* Read the xref:../../service_mesh/v2x/preparing-ossm-installation.adoc#preparing-ossm-installation[Preparing to install {ProductName}] process.\n* An account with the `cluster-admin` role. If you use {product-dedicated}, you must have an account with the `dedicated-admin` role.\n\nThe following steps show how to install a basic instance of {ProductName} on {product-title}.\n\ninclude::modules/ossm-installation-activities.adoc[leveloffset=+1]\n\n[WARNING]\n====\nDo not install Community versions of the Operators. Community Operators are not supported.\n====\n\ninclude::modules/ossm-install-ossm-operator.adoc[leveloffset=+1]\n\n== Next steps\n\nCreate a `ServiceMeshControlPlane` resource to configure the components of {ProductShortName}. For more information, see xref:../../service_mesh/v2x/ossm-create-smcp.adoc#ossm-create-smcp[Creating the ServiceMeshControlPlane].\n" }, { "from": "gpt", "value": ":_content-type: ASSEMBLY\n[id=\"installing-ossm\"]\n= Installing the Operators\ninclude::modules/ossm-document-attributes.adoc[]\n:context: installing-ossm\n\ntoc::[]\n\nTo install {ProductName}, first install the required Operators on {product-title} and then create a `ServiceMeshControlPlane` resource to deploy the control plane.\n\n[NOTE]\n====\nThis basic installation is configured based on the default OpenShift settings and is not designed for production use.  Use this default installation to verify your installation, and then configure your service mesh for your specific environment.\n====\n\n.Prerequisites\n* Read the xref:../../service_mesh/v2x/preparing-ossm-installation.adoc#preparing-ossm-installation[Preparing to install {ProductName}] process.\n* An account with the `cluster-admin` role. If you use {product-dedicated}, you must have an account with the `dedicated-admin` role.\n\nThe following steps show how to install a basic instance of {ProductName} on {product-title}.\n\ninclude::modules/ossm-installation-activities.adoc[leveloffset=+1]\n\n[WARNING]\n====\nDo not install Community versions of the Operators. Community Operators are not supported.\n====\n\ninclude::modules/ossm-install-ossm-operator.adoc[leveloffset=+1]\n\n== Next steps\n\nCreate a `ServiceMeshControlPlane` resource to configure the components of {ProductShortName}. For more information, see xref:../../service_mesh/v2x/ossm-create-smcp.adoc#ossm-create-smcp[Creating the ServiceMeshControlPlane].\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Docs: Add query to Query DSL documentation\n\nWe already introduced the MatchNoneQueryBuilder query that does not\nreturn any documents, mainly because we needed it for internal\nrepresentation of the NONE option in the IndicesQueryBuilder.\nHowever, the query was requested at least once also for the query dsl,\nand since we can parser it already we should document it as\n`match_none` query in the relevant reference docs as well.\n" }, { "from": "human", "value": "[[query-dsl-match-all-query]]\n== Match All Query\n\nThe most simple query, which matches all documents, giving them all a `_score`\nof `1.0`.\n\n[source,js]\n--------------------------------------------------\n{ \"match_all\": {} }\n--------------------------------------------------\n\nThe `_score` can be changed with the `boost` parameter:\n\n[source,js]\n--------------------------------------------------\n{ \"match_all\": { \"boost\" : 1.2 }}\n--------------------------------------------------\n" }, { "from": "gpt", "value": "[[query-dsl-match-all-query]]\n== Match All Query\n\nThe most simple query, which matches all documents, giving them all a `_score`\nof `1.0`.\n\n[source,js]\n--------------------------------------------------\n{ \"match_all\": {} }\n--------------------------------------------------\n\nThe `_score` can be changed with the `boost` parameter:\n\n[source,js]\n--------------------------------------------------\n{ \"match_all\": { \"boost\" : 1.2 }}\n--------------------------------------------------\n\n[[query-dsl-match-none-query]]\n[float]\n== Match None Query\n\nThis is the inverse of the `match_all` query, which matches no documents.\n\n[source,js]\n--------------------------------------------------\n{ \"match_none\": {} }\n--------------------------------------------------\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Comment out dead links in docs\n" }, { "from": "human", "value": "= Edge Documentation\n\nEdge is a starting point for creating Clojure projects.\nNot sure if Edge is for you? See <<why-edge.adoc#,Why Edge?>>.\n\n== First Steps\n\nAre you new to Edge?\nThis is the place to start!\n\nInstall the tools::\n. link:https://clojure.org/guides/getting_started[Install clj] (<<windows.adoc#,Instructions for Windows users>>)\n. <<editor.adoc#,Set up your editor for Clojure>>\n\nDownload Edge::\n. <<adaptation.adoc#,Adapting Edge for new projects>>\n\n// Paste Tutorial::\n// . Creating a new project\n// . Create paste\n// . Deploy to Elastic Beanstalk\n\nUseful References::\n. <<dev-guide.adoc#,Dev guide>>\n\n== Topical Guides\n\n. Configuration\n. Components\n. <<dev-guide.adoc#,Dev guide>>\n\n== Release\n\nReleasing your application is an important part of the lifecycle of your project. \nEdge provides multiple mechanisms for making releasing easier.\n\n// * Packed and unpacked deployments\n* <<uberjar.adoc#,Producing an Uberjar>>\n* <<elastic-beanstalk.adoc#,Using the Elastic Beanstalk Quickstart>>\n* <<socket-repl.adoc#,Setting up a socket REPL>>\n\n== The Edge Project\n\n* <<why-edge.adoc#,Why Edge?>>\n* <<guidelines.adoc#,Contributing Guidelines>>\n* Getting help\n* How to get involved\n* License\n" }, { "from": "gpt", "value": "= Edge Documentation\n\nEdge is a starting point for creating Clojure projects.\nNot sure if Edge is for you? See <<why-edge.adoc#,Why Edge?>>.\n\n== First Steps\n\nAre you new to Edge?\nThis is the place to start!\n\nInstall the tools::\n. link:https://clojure.org/guides/getting_started[Install clj] (<<windows.adoc#,Instructions for Windows users>>)\n. <<editor.adoc#,Set up your editor for Clojure>>\n\nDownload Edge::\n. <<adaptation.adoc#,Adapting Edge for new projects>>\n\n// Paste Tutorial::\n// . Creating a new project\n// . Create paste\n// . Deploy to Elastic Beanstalk\n\nUseful References::\n. <<dev-guide.adoc#,Dev guide>>\n\n== Topical Guides\n\n//. Configuration\n//. Components\n. <<dev-guide.adoc#,Dev guide>>\n\n== Release\n\nReleasing your application is an important part of the lifecycle of your project. \nEdge provides multiple mechanisms for making releasing easier.\n\n// * Packed and unpacked deployments\n* <<uberjar.adoc#,Producing an Uberjar>>\n* <<elastic-beanstalk.adoc#,Using the Elastic Beanstalk Quickstart>>\n* <<socket-repl.adoc#,Setting up a socket REPL>>\n\n== The Edge Project\n\n* <<why-edge.adoc#,Why Edge?>>\n* <<guidelines.adoc#,Contributing Guidelines>>\n//* Getting help\n//* How to get involved\n//* License\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Add shield link to Stack Overflow\n" }, { "from": "human", "value": "= JBoss Fuse\n:toc:\n:icons: font\n\nRed Hat JBoss Fuse is a lightweight, flexible integration platform that enables rapid integration\nacross the extended enterprise—on-premise or in the cloud.\n\nBased on Apache Camel, Fuse leverages pattern-based integration, a rich connector catalog, and extensive\ndata transformation capabilities to enables users to integrate anything.\n\n== Development Process\n\nThere are two ways of building JBoss Fuse 7:\n\n* using Jenkins pipeline\n* using plain old Maven\n\n=== Building using Jenkins Pipeline\n\nTO DESCRIBE\n\n=== Building locally using Maven\n\nBasic invocation of `mvn clean install` will build Fuse minimal and full distro. Minimal distribution is\nmaintained only for presentation and test purposes.\n\nWe can enable additional profiles:\n\n* `mvn -Pcs validate` - Run checkstyle checks\n* `mvn -Plicense process-sources` - regenerate license headers in configured files using `license-header.txt`\n* `mvn -Ptest verify` - run integration test inside `fuse-itests` module\n\n(In order to generate license headers and/or perform checkstyle checks, it's recommended to use `-Plicense,test` and\n`-Pcs,test` respectively).\n" }, { "from": "gpt", "value": "image:https://img.shields.io/badge/stack%20overflow-jbossfuse-red.svg?colorB=cc0000[link=\"https://stackoverflow.com/questions/tagged/jbossfuse\"]\n\n= JBoss Fuse\n:toc:\n:icons: font\n\nRed Hat JBoss Fuse is a lightweight, flexible integration platform that enables rapid integration\nacross the extended enterprise—on-premise or in the cloud.\n\nBased on Apache Camel, Fuse leverages pattern-based integration, a rich connector catalog, and extensive\ndata transformation capabilities to enables users to integrate anything.\n\n== Development Process\n\nThere are two ways of building JBoss Fuse 7:\n\n* using Jenkins pipeline\n* using plain old Maven\n\n=== Building using Jenkins Pipeline\n\nTO DESCRIBE\n\n=== Building locally using Maven\n\nBasic invocation of `mvn clean install` will build Fuse minimal and full distro. Minimal distribution is\nmaintained only for presentation and test purposes.\n\nWe can enable additional profiles:\n\n* `mvn -Pcs validate` - Run checkstyle checks\n* `mvn -Plicense process-sources` - regenerate license headers in configured files using `license-header.txt`\n* `mvn -Ptest verify` - run integration test inside `fuse-itests` module\n\n(In order to generate license headers and/or perform checkstyle checks, it's recommended to use `-Plicense,test` and\n`-Pcs,test` respectively).\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Remove echo from mv script\n" }, { "from": "human", "value": "Quick-n-dirty PHP router\n========================\n\n[quote]\nA *transitionary* router to help upgrading an old PHP codebase to a more modern\nset of standards.\n\nMore information is available\nhttp://ciarand.me/posts/refactoring-php-part-1/[on my blog].\n\nUsage\n-----\n\nRequire it in your project:\n\n[source, bash]\ncomposer require 'ciarand\\quick-n-dirty-router' 'dev-master'\n\nMove all your files into a handlers dir:\n\n[source, bash]\n----\n# assuming you're in the webroot\nmkdir ../handlers/\nfind . -name '*.php' | while read file; do\n echo mv \"$file\" \"../handlers/$file\";\ndone;\n----\n\nPut this in an `index.php` file in your web root:\n\n[source, php]\n----\n<?php // index.php\n\nrequire \"vendor/autoload.php\";\n\nuse RestlessCo\\Router;\n\n// where you keep your scripts (outside of the public web dir)\n$handler_dir = __DIR__ \"/../handler\";\n\n// the script you'd like to handle 404s with\n$missing_script = $handler_dir . \"/404.php\";\n\n$requested_path = $handler_dir . \"/\" . Router::scriptForEnv($_SERVER);\n\nreturn require (file_exists($requested_path))\n ? $requested_path\n : $missing_script;\n----\n\nTests\n-----\n\n[source, bash]\nvendor/bin/phpunit\n\nLicense\n-------\nMIT, full text is in the link:LICENSE[LICENSE file].\n" }, { "from": "gpt", "value": "Quick-n-dirty PHP router\n========================\n\n[quote]\nA *transitionary* router to help upgrading an old PHP codebase to a more modern\nset of standards.\n\nMore information is available\nhttp://ciarand.me/posts/refactoring-php-part-1/[on my blog].\n\nUsage\n-----\n\nRequire it in your project:\n\n[source, bash]\ncomposer require 'ciarand\\quick-n-dirty-router' 'dev-master'\n\nMove all your files into a handlers dir:\n\n[source, bash]\n----\n# assuming you're in the webroot\nmkdir ../handlers/\nfind . -name '*.php' | while read file; do\n mv \"$file\" \"../handlers/$file\";\ndone;\n----\n\nPut this in an `index.php` file in your web root:\n\n[source, php]\n----\n<?php // index.php\n\nrequire \"vendor/autoload.php\";\n\nuse RestlessCo\\Router;\n\n// where you keep your scripts (outside of the public web dir)\n$handler_dir = __DIR__ \"/../handler\";\n\n// the script you'd like to handle 404s with\n$missing_script = $handler_dir . \"/404.php\";\n\n$requested_path = $handler_dir . \"/\" . Router::scriptForEnv($_SERVER);\n\nreturn require (file_exists($requested_path))\n ? $requested_path\n : $missing_script;\n----\n\nTests\n-----\n\n[source, bash]\nvendor/bin/phpunit\n\nLicense\n-------\nMIT, full text is in the link:LICENSE[LICENSE file].\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Fix relative link for AsciiDoc\n" }, { "from": "human", "value": "= Parity\n\nifdef::env-github,env-browser[:outfilesuffix: .adoc]\nifndef::rootdir[:rootdir: ../]\n:imagesdir: {rootdir}/images\n\nhttps://ethcore.io/parity.html[Parity] is an Ethereum client\nactively developed as an open source project on https://github.com/ethcore/parity[github].\n\nThere are a default configuration (link:parity/default.toml[parity/default.toml])\nused by default by Vagrant.\n\n----\nvagrant up\n----\n\nHowever, you can free to use your own alternative variants, for example:\n\n----\nenv CONFIG=with_dapps CHAIN=example1 vagrant up\n----\n\nor even:\n\n----\nenv CONFIG=with_dapps CHAIN=example1 vagrant provision\n----\n\nIn these examples are used simple name conventions.\n'`with_dapps`' implies the existence of configuration file at 'parity/with_dapps.toml'.\nFor '`with_dapps`' it corresponds to 'parity/chain/example1.json' respectively.\n" }, { "from": "gpt", "value": "= Parity\n\nifdef::env-github,env-browser[:outfilesuffix: .adoc]\nifndef::rootdir[:rootdir: ../]\n:imagesdir: {rootdir}/images\n\nhttps://ethcore.io/parity.html[Parity] is an Ethereum client\nactively developed as an open source project on https://github.com/ethcore/parity[github].\n\nThere are a default configuration (link:../parity/default.toml[parity/default.toml])\nused by default by Vagrant.\n\n----\nvagrant up\n----\n\nHowever, you are free to use your own alternative variants, for example:\n\n----\nenv CONFIG=with_dapps CHAIN=example1 vagrant up\n----\n\nor even:\n\n----\nenv CONFIG=with_dapps CHAIN=example1 vagrant provision\n----\n\nIn these examples are used simple name conventions.\n'`with_dapps`' implies the existence of configuration file at 'parity/with_dapps.toml'.\nFor '`with_dapps`' it corresponds to 'parity/chain/example1.json' respectively.\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Clean up documentation for basic command line guide for accuracy and aeshetics.\n" }, { "from": "human", "value": "Basic Command Line Guide\n------------------------\n\nThis is a guide for common command line commands.\n\n.sudo\n----\nAllows a user with proper permissions (such as an admin) to execute commands. You will need to enter a password for the command to run.\n\n$ sudo <command>\n\ne.g.\n\n$ make install\n\n(usually will output an error, so use sudo)\n\n$ sudo make install\n\n(you will then be prompted to input the admins password)\n----\n\n.cd\n----\nUsed to navigate through directories\n\n$ cd <directoryname>\n\n$ cd .. (navigates to previous directory)\n----\n\n.ls\n----\nUsed to list the contents in the current directory\n\n$ ls \n----\n\n.rm\n----\nUsed to delete files or directories\n\n$ rm <file_name/directory_name>\n----\n\n.cp\n----\nUsed to make copies of files or directories\n\n$ cp <file_name/directory_name>\n----\n\n.mv\n----\nUsed to move or rename files\n\n$ mv <file> <target_directory> (Moves file to target directory)\n\n$ mv <old_file_name> <new_file_name> (Changes name of the file)\n----\n\n.pwd\n----\nUsed to print the name of the working directory. i.e. Show where you are.\n\n$ pwd\n----\n\n\n" }, { "from": "gpt", "value": "= Basic Command Line Guide\n\nThis is a guide for common command line commands.\n\n\n. `cd` is used to navigate through directories.\nThe usage is `cd <directory>`.\nYou can use `..` to go up a level.\n+\n.Example\n----\ncd Documents/\ncd ..\n----\n\n. `ls` is used to list the contents in the current directory.\nThis command does not normally need any arguments.\n\n. `rm` is used to delete files or directories.\nThe usage is `rm <file>`.\nTo delete directories, add the recursive flag `-r`.\n+\n.Example\n----\nrm hello.txt\nrm -r Documents/\n----\n\n. `cp` is used to make copies of files or directories.\nThe usage is `cp <file> <destination>`.\nTo copy directories, add the recursive flag `-r`.\n+\n.Copying a File Example\n----\ncp hello.txt Documents/hello.txt\n----\n+\n.Copying a Directory Example\n----\ncp -r Documents/ Downloads/\n----\n\n. `mv` is used to move or rename files\nThe usage is `mv <file> <destination>`.\nUnlike the other commands, no extra flag is needed to move directories.\n+\n.Moving a File Example\n----\nmv hello.txt Documents/\n----\n+\n.Renaming a File Example\n----\nmv hello.txt world.txt\n----\n\n. `pwd` is used to print the name of the working directory, in other words, show where you are.\nThis command does not normally need any arguments.\n\n\n. `sudo` can precede a command to act with administrative rights.\nIf you run a command with a permission error, a common fix is to run the same command but with sudo.\n+\n.Example\n----\nsudo mv /opt/soarr /opt/soar\n----\n+\nBut be warned, because sudo elevates privileges, it can be a security hazard!\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Update docs in the documentation project\n" }, { "from": "human", "value": "---\nlayout: documentation\n---\n:title: Muon Java\n:date: 2015-01-05 11:40\n:comments: true\n:sharing: true\n:source-highlighter: pygments\n:toc: right\n:toc-title: Muon Java\n:toclevels: 6\n\n# Muon Java\n\nMuon Java is the implementation of the Muon microservices toolkit, for the Java platform. Muon enables you to easily build\nmicroservices in many languages that have richer, more performant and fully reactive communication semantics and gain access\nto Muon compatible services written in other languages without having to sacrifice the rich semantics or performance.\n\n\n.Contributing\n****\nMuon and the ecosystem services are open source.\nContribute to Muon Java and this documentation at http://github.com/muoncore/muon-java\n****\n\ninclude::java/SNAPSHOT/core.adoc[leveloffset=+1]\ninclude::java/SNAPSHOT/eventsystem.adoc[leveloffset=+1]\ninclude::java/SNAPSHOT/spring.adoc[leveloffset=+1]\ninclude::java/SNAPSHOT/extension.adoc[leveloffset=+1]\n" }, { "from": "gpt", "value": "---\nlayout: documentation\n---\n:title: Muon Java\n:date: 2015-01-05 11:40\n:comments: true\n:sharing: true\n:source-highlighter: pygments\n:toc: right\n:toc-title: Muon Java\n:toclevels: 6\n\n:includedir: .\nifdef::env-doc[]\n:includedir: submodules/java/doc\nendif::[]\n\n## Muon Java\n\nMuon Java is the implementation of the Muon microservices toolkit, for the Java platform. Muon enables you to easily build\nmicroservices in many languages that have richer, more performant and fully reactive communication semantics and gain access\nto Muon compatible services written in other languages without having to sacrifice the rich semantics or performance.\n\n.Contributing\n****\nMuon and the ecosystem services are open source.\nContribute to Muon Java and this documentation at http://github.com/muoncore/muon-java\n****\n\ninclude::{includedir}/core.adoc[leveloffset=+1]\ninclude::{includedir}/eventsystem.adoc[leveloffset=+1]\ninclude::{includedir}/spring.adoc[leveloffset=+1]\ninclude::{includedir}/extension.adoc[leveloffset=+1]\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Move Travis CI badge again\n" }, { "from": "human", "value": "# Janitor - Keeping your Pocket clean since 2016\nimage:https://travis-ci.org/techdev-solutions/janitor.svg?branch=master[\"Build Status\",link=\"https://travis-ci.org/techdev-solutions/janitor\"]\n\nAn application to perform cleanup work using the https://getpocket.com[Pocket API].\n\n## API Documentation\nThe documentation for the Kotlin API bindings can be found https://techdev-solutions.github.io/janitor/pocket-api/[here]." }, { "from": "gpt", "value": "# Janitor image:https://travis-ci.org/techdev-solutions/janitor.svg?branch=master[\"Build Status\",link=\"https://travis-ci.org/techdev-solutions/janitor\"]\n\nAn application to perform cleanup work using the https://getpocket.com[Pocket API].\n\n## API Documentation\nThe documentation for the Kotlin API bindings can be found https://techdev-solutions.github.io/janitor/pocket-api/[here]." } ]
asciidoc
chosen
[ { "from": "system", "value": "Remove links to nonexistent migration docs\n" }, { "from": "human", "value": "[[breaking-changes]]\n= Breaking changes\n\n[partintro]\n--\nThis section discusses the changes that you need to be aware of when migrating\nyour application from one version of Elasticsearch to another.\n\nAs a general rule:\n\n* Migration between major versions -- e.g. `1.x` to `2.x` --\n requires a <<restart-upgrade,full cluster restart>>.\n\n* Migration between minor versions -- e.g. `1.x` to `1.y` -- can be\n performed by <<rolling-upgrades,upgrading one node at a time>>.\n\nSee <<setup-upgrade>> for more info.\n--\ninclude::migrate_5_0.asciidoc[]\n\ninclude::migrate_2_3.asciidoc[]\n\ninclude::migrate_2_2.asciidoc[]\n\ninclude::migrate_2_1.asciidoc[]\n\ninclude::migrate_2_0.asciidoc[]\n\ninclude::migrate_1_6.asciidoc[]\n\ninclude::migrate_1_4.asciidoc[]\n\ninclude::migrate_1_0.asciidoc[]\n" }, { "from": "gpt", "value": "[[breaking-changes]]\n= Breaking changes\n\n[partintro]\n--\nThis section discusses the changes that you need to be aware of when migrating\nyour application from one version of Elasticsearch to another.\n\nAs a general rule:\n\n* Migration between major versions -- e.g. `1.x` to `2.x` --\n requires a <<restart-upgrade,full cluster restart>>.\n\n* Migration between minor versions -- e.g. `1.x` to `1.y` -- can be\n performed by <<rolling-upgrades,upgrading one node at a time>>.\n\nSee <<setup-upgrade>> for more info.\n--\ninclude::migrate_5_0.asciidoc[]\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "[DOCS] Update to the new {index}/_mapping/{type} format" }, { "from": "human", "value": "[[indices-get-mapping]]\n== Get Mapping\n\nThe get mapping API allows to retrieve mapping definitions for an index or\nindex/type.\n\n[source,js]\n--------------------------------------------------\ncurl -XGET 'http://localhost:9200/twitter/_mapping/tweet'\n--------------------------------------------------\n\n[float]\n=== Multiple Indices and Types\n\nThe get mapping API can be used to get more than one index or type\nmapping with a single call. General usage of the API follows the\nfollowing syntax: `host:port/{index}/{type}/_mapping` where both\n`{index}` and `{type}` can accept a comma-separated list of names. To\nget mappings for all indices you can use `_all` for `{index}`. The\nfollowing are some examples:\n\n[source,js]\n--------------------------------------------------\ncurl -XGET 'http://localhost:9200/_mapping/twitter,kimchy'\n\ncurl -XGET 'http://localhost:9200/_all/_mapping/tweet,book'\n--------------------------------------------------\n\nIf you want to get mappings of all indices and types then the following\ntwo examples are equivalent:\n\n[source,js]\n--------------------------------------------------\ncurl -XGET 'http://localhost:9200/_all/_mapping'\n\ncurl -XGET 'http://localhost:9200/_mapping'\n--------------------------------------------------\n" }, { "from": "gpt", "value": "[[indices-get-mapping]]\n== Get Mapping\n\nThe get mapping API allows to retrieve mapping definitions for an index or\nindex/type.\n\n[source,js]\n--------------------------------------------------\ncurl -XGET 'http://localhost:9200/twitter/_mapping/tweet'\n--------------------------------------------------\n\n[float]\n=== Multiple Indices and Types\n\nThe get mapping API can be used to get more than one index or type\nmapping with a single call. General usage of the API follows the\nfollowing syntax: `host:port/{index}/_mapping/{type}` where both\n`{index}` and `{type}` can accept a comma-separated list of names. To\nget mappings for all indices you can use `_all` for `{index}`. The\nfollowing are some examples:\n\n[source,js]\n--------------------------------------------------\ncurl -XGET 'http://localhost:9200/_mapping/twitter,kimchy'\n\ncurl -XGET 'http://localhost:9200/_all/_mapping/tweet,book'\n--------------------------------------------------\n\nIf you want to get mappings of all indices and types then the following\ntwo examples are equivalent:\n\n[source,js]\n--------------------------------------------------\ncurl -XGET 'http://localhost:9200/_all/_mapping'\n\ncurl -XGET 'http://localhost:9200/_mapping'\n--------------------------------------------------\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Update the VR page to something more recent" }, { "from": "human", "value": "= virtualreality\n:author: \n:revnumber: \n:revdate: 2016/03/17 20:48\n:relfileprefix: ../\n:imagesdir: ..\nifdef::env-github,env-browser[:outfilesuffix: .adoc]\n\n\n\n== OpenCV\n\n\n== JavaCV\n\n\n== Hand gesture links:\n\nlink:http://www.movesinstitute.org/~kolsch/HandVu/HandVu.html[http://www.movesinstitute.org/~kolsch/HandVu/HandVu.html]\n" }, { "from": "gpt", "value": "= virtualreality\n:author: \n:revnumber: \n:revdate: 2016/10/23 11:22\n:relfileprefix: ../\n:imagesdir: ..\nifdef::env-github,env-browser[:outfilesuffix: .adoc]\n\njMonkeyEngine 3 has a wide range of support for Virtual Reality (VR). The known supported systems are:\nHTC Vive and systems supporting SteamVR/OpenVR\nRazer HDK and systems supporting OSVR\nGoogle Cardboard / GoogleVR\nLegacy: Oculus Rift 0.5\n\nIn any form of VR, a library is needed as support is not built in to the core of jMonkeyEngine3.\n\n== jMonkeyVR (https://github.com/phr00t/jMonkeyVR)\nThis is the main library to use when you want wide VR support. It has full OpenVR/SteamVR support and OSVR support in the making (as of Oct 2016)\n\n== jme3-osvr (https://github.com/neph1/jme3-osvr)\nTogether with https://github.com/neph1/OSVR-Java provides rudimentary OSVR support (extended mode). Examples: https://github.com/neph1/jme3-osvr-examples\n\n== gvr-android-jme (https://github.com/nordfalk/gvr-android-jme)\nGoogle Cardboard VR SDK 1.0 integration\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Add migration note on get alias and get mapping APIs\n\nThese APIs now return 404 if any requested alias or type, respectively,\ndoes not exist. This commit adds a note to the migration docs regarding\nthis change.\n" }, { "from": "human", "value": "[[breaking-changes-5.5]]\n== Breaking changes in 5.5\n\n[[breaking_55_plugins]]\n[float]\n=== Ignoring hidden folders\n\nPrevious versions of Elasticsearch would skip hidden files and directories when\nscanning the plugins folder. This allows broken plugin installs to remain\nundetected so this leniency has been removed.\n\n[[breaking_55_packaging_changes]]\n[float]\n=== Skip setting kernel parameters\n\nElasticsearch attempts to set the kernel parameter `vm.max_map_count` during\ninstallation. In some environments this can fail. Previous versions of\nElasticsearch accepted a flag `ES_SKIP_SET_KERNEL_PARAMTERS` to skip this\nwriting. Starting in Elasticsearch 5.5.0, this flag has been removed and\ninstead the same functionality can be achieved by masking the `systemd-sysctl`\nservice.\n" }, { "from": "gpt", "value": "[[breaking-changes-5.5]]\n== Breaking changes in 5.5\n\n[[breaking_55_plugins]]\n[float]\n=== Ignoring hidden folders\n\nPrevious versions of Elasticsearch would skip hidden files and directories when\nscanning the plugins folder. This allows broken plugin installs to remain\nundetected so this leniency has been removed.\n\n[[breaking_55_packaging_changes]]\n[float]\n=== Skip setting kernel parameters\n\nElasticsearch attempts to set the kernel parameter `vm.max_map_count` during\ninstallation. In some environments this can fail. Previous versions of\nElasticsearch accepted a flag `ES_SKIP_SET_KERNEL_PARAMTERS` to skip this\nwriting. Starting in Elasticsearch 5.5.0, this flag has been removed and\ninstead the same functionality can be achieved by masking the `systemd-sysctl`\nservice.\n\n[[breaking_55_rest_changes]]\n[float]\n=== 404s on get alias and get mapping APIs\n\nPrevious versions of Elasticsearch did not return a 404 in all cases when a\nrequested alias or a requested type did not exist on the get alias or get\nmapping APIs, respectively. Starting in Elasticsearch 5.5.0, these APIs\nconsistently return a 404 in any case when an alias or type is requested that\ndoes not exist.\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Add migration note on private temporary directory\n\nThis commit adds a note to the migration docs on the packaging change to\r\nalways use a private temporary directory.\r\n\r\nRelates #27624\r\n" }, { "from": "human", "value": "[[breaking-changes-6.2]]\n== Breaking changes in 6.2\n\n[[breaking_62_packaging]]\n[float]\n=== All permission bootstrap check\n\nElasticsearch installs a security manager during bootstrap to mitigate the scope\nof exploits in the JDK, in third-party dependencies, and in Elasticsearch itself\nas well as to sandbox untrusted plugins. A custom security policy can be applied\nand one permission that can be added to this policy is\n`java.security.AllPermission`. However, this effectively disables the security\nmanager. As such, granting this permission in production mode is now forbidden\nvia the <<all-permission-check, all permission bootstrap check>>.\n" }, { "from": "gpt", "value": "[[breaking-changes-6.2]]\n== Breaking changes in 6.2\n\n[[breaking_62_packaging]]\n[float]\n=== All permission bootstrap check\n\nElasticsearch installs a security manager during bootstrap to mitigate the scope\nof exploits in the JDK, in third-party dependencies, and in Elasticsearch itself\nas well as to sandbox untrusted plugins. A custom security policy can be applied\nand one permission that can be added to this policy is\n`java.security.AllPermission`. However, this effectively disables the security\nmanager. As such, granting this permission in production mode is now forbidden\nvia the <<all-permission-check, all permission bootstrap check>>.\n\n=== Private temporary directory\n\nOn Linux, previous versions of Elasticsearch defaulted to using `/tmp` as the\ntemporary directory for the process. However, `/tmp` is public so we have\nelected to change the packaging so that we use a private temporary directory. If\nyou are upgrading from a previous version of Elasticsearch and preserve your\nexisting <<jvm-options,`jvm.options`>>, you should add the line\n`-Djava.io.tmpdir=${ES_TMPDIR}`. It is safe to do this on all OS as we preserve\nusing a private temporary directory on non-Linux systems with the same\nmechanism.\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Fix a broken reference to types removal in the Java API docs.\n" }, { "from": "human", "value": "[[java-query-dsl-type-query]]\n==== Type Query\n\ndeprecated[7.0.0, Types are being removed, prefer filtering on a field instead. For more information, please see <<removal-of-types>>.]\n\nSee {ref}/query-dsl-type-query.html[Type Query]\n\n[\"source\",\"java\",subs=\"attributes,callouts,macros\"]\n--------------------------------------------------\ninclude-tagged::{query-dsl-test}[type]\n--------------------------------------------------\n<1> type\n" }, { "from": "gpt", "value": "[[java-query-dsl-type-query]]\n==== Type Query\n\ndeprecated[7.0.0, Types are being removed, prefer filtering on a field instead. For more information, please see {ref}/removal-of-types.html[Removal of mapping types].]\n\nSee {ref}/query-dsl-type-query.html[Type Query]\n\n[\"source\",\"java\",subs=\"attributes,callouts,macros\"]\n--------------------------------------------------\ninclude-tagged::{query-dsl-test}[type]\n--------------------------------------------------\n<1> type\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Add a blurb about SSL setup\n" }, { "from": "human", "value": "= Explorer for Hawkular\n\nhttp://hawkular.org/[Hawkular] is a set of components for Monitoring.\nThis explorer connects to a Hawkular server and allows to browse trough\ninventory and view entities, graph metrics.\n\n.Main screen with availability pop-up\nifndef::env-github[]\nimage::docs/screenshot.png[]\nendif::[]\nifdef::env-github[]\nimage::https://github.com/pilhuhn/hawkfx/blob/master/docs/screenshot.png[]\nendif::[]\n\nOther functions include listing of all metrics on a system, listing of Events/Alerts, injection of Events,\ninjection of metric values and a few more things.\n\n\n.Events list\nifndef::env-github[]\nimage::docs/event_list.png[]\nendif::[]\nifdef::env-github[]\nimage::https://github.com/pilhuhn/hawkfx/blob/master/docs/event_list.png[]\nendif::[]\n\n\n== Running\n\nThe explorer requires JRuby in version 9+ and Java8\n\nIf you are using `rvm` you can select it via\n\n`rvm use jruby-9.1.5.0`\n\ninstall and use `bundler` to install the required gems\n\n`gem install bundler`\n\n`bundle install`\n\nthen run\n\n`jruby -S -G hawkfx.rb`" }, { "from": "gpt", "value": "= Explorer for Hawkular\n\nhttp://hawkular.org/[Hawkular] is a set of components for Monitoring.\nThis explorer connects to a Hawkular server and allows to browse trough\ninventory and view entities, graph metrics.\n\n.Main screen with availability pop-up\nifndef::env-github[]\nimage::docs/screenshot.png[]\nendif::[]\nifdef::env-github[]\nimage::https://github.com/pilhuhn/hawkfx/blob/master/docs/screenshot.png[]\nendif::[]\n\nOther functions include listing of all metrics on a system, listing of Events/Alerts, injection of Events,\ninjection of metric values and a few more things.\n\n\n.Events list\nifndef::env-github[]\nimage::docs/event_list.png[]\nendif::[]\nifdef::env-github[]\nimage::https://github.com/pilhuhn/hawkfx/blob/master/docs/event_list.png[]\nendif::[]\n\n\n== Running\n\nThe explorer requires JRuby in version 9+ and Java8\n\nIf you are using `rvm` you can select it via\n\n`rvm use jruby-9.1.5.0`\n\ninstall and use `bundler` to install the required gems\n\n`gem install bundler`\n\n`bundle install`\n\nthen run\n\n`jruby -S -G hawkfx.rb`\n\n=== Using SSL\n\nThere is a blog post on the Hawkular\nblog http://www.hawkular.org/blog/2016/09/14/consuming-hawkular-api-over-ssl.html[that describes how to use SSL] in Hawkular and HawkFX." } ]
asciidoc
chosen
[ { "from": "system", "value": "Remove JediTerm build instruction (since terminal jar is provided in repository)\n" }, { "from": "human", "value": "= jediterm-forge\n\nProvides https://github.com/forge/core/[JBoss Forge] console component based on \nhttps://github.com/JetBrains/jediterm[JediTerm] for Swing applications, like IntelliJ, NetBeans, \nor standalone shell client.\n\n== Build\n\n* Clone this JediTerm repository fork:\n** `git clone https://github.com/adamwy/jediterm.git`\n** `git checkout 2.0`\n* Build default JediTerm target with http://ant.apache.org/bindownload.cgi[Apache Ant]:\n** `ant`\n* Install JediTerm as Maven artifact:\n** `./install-jediterm-artifact \"jediterm/build/jediterm-terminal-2.0.jar\"`\n* Build and install this project to local Maven repository:\n** `mvn clean install`\n" }, { "from": "gpt", "value": "= jediterm-forge\n\nProvides https://github.com/forge/core/[JBoss Forge] console component based on \nhttps://github.com/JetBrains/jediterm[JediTerm] for Swing applications, like IntelliJ, NetBeans, \nor standalone shell client.\n\n== Build\n\n* Clone this JediTerm repository fork:\n** `git clone https://github.com/adamwy/jediterm.git`\n** `git checkout 2.0`\n* Install JediTerm as Maven artifact:\n** `./install-jediterm-artifact \"jediterm/build/jediterm-terminal-2.0.jar\"`\n* Build and install this project to local Maven repository:\n** `mvn clean install`\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Add note on Kibana OAuth refresh\n" }, { "from": "human", "value": "// Module included in the following assemblies:\n//\n// * logging/cluster-logging-kibana-interface.adoc\n\n[id=\"cluster-logging-kibana-interface-launch_{context}\"]\n= Launching Kibana\n\nKibana is a browser-based console to query, discover, and visualize your logs through histograms, line graphs,\npie charts, heat maps, built-in geospatial support, and other visualizations.\n\n.Procedure\n\nTo launch Kibana:\n\n. In the {product-title} console, click *Monitoring* -> *Logging*.\n\n. Log in using the same credentials you use to log in to the {product-title} console.\n+\nThe Kibana interface launches. You can now:\n+\n* Search and browse your data using the Discover page.\n* Chart and map your data using the Visualize page.\n* Create and view custom dashboards using the Dashboard page.\n+\nUse and configuration of the Kibana interface is beyond the scope of this documentation. For more information,\non using the interface, see the link:https://www.elastic.co/guide/en/kibana/5.6/connect-to-elasticsearch.html[Kibana documentation].\n" }, { "from": "gpt", "value": "// Module included in the following assemblies:\n//\n// * logging/cluster-logging-kibana-interface.adoc\n\n[id=\"cluster-logging-kibana-interface-launch_{context}\"]\n= Launching Kibana\n\nKibana is a browser-based console to query, discover, and visualize your logs through histograms, line graphs,\npie charts, heat maps, built-in geospatial support, and other visualizations.\n\n.Procedure\n\nTo launch Kibana:\n\n. In the {product-title} console, click *Monitoring* -> *Logging*.\n\n. Log in using the same credentials you use to log in to the {product-title} console.\n+\nThe Kibana interface launches. You can now:\n+\n* Search and browse your data using the Discover page.\n* Chart and map your data using the Visualize page.\n* Create and view custom dashboards using the Dashboard page.\n+\nUse and configuration of the Kibana interface is beyond the scope of this documentation. For more information,\non using the interface, see the link:https://www.elastic.co/guide/en/kibana/5.6/connect-to-elasticsearch.html[Kibana documentation].\n\n[NOTE]\n====\nIf you get a *security_exception* error in the Kibana console and cannot access your Kibana indices, you might have an expired OAuth token. If you see this error, log out of the Kibana console, and then log back in. This refreshes your OAuth tokens and you should be able to access your indices.\n====\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Add note about potential slow hosted app startup\n" }, { "from": "human", "value": "= Spring Boot and Two DataSources\n\nThis project demonstrates how to use two `DataSource` s with Spring Boot. \n\nIt utilizes: \n\n* Spring Data https://github.com/spring-projects/spring-data-jpa[JPA] / https://github.com/spring-projects/spring-data-rest[REST]\n* https://github.com/flyway/flyway[Flyway] migrations for the two `DataSource` s\n* Separate Hibernate properties for each `DataSource` defined in the application.yml\n* https://github.com/thymeleaf/thymeleaf[Thymeleaf] 3\n* https://github.com/DataTables/DataTablesSrc[DataTables]\n* Unit tests for components" }, { "from": "gpt", "value": "= Spring Boot and Two DataSources\n\nThis project demonstrates how to use two `DataSource` s with Spring Boot. \n\nIt utilizes: \n\n* Spring Data https://github.com/spring-projects/spring-data-jpa[JPA] / https://github.com/spring-projects/spring-data-rest[REST]\n* https://github.com/flyway/flyway[Flyway] migrations for the two `DataSource` s\n* Separate Hibernate properties for each `DataSource` defined in the application.yml\n* https://github.com/thymeleaf/thymeleaf[Thymeleaf] 3\n* https://github.com/DataTables/DataTablesSrc[DataTables]\n* Unit tests for components\n\n Note: It may take a few seconds for the app to start if no one has not accessed it recently" } ]
asciidoc
chosen
[ { "from": "system", "value": "Add links for example output files\n" }, { "from": "human", "value": "== asciidoctor-fopub-embed-svg-example\n\nAn example for embedding SVG images in an asciidoctor document.\n\n=== Setup\n\n----\nbundle install\n----\n\n=== How to make the target pdf file\n\nRunning the command below make the target pdf file named \"class_diagram.pdf\" out of \"class_diagram.adoc\"\n\n----\nbundle exec rake\n----\n\n=== Watch *.adoc files\n\nRunning the command below, rake watches \"\\*.adoc\" files and build \"*.adoc\" files whenever they are modified.\n\n----\nbundle exec rake watch\n----\n" }, { "from": "gpt", "value": "== asciidoctor-fopub-embed-svg-example\n\nAn example for embedding SVG images in an asciidoctor document.\n\n=== Setup\n\n----\nbundle install\n----\n\n=== How to make the target pdf file\n\nRunning the command below make the target pdf file named \"class_diagram.pdf\" out of \"class_diagram.adoc\"\n\n----\nbundle exec rake\n----\n\n=== Watch *.adoc files\n\nRunning the command below, rake watches \"\\*.adoc\" files and build \"*.adoc\" files whenever they are modified.\n\n----\nbundle exec rake watch\n----\n\n=== Example output files\n\nYou can see the example output files at\n\n* https://hnakamur.github.io/asciidoctor-fopub-embed-svg-example/class_diagram.pdf[class_diagram.pdf]\n* https://hnakamur.github.io/asciidoctor-fopub-embed-svg-example/class_diagram.html[class_diagram.html]\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Add CONSOLE to from/size docs\n\nRelates to #18160\n" }, { "from": "human", "value": "[[search-request-from-size]]\n=== From / Size\n\nPagination of results can be done by using the `from` and `size`\nparameters. The `from` parameter defines the offset from the first\nresult you want to fetch. The `size` parameter allows you to configure\nthe maximum amount of hits to be returned.\n\nThough `from` and `size` can be set as request parameters, they can also\nbe set within the search body. `from` defaults to `0`, and `size`\ndefaults to `10`.\n\n[source,js]\n--------------------------------------------------\n{\n \"from\" : 0, \"size\" : 10,\n \"query\" : {\n \"term\" : { \"user\" : \"kimchy\" }\n }\n}\n--------------------------------------------------\n\nNote that `from` + `size` can not be more than the `index.max_result_window`\nindex setting which defaults to 10,000. See the <<search-request-scroll,Scroll>> or <<search-request-search-after,Search After>>\nAPI for more efficient ways to do deep scrolling.\n" }, { "from": "gpt", "value": "[[search-request-from-size]]\n=== From / Size\n\nPagination of results can be done by using the `from` and `size`\nparameters. The `from` parameter defines the offset from the first\nresult you want to fetch. The `size` parameter allows you to configure\nthe maximum amount of hits to be returned.\n\nThough `from` and `size` can be set as request parameters, they can also\nbe set within the search body. `from` defaults to `0`, and `size`\ndefaults to `10`.\n\n[source,js]\n--------------------------------------------------\nGET /_search\n{\n \"from\" : 0, \"size\" : 10,\n \"query\" : {\n \"term\" : { \"user\" : \"kimchy\" }\n }\n}\n--------------------------------------------------\n// CONSOLE\n\n\nNote that `from` + `size` can not be more than the `index.max_result_window`\nindex setting which defaults to 10,000. See the <<search-request-scroll,Scroll>> or <<search-request-search-after,Search After>>\nAPI for more efficient ways to do deep scrolling.\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Add Patreon badge to readme\n" }, { "from": "human", "value": "= griffon-configuration-plugin\n:linkattrs:\n:project-name: griffon-configuration-plugin\n\nimage:http://img.shields.io/travis/griffon-plugins/{project-name}/master.svg[\"Build Status\", link=\"https://travis-ci.org/griffon-plugins/{project-name}\"]\nimage:http://img.shields.io/coveralls/griffon-plugins/{project-name}/master.svg[\"Coverage Status\", link=\"https://coveralls.io/r/griffon-plugins/{project-name}\"]\nimage:http://img.shields.io/badge/license-ASF2-blue.svg[\"Apache License 2\", link=\"http://www.apache.org/licenses/LICENSE-2.0.txt\"]\nimage:https://api.bintray.com/packages/griffon/griffon-plugins/{project-name}/images/download.svg[link=\"https://bintray.com/griffon/griffon-plugins/{project-name}/_latestVersion\"]\n\n---\n\nProvides additional configuration formats such as YAML and JSON.\n\nRefer to the link:http://griffon-plugins.github.io/griffon-configuration-plugin/[plugin guide, window=\"_blank\"] for\nfurther information on configuration and usage.\n" }, { "from": "gpt", "value": "= griffon-configuration-plugin\n:linkattrs:\n:project-name: griffon-configuration-plugin\n\nimage:http://img.shields.io/travis/griffon-plugins/{project-name}/master.svg[\"Build Status\", link=\"https://travis-ci.org/griffon-plugins/{project-name}\"]\nimage:http://img.shields.io/coveralls/griffon-plugins/{project-name}/master.svg[\"Coverage Status\", link=\"https://coveralls.io/r/griffon-plugins/{project-name}\"]\nimage:http://img.shields.io/badge/license-ASF2-blue.svg[\"Apache License 2\", link=\"http://www.apache.org/licenses/LICENSE-2.0.txt\"]\nimage:https://api.bintray.com/packages/griffon/griffon-plugins/{project-name}/images/download.svg[link=\"https://bintray.com/griffon/griffon-plugins/{project-name}/_latestVersion\"]\n\n---\n\nimage:https://img.shields.io/gitter/room/griffon/griffon.js.svg[link=\"https://gitter.im/griffon/griffon]\nimage:https://img.shields.io/badge/donations-Patreon-orange.svg[https://www.patreon.com/user?u=6609318]\n\n---\n\nProvides additional configuration formats such as YAML and JSON.\n\nRefer to the link:http://griffon-plugins.github.io/griffon-configuration-plugin/[plugin guide, window=\"_blank\"] for\nfurther information on configuration and usage.\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Upgrade Travis build status badge" }, { "from": "human", "value": "= Opal Runtime for Node.js\n\nifdef::env-github[]\nimage:https://img.shields.io/travis/mogztter/opal-runtime/master.svg[Travis build status, link=https://travis-ci.org/mogztter/opal-runtime]\nimage:http://img.shields.io/npm/v/opal-runtime.svg[npm version, link=https://www.npmjs.org/package/opal-runtime]\nendif::[]\n\n== Usage\n\n```javascript\nvar Opal = require('opal-runtime').Opal;\n// Now let's have fun with Opal!\n```\n" }, { "from": "gpt", "value": "= Opal Runtime for Node.js\n\nifdef::env-github[]\nimage:https://img.shields.io/travis/Mogztter/opal-node-runtime/master.svg[Travis build status, link=https://travis-ci.org/Mogztter/opal-node-runtime]\nimage:http://img.shields.io/npm/v/opal-runtime.svg[npm version, link=https://www.npmjs.org/package/opal-runtime]\nendif::[]\n\n== Usage\n\n```javascript\nvar Opal = require('opal-runtime').Opal;\n// Now let's have fun with Opal!\n```\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Add IMPORTANT section for Release status\n" }, { "from": "human", "value": "= flexy-pool\nAuthor <[email protected]>\nv1.0.0, 2014-02-25\n\n:homepage: http://vladmihalcea.com/\n\n== Introduction\n\nThe flexy-pool library brings adaptability to a given Connection Pool, allowing it to resize on demand.\nThis is very handy since most connection pools offer a limited set of dynamic configuration strategies.\n\n== Features \n\n* extensive connection pool support\n** http://docs.codehaus.org/display/BTM/Home[Bitronix Transaction Manager]\n** http://commons.apache.org/proper/commons-dbcp/[Apache DBCP]\n** http://commons.apache.org/proper/commons-dbcp/[Apache DBCP2]\n** http://www.mchange.com/projects/c3p0/[C3P0]\n** http://jolbox.com/[BoneCP]\n** http://brettwooldridge.github.io/HikariCP/[HikariCP]\n* statistics support\n** source connection acquiring time histogram\n** total connection acquiring time histogram\n** retries attempts histogram\n** maximum CP size histogram\n** connection request count histogram\n** connection lease time histogram\n\n== Documentation \n\n. https://github.com/vladmihalcea/flexy-pool/wiki/Installation-Guide[Installation Guide]\n. https://github.com/vladmihalcea/flexy-pool/wiki/User-Guide[User Guide]\n\nIMPORTANT: Flexy Pool has not yet been released to OSS Maven Repository.\n" }, { "from": "gpt", "value": "= flexy-pool\nAuthor <[email protected]>\nv1.0.0, 2014-02-25\n\n:homepage: http://vladmihalcea.com/\n\n== Introduction\n\nThe flexy-pool library brings adaptability to a given Connection Pool, allowing it to resize on demand.\nThis is very handy since most connection pools offer a limited set of dynamic configuration strategies.\n\n== Features \n\n* extensive connection pool support\n** http://docs.codehaus.org/display/BTM/Home[Bitronix Transaction Manager]\n** http://commons.apache.org/proper/commons-dbcp/[Apache DBCP]\n** http://commons.apache.org/proper/commons-dbcp/[Apache DBCP2]\n** http://www.mchange.com/projects/c3p0/[C3P0]\n** http://jolbox.com/[BoneCP]\n** http://brettwooldridge.github.io/HikariCP/[HikariCP]\n* statistics support\n** source connection acquiring time histogram\n** total connection acquiring time histogram\n** retries attempts histogram\n** maximum CP size histogram\n** connection request count histogram\n** connection lease time histogram\n\n== Documentation \n\n. https://github.com/vladmihalcea/flexy-pool/wiki/Installation-Guide[Installation Guide]\n. https://github.com/vladmihalcea/flexy-pool/wiki/User-Guide[User Guide]\n\n[IMPORTANT]\nFlexy Pool has not yet been released to OSS Maven Repository.\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Add angular2-ui as a dependency to spring-boot-server\n" }, { "from": "human", "value": "= Spring Boot, AngularJS 2, Kotlin demo\n\nThis is a demo which shows how to use Spring Boot, Spring Data JPA and AngularJS 2.\nThe demo is an implementation of Angular's Tour of Heroes tutorial. The demo is written in Kotlin and TypeScript.\n\nThe Heroes are stored in an in-memory H2 database. The H2 schema is managed by Flyway.\n\nYou can start the AngularJS 2 App and access it on `http://localhost:3000`:\n\n`gradlew npmStartServer`\n\nYou can start the Spring Boot App and access it on `http://localhost:8090`:\n\n`gradlew bootRun`\n\nWebpack proxies the request from Angular2 to Spring Boot.\n\nTODOs:\n\n- [x] Implement Angular's Tour of Heroes tutorial\n- Add Angular2 app as a static resource to the Spring Boot app\n- Implement authentication/authorization with Spring Security and JWT Tokens\n\n== Angular's Tour of Heroes\n\nimage:docs/tutorial.JPG[]\n" }, { "from": "gpt", "value": "= Spring Boot, AngularJS 2, Kotlin demo\n\nThis is a demo which shows how to use Spring Boot, Spring Data JPA and AngularJS 2.\nThe demo is an implementation of Angular's Tour of Heroes tutorial. The demo is written in Kotlin and TypeScript.\n\nThe Heroes are stored in an in-memory H2 database. The H2 schema is managed by Flyway.\n\nYou can start the Angular2 App and access it on `http://localhost:3000`:\n\n`gradlew npmStartServer`\n\nYou can start the Spring Boot App and access it on `http://localhost:8090`:\n\n`gradlew bootRun`\n\nWebpack proxies the request from Angular2 to Spring Boot.\n\nYou can package the Spring Boot App and Angular2 App into one Jar file:\n\n`gradlew build`\n\n\nThe Angular2 App is served as static content under `http://localhost:8090/index.html, when you start the application`:\n\n`java -jar spring-boot-server-0.1.0.jar`\n\nTODOs:\n\n- [x] Implement Angular's Tour of Heroes tutorial\n- Add Angular2 app as a static resource to the Spring Boot app\n- Implement authentication/authorization with Spring Security and JWT Tokens\n\n== Angular's Tour of Heroes\n\nimage:docs/tutorial.JPG[]\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Add Flexy Pool Architecture diagram" }, { "from": "human", "value": "= Flexy Pool\nAuthor <[email protected]>\nv1.0.0, 2014-02-25\n\n:homepage: http://vladmihalcea.com/\n\n== Introduction\n\nThe Flexy Pool library brings adaptability to a given Connection Pool, allowing it to resize on demand.\nThis is very handy since most connection pools offer a limited set of dynamic configuration strategies.\n\n== Features \n\n* extensive connection pool support\n** http://docs.codehaus.org/display/BTM/Home[Bitronix Transaction Manager]\n** http://commons.apache.org/proper/commons-dbcp/[Apache DBCP]\n** http://commons.apache.org/proper/commons-dbcp/[Apache DBCP2]\n** http://www.mchange.com/projects/c3p0/[C3P0]\n** http://jolbox.com/[BoneCP]\n** http://brettwooldridge.github.io/HikariCP/[HikariCP]\n* statistics support\n** source connection acquiring time histogram\n** total connection acquiring time histogram\n** retries attempts histogram\n** maximum CP size histogram\n** connection request count histogram\n** connection lease time histogram\n\n== Documentation \n\n. https://github.com/vladmihalcea/flexy-pool/wiki/Installation-Guide[Installation Guide]\n. https://github.com/vladmihalcea/flexy-pool/wiki/User-Guide[User Guide]\n" }, { "from": "gpt", "value": "= Flexy Pool\nAuthor <[email protected]>\nv1.0.0, 2014-02-25\n\n:homepage: http://vladmihalcea.com/\n\n== Introduction\n\nThe Flexy Pool library brings adaptability to a given Connection Pool, allowing it to resize on demand.\nThis is very handy since most connection pools offer a limited set of dynamic configuration strategies.\n\nimage::https://raw.githubusercontent.com/wiki/vladmihalcea/flexy-pool/image/architecture/FlexyPoolArchitecture.gif[Flexy Pool Architecture]\n\n== Features \n\n* extensive connection pool support\n** http://docs.codehaus.org/display/BTM/Home[Bitronix Transaction Manager]\n** http://commons.apache.org/proper/commons-dbcp/[Apache DBCP]\n** http://commons.apache.org/proper/commons-dbcp/[Apache DBCP2]\n** http://www.mchange.com/projects/c3p0/[C3P0]\n** http://jolbox.com/[BoneCP]\n** http://brettwooldridge.github.io/HikariCP/[HikariCP]\n* statistics support\n** source connection acquiring time histogram\n** total connection acquiring time histogram\n** retries attempts histogram\n** maximum CP size histogram\n** connection request count histogram\n** connection lease time histogram\n\n== Documentation \n\n. https://github.com/vladmihalcea/flexy-pool/wiki/Installation-Guide[Installation Guide]\n. https://github.com/vladmihalcea/flexy-pool/wiki/User-Guide[User Guide]\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Fix ISSUE-23216 - add RHEL note\n" }, { "from": "human", "value": "// Module included in the following assemblies:\n//\n// * storage/persistent_storage/persistent-storage-nfs.adoc\n\n[id=\"nfs-selinux_{context}\"]\n= SELinux\n\nBy default, SELinux does not allow writing from a Pod to a remote\nNFS server. The NFS volume mounts correctly, but is read-only.\n\nTo enable writing to a remote NFS server, follow the below procedure.\n\n.Prerequisites\n\n* The `container-selinux` package must be installed. This package provides\nthe `virt_use_nfs` SELinux boolean.\n\n.Procedure\n\n* Enable the `virt_use_nfs` boolean using the following command.\nThe `-P` option makes this boolean persistent across reboots.\n+\n[source,terminal]\n----\n# setsebool -P virt_use_nfs 1\n----\n" }, { "from": "gpt", "value": "// Module included in the following assemblies:\n//\n// * storage/persistent_storage/persistent-storage-nfs.adoc\n\n[id=\"nfs-selinux_{context}\"]\n= SELinux\n\n{op-system-base-full} and {op-system-first} systems are configured to use SELinux on remote NFS servers by default.\n\nFor non-{op-system-base} and non-{op-system} systems, SELinux does not allow writing from a pod to a remote NFS server. The NFS volume mounts correctly but it is read-only. You will need to enable the correct SELinux permissions by using the following procedure.\n\n.Prerequisites\n\n* The `container-selinux` package must be installed. This package provides the `virt_use_nfs` SELinux boolean.\n\n.Procedure\n\n* Enable the `virt_use_nfs` boolean using the following command. The `-P` option makes this boolean persistent across reboots.\n+\n[source,terminal]\n----\n# setsebool -P virt_use_nfs 1\n----\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Add message types to example\n" }, { "from": "human", "value": "= Getting Started\nGary Brown\n2015-04-21\n:description: How to get started with Scribble\n:jbake-type: page\n:jbake-status: published\n:icons: font\n\n== Getting Started with Scribble.\n\n----\nmodule examples;\n\nglobal protocol HelloWorld(role Me, role World) {\n\thello(Greetings) from Me to World;\n\tchoice at World {\n\t\tgoodMorning(Compliments) from World to Me;\n\t} or {\n\t\tgoodAfternoon(Salutations) from World to Me;\n\t}\n}\n----\n\nA very simply example, but this illustrates the basic syntax for a hello world interaction, where a party performing the role Me sends a message of type Greetings to another party performing the role 'World', who subsequently makes a decision which determines which path of the choice will be followed, resulting in a GoodMorning or GoodAfternoon message being exchanged.\n\n\n=== Working on Scribble source\n\nIf you are interested in building Scribble from source, then check out\nlink:/docs/dev/developer-guide.html[Development resources]\n\n\n\n\n" }, { "from": "gpt", "value": "= Getting Started\nGary Brown\n2015-04-21\n:description: How to get started with Scribble\n:jbake-type: page\n:jbake-status: published\n:icons: font\n\n== Getting Started with Scribble.\n\n----\nmodule example;\n\ntype <xsd> \"{http://www.acme.com/types}Greetings\" from \"http://www.acme.com/types/Greetings.xsd\" as Greetings;\ntype <xsd> \"{http://www.acme.com/types}Compliments\" from \"http://www.acme.com/types/Compliments.xsd\" as Compliments;\ntype <xsd> \"{http://www.acme.com/types}Salutations\" from \"http://www.acme.com/types/Salutations.xsd\" as Salutations;\n\nglobal protocol HelloWorld(role Me, role World) {\n\thello(Greetings) from Me to World;\n\tchoice at World {\n\t\tgoodMorning(Compliments) from World to Me;\n\t} or {\n\t\tgoodAfternoon(Salutations) from World to Me;\n\t}\n}\n----\n\nA very simply example, but this illustrates the basic syntax for a hello world interaction, where a party performing the role Me sends a message of type Greetings to another party performing the role 'World', who subsequently makes a decision which determines which path of the choice will be followed, resulting in a _goodMorning_ or _goodAfternoon_ message being exchanged.\n\n\n=== Working on Scribble source\n\nIf you are interested in building Scribble from source, then check out\nlink:/docs/dev/developer-guide.html[Development resources]\n\n\n\n\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Make it obvious the docs are about Discovery\n" }, { "from": "human", "value": "= Spring Cloud for Cloud Foundry\n\ninclude::intro.adoc[]\n\nThe first time you use it the discovery client might be slow owing to\nthe fact that it has to get an access token from Cloud Foundry.\n\n== Quickstart\n\ninclude::quickstart.adoc[]\n\n=== Single Sign On\n\nNOTE: All of the OAuth2 SSO and resource server features moved to Spring Boot\nin version 1.3. You can find documentation in the \nhttp://docs.spring.io/spring-boot/docs/current/reference/htmlsingle/[Spring Boot user guide].\n\nThis project provides automatic binding from CloudFoundry service\ncredentials to the Spring Boot features. If you have a CloudFoundry\nservice called \"sso\", for instance, with credentials containing\n\"client_id\", \"client_secret\" and \"auth_domain\", it will bind\nautomatically to the Spring OAuth2 client that you enable with\n`@EnableOAuth2Sso` (from Spring Boot). The name of the service can be\nparameterized using `spring.oauth2.sso.serviceId`.\n\n" }, { "from": "gpt", "value": "= Spring Cloud for Cloud Foundry\n\ninclude::intro.adoc[]\n\nThe first time you use it the discovery client might be slow owing to\nthe fact that it has to get an access token from Cloud Foundry.\n\n== Discovery\n\ninclude::quickstart.adoc[]\n\n== Single Sign On\n\nNOTE: All of the OAuth2 SSO and resource server features moved to Spring Boot\nin version 1.3. You can find documentation in the \nhttp://docs.spring.io/spring-boot/docs/current/reference/htmlsingle/[Spring Boot user guide].\n\nThis project provides automatic binding from CloudFoundry service\ncredentials to the Spring Boot features. If you have a CloudFoundry\nservice called \"sso\", for instance, with credentials containing\n\"client_id\", \"client_secret\" and \"auth_domain\", it will bind\nautomatically to the Spring OAuth2 client that you enable with\n`@EnableOAuth2Sso` (from Spring Boot). The name of the service can be\nparameterized using `spring.oauth2.sso.serviceId`.\n\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Include Hello World Youtube video" }, { "from": "human", "value": "= Minimal-J\n\nJava - but small.\n\nimage::doc/frontends.png[]\n\nMinimal-J applications are\n\n* Responsive to use on every device\n* Straight forward to specify and implement and therefore\n* Easy to plan and manage\n\n=== Idea\n\nBusiness applications tend to get complex and complicated. Minimal-J prevents this by setting clear rules how an application should behave and how it should be implemented.\n\nMinimal applications may not always look the same. But the UI concepts never change. There are no surprises for the user.\n\n== Technical Features\n\n* Independent of the used UI technology. Implementations for Web / Mobile / Desktop.\n* ORM persistence layer for Maria DB or in memory DB. Transactions and Authorization supported.\n* Small: The minimalj.jar is still < 1MB\n* Very few dependencies\n* Applications can run standalone (like SpringBoot)\n\n== Documentation\n\n* link:doc/user_guide/user_guide.adoc[Minimal user guide] User guide for Minimal-J applications.\n* link:doc/topics.adoc[Tutorial and examples] Informations for developers.\n* link:doc/release_notes.adoc[Release Notes]\n\n== Hello World\n\nHow to implement Hello World in Minimal-J:\n\nlink:_includes/ex-video.adoc[]\n\nvideo::0VHz7gv6TpA[youtube]\n\n=== Contact\n\n* Bruno Eberhard, mailto:[email protected][[email protected]] " }, { "from": "gpt", "value": "= Minimal-J\n\nJava - but small.\n\nimage::doc/frontends.png[]\n\nMinimal-J applications are\n\n* Responsive to use on every device\n* Straight forward to specify and implement and therefore\n* Easy to plan and manage\n\n=== Idea\n\nBusiness applications tend to get complex and complicated. Minimal-J prevents this by setting clear rules how an application should behave and how it should be implemented.\n\nMinimal applications may not always look the same. But the UI concepts never change. There are no surprises for the user.\n\n== Technical Features\n\n* Independent of the used UI technology. Implementations for Web / Mobile / Desktop.\n* ORM persistence layer for Maria DB or in memory DB. Transactions and Authorization supported.\n* Small: The minimalj.jar is still < 1MB\n* Very few dependencies\n* Applications can run standalone (like SpringBoot)\n\n== Documentation\n\n* link:doc/user_guide/user_guide.adoc[Minimal user guide] User guide for Minimal-J applications.\n* link:doc/topics.adoc[Tutorial and examples] Informations for developers.\n* link:doc/release_notes.adoc[Release Notes]\n\n== Hello World\n\nlink:http://www.youtube.com/watch?v=0VHz7gv6TpA[How to implement Hello World in Minimal-J]\n\n\n=== Contact\n\n* Bruno Eberhard, mailto:[email protected][[email protected]] " } ]
asciidoc
chosen
[ { "from": "system", "value": "Add `deprecated` to extension status note\n\nAlso fix the Extension status URL\n" }, { "from": "human", "value": "[NOTE]\n====\nThis technology is considered {extension-status}.\n\nifeval::[\"{extension-status}\" == \"experimental\"]\nIn _experimental_ mode, early feedback is requested to mature the idea.\nThere is no guarantee of stability nor long term presence in the platform until the solution matures.\nFeedback is welcome on our https://groups.google.com/d/forum/quarkus-dev[mailing list] or as issues in our https://github.com/quarkusio/quarkus/issues[GitHub issue tracker].\nendif::[]\nifeval::[\"{extension-status}\" == \"preview\"]\nIn _preview_, backward compatibility and presence in the ecosystem is not guaranteed.\nSpecific improvements might require changing configuration or APIs, and plans to become _stable_ are under way.\nFeedback is welcome on our https://groups.google.com/d/forum/quarkus-dev[mailing list] or as issues in our https://github.com/quarkusio/quarkus/issues[GitHub issue tracker].\nendif::[]\nifeval::[\"{extension-status}\" == \"stable\"]\nBeing _stable_, backward compatibility and presence in the ecosystem are taken very seriously.\nendif::[]\n\nFor a full list of possible statuses, check our https://quarkus.io/faq/#extension-status[FAQ entry].\n====\n" }, { "from": "gpt", "value": "[NOTE]\n====\nThis technology is considered {extension-status}.\n\nifeval::[\"{extension-status}\" == \"experimental\"]\nIn _experimental_ mode, early feedback is requested to mature the idea.\nThere is no guarantee of stability nor long term presence in the platform until the solution matures.\nFeedback is welcome on our https://groups.google.com/d/forum/quarkus-dev[mailing list] or as issues in our https://github.com/quarkusio/quarkus/issues[GitHub issue tracker].\nendif::[]\nifeval::[\"{extension-status}\" == \"preview\"]\nIn _preview_, backward compatibility and presence in the ecosystem is not guaranteed.\nSpecific improvements might require changing configuration or APIs, and plans to become _stable_ are under way.\nFeedback is welcome on our https://groups.google.com/d/forum/quarkus-dev[mailing list] or as issues in our https://github.com/quarkusio/quarkus/issues[GitHub issue tracker].\nendif::[]\nifeval::[\"{extension-status}\" == \"stable\"]\nBeing _stable_, backward compatibility and presence in the ecosystem are taken very seriously.\nendif::[]\nifeval::[\"{extension-status}\" == \"deprecated\"]\nBeing _deprecated_ means that this extension is likely to be replaced or removed in a future version of Quarkus.\nendif::[]\n\nFor a full list of possible statuses, check our https://quarkus.io/faq/#what-are-the-extension-statuses[FAQ entry].\n====\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Add link to github in rendered docs\n" }, { "from": "human", "value": "== Spring Cloud Starters\n\ninclude::intro.adoc[]\n\n\n" }, { "from": "gpt", "value": "== Spring Cloud Starters\n:github: https://github.com/spring-cloud/spring-cloud-starters\n:githubmaster: {github}/tree/master\n:docslink: {githubmaster}/docs/src/main/asciidoc\n\ninclude::intro.adoc[]\n\ninclude::https://raw.githubusercontent.com/spring-cloud/spring-cloud-build/master/docs/src/main/asciidoc/contributing-docs.adoc[]\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Add reference to unique constraints to unique index chapter.\n" }, { "from": "human", "value": "[[rest-api-unique-indexes]]\n== Unique Indexing ==\n\nFor uniqueness enforcements, there are two modes:\n\n* URL Parameter `unique=get_or_create`: Create a new node/relationship and index it if no existing one can be found. If an existing node/relationship is found, discard the sent data and return the existing node/relationship.\n\n* URL Parameter `unique=create_or_fail`: Create a new node/relationship if no existing one can be found in the index. If an existing node/relationship is found, return a conflict error.\n\nFor more information, see <<transactions-unique-nodes>>.\n\ninclude::get-or-create-unique-node-(create).asciidoc[]\n\ninclude::get-or-create-unique-node-(existing).asciidoc[]\n\ninclude::create-a-unique-node-or-return-fail-(create).asciidoc[]\n\ninclude::create-a-unique-node-or-return-fail-(fail).asciidoc[]\n\ninclude::get-or-create-unique-relationship-(create).asciidoc[]\n\ninclude::get-or-create-unique-relationship-(existing).asciidoc[]\n\ninclude::create-a-unique-relationship-or-return-fail-(create).asciidoc[]\n\ninclude::create-a-unique-relationship-or-return-fail-(fail).asciidoc[]\n\n\n" }, { "from": "gpt", "value": "[[rest-api-unique-indexes]]\n== Unique Indexing ==\n\nNOTE: As of Neo4j 2.0, unique constraints have been added. These make Neo4j enforce the uniqueness,\nguaranteeing that uniqueness is maintained. See <<graphdb-neo4j-schema-constraints>> for details about this.\n\nFor uniqueness enforcements, there are two modes:\n\n* URL Parameter `unique=get_or_create`: Create a new node/relationship and index it if no existing one can be found.\nIf an existing node/relationship is found, discard the sent data and return the existing node/relationship.\n\n* URL Parameter `unique=create_or_fail`: Create a new node/relationship if no existing one can be found in the index.\nIf an existing node/relationship is found, return a conflict error.\n\nFor more information, see <<transactions-unique-nodes>>.\n\ninclude::get-or-create-unique-node-(create).asciidoc[]\n\ninclude::get-or-create-unique-node-(existing).asciidoc[]\n\ninclude::create-a-unique-node-or-return-fail-(create).asciidoc[]\n\ninclude::create-a-unique-node-or-return-fail-(fail).asciidoc[]\n\ninclude::get-or-create-unique-relationship-(create).asciidoc[]\n\ninclude::get-or-create-unique-relationship-(existing).asciidoc[]\n\ninclude::create-a-unique-relationship-or-return-fail-(create).asciidoc[]\n\ninclude::create-a-unique-relationship-or-return-fail-(fail).asciidoc[]\n\n\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Remove ipi/upi dichotomy for machine management capabilities\n\nUsage of the machine API does NOT mandate that the customer used IPI. The only requirement is that the customer delegates the ability to create lifecycle machines to the product.\n" }, { "from": "human", "value": "// Module included in the following assemblies:\n//\n// * machine_management/applying_autoscaling.adoc\n// * machine_management/creating-infrastructure-machinesets.adoc\n// * machine_management/creating-machinesets.adoc\n// * machine_management/deploying-machine-health-checks.adoc\n// * machine_management/manually-scaling-machinesets.adoc\n\n[IMPORTANT]\n====\nThis process is not applicable to clusters that use user-provisioned\ninfrastructure. Because you manually provisioned your machines yourself, you\ncannot use the advanced machine management and scaling capabilities\nthat an installer-provisioned infrastructure cluster offers.\n====\n" }, { "from": "gpt", "value": "// Module included in the following assemblies:\n//\n// * machine_management/applying_autoscaling.adoc\n// * machine_management/creating-infrastructure-machinesets.adoc\n// * machine_management/creating-machinesets.adoc\n// * machine_management/deploying-machine-health-checks.adoc\n// * machine_management/manually-scaling-machinesets.adoc\n\n[IMPORTANT]\n====\nThis process is not applicable to clusters where you manually provisioned the machines yourself. You\ncan use the advanced machine management and scaling capabilities only in clusters where the machine API is operational.\n====\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Fix doc filename instead of title shown in nav menu\n" }, { "from": "human", "value": "== Using a pre-release version\n\nPre-release versions of `AsciidoctorJ` are published to oss.sonatype.org.\nThe exact location of the repository will be announced.\nFinal releases are released to both Maven Central.\n\nHere's how to use a pre-release version in Maven:\n\n[source, xml]\n----\n <repositories>\n <repository>\n <id>staging</id>\n <url>https://oss.sonatype.org/content/repositories/orgasciidoctor-1234</url> <1>\n <releases>\n <enabled>true</enabled>\n </releases>\n </repository>\n </repositories>\n----\n<1> The exact URL differs for every build\n" }, { "from": "gpt", "value": "= Using a pre-release version\n\nPre-release versions of `AsciidoctorJ` are published to oss.sonatype.org.\nThe exact location of the repository will be announced.\nFinal releases are released to both Maven Central.\n\nHere's how to use a pre-release version in Maven:\n\n[source, xml]\n----\n <repositories>\n <repository>\n <id>staging</id>\n <url>https://oss.sonatype.org/content/repositories/orgasciidoctor-1234</url> <1>\n <releases>\n <enabled>true</enabled>\n </releases>\n </repository>\n </repositories>\n----\n<1> The exact URL differs for every build\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Use version in API link\n" }, { "from": "human", "value": "= Gradle Docker Plugin User Guide & Examples\n:uri-bmuschko-blog: https://bmuschko.com/blog\n:uri-ghpages: https://bmuschko.github.io/gradle-docker-plugin\n:uri-github: https://github.com/bmuschko/gradle-docker-plugin\n:uri-gradle-docs: https://docs.gradle.org/current\n\ninclude::00-intro.adoc[]\n\ninclude::10-remote-api-plugin.adoc[]\n\ninclude::20-java-application-plugin.adoc[]\n\ninclude::30-spring-boot-application-plugin.adoc[]\n\ninclude::40-changes.adoc[]" }, { "from": "gpt", "value": "= Gradle Docker Plugin User Guide & Examples\n:uri-bmuschko-blog: https://bmuschko.com/blog\n:uri-ghpages: https://bmuschko.github.io/gradle-docker-plugin/{gradle-project-version}\n:uri-github: https://github.com/bmuschko/gradle-docker-plugin\n:uri-gradle-docs: https://docs.gradle.org/current\n\ninclude::00-intro.adoc[]\n\ninclude::10-remote-api-plugin.adoc[]\n\ninclude::20-java-application-plugin.adoc[]\n\ninclude::30-spring-boot-application-plugin.adoc[]\n\ninclude::40-changes.adoc[]" } ]
asciidoc
chosen
[ { "from": "system", "value": "Fix documentation bug for unique indexes in REST.\n" }, { "from": "human", "value": "[[rest-api-unique-indexes]]\n== Unique Indexes ==\n\nFor uniqueness enforcements, there are two modes:\n\n* URL Parameter `unique=get_or_create`: Create a new node/relationship and index it if no existing one can be found. If an existing node/relationship is found, discard the sent data and return the existing node/relationship.\n\n* URL Parameter `unique=create_or_fail`: Create a new node/relationship if no existing one can be found in the index. If an existing node/relationship is found, return a conflict error.\n\nFor more information, see <<transactions-unique-nodes>>.\n\ninclude::get-or-create-unique-node-(create).asciidoc[]\n\ninclude::get-or-create-unique-node-(existing).asciidoc[]\n\ninclude::create-a-unique-node-or-return-fail-(create).asciidoc[]\n\ninclude::create-a-unique-node-or-return-fail-(fail).asciidoc[]\n\ninclude::get-or-create-unique-relationship-(create).asciidoc[]\n\ninclude::get-or-create-unique-relationship-(existing).asciidoc[]\n\ninclude::create-a-unique-relationship-or-return-fail-(create).asciidoc[]\n\ninclude::create-a-unique-relationship-or-return-fail-(fail).asciidoc[]\n\n\n" }, { "from": "gpt", "value": "[[rest-api-unique-indexes]]\n== Unique Indexes ==\n\nFor uniqueness enforcements, there are two modes:\n\n* URL Parameter `uniqueness=get_or_create`: Create a new node/relationship and index it if no existing one can be found. If an existing node/relationship is found, discard the sent data and return the existing node/relationship.\n\n* URL Parameter `uniqueness=create_or_fail`: Create a new node/relationship if no existing one can be found in the index. If an existing node/relationship is found, return a conflict error.\n\nFor more information, see <<transactions-unique-nodes>>.\n\ninclude::get-or-create-unique-node-(create).asciidoc[]\n\ninclude::get-or-create-unique-node-(existing).asciidoc[]\n\ninclude::create-a-unique-node-or-return-fail-(create).asciidoc[]\n\ninclude::create-a-unique-node-or-return-fail-(fail).asciidoc[]\n\ninclude::get-or-create-unique-relationship-(create).asciidoc[]\n\ninclude::get-or-create-unique-relationship-(existing).asciidoc[]\n\ninclude::create-a-unique-relationship-or-return-fail-(create).asciidoc[]\n\ninclude::create-a-unique-relationship-or-return-fail-(fail).asciidoc[]\n\n\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Update breaking changes for docker images\n" }, { "from": "human", "value": "= Breaking changes in 2.10\n:page-sidebar: am_2_x_sidebar\n:page-permalink: am/2.x/am_breaking_changes_2.10.html\n:page-folder: am/installation-guide\n:page-layout: am\n\n== JAR Manifest file has changed\n\nStarting from version 2.10.10, the Main-Class of the entry point of the runnable management and gateway application has changed due to typo error.\n\n----\nAM Management API\n\n - <mainClass>io.gravitee.am.management.standalone.boostrap.Bootstrap</mainClass>\n + <mainClass>io.gravitee.am.management.standalone.bootstrap.Bootstrap</mainClass>\n\nAM Gateway\n\n - <mainClass>io.gravitee.am.gateway.boostrap.Bootstrap</mainClass>\n + <mainClass>io.gravitee.am.gateway.bootstrap.Bootstrap</mainClass>\n----\n\nIt could have an impact if you are using your own custom build/launch script with command like `java -jar gravitee-am-*-bootstrap.jar`.\nBe sure to test this scenario before going further in your production environment.\n\nNOTE: There is no impact if you rely on the built-in gravitee(.bat) script.\n\n" }, { "from": "gpt", "value": "= Breaking changes in 2.10\n:page-sidebar: am_2_x_sidebar\n:page-permalink: am/2.x/am_breaking_changes_2.10.html\n:page-folder: am/installation-guide\n:page-layout: am\n\n== JAR Manifest file has changed\n\nStarting from version 2.10.10, the Main-Class of the entry point of the runnable management and gateway application has changed due to typo error.\n\n----\nAM Management API\n\n - <mainClass>io.gravitee.am.management.standalone.boostrap.Bootstrap</mainClass>\n + <mainClass>io.gravitee.am.management.standalone.bootstrap.Bootstrap</mainClass>\n\nAM Gateway\n\n - <mainClass>io.gravitee.am.gateway.boostrap.Bootstrap</mainClass>\n + <mainClass>io.gravitee.am.gateway.bootstrap.Bootstrap</mainClass>\n----\n\nIt could have an impact if you are using your own custom build/launch script with command like `java -jar gravitee-am-*-bootstrap.jar`.\nBe sure to test this scenario before going further in your production environment.\n\nNOTE: There is no impact if you rely on the built-in gravitee(.bat) script.\n\n\n== Docker images\nAll the UI based docker images (console) are now using HTTP port 8080 and HTTPS port 8443 by default to\navoid the need of a root account to use 80/443 to improve the security of our platform." } ]
asciidoc
chosen
[ { "from": "system", "value": "Improve doc for containerized native build\n\nOn macOS and Windows the following command fails:\n`quarkus build --native -Dquarkus.native.container-build=true`.\nReason is that the Quarkus CLI runs integration tests against binary built in Linux container.\nThe produced binary is ELF not runnable under another OS.\n\nSigned-off-by: Matej Vasek <[email protected]>\n" }, { "from": "human", "value": "[source, bash, subs=attributes+, role=\"primary asciidoc-tabs-sync-cli\"]\n.CLI\n----\nifdef::build-additional-parameters[]\nquarkus build --native -Dquarkus.native.container-build=true {build-additional-parameters}\nendif::[]\nifndef::build-additional-parameters[]\nquarkus build --native -Dquarkus.native.container-build=true\nendif::[]\n----\nifndef::devtools-no-maven[]\nifdef::devtools-wrapped[+]\n[source, bash, subs=attributes+, role=\"secondary asciidoc-tabs-sync-maven\"]\n.Maven\n----\nifdef::build-additional-parameters[]\n./mvnw package -Dnative -Dquarkus.native.container-build=true {build-additional-parameters}\nendif::[]\nifndef::build-additional-parameters[]\n./mvnw package -Dnative -Dquarkus.native.container-build=true\nendif::[]\n----\nendif::[]\nifndef::devtools-no-gradle[]\nifdef::devtools-wrapped[+]\n[source, bash, subs=attributes+, role=\"secondary asciidoc-tabs-sync-gradle\"]\n.Gradle\n----\nifdef::build-additional-parameters[]\n./gradlew build -Dquarkus.package.type=native -Dquarkus.native.container-build=true {build-additional-parameters}\nendif::[]\nifndef::build-additional-parameters[]\n./gradlew build -Dquarkus.package.type=native -Dquarkus.native.container-build=true\nendif::[]\n----\nendif::[]" }, { "from": "gpt", "value": "[source, bash, subs=attributes+, role=\"primary asciidoc-tabs-sync-cli\"]\n.CLI\n----\nifdef::build-additional-parameters[]\nquarkus build --native --no-tests -Dquarkus.native.container-build=true {build-additional-parameters}\nendif::[]\nifndef::build-additional-parameters[]\nquarkus build --native --no-tests -Dquarkus.native.container-build=true\nendif::[]\n# The --no-tests flag is required only on Windows and macOS.\n----\nifndef::devtools-no-maven[]\nifdef::devtools-wrapped[+]\n[source, bash, subs=attributes+, role=\"secondary asciidoc-tabs-sync-maven\"]\n.Maven\n----\nifdef::build-additional-parameters[]\n./mvnw package -Dnative -Dquarkus.native.container-build=true {build-additional-parameters}\nendif::[]\nifndef::build-additional-parameters[]\n./mvnw package -Dnative -Dquarkus.native.container-build=true\nendif::[]\n----\nendif::[]\nifndef::devtools-no-gradle[]\nifdef::devtools-wrapped[+]\n[source, bash, subs=attributes+, role=\"secondary asciidoc-tabs-sync-gradle\"]\n.Gradle\n----\nifdef::build-additional-parameters[]\n./gradlew build -Dquarkus.package.type=native -Dquarkus.native.container-build=true {build-additional-parameters}\nendif::[]\nifndef::build-additional-parameters[]\n./gradlew build -Dquarkus.package.type=native -Dquarkus.native.container-build=true\nendif::[]\n----\nendif::[]" } ]
asciidoc
chosen
[ { "from": "system", "value": "Use correct Asciidoc headers (not Markdown)\n" }, { "from": "human", "value": "== Rook - Sane, smart, fast, Clojure web services\n\nimage:http://clojars.org/io.aviso/rook/latest-version.svg[Clojars Project, link=\"http://clojars.org/io.aviso/rook\"]\n\nimage:https://drone.io/github.com/AvisoNovate/rook/status.png[Build Status, link=\"https://drone.io/github.com/AvisoNovate/rook\"]\n\nRook is a set of middleware and handlers to enable metadata-based routing for https://github.com/ring-clojure/ring[Ring] web applications.\n\nThe intention is to expose a Clojure _namespace_ as a web service _resource_; there's a default mapping of HTTP verbs\nand paths to function names; these can be extended or overridden by metadata on the functions in the namespace.\n\nThe end result is that a compliant web service resource can be created in very little code.\n\nRook also supports Swagger 2.0: a detailed JSON description of your web service is generated directly from the functions and metadata.\n\nRook is available under the terms of the Apache Software License 2.0.\n\nlink:https://portal.aviso.io/#/document/open-source/rook/Current[Full Documentation]\n" }, { "from": "gpt", "value": "## Rook - Sane, smart, fast, Clojure web services\n\nimage:http://clojars.org/io.aviso/rook/latest-version.svg[Clojars Project, link=\"http://clojars.org/io.aviso/rook\"]\n\nimage:https://drone.io/github.com/AvisoNovate/rook/status.png[Build Status, link=\"https://drone.io/github.com/AvisoNovate/rook\"]\n\nRook is a set of middleware and handlers to enable metadata-based routing for https://github.com/ring-clojure/ring[Ring] web applications.\n\nThe intention is to expose a Clojure _namespace_ as a web service _resource_; there's a default mapping of HTTP verbs\nand paths to function names; these can be extended or overridden by metadata on the functions in the namespace.\n\nThe end result is that a compliant web service resource can be created in very little code.\n\nRook also supports Swagger 2.0: a detailed JSON description of your web service is generated directly from the functions and metadata.\n\nRook is available under the terms of the Apache Software License 2.0.\n\nlink:https://portal.aviso.io/#/document/open-source/rook/Current[Full Documentation]\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Update the metric names for OVN-Kubernetes\n" }, { "from": "human", "value": "[id=\"learn-about-ovn-kubernetes\"]\n= About the OVN-Kubernetes Pod network provider\ninclude::modules/common-attributes.adoc[]\n:context: learn-about-ovn-kubernetes\n\ntoc::[]\n\ninclude::modules/nw-ovn-technology-preview.adoc[leveloffset=+1]\n\n[id=\"ovn-kubernetes-metrics\"]\n== Exposed metrics\n\nThe OVN-Kubernetes network provider exposes certain metrics for use by the Prometheus-based {product-title} cluster monitoring stack.\n\n// openshift/ovn-kubernetes => go-controller/pkg/ovn/metrics.go\n\n.Metrics exposed by OVN-Kubernetes\n[cols=\"2a,8a\",options=\"header\"]\n|===\n|Name |Description\n\n|`nb_e2e_timestamp`\n|A timestamp persisted to the Open vSwitch (OVN) northbound database and updated frequently.\n\n|`pod_creation_latency_seconds`\n|The latency between when a Pod is created and when the Pod is annotated by OVN-Kubernetes. The higher the latency, the more time that elapses before a Pod is available for network connectivity.\n\n|===\n" }, { "from": "gpt", "value": "[id=\"learn-about-ovn-kubernetes\"]\n= About the OVN-Kubernetes Pod network provider\ninclude::modules/common-attributes.adoc[]\n:context: learn-about-ovn-kubernetes\n\ntoc::[]\n\ninclude::modules/nw-ovn-technology-preview.adoc[leveloffset=+1]\n\n[id=\"ovn-kubernetes-metrics\"]\n== Exposed metrics\n\nThe OVN-Kubernetes network provider exposes certain metrics for use by the Prometheus-based {product-title} cluster monitoring stack.\n\n// openshift/ovn-kubernetes => go-controller/pkg/ovn/metrics.go\n\n.Metrics exposed by OVN-Kubernetes\n[cols=\"2a,8a\",options=\"header\"]\n|===\n|Name |Description\n\n|`ovnkube_master_nb_e2e_timestamp`\n|A timestamp persisted to the Open vSwitch (OVN) northbound database and updated frequently.\n\n|`ovnkube_master_pod_creation_latency_seconds`\n|The latency between when a Pod is created and when the Pod is annotated by OVN-Kubernetes. The higher the latency, the more time that elapses before a Pod is available for network connectivity.\n\n|===\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Document how to access incoming query parameters\n\nIt wasn't clear how to use values of query parameters of the incoming request." }, { "from": "human", "value": "= Transform Query Parameters Policy\n\nifdef::env-github[]\nimage:https://ci.gravitee.io/buildStatus/icon?job=gravitee-io/gravitee-policy-transformqueryparams/master[\"Build status\", link=\"https://ci.gravitee.io/job/gravitee-io/job/gravitee-policy-transformqueryparams/\"]\nimage:https://badges.gitter.im/Join Chat.svg[\"Gitter\", link=\"https://gitter.im/gravitee-io/gravitee-io?utm_source=badge&utm_medium=badge&utm_campaign=pr-badge&utm_content=badge\"]\nendif::[]\n\n== Phase\n\n[cols=\"2*\", options=\"header\"]\n|===\n^|onRequest\n^|onResponse\n\n^.^| X\n^.^|\n\n|===\n\n== Description\n\nThe Transform Query Parameters policy is used to override incoming request HTTP query parameters.\nYou can override these HTTP query parameters by:\n\n* Clearing all existing query parameters\n* Adding or updating list of query parameters\n* Removing them individually\n\n[source, json]\n.Sample\n----\n\"transform-queryparams\": {\n \"addQueryParameters\": [\n {\n \"name\": \"myParam\",\n \"value\": \"{#request.id}\"\n }\n ],\n \"removeQueryParameters\": [\n \"secretParam\"\n ]\n}\n----" }, { "from": "gpt", "value": "= Transform Query Parameters Policy\n\nifdef::env-github[]\nimage:https://ci.gravitee.io/buildStatus/icon?job=gravitee-io/gravitee-policy-transformqueryparams/master[\"Build status\", link=\"https://ci.gravitee.io/job/gravitee-io/job/gravitee-policy-transformqueryparams/\"]\nimage:https://badges.gitter.im/Join Chat.svg[\"Gitter\", link=\"https://gitter.im/gravitee-io/gravitee-io?utm_source=badge&utm_medium=badge&utm_campaign=pr-badge&utm_content=badge\"]\nendif::[]\n\n== Phase\n\n[cols=\"2*\", options=\"header\"]\n|===\n^|onRequest\n^|onResponse\n\n^.^| X\n^.^|\n\n|===\n\n== Description\n\nThe Transform Query Parameters policy is used to override incoming request HTTP query parameters.\nYou can override these HTTP query parameters by:\n\n* Clearing all existing query parameters\n* Adding or updating list of query parameters\n* Removing them individually\n\nThe values of query parameters of the incoming request are accessible via the `{#request.params['query_parameter_name']}` construct. \n\nThe example below shows how to add the ID of the incoming request to the outgoing request.\n\n[source, json]\n.Sample\n----\n\"transform-queryparams\": {\n \"addQueryParameters\": [\n {\n \"name\": \"myParam\",\n \"value\": \"{#request.id}\"\n }\n ],\n \"removeQueryParameters\": [\n \"secretParam\"\n ]\n}\n----\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Update readme with demo link\n" }, { "from": "human", "value": "# react-popover\n:toc: macro\n\nA specification backed popover for react\n\n\ntoc::[]\n\n\n## Installation\n\n----\nnpm install react-popover\n----\n\n\n## Examples\n\nThe `examples` folder contains basic applications of this component. To try them locally run `npm start`.\n\n\n## API\n\nTODO\n" }, { "from": "gpt", "value": "# react-popover\n:toc: macro\n\nA specification backed popover for react. See [demo](https://littlebits.github.io/react-popover/build).\n\n\ntoc::[]\n\n\n## Installation\n\n----\nnpm install react-popover\n----\n\n\n## Examples\n\nThe `examples` folder contains basic applications of this component. To try them locally run `npm start`.\n\n\n## API\n\nTODO\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Fix broken link to OmniJ Project README" }, { "from": "human", "value": "= OmniJ Developer's Guide\nSean Gilligan\nv0.1, July 30, 2015: Early draft\n:numbered:\n:toc:\n:toclevels: 3\n:linkattrs:\n:imagesdir: images\n\nParagraph TBD.\n\n== Introduction to OmniJ\n\nThis section is TBD. For now the project http://github.com/OmniLayer/OmniJ/README.adoc[README] is the best place to get started.\n\n== JSON-RPC Clients\n\n[plantuml, diagram-classes, svg]\n....\nskinparam packageStyle Rect\nskinparam shadowing false\nhide empty members\n\nnamespace com.msgilligan.bitcoin.rpc {\n class RPCClient\n RPCClient <|-- class DynamicRPCClient << Groovy >>\n RPCClient <|-- BitcoinClient\n BitcoinClient <|-- class BitcoinCLIClient << Groovy >>\n}\n\nnamespace foundation.omni.rpc {\n com.msgilligan.bitcoin.rpc.BitcoinClient <|-- OmniClient\n OmniClient <|-- OmniExtendedClient\n OmniExtendedClient <|-- class OmniCLIClient << Groovy >>\n}\n....\n\n\n\n" }, { "from": "gpt", "value": "= OmniJ Developer's Guide\nSean Gilligan\nv0.1, July 30, 2015: Early draft\n:numbered:\n:toc:\n:toclevels: 3\n:linkattrs:\n:imagesdir: images\n\nParagraph TBD.\n\n== Introduction to OmniJ\n\nThis section is TBD. For now the project https://github.com/OmniLayer/OmniJ#omnij-project[README] is the best place to get started.\n\n== JSON-RPC Clients\n\n[plantuml, diagram-classes, svg]\n....\nskinparam packageStyle Rect\nskinparam shadowing false\nhide empty members\n\nnamespace com.msgilligan.bitcoin.rpc {\n class RPCClient\n RPCClient <|-- class DynamicRPCClient << Groovy >>\n RPCClient <|-- BitcoinClient\n BitcoinClient <|-- class BitcoinCLIClient << Groovy >>\n}\n\nnamespace foundation.omni.rpc {\n com.msgilligan.bitcoin.rpc.BitcoinClient <|-- OmniClient\n OmniClient <|-- OmniExtendedClient\n OmniExtendedClient <|-- class OmniCLIClient << Groovy >>\n}\n....\n\n\n\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Make the glossary a bit better\n" }, { "from": "human", "value": "= Internal glossary\n\nThis is not a document to be published outside.\nThis is a collection of preferred term in the documentation and website.\nPlease stay witin these terms for consistency.\n\nLive reload:: for our `quarkus:dev` capability\nGraalVM:: preferred term for the VM creating native executable. No space.\nSubstrate VM:: non-preferred. Only if you want to clarify which part of GraalVM we use.\nNative Executable:: the executable that is compiled to native 1s and 0s\nDocker image:: for the actual `Dockerfile` definition and when the tool chain is involved\nContainer:: when we discuss Quarkus running in... containers\nSupersonic Subatomic Java:: our tagline\nKubernetes Native Java:: our preferred tagline to say that we rock for containers \nDeveloper Joy:: for everything going from live reload to the opinionated layer to a single config file\nUnify Imperative and Reactive:: imperative and reactive. 'Nuff said.\nBest of breed frameworks and standards:: when we explain our stack\n" }, { "from": "gpt", "value": "= Internal glossary\n\nThis is not a document to be published outside.\nThis is a collection of preferred term in the documentation and website.\nPlease stay within these terms for consistency.\n\n* Live reload:: for our `quarkus:dev` capability\n* GraalVM:: preferred term for the VM creating native executable. No space.\n* Substrate VM:: non-preferred. Only if you want to clarify which part of GraalVM we use.\n* Native Executable:: the executable that is compiled to native 1s and 0s\n* Docker image:: for the actual `Dockerfile` definition and when the tool chain is involved\n* Container:: when we discuss Quarkus running in... containers\n* Supersonic Subatomic Java:: our tagline\n* Kubernetes Native Java:: our preferred tagline to say that we rock for containers\n* Developer Joy:: for everything going from live reload to the opinionated layer to a single config file\n* Unify Imperative and Reactive:: imperative and reactive. 'Nuff said.\n* Best of breed frameworks and standards:: when we explain our stack\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Fix link to Upgrading From 1.x in multi-page documentation\n\nSee gh-30890\n" }, { "from": "human", "value": "[[documentation.upgrading]]\n== Upgrading From an Earlier Version\n\nYou should always ensure that you are running a {github-wiki}/Supported-Versions[supported version] of Spring Boot.\n\nDepending on the version that you are upgrading to, you can find some additional tips here:\n\n* *From 1.x:* <<actuator#upgrading.from-1x, Upgrading from 1.x>>\n* *To a new feature release:* <<upgrading#upgrading.to-feature, Upgrading to New Feature Release>>\n* *Spring Boot CLI:* <<upgrading#upgrading.cli, Upgrading the Spring Boot CLI>>\n\n" }, { "from": "gpt", "value": "[[documentation.upgrading]]\n== Upgrading From an Earlier Version\n\nYou should always ensure that you are running a {github-wiki}/Supported-Versions[supported version] of Spring Boot.\n\nDepending on the version that you are upgrading to, you can find some additional tips here:\n\n* *From 1.x:* <<upgrading#upgrading.from-1x, Upgrading from 1.x>>\n* *To a new feature release:* <<upgrading#upgrading.to-feature, Upgrading to New Feature Release>>\n* *Spring Boot CLI:* <<upgrading#upgrading.cli, Upgrading the Spring Boot CLI>>\n\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Remove group azure as its not from azure\n" }, { "from": "human", "value": "= Hashicorp Vault Component\n:doctitle: Hashicorp Vault\n:shortname: hashicorp-vault\n:artifactid: camel-hashicorp-vault\n:description: Manage secrets in Hashicorp Vault Service\n:since: 3.18\n:supportlevel: Preview\n:component-header: Only producer is supported\n//Manually maintained attributes\n:group: Azure\n:camel-spring-boot-name: hashicorp-vault\n\n*Since Camel {since}*\n\n*{component-header}*\n\nThe hashicorp-vault component that integrates https://www.vaultproject.io/[Hashicorp Vault]. \n\n\n== URI Format\n\n[source,xml]\n------------------------------------------------------------\n<dependency>\n <groupId>org.apache.camel</groupId>\n <artifactId>camel-hashicorp-vault</artifactId>\n <version>x.x.x</version>\n <!-- use the same version as your Camel core version -->\n</dependency>\n------------------------------------------------------------\n\n\n// component-configure options: START\n\n// component-configure options: END\n\n// component options: START\ninclude::partial$component-configure-options.adoc[]\ninclude::partial$component-endpoint-options.adoc[]\n// component options: END\n\n// endpoint options: START\n\n// endpoint options: END\n" }, { "from": "gpt", "value": "= Hashicorp Vault Component\n:doctitle: Hashicorp Vault\n:shortname: hashicorp-vault\n:artifactid: camel-hashicorp-vault\n:description: Manage secrets in Hashicorp Vault Service\n:since: 3.18\n:supportlevel: Preview\n:component-header: Only producer is supported\n//Manually maintained attributes\n:camel-spring-boot-name: hashicorp-vault\n\n*Since Camel {since}*\n\n*{component-header}*\n\nThe hashicorp-vault component that integrates https://www.vaultproject.io/[Hashicorp Vault]. \n\n\n== URI Format\n\n[source,xml]\n------------------------------------------------------------\n<dependency>\n <groupId>org.apache.camel</groupId>\n <artifactId>camel-hashicorp-vault</artifactId>\n <version>x.x.x</version>\n <!-- use the same version as your Camel core version -->\n</dependency>\n------------------------------------------------------------\n\n\n// component-configure options: START\n\n// component-configure options: END\n\n// component options: START\ninclude::partial$component-configure-options.adoc[]\ninclude::partial$component-endpoint-options.adoc[]\n// component options: END\n\n// endpoint options: START\n\n// endpoint options: END\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Mark completed items as DONE instead of using line-through\n\nGitHub's flavor of asciidoc does not support \"line-through\" for\nstriking out text. Added explicit \"DONE\" indication instead.\n" }, { "from": "human", "value": "= To-Do List\n\n- mock_cone_detector creates infinite area and overflows h\n- [line-through]#new waypoints shorter than old don't delete existing waypoints#\n- adjust waypoints for start position and cone position\n- cone area goes down when very close to cone\n- [line-through]#parameterize throttle and steering PWM values#\n- touch sensor does not work\n- cone detection in bright light does not work\n- GUIDED mode does not work\n- Encode PWM values or range set to use in the waypoints file\n- If waypoint encountered before cone is seen, rover goes into HOLD mode\nwith no recovery. Needs to be fixed.\n\n== Possible To-Do\n\n- [line-through]#Change from using WP_SPEED to CRUISE_SPEED. (Seems to be used by Vicky,\nwhile WP_SPEED is not.)#\n- Have a way of manually triggering parameter reload\n\n== Notes\n\nMAV_CMD_DO_SET_HOME appears to reset the map origin, as well as zero the\noffset between the map origin and base_link (for /mavros/local_position/pose\nand /mavros/local_position/odom).\n" }, { "from": "gpt", "value": "= To-Do List\n\n- mock_cone_detector creates infinite area and overflows h\n- (*DONE*) new waypoints shorter than old don't delete existing waypoints\n- adjust waypoints for start position and cone position\n- cone area goes down when very close to cone\n- (*DONE*) parameterize throttle and steering PWM values\n- touch sensor does not work\n- cone detection in bright light does not work\n- GUIDED mode does not work\n- Encode PWM values or range set to use in the waypoints file\n- If waypoint encountered before cone is seen, rover goes into HOLD mode\nwith no recovery. Needs to be fixed.\n\n== Possible To-Do\n\n- (*DONE*) Change from using WP_SPEED to CRUISE_SPEED. (Seems to be used by Vicky,\nwhile WP_SPEED is not.)\n- Have a way of manually triggering parameter reload\n\n== Notes\n\nMAV_CMD_DO_SET_HOME appears to reset the map origin, as well as zero the\noffset between the map origin and base_link (for /mavros/local_position/pose\nand /mavros/local_position/odom).\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Fix spelling error in endpoint\n" }, { "from": "human", "value": "=== Get the Clusters that Have Command Available\n\n==== Description\n\nSearch the clusters that the given command is linked to.\n\n==== Endpoint\n\n`GET /api/v3/applications/{id}/commands`\n\n:snippet-base: {snippets}/command-rest-controller-integration-test/can-get-clusters-for-command/8\n:id-base: get-clusters-for-command\n:!request-headers:\n:request-path-params: {snippet-base}/path-parameters.adoc\n:request-query-params: {snippet-base}/request-parameters.adoc\n:!request-fields:\n:curl-request: {snippet-base}/curl-request.adoc\n:httpie-request: {snippet-base}/httpie-request.adoc\n:response-headers: {snippet-base}/response-headers.adoc\n:response-fields: {snippet-base}/response-fields.adoc\n:!response-links:\n:http-request: {snippet-base}/http-request.adoc\n:http-response: {snippet-base}/http-response.adoc\n\ninclude::../_apiTemplate.adoc[]\n" }, { "from": "gpt", "value": "=== Get the Clusters that Have Command Available\n\n==== Description\n\nSearch the clusters that the given command is linked to.\n\n==== Endpoint\n\n`GET /api/v3/commands/{id}/clusters`\n\n:snippet-base: {snippets}/command-rest-controller-integration-test/can-get-clusters-for-command/8\n:id-base: get-clusters-for-command\n:!request-headers:\n:request-path-params: {snippet-base}/path-parameters.adoc\n:request-query-params: {snippet-base}/request-parameters.adoc\n:!request-fields:\n:curl-request: {snippet-base}/curl-request.adoc\n:httpie-request: {snippet-base}/httpie-request.adoc\n:response-headers: {snippet-base}/response-headers.adoc\n:response-fields: {snippet-base}/response-fields.adoc\n:!response-links:\n:http-request: {snippet-base}/http-request.adoc\n:http-response: {snippet-base}/http-response.adoc\n\ninclude::../_apiTemplate.adoc[]\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Update changelog related to jws changes.\n" }, { "from": "human", "value": "= Changelog\n\n== Version 0.3.0\n\nDate: 2014-01-18\n\n- First version splitted from monolitic buddy package.\n- No changes from original version.\n" }, { "from": "gpt", "value": "= Changelog\n\n== Version 0.4.0\n\nDate: unreleased\n\n- Add encode/decode functions to JWS/JWT implementation. Them instead of return\n plain value, return a monadic either. That allows granular error reporting\n instead something like nil that not very useful. The previous sign/unsign\n are conserved for backward compatibility but maybe in future will be removed.\n\n-\n\n== Version 0.3.0\n\nDate: 2014-01-18\n\n- First version splitted from monolitic buddy package.\n- No changes from original version.\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Add credits section for included content.\n" }, { "from": "human", "value": "== dda - Django Desktop Application\n\n=== Introduction\n\nDDA is an example project to demonstrate how to use Django as a desktop\napplication. It provides an alternative to using a GUI framework such as\nhttps://docs.python.org/3.4/library/tkinter.html[Tkinter] or\nhttp://sourceforge.net/projects/pyqt/[PyQt].\n\nIt is a work in progress.\n" }, { "from": "gpt", "value": "== dda - Django Desktop Application\n\n=== Introduction\n\nDDA is an example project to demonstrate how to use Django as a desktop\napplication. It provides an alternative to using a GUI framework such as\nhttps://docs.python.org/3.4/library/tkinter.html[Tkinter] or\nhttp://sourceforge.net/projects/pyqt/[PyQt].\n\nIt is a work in progress.\n\n=== Credits\n\nTo allow the Django Desktop Application to access common javascript and css\nfiles used by the web pages (without having to hit the network), I have\nincluded copies in this distribution. Here is a list:\n\n - http://jquery.com/[jQuery (Version 2.1.1)]\n - http://getbootstrap.com/[Bootstrap (Version 3.3.1)]\n - http://glyphicons.com/[Glyphicons] (Graciously donated to Bootstrap)\n - http://silviomoreto.github.io/bootstrap-select/[Bootstrap Select (Version 1.6.3)]\n\nThey are all awesome.\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Add docs for how to use the capsule command" }, { "from": "human", "value": "= Creating an Uberjar\n\nEdge provides a script for running pack to build an uberjar.\nTo use it, you can simply run the below from your project sub-directory (the same folder as your deps.edn).\n\n[source,shell]\n----\n$ ../bin/onejar -A:prod --args '-m edge.main' project.jar\n----\n\nThe `-A:prod` indicates an alias you would like to have its `:extra-deps` and `:paths` included in your resulting jar.\n`--args` are default arguments to your jar, in this case we are specifying that the application should run edge.main, part of the edge production modules.\n\nYou can run this jar in production quite easily:\n\n[source,shell]\n----\n$ java -Xmx1G -jar project.jar\n----\n\nWe recommend that you specify the memory usage of your JVM, as the default on Java 6+ is usually insufficient for hosts run only this JVM process.\nA rule of thumb is to use 2/3rds of the memory of your host.\n" }, { "from": "gpt", "value": "= Creating an Uberjar\n\n== OneJar\n\nEdge provides a script for running pack to build an uberjar.\nTo use it, you can simply run the below from your project sub-directory (the same folder as your deps.edn).\n\n[source,shell]\n----\n$ ../bin/onejar -A:prod --args '-m edge.main' project.jar\n----\n\nThe `-A:prod` indicates an alias you would like to have its `:extra-deps` and `:paths` included in your resulting jar.\n`--args` are default arguments to your jar, in this case we are specifying that the application should run edge.main, part of the edge production modules.\n\n== Capsule \n\nBuild a capsule uberjar using this command: \n\n[source,shell]\n----\n$ ../bin/capsule -m edge.main -e 'target/prod' -A:prod project.jar\n----\n\nIf you are using clojurescript in your project make sure you run this command first: \n\n[source,shell]\n----\n$ clojure -A:build:build/once\n----\n\nIt will build the clojurescript files and put them in the correct folder to be included in the uberjar. \n\n== Running the jar\n\nYou can run the produced jar in production quite easily:\n\n[source,shell]\n----\n$ java -Xmx1G -jar project.jar\n----\n\nWe recommend that you specify the memory usage of your JVM, as the default on Java 6+ is usually insufficient for hosts run only this JVM process.\nA rule of thumb is to use 2/3rds of the memory of your host.\n\n\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Update readmy with dependency versions" }, { "from": "human", "value": "= Callfire API v2 REST client\n\nC# client for Callfire platform API version 2. See link:https://developers.callfire.com/callfire-api-client-csharp.html[Getting Started]\n page for setup instructions.\n\n.*Requirements:*\n* .NET 4.6+\n\n.*Dependencies:*\n* Newtonsoft.Json >= 7.0.1\n* RestSharp >= 106.4.2\n\n.*Table of contents*\n* link:https://developers.callfire.com/callfire-api-client-csharp.html[Getting Started]\n* link:https://developers.callfire.com/docs.html[REST endpoints documentation and api code samples]\n* Have a question ?\n** link:https://developers.callfire.com/chat.html[Public chat room]\n** link:http://stackoverflow.com/questions/tagged/callfire[Ask on stackoverflow]\n** link:https://answers.callfire.com/hc/en-us[Call Us]\n" }, { "from": "gpt", "value": "= Callfire API v2 REST client\n\nC# client for Callfire platform API version 2. See link:https://developers.callfire.com/callfire-api-client-csharp.html[Getting Started]\n page for setup instructions.\n\n.*Requirements:*\n* .NET 4.6+\n\n.*Dependencies:*\n* Newtonsoft.Json >= 12.0.2\n* RestSharp = 106.6.4 (RestSharp have a bug with serialization to camelCase json string, wait for fix https://github.com/restsharp/RestSharp/issues/1251 before upgrade to newer)\n\n.*Table of contents*\n* link:https://developers.callfire.com/callfire-api-client-csharp.html[Getting Started]\n* link:https://developers.callfire.com/docs.html[REST endpoints documentation and api code samples]\n* Have a question ?\n** link:https://developers.callfire.com/chat.html[Public chat room]\n** link:http://stackoverflow.com/questions/tagged/callfire[Ask on stackoverflow]\n** link:https://answers.callfire.com/hc/en-us[Call Us]\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Update to OptaPlanner scrum board" }, { "from": "human", "value": "= Report an issue\n:awestruct-layout: normalBase\n:showtitle:\n\n== Issue tracker\n\nWe welcome issue reports (bugs, improvements, new feature requests, ...) in our issue tracker:\n\n*Show https://issues.jboss.org/browse/drools[the JIRA issue tracker].*\n\nLog in and click on the button _Create Issue_ to report a bug, improvement or feature request.\n\n== Pull requests on GitHub\n\nWant to fix the issue yourself? Fork https://github.com/droolsjbpm[the git repository] and send in a pull request.\nWe usually process all pull requests within a few days.\n\n== Kanban boards\n\n* https://issues.jboss.org/secure/RapidBoard.jspa?rapidView=4016[Drools]\n* https://issues.jboss.org/secure/RapidBoard.jspa?rapidView=3828[OptaPlanner]\n* https://issues.jboss.org/secure/RapidBoard.jspa?rapidView=3972[jBPM]\n* https://issues.jboss.org/secure/RapidBoard.jspa?rapidView=3462[Guvnor]\n* AppFormer (todo)\n* https://issues.jboss.org/secure/RapidBoard.jspa?rapidView=3838[Designer.NEXT]\n" }, { "from": "gpt", "value": "= Report an issue\n:awestruct-layout: normalBase\n:showtitle:\n\n== Issue tracker\n\nWe welcome issue reports (bugs, improvements, new feature requests, ...) in our issue tracker:\n\n*Show https://issues.jboss.org/browse/drools[the JIRA issue tracker].*\n\nLog in and click on the button _Create Issue_ to report a bug, improvement or feature request.\n\n== Pull requests on GitHub\n\nWant to fix the issue yourself? Fork https://github.com/droolsjbpm[the git repository] and send in a pull request.\nWe usually process all pull requests within a few days.\n\n== Kanban boards\n\n* https://issues.jboss.org/secure/RapidBoard.jspa?rapidView=4016[Drools]\n* https://issues.jboss.org/secure/RapidBoard.jspa?rapidView=4032[OptaPlanner]\n* https://issues.jboss.org/secure/RapidBoard.jspa?rapidView=3972[jBPM]\n* https://issues.jboss.org/secure/RapidBoard.jspa?rapidView=3462[Guvnor]\n* AppFormer (todo)\n* https://issues.jboss.org/secure/RapidBoard.jspa?rapidView=3838[Designer.NEXT]\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Fix the documentation of REST API link\n" }, { "from": "human", "value": "= Components\n:page-sidebar: apim_sidebar\n:page-permalink: apim_overview_components.html\n:page-folder: apim/overview\n\nNext sections are describing top components which are part of Gravitee.io.\n\n== Gateway\nThe gateway is the core component of the Gravitee.io platform. You can compare it to a \"smart\" proxy to understand its goal.\n\nUnlike traditional HTTP proxy, the gateway is able to apply <<apim_overview_plugins.adoc#gravitee-plugins-policies, policies>> (ie. rules) to\nboth HTTP requests and responses according to your needs, meaning that you can enhance requests and responses processing\nby adding transformation, security, and many other crazy features!\n\n.Gravitee.io - Gateway Internal\nimage::architecture/graviteeio-gateway-internal.png[Gravitee.io - Gateway Internal]\n\n.Gravitee.io - Gateway Architecture\nimage::architecture/graviteeio-gateway-architecture.png[Gravitee.io - Gateway Architecture]\n\n== Management API\nA restful API providing a bunch of services to manage and configure the global platform. All exposed\nservices are constraint by authentication and authorization rules.\nYou can find more information by having a look to the <<gravitee-management-api-overview, Management API>> section.\n\n== Management UI\nThis is the Web UI for <<gravitee-components-mgmt-api, Gravitee.io Management API>>. This UI acts as a management tool\nfor API Publishers but also as a portal for your API Consumers.\n\n" }, { "from": "gpt", "value": "= Components\n:page-sidebar: apim_sidebar\n:page-permalink: apim_overview_components.html\n:page-folder: apim/overview\n\nNext sections are describing top components which are part of Gravitee.io.\n\n== Gateway\nThe gateway is the core component of the Gravitee.io platform. You can compare it to a \"smart\" proxy to understand its goal.\n\nUnlike traditional HTTP proxy, the gateway is able to apply <<apim_overview_plugins.adoc#gravitee-plugins-policies, policies>> (ie. rules) to\nboth HTTP requests and responses according to your needs, meaning that you can enhance requests and responses processing\nby adding transformation, security, and many other crazy features!\n\n.Gravitee.io - Gateway Internal\nimage::architecture/graviteeio-gateway-internal.png[Gravitee.io - Gateway Internal]\n\n.Gravitee.io - Gateway Architecture\nimage::architecture/graviteeio-gateway-architecture.png[Gravitee.io - Gateway Architecture]\n\n== Management API\nA restful API providing a bunch of services to manage and configure the global platform. All exposed\nservices are constraint by authentication and authorization rules.\nYou can find more information by having a look to the link:/apim_installguide_management_api_documentation.html[Management API > Documentation] section.\n\n== Management UI\nThis is the Web UI for <<gravitee-components-mgmt-api, Gravitee.io Management API>>. This UI acts as a management tool\nfor API Publishers but also as a portal for your API Consumers.\n\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Remove done job on todo list\n" }, { "from": "human", "value": "= Talend component for Neo4j\nBenoit Simard <[email protected]>\nV1.0\n:experimental:\n:toc:\n:toc-placement: preamble\n:toc-title: pass:[<b>Table of Contents</b>]\n:outfilesuffix-old: {outfilesuffix}\nifdef::env-github[:outfilesuffix: .adoc]\nifndef::env-github[]\n:idprefix:\n:idseparator: -\nendif::[]\n\nimage:https://travis-ci.org/sim51/neo4j-talend-component.svg?branch=master[\"Build Status\", link=\"https://travis-ci.org/sim51/neo4j-talend-component\"]\n\nimage:https://coveralls.io/repos/github/sim51/neo4j-talend-component/badge.svg?branch=master[\"Coverage Status\", link=\"https://coveralls.io/github/sim51/neo4j-talend-component?branch=master\"]\n\nProject website : http://sim51.github.io/neo4j-talend-component/\n\n== TODO\n\n* Remake packaging phase for the assembly (there is some cobertura thing into compile class on the archive)\n\n" }, { "from": "gpt", "value": "= Talend component for Neo4j\nBenoit Simard <[email protected]>\nV1.0\n:experimental:\n:toc:\n:toc-placement: preamble\n:toc-title: pass:[<b>Table of Contents</b>]\n:outfilesuffix-old: {outfilesuffix}\nifdef::env-github[:outfilesuffix: .adoc]\nifndef::env-github[]\n:idprefix:\n:idseparator: -\nendif::[]\n\nimage:https://travis-ci.org/sim51/neo4j-talend-component.svg?branch=master[\"Build Status\", link=\"https://travis-ci.org/sim51/neo4j-talend-component\"]\n\nimage:https://coveralls.io/repos/github/sim51/neo4j-talend-component/badge.svg?branch=master[\"Coverage Status\", link=\"https://coveralls.io/github/sim51/neo4j-talend-component?branch=master\"]\n\nProject website : http://sim51.github.io/neo4j-talend-component/\n\n== TODO\n\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Add deleted html example link\n" }, { "from": "human", "value": "== asciidoctor-fopub-embed-svg-example\n\nAn example for embedding SVG images in an asciidoctor document.\n\n=== Setup\n\n----\nbundle install\n----\n\n=== How to make the target pdf file\n\nRunning the command below make the target pdf file named \"class_diagram.pdf\" out of \"class_diagram.adoc\"\n\n----\nbundle exec rake\n----\n\n=== Watch *.adoc files\n\nRunning the command below, rake watches \"\\*.adoc\" files and build \"*.adoc\" files whenever they are modified.\n\n----\nbundle exec rake watch\n----\n\n=== Example output files\n\nYou can see the example output files at\n\n* https://hnakamur.github.io/asciidoctor-fopub-embed-svg-example/class_diagram.pdf[class_diagram.pdf]\n\n=== Note\n\nYou must be add `width` and `height` attributes to the root `<svg>` element.\nWithout these, you got redundant spaces between the embedded image and the following text.\n" }, { "from": "gpt", "value": "== asciidoctor-fopub-embed-svg-example\n\nAn example for embedding SVG images in an asciidoctor document.\n\n=== Setup\n\n----\nbundle install\n----\n\n=== How to make the target pdf file\n\nRunning the command below make the target pdf file named \"class_diagram.pdf\" out of \"class_diagram.adoc\"\n\n----\nbundle exec rake\n----\n\n=== Watch *.adoc files\n\nRunning the command below, rake watches \"\\*.adoc\" files and build \"*.adoc\" files whenever they are modified.\n\n----\nbundle exec rake watch\n----\n\n=== Example output files\n\nYou can see the example output files at\n\n* https://hnakamur.github.io/asciidoctor-fopub-embed-svg-example/class_diagram.pdf[class_diagram.pdf]\n* https://hnakamur.github.io/asciidoctor-fopub-embed-svg-example/class_diagram.html[class_diagram.html]\n\n=== Note\n\nYou must be add `width` and `height` attributes to the root `<svg>` element.\nWithout these, you got redundant spaces between the embedded image and the following text.\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Update code signing commands for macOS" }, { "from": "human", "value": "== Source release to pypi\n\n python setup.py release --skip-tests --pypi\n\n== Sign source release for GitHub\n\n gpg --detach-sign dist/yubikey-manager-VERSION.tar.gz\n\n== Releasing on macOS\n\n python setup.py executable\n resources/fix_libusb_dep_app_bundle.sh dist/YubiKey\\ Manager.app/\n codesign -s 'Developer ID Application' dist/YubiKey\\ Manager.app/ --deep\n sudo pkgbuild --install-location /Applications --component dist/YubiKey\\ Manager.app/ dist/installer.pkg\n productsign -s 'Developer ID Installer' dist/installer.pkg dist/yubikey-manager-VERSION-mac.pkg\n gpg --detach-sign dist/yubikey-manager-VERSION-mac.pkg\n\n== Releasing on Windows\n\n python setup.py executable\n gpg --detach-sign dist/yubikey-manager-VERSION-win.exe\n" }, { "from": "gpt", "value": "== Source release to pypi\n\n python setup.py release --skip-tests --pypi\n\n== Sign source release for GitHub\n\n gpg --detach-sign dist/yubikey-manager-VERSION.tar.gz\n\n== Releasing on macOS\n\n python setup.py executable\n resources/fix_libusb_dep_app_bundle.sh dist/YubiKey\\ Manager.app/\n codesign --deep --force --verify --verbose --sign 'Developer ID Application' dist/YubiKey\\ Manager.app/\n sudo pkgbuild --sign 'Developer ID Installer' --install-location /Applications --component dist/YubiKey\\ Manager.app/ dist/installer.pkg\n gpg --detach-sign dist/yubikey-manager-VERSION-mac.pkg\n\n== Releasing on Windows\n\n python setup.py executable\n gpg --detach-sign dist/yubikey-manager-VERSION-win.exe\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Update main document of 'Implementation protocol'.\n" }, { "from": "human", "value": "= Implementation protocols\nFranz Wilhelmstötter <[email protected]>\n:Author Initials: FWi\n:toc:\n:icons:\n:numbered:\n:website: http://jenetics.io/\n\nWhile implementing the library, a lot of concrete implementation decisions have been made. This document gives a more detailed description of some selected implementation details.\n\ninclude::CreateRandomSeed.adoc[]\n\ninclude::OptimizeMinimum.adoc[]\n" }, { "from": "gpt", "value": "= Implementation protocols\nFranz Wilhelmstötter <[email protected]>\n:Author Initials: FWi\n:toc:\n:icons:\n:numbered:\n:website: http://jenetics.io/\n\nWhile implementing the library, a lot of concrete implementation decisions have been made. This document gives a more detailed description of selected implementation details.\n\ninclude::CreateRandomSeed.adoc[]\n\ninclude::OptimizeMinimum.adoc[]\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Docs: Fix :issue: and :pull: asciidoc replacements\n" }, { "from": "human", "value": "[[elasticsearch-reference]]\n= Elasticsearch Reference\n\n:version: 2.2.0\n:branch: 2.0\n:major-version: 2.x\n:jdk: 1.8.0_25\n:defguide: https://www.elastic.co/guide/en/elasticsearch/guide/current\n:plugins: https://www.elastic.co/guide/en/elasticsearch/plugins/2.x\n:issue: https://github.com/elastic/elasticsearch/issues\n:pull: https://github.com/elastic/elasticsearch/pull\n\ninclude::getting-started.asciidoc[]\n\ninclude::setup.asciidoc[]\n\ninclude::migration/index.asciidoc[]\n\ninclude::api-conventions.asciidoc[]\n\ninclude::docs.asciidoc[]\n\ninclude::search.asciidoc[]\n\ninclude::aggregations.asciidoc[]\n\ninclude::indices.asciidoc[]\n\ninclude::cat.asciidoc[]\n\ninclude::cluster.asciidoc[]\n\ninclude::query-dsl.asciidoc[]\n\ninclude::mapping.asciidoc[]\n\ninclude::analysis.asciidoc[]\n\ninclude::modules.asciidoc[]\n\ninclude::index-modules.asciidoc[]\n\ninclude::testing.asciidoc[]\n\ninclude::glossary.asciidoc[]\n\ninclude::release-notes.asciidoc[]\n\ninclude::redirects.asciidoc[]\n\n\n\n\n" }, { "from": "gpt", "value": "[[elasticsearch-reference]]\n= Elasticsearch Reference\n\n:version: 2.2.0\n:branch: 2.x\n:major-version: 2.x\n:jdk: 1.8.0_25\n:defguide: https://www.elastic.co/guide/en/elasticsearch/guide/current\n:plugins: https://www.elastic.co/guide/en/elasticsearch/plugins/2.x\n:issue: https://github.com/elastic/elasticsearch/issues/\n:pull: https://github.com/elastic/elasticsearch/pull/\n\ninclude::getting-started.asciidoc[]\n\ninclude::setup.asciidoc[]\n\ninclude::migration/index.asciidoc[]\n\ninclude::api-conventions.asciidoc[]\n\ninclude::docs.asciidoc[]\n\ninclude::search.asciidoc[]\n\ninclude::aggregations.asciidoc[]\n\ninclude::indices.asciidoc[]\n\ninclude::cat.asciidoc[]\n\ninclude::cluster.asciidoc[]\n\ninclude::query-dsl.asciidoc[]\n\ninclude::mapping.asciidoc[]\n\ninclude::analysis.asciidoc[]\n\ninclude::modules.asciidoc[]\n\ninclude::index-modules.asciidoc[]\n\ninclude::testing.asciidoc[]\n\ninclude::glossary.asciidoc[]\n\ninclude::release-notes.asciidoc[]\n\ninclude::redirects.asciidoc[]\n\n\n\n\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Update docs to reflect dev->repl\n" }, { "from": "human", "value": "= Edge App\n\nUntil CIDER 0.17 has been released (which brings improved support for\nClojure CLI), please use a terminal to run `clj`.\n\nEnsure that you are in the `app` sub-directory.\n\n----\n$ ./bin/dev\ndev=> (go)\n----\n\nThis starts an nrepl-server, the port is announced on the command line.\n\n== Cursive support\n\nCurrent versions of Cursive don't support `deps.edn`, but will work with a `pom.xml`.\n\nYou can generate the `pom.xml` with the following command:\n\n----\n$ clojure -Spom\n----\n" }, { "from": "gpt", "value": "= Edge App\n\nUntil CIDER 0.17 has been released (which brings improved support for\nClojure CLI), please use a terminal to run `clj`.\n\nEnsure that you are in the `app` sub-directory.\n\n----\n$ ./bin/repl\ndev=> (go)\n----\n\nThis starts an nrepl-server, the port is announced on the command line.\n\n== Cursive support\n\nCurrent versions of Cursive don't support `deps.edn`, but will work with a `pom.xml`.\n\nYou can generate the `pom.xml` with the following command:\n\n----\n$ clojure -Spom\n----\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Document default user is root\n" }, { "from": "human", "value": "\n# Conventions over configuration\n\nThis document covers conventions used in Dogen that may help you build Dockerfiles faster and easier.\n\n## Handling YUM/DNF repo files\n\nIn case you add YUM/DNF repo files that you want to put to the `/etc/yum.repos.d` directory\nto enable additional repositories Dogen will handle it for you automatically.\n\nIf Dogen will detect files\nwith `.repo` extension in the `scripts` directory located under the target (output) directory Dogen will\nautomatically copy all these files to `/etc/yum.repos.d` directory and enable them to be used while\ninstalling packages listed in the `packages` section.\n\nAfter installation Dogen removes newly added repo files from the `/etc/yum.repos.d` directory *automatically*.\nIf you do not want to have these files removed after installation -- you need to make your repo files\npart of some package and install them manually.\n\n### Things to note\n\n1. This feature covers only the situation where you want to add a custom repo file at *build\n time* but you do not want it to be enabled in containers.\n2. Repo file name should be the *same* as the repo id in the repository\n (the name between square brackets).\n3. There should be *only one* repository per file.\n4. Only added repositories will be enabled during install of packages, all other repositories (including default)\n will be *disabled*.\n\n## Running scripts\n\n### Default script name\n\nIf you do not specify an `exec` key for a script, dogen will check for a script named `run`\nand execute it if it exists.\n" }, { "from": "gpt", "value": "\n# Conventions over configuration\n\nThis document covers conventions used in Dogen that may help you build Dockerfiles faster and easier.\n\n## Handling YUM/DNF repo files\n\nIn case you add YUM/DNF repo files that you want to put to the `/etc/yum.repos.d` directory\nto enable additional repositories Dogen will handle it for you automatically.\n\nIf Dogen will detect files\nwith `.repo` extension in the `scripts` directory located under the target (output) directory Dogen will\nautomatically copy all these files to `/etc/yum.repos.d` directory and enable them to be used while\ninstalling packages listed in the `packages` section.\n\nAfter installation Dogen removes newly added repo files from the `/etc/yum.repos.d` directory *automatically*.\nIf you do not want to have these files removed after installation -- you need to make your repo files\npart of some package and install them manually.\n\n### Things to note\n\n1. This feature covers only the situation where you want to add a custom repo file at *build\n time* but you do not want it to be enabled in containers.\n2. Repo file name should be the *same* as the repo id in the repository\n (the name between square brackets).\n3. There should be *only one* repository per file.\n4. Only added repositories will be enabled during install of packages, all other repositories (including default)\n will be *disabled*.\n\n## Running scripts\n\n### Default script name\n\nIf you do not specify an `exec` key for a script, dogen will check for a script named `run`\nand execute it if it exists.\n\n### Default script execution user\n\nDogen will execute all scripts as `root` unless you have specified a user with the `user` key.\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Fix the addition of Rabbit DLQ documentation\n" }, { "from": "human", "value": "include::{stream-docs-basedir}/core/spring-cloud-stream-core-docs/src/main/asciidoc/spring-cloud-stream-overview.adoc[]\n\n= Binder Implementations\n\n== Apache Kafka Binder\n\ninclude::{stream-docs-basedir}/kafka/spring-cloud-stream-binder-kafka-docs/src/main/asciidoc/overview.adoc[leveloffset=+1]\n\n== RabbitMQ Binder\n\ninclude::{stream-docs-basedir}/rabbit/spring-cloud-stream-binder-rabbit-docs/src/main/asciidoc/overview.adoc[leveloffset=+1]\n" }, { "from": "gpt", "value": "include::{stream-docs-basedir}/core/spring-cloud-stream-core-docs/src/main/asciidoc/spring-cloud-stream-overview.adoc[]\n\n= Binder Implementations\n\n== Apache Kafka Binder\n\ninclude::{stream-docs-basedir}/kafka/spring-cloud-stream-binder-kafka-docs/src/main/asciidoc/overview.adoc[leveloffset=+1]\n\n== RabbitMQ Binder\n\ninclude::{stream-docs-basedir}/rabbit/spring-cloud-stream-binder-rabbit-docs/src/main/asciidoc/overview.adoc[leveloffset=+1]\ninclude::{stream-docs-basedir}/rabbit/spring-cloud-stream-binder-rabbit-docs/src/main/asciidoc/dlq.adoc[leveloffset=+1]\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Fix example for ingress sharding by namespace\n\nFollow-up to commit 69c879f8edaceb4f679eca3f475f0d9851e460c6.\n\n* modules/nw-ingress-sharding-namespace-labels.adoc: Change routeSelector\nto namespaceSelector.\n" }, { "from": "human", "value": "// Module included in the following assemblies:\n//\n// * configuring_ingress_cluster_traffic/configuring-ingress-cluster-traffic-ingress-controller.adoc\n// * ingress-operator.adoc\n\n[id=\"nw-ingress-sharding-namespace-labels_{context}\"]\n= Configuring ingress controller sharding by using namespace labels\n\nIngress Controller sharding by using namespace labels means that the Ingress\nController serves any route in any namespace that is selected by the namespace\nselector.\n\nIngress Controller sharding is useful when balancing incoming traffic load among\na set of Ingress Controllers and when isolating traffic to a specific Ingress\nController. For example, company A goes to one Ingress Controller and company B\nto another.\n\n.Procedure\n\n. Edit the `router-internal.yaml` file:\n+\n----\n# cat router-internal.yaml\napiVersion: v1\nitems:\n- apiVersion: operator.openshift.io/v1\n kind: IngressController\n metadata:\n name: sharded\n namespace: openshift-ingress-operator\n spec:\n domain: <apps-sharded.basedomain.example.net>\n nodePlacement:\n nodeSelector:\n matchLabels:\n node-role.kubernetes.io/worker: \"\"\n routeSelector:\n matchLabels:\n type: sharded\n status: {}\nkind: List\nmetadata:\n resourceVersion: \"\"\n selfLink: \"\"\n----\n\n. Apply the Ingress Controller `router-internal.yaml` file:\n+\n----\n# oc apply -f router-internal.yaml\n----\n+\nThe Ingress Controller selects routes in any namespace that is selected by the\nnamespace selector that have the label `type: sharded`.\n" }, { "from": "gpt", "value": "// Module included in the following assemblies:\n//\n// * configuring_ingress_cluster_traffic/configuring-ingress-cluster-traffic-ingress-controller.adoc\n// * ingress-operator.adoc\n\n[id=\"nw-ingress-sharding-namespace-labels_{context}\"]\n= Configuring ingress controller sharding by using namespace labels\n\nIngress Controller sharding by using namespace labels means that the Ingress\nController serves any route in any namespace that is selected by the namespace\nselector.\n\nIngress Controller sharding is useful when balancing incoming traffic load among\na set of Ingress Controllers and when isolating traffic to a specific Ingress\nController. For example, company A goes to one Ingress Controller and company B\nto another.\n\n.Procedure\n\n. Edit the `router-internal.yaml` file:\n+\n----\n# cat router-internal.yaml\napiVersion: v1\nitems:\n- apiVersion: operator.openshift.io/v1\n kind: IngressController\n metadata:\n name: sharded\n namespace: openshift-ingress-operator\n spec:\n domain: <apps-sharded.basedomain.example.net>\n nodePlacement:\n nodeSelector:\n matchLabels:\n node-role.kubernetes.io/worker: \"\"\n namespaceSelector:\n matchLabels:\n type: sharded\n status: {}\nkind: List\nmetadata:\n resourceVersion: \"\"\n selfLink: \"\"\n----\n\n. Apply the Ingress Controller `router-internal.yaml` file:\n+\n----\n# oc apply -f router-internal.yaml\n----\n+\nThe Ingress Controller selects routes in any namespace that is selected by the\nnamespace selector that have the label `type: sharded`.\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Update user guide copyright date.\n" }, { "from": "human", "value": "== License\n\n+++<a href=\"http://deepsymmetry.org\"><img src=\"_images/DS-logo-bw-200-padded-left.png\" align=\"right\" alt=\"Deep Symmetry logo\" width=\"216\" height=\"123\"></a>+++\nCopyright © 2015&ndash;2019 http://deepsymmetry.org[Deep Symmetry, LLC]\n\nDistributed under the https://opensource.org/licenses/EPL-2.0[Eclipse\nPublic License 2.0]. By using this software in any fashion, you are\nagreeing to be bound by the terms of this license. You must not remove\nthis notice, or any other, from this software. A copy of the license\ncan be found in\nhttps://github.com/Deep-Symmetry/afterglow/blob/master/LICENSE[LICENSE]\nwithin this project.\n" }, { "from": "gpt", "value": "== License\n\n+++<a href=\"http://deepsymmetry.org\"><img src=\"_images/DS-logo-bw-200-padded-left.png\" align=\"right\" alt=\"Deep Symmetry logo\" width=\"216\" height=\"123\"></a>+++\nCopyright © 2015&ndash;2020 http://deepsymmetry.org[Deep Symmetry, LLC]\n\nDistributed under the https://opensource.org/licenses/EPL-2.0[Eclipse\nPublic License 2.0]. By using this software in any fashion, you are\nagreeing to be bound by the terms of this license. You must not remove\nthis notice, or any other, from this software. A copy of the license\ncan be found in\nhttps://github.com/Deep-Symmetry/afterglow/blob/master/LICENSE[LICENSE]\nwithin this project.\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Correct typo in common warning" }, { "from": "human", "value": "WARNING: Be sure to add the dependency to the `dependencies` section of the `build.gradle` file, and not the `buildscript`/`dependencies` section. The former is for application dependencies (needed for at compile time, runtime, or testing), whereas the `buildscript` `dependencies` is for those needed as part of the Gradle build process (managing static assets, for example).\n" }, { "from": "gpt", "value": "WARNING: Be sure to add the dependency to the `dependencies` section of the `build.gradle` file, and not the `buildscript`/`dependencies` section. The former is for application dependencies (needed at compile time, runtime, or for testing), whereas the `buildscript` `dependencies` is for those needed as part of the Gradle build process (managing static assets, for example).\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Add a thank-you message to the release notes.\n" }, { "from": "human", "value": "== New Releases Restoring Compatibility\n\nRevapi has started to use Java 9 for building while maintaining compatibility with Java 8 using the Animal Sniffer Maven\nplugin. This has proven to be problematic and the link:20171015-releases.html[last] releases could not be used with\nJava 8. This should be now fixed and compatibility with Java 8 restored.\n\nThe following were released:\n\n* Revapi API 0.8.4\n* Revapi Basic Features 0.6.3\n* Revapi Ant Task 0.6.3\n* Revapi Java SPI 0.14.0\n* Revapi Java Extension 0.14.3\n* Revapi Text Reporter 0.8.2\n* Revapi Standalone 0.6.4\n* Revapi Maven Plugin 0.9.4\n\nIn addition to the above fix, a new artifact is now released `revapi-java-test-support` starting at version `0.1.0`.\nThis artifact is supposed to help custom extensions with testing by providing simple utilities to build, compile and\ninspect custom jar files.\n" }, { "from": "gpt", "value": "== New Releases Restoring Compatibility\n\nRevapi has started to use Java 9 for building while maintaining compatibility with Java 8 using the Animal Sniffer Maven\nplugin. This has proven to be problematic and the link:20171015-releases.html[last] releases could not be used with\nJava 8. This should be now fixed and compatibility with Java 8 restored.\n\nThe following were released:\n\n* Revapi API 0.8.4\n* Revapi Basic Features 0.6.3\n* Revapi Ant Task 0.6.3\n* Revapi Java SPI 0.14.0\n* Revapi Java Extension 0.14.3\n* Revapi Text Reporter 0.8.2\n* Revapi Standalone 0.6.4\n* Revapi Maven Plugin 0.9.4\n\nIn addition to the above fix, a new artifact is now released `revapi-java-test-support` starting at version `0.1.0`.\nThis artifact is supposed to help custom extensions with testing by providing simple utilities to build, compile and\ninspect custom jar files.\n\nThanks go out to Petr Široký who found, diagnosed and fixed the compatibility issue. Petr, you're a star!\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Update readme to include installation. Add bundler: exec needs a command to run to the Converting to HTML...\n -- HTML output at output/master.html\nConverting to EPub...\n -- Epub output at output/master.epub\nConverting to Mobi (kf8)...\n -- Mobi output at output/master.mobi\nConverting to PDF... (this one takes a while)\n -- PDF output at output/master.pdf command.\n" }, { "from": "human", "value": "= AsciiDoc Book Starter\n\nThis is the starter folder structure for writing book with AsciiDoc.\n\nInspired by https://github.com/progit/progit2[Pro Git 2].\n\n== Building\n\nrun `rake` to build book.\n\nrun `rake book:build_chapter_html` to build each chapter into HTML file.\n\n== Lincense\n\nThe MIT License\n" }, { "from": "gpt", "value": "= AsciiDoc Book Starter\n\nThis is the starter folder structure for writing book with AsciiDoc.\n\nInspired by https://github.com/progit/progit2[Pro Git 2].\n\n== Installation\n\nInstall the gems via `bundle install`.\n\n== Building\n\nrun `bundle exec rake` to build book.\n\nrun `bundle exec rake book:build_chapter_html` to build each chapter into HTML file.\n\n== Lincense\n\nThe MIT License\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Update documentation version attributes [ci skip]\n" }, { "from": "human", "value": ":moduledir: ..\n:attachmentsdir: {moduledir}/assets/attachments\n:examplesdir: {moduledir}/examples\n:imagesdir: {moduledir}/assets/images\n:partialsdir: {moduledir}/pages/_partials\n\n:debezium-version: 0.10.0.Beta4\n:debezium-dev-version: 0.10\n:debezium-kafka-version: 2.3.0\n:debezium-docker-label: 0.10\n:install-version: 0.10\n:confluent-platform-version: 5.1.2\n:strimzi-version: 0.9.0\n" }, { "from": "gpt", "value": ":moduledir: ..\n:attachmentsdir: {moduledir}/assets/attachments\n:examplesdir: {moduledir}/examples\n:imagesdir: {moduledir}/assets/images\n:partialsdir: {moduledir}/pages/_partials\n\n:debezium-version: 1.0.0.Beta2\n:debezium-dev-version: 1.0\n:debezium-kafka-version: 2.3.0\n:debezium-docker-label: 1.0\n:install-version: 1.0\n:confluent-platform-version: 5.1.2\n:strimzi-version: 0.9.0\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Add Neo4j-OGM version in short description.\n" }, { "from": "human", "value": "= Neo4j-OGM - An Object Graph Mapping Library for Neo4j\n\n\nifdef::backend-html5[(C) {copyright}]\nifndef::backend-pdf[]\n\nLicense: link:{common-license-page-uri}[Creative Commons 4.0]\nendif::[]\nifdef::backend-pdf[]\n(C) {copyright}\n\nLicense: <<license, Creative Commons 4.0>>\nendif::[]\n\n[abstract]\n--\nThis is the Neo4j object-graph mapping (OGM) manual, authored by the Neo4j team.\n--\n\nThe three parts of the manual are:\n\n* <<introduction>> -- Introducing graph database concepts, Neo4j and object-graph mapping.\n* <<tutorial>> -- Follow along as you get started using Neo4j-OGM.\n* <<reference>> -- Reference documentation for Neo4j-OGM.\n\nBut before starting, let's see the most important new features.\n\ninclude::introduction/index.adoc[leveloffset=+1]\n\ninclude::tutorial/index.adoc[leveloffset=+1]\n\ninclude::reference/index.adoc[leveloffset=+1]\n\n:numbered!:\n\ninclude::appendix/migration-2.1-to-3.0.adoc[leveloffset=+1]\n\ninclude::appendix/design-considerations.adoc[leveloffset=+1]\n\ninclude::appendix/faq.adoc[leveloffset=+1]\n\n\nifdef::backend-pdf[]\ninclude::{license-dir}/license.adoc[leveloffset=1]\nendif::[]\n" }, { "from": "gpt", "value": "= Neo4j-OGM - An Object Graph Mapping Library for Neo4j\n\n\nifdef::backend-html5[(C) {copyright}]\nifndef::backend-pdf[]\n\nLicense: link:{common-license-page-uri}[Creative Commons 4.0]\nendif::[]\nifdef::backend-pdf[]\n(C) {copyright}\n\nLicense: <<license, Creative Commons 4.0>>\nendif::[]\n\n[abstract]\n--\nThis is the manual of Neo4j object-graph mapping (OGM) v{ogm-doc-version}, authored by the Neo4j team.\n--\n\nThe three parts of the manual are:\n\n* <<introduction>> -- Introducing graph database concepts, Neo4j and object-graph mapping.\n* <<tutorial>> -- Follow along as you get started using Neo4j-OGM.\n* <<reference>> -- Reference documentation for Neo4j-OGM.\n\nBut before starting, let's see the most important new features.\n\ninclude::introduction/index.adoc[leveloffset=+1]\n\ninclude::tutorial/index.adoc[leveloffset=+1]\n\ninclude::reference/index.adoc[leveloffset=+1]\n\n:numbered!:\n\ninclude::appendix/migration-2.1-to-3.0.adoc[leveloffset=+1]\n\ninclude::appendix/design-considerations.adoc[leveloffset=+1]\n\ninclude::appendix/faq.adoc[leveloffset=+1]\n\n\nifdef::backend-pdf[]\ninclude::{license-dir}/license.adoc[leveloffset=1]\nendif::[]\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Split into JSF and REST\n" }, { "from": "human", "value": "= TodoService - World's least awesome Web View and REST Service for TODO\n\nThis will be a nice working TODO app by 2022..\nFor now it provides a trivial ToDo lister/editor for one person via HTML (JSF).\n\nUses ../TodoModel for its JPA Data; do mvn install in that project first.\n\nIt will someday soon also provide a REST interface, which will be awesome enough\nthat there will also be an Android Synch Service to go with it.\n\nYou need to copy config-sample.properties to config.properties, and edit it.\n\nAssuming you're using Wildfly 8 or 9 and PostGres 9,, just set up the datasource and do `mvn wildfly:deploy`\n\n\nThings do do.\n\nBetter UI!\n\nMove away from container-managed security so it can more easily become multi-user.\n" }, { "from": "gpt", "value": "= TodoService - World's least awesome REST Service for TODO\n\nThis will be a nice working TODO app by 2022..\nFor now it provides a trivial ToDo lister/editor for one person via HTML (JSF).\n\nUses ../TodoModel for its JPA Data; do mvn install in that project first.\n\nYou need to copy config-sample.properties to config.properties, and edit it.\n\nAssuming you're using Wildfly 8 or 9 and PostGreSQL 9, just set up the datasource and do `mvn wildfly:deploy`\n\n== Things do do.\n\nMove away from container-managed security so it can more easily become multi-user.\n\n== BUGS\n\nThe only reason this is split off from TodoJSF is that Java EE doesn't allow two\nlogin-config elements, we need FORM for the JSF and BASIC for the REST. Sigh...\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Update the readme for github\n" }, { "from": "human", "value": "= Docker test training\n\n== Prerequisites\n* Hebben ze docker geinstalleerd? Zo nee, doen!\n\n== Doel van de presentatie\n* Basisconcepten van Docker uitleggen\n* Simpele voorbeelden van containers\n* Workshop met als doel zelf webpagina met DB backend testen mbv Docker\n\n=== Vragen wat mensen nog meer zouden willen weten\n\n== Wat is docker?\n* Docker vs VM: de verschillen (-> networking)\n* Docker distributie model (git-achtig)\n* Docker client/server\n* Container file systeem -> volumes\n\n== Simpele voorbeelden:\n* docker run hello-world\n* docker run bash -> docker inspect, docker ps\n* docker run nginx -> in browser laten zien\n* docker run pgsql -> inloggen met pgadmin\n\n== Labs\n* start nginx container met statische inhoud\n** volume mounten, aanpassen etc\n* nginx + selenium, klein testje (java omgeving?)\n* nginx + selenium + database, test dynamische pagina (docker link)\n\n== Extra credit lab\n* Met docker build pgsql image aanpassen dat ie automatisch baseline data inlaadt, die weer gebruiken in de test\n\n" }, { "from": "gpt", "value": "= Docker for testers\n\nThis repository contains training material for testers that want to get started with http://www.docker.com[Docker]. It was developed by engineers at http://www.bol.com[bol.com].\n\n== Building the course material\nNOTE: You must have a recent version of the http://www.oracle.com/technetwork/java/javase/downloads/index.html[JDK] installed to build the course material.\n\nThe course material is written in http://asciidoctor.org/[Asciidoctor] and can be built using the following command:\n\n ./gradlew assemble\n\nThis command will generate a zip file containing all the material in the `workshop/build/distributions` directory." } ]
asciidoc
chosen
[ { "from": "system", "value": "Fix typo; while here, tweak wording, formatting.\n" }, { "from": "human", "value": "= SpringBoot WebApp Demo\n\nhttps://projects.spring.io/spring-boot[Spring Boot]\nlooks like a nice way to get started.\nThis is a trivial webapp created using SpringBoot.\n\n== HowTo\n\n\tmvn spring-boot:run\n\nOr, in Eclipse with Spring plugin (or \"Spring Tool Suite\"), \nright-click project or Application.java, \nthen RunAs->Spring Boot Application\n\nSeconds later, connect to \n\nhttp://localhost:8000/\n\nNote: that is port 8000, _not_ the usual 8080, to avoid conflicts.\nChange this in application.properties if you don't like it.\n\nFill in the form and click Submit.\n\n== War Packaging for EE App Server Deployment\n\nThis version has the modifications described in\nthe section entitled \"Traditional Deployment\"--\"Create a deployable war file\" in\nthe spring-boot reference manual (Section 85.1 as of\nthis writing).\nit still works in Maven and Eclipse with these settings,\nand the packaged was file has a jboss-web.xml config file\nso it deploys as /springbootdemos when you simply copy the war file to\nWILDFLYHOME/standalone/deployments.\n" }, { "from": "gpt", "value": "= SpringBoot WebApp Demo\n\nhttps://projects.spring.io/spring-boot[Spring Boot]\nlooks like a nice way to get started.\nThis is a trivial webapp created using SpringBoot.\n\n== HowTo\n\n\tmvn spring-boot:run\n\nOr, in Eclipse with Spring plugin (or \"Spring Tool Suite\"), \nright-click project or Application.java, \nthen RunAs->Spring Boot Application\n\nSeconds later, connect to \n\nhttp://localhost:8000/\n\nNote: that is port 8000, _not_ the usual 8080, to avoid conflicts.\nChange this in application.properties if you don't like it.\n\nFill in the form and click Submit.\n\n== War Packaging for EE App Server Deployment\n\nThis version has the modifications described in\nthe section entitled \"Traditional Deployment\"--\"Create a deployable war file\" in\nthe spring-boot reference manual (Section 85.1 as of\nthis writing).\nit still works in Maven (`mvn package`) and can be run in Eclipse with these settings.\nFor WildFly users, the packaged _war_ file has a jboss-web.xml config file\nso it deploys as _/springbootdemos_ when you simply copy the _war_ file to\nWILDFLYHOME/standalone/deployments.\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Add badge for SonarQube integration" }, { "from": "human", "value": "= Blueprint\n:author: Hafid Haddouti\n\nimage:https://travis-ci.org/haf-tech/blueprint.svg?branch=master[\"Build Status\", link=\"https://travis-ci.org/haf-tech/blueprint\"] \nimage:https://img.shields.io/badge/License-Apache%202.0-blue.svg[\"License\", link=\"https://opensource.org/licenses/Apache-2.0\"]\n\n\nBlueprint is a playground for illustrating different paradigms.\n\nIn the meantime the following concepts are integrated or planned:\n\n- Spring Boot\n- AsciiDoctor integration, with UML and different outputs\n- Onion architecture\n- Docker build/push\n\nNext:\n\n- reactive programming\n\nAn up to date documentation is https://haf-tech.github.io/blueprint/[online] available.\n" }, { "from": "gpt", "value": "= Blueprint\n:author: Hafid Haddouti\n\nimage:https://travis-ci.org/haf-tech/blueprint.svg?branch=master[\"Build Status\", link=\"https://travis-ci.org/haf-tech/blueprint\"] \nimage:https://img.shields.io/badge/License-Apache%202.0-blue.svg[\"License\", link=\"https://opensource.org/licenses/Apache-2.0\"]\nimage:https://sonarqube.com/api/badges/gate?key=com.haddouti.pg%3Ablueprint-parent[\"Sonar Quality Gate\", link=\"https://sonarqube.com/api/badges/gate?key=com.haddouti.pg%3Ablueprint-parent\"]\n\nBlueprint is a playground for illustrating different paradigms.\n\nIn the meantime the following concepts are integrated or planned:\n\n- Spring Boot\n- AsciiDoctor integration, with UML and different outputs\n- Onion architecture\n- Docker build/push\n\nNext:\n\n- reactive programming\n\nAn up to date documentation is https://haf-tech.github.io/blueprint/[online] available.\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Rebuild master with coverage report\n" }, { "from": "human", "value": "= Tiny Request Limits Library designed for Java 8\n\n:hardbreaks:\n\nimage:https://travis-ci.org/storozhukBM/requestlimit.svg?branch=master[\"Build Status\", link=\"https://travis-ci.org/storozhukBM/requestlimit\"] image:https://coveralls.io/repos/github/storozhukBM/requestlimit/badge.svg[\"Coverage Status\", link=\"https://coveralls.io/github/storozhukBM/requestlimit\"]" }, { "from": "gpt", "value": "= Tiny Request Limits Library designed for Java 8\n\n:hardbreaks:\n\n\nimage:https://travis-ci.org/storozhukBM/requestlimit.svg?branch=master[\"Build Status\", link=\"https://travis-ci.org/storozhukBM/requestlimit\"] image:https://coveralls.io/repos/github/storozhukBM/requestlimit/badge.svg[\"Coverage Status\", link=\"https://coveralls.io/github/storozhukBM/requestlimit\"]" } ]
asciidoc
chosen
[ { "from": "system", "value": "Update release process for Gitlab Maven repo" }, { "from": "human", "value": "= OmniJ Release Process\n\n== Main Release Process\n\n\n. Official builds are now done with JDK 11. (Although most release JARs are JDK 8-compatible)\n.. `sdk use java 11.0.10.hs-adpt`\n. Update `CHANGELOG.adoc`\n. Set versions\n.. `gradle.properties`\n.. omnij-dsl `ExtensionModule`\n.. `README.adoc`\n. Commit version bump and changelog.\n. Full build, test\n.. `./gradlew clean buildCI regTest`\n. Tag: `git tag -a v0.x.y -m \"Release 0.x.y\"`\n. Push: `git push --tags origin master`\n. Publish to Bintray:\n.. `./gradlew bintrayUpload`\n.. Confirm publish of artifacts in Bintray Web UI.\n. Publish Github pages site including JavaDoc and GroovyDoc\n.. `./gradlew publishSite`\n. Publish to Github Packages\n.. `./gradlew publish`\n. Update Github https://github.com/OmniLayer/OmniJ/releases[OmniJ Releases] page.\n\n== Announcements\n\n. Not yet.\n\n== After release\n\n. Set versions back to -SNAPSHOT\n.. `gradle.properties`\n.. omnij-dsl `ExtensionModule`\n.. `CHANGELOG.adoc`\n.. *Not* `README.adoc` -- it should match release version\n. Commit and push to master\n\n\n== Publish Snapshot release to Github Maven repo\n\n. `.gradlew clean buildCI regTest publish`\n\n== Install Snapshot to local maven repo\n\n. `./gradlew clean buildCI publishToMavenLocal`\n\n\n\n" }, { "from": "gpt", "value": "= OmniJ Release Process\n\n== Main Release Process\n\n\n. Official builds are now done with JDK 11. (Although most release JARs are JDK 8-compatible)\n.. `sdk use java 11.0.10.hs-adpt`\n. Update `CHANGELOG.adoc`\n. Set versions\n.. `gradle.properties`\n.. omnij-dsl `ExtensionModule`\n.. `README.adoc`\n. Commit version bump and changelog.\n. Full build, test\n.. `./gradlew clean buildCI regTest`\n. Tag: `git tag -a v0.x.y -m \"Release 0.x.y\"`\n. Push: `git push --tags origin master`\n. Publish to Gitlab Maven repo:\n.. `./gradlew publish`\n. Publish Github pages site including JavaDoc and GroovyDoc\n.. `./gradlew publishSite`\n. Update Github https://github.com/OmniLayer/OmniJ/releases[OmniJ Releases] page.\n\n== Announcements\n\n. Not yet.\n\n== After release\n\n. Set versions back to -SNAPSHOT\n.. `gradle.properties`\n.. omnij-dsl `ExtensionModule`\n.. `CHANGELOG.adoc`\n.. *Not* `README.adoc` -- it should match release version\n. Commit and push to master\n\n\n== Publish Snapshot release to Github Maven repo\n\n. `.gradlew clean buildCI regTest publish`\n\n== Install Snapshot to local maven repo\n\n. `./gradlew clean buildCI publishToMavenLocal`\n\n\n\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "[Docs] Add a sidebar about gems that were extracted from RuboCop\n" }, { "from": "human", "value": "= Installation\n\nRuboCop's installation is pretty standard:\n\n[source,sh]\n----\n$ gem install rubocop\n----\n\nIf you'd rather install RuboCop using `bundler`, don't require it in your `Gemfile`:\n\n[source,rb]\n----\ngem 'rubocop', require: false\n----\n\nRuboCop's development is moving at a very rapid pace and there are\noften backward-incompatible changes between minor releases (since we\nhaven't reached version 1.0 yet). To prevent an unwanted RuboCop update you\nmight want to use a conservative version locking in your `Gemfile`:\n\n[source,rb]\n----\ngem 'rubocop', '~> 0.85.1', require: false\n----\n\nNOTE: You can check out our progress on the road to version 1.0 https://github.com/rubocop-hq/rubocop/milestone/4[here].\nYou can also help us get there faster! :-)\n" }, { "from": "gpt", "value": "= Installation\n\nRuboCop's installation is pretty standard:\n\n[source,sh]\n----\n$ gem install rubocop\n----\n\nIf you'd rather install RuboCop using `bundler`, don't require it in your `Gemfile`:\n\n[source,rb]\n----\ngem 'rubocop', require: false\n----\n\nRuboCop's development is moving at a very rapid pace and there are\noften backward-incompatible changes between minor releases (since we\nhaven't reached version 1.0 yet). To prevent an unwanted RuboCop update you\nmight want to use a conservative version locking in your `Gemfile`:\n\n[source,rb]\n----\ngem 'rubocop', '~> 0.85.1', require: false\n----\n\nNOTE: You can check out our progress on the road to version 1.0 https://github.com/rubocop-hq/rubocop/milestone/4[here].\nYou can also help us get there faster! :-)\n\n.A Modular RuboCop\n****\nOriginally RuboCop bundled cops focused on performance and Ruby on Rails, but those were\nextracted into their own gems eventually:\n\n* the performance cops were extracted in `rubocop-performance` and were removed from the main gem in 0.68\n* the Ruby on Rails cops were extracted in `rubocop-rails` and were removed from the main gem in 0.72\n\nYou'll need to install those gems separately if you'd like to use them.\nSee https://metaredux.com/posts/2019/05/22/a-modular-rubocop.html[this article] for more details.\n\nAdditionally, RuboCop's AST logic was extracted to `rubocop-ast` in RuboCop 0.84. While this change is\ntransparent to RuboCop's users (`rubocop-ast` is a runtime dependency of `rubocop`), it's good to be\naware that you can leverage RuboCop's AST node extensions and AST node pattern matching outside of RuboCop.\n****\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Add tests for stem macros\n" }, { "from": "human", "value": "= _Precompiled_ Math!\n:math:\n:imagesoutdir: generated_images\n:imagesdir: images\n\n## Equations!\n\n[latexmath]\n++++\nk_{n+1} = n^2 + k_n^2 - k_{n-1}\n++++\n\nSome useful text! Formula for quadratic root:\n\n[latexmath]\n++++\nx = \\frac{-b \\pm \\sqrt{b^2 - 4ac}}{2a}\n++++\n\nInline equation works too! latexmath:[a^2+b^2=c^2]. Pretty nice, huh?\n" }, { "from": "gpt", "value": "= _Precompiled_ Math!\n:math:\n:imagesoutdir: generated_images\n:imagesdir: images\n:stem: latexmath\n\n## Equations!\n\n[latexmath]\n++++\nk_{n+1} = n^2 + k_n^2 - k_{n-1}\n++++\n\nSome useful text! Formula for quadratic root:\n\n[stem]\n++++\nx = \\frac{-b \\pm \\sqrt{b^2 - 4ac}}{2a}\n++++\n\nInline equation works too! latexmath:[a^2+b^2=c^2]. Or as stem\nstem:[a^2+b^2=c^2]. Pretty nice, huh?\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Update changelog with release date of 0.4.0\n" }, { "from": "human", "value": "= Changelog\n\n== Version 0.4.0\n\nDate: unreleased\n\n- Add encode/decode functions to JWS/JWT implementation. Them instead of return\n plain value, return a monadic either. That allows granular error reporting\n instead something like nil that not very useful. The previous sign/unsign\n are conserved for backward compatibility but maybe in future will be removed.\n- Rename parameter `maxage` to `max-age` on jws implementation. This change\n introduces a little backward incompatibility.\n- Add \"compact\" signing implementation as replacemen of django based one.\n- Django based generic signing is removed.\n- Update buddy-core version to 0.4.0\n\n\n== Version 0.3.0\n\nDate: 2014-01-18\n\n- First version splitted from monolitic buddy package.\n- No changes from original version.\n" }, { "from": "gpt", "value": "= Changelog\n\n== Version 0.4.0\n\nDate: 2015-02-22\n\n- Add encode/decode functions to JWS/JWT implementation. Them instead of return\n plain value, return a monadic either. That allows granular error reporting\n instead something like nil that not very useful. The previous sign/unsign\n are conserved for backward compatibility but maybe in future will be removed.\n- Rename parameter `maxage` to `max-age` on jws implementation. This change\n introduces a little backward incompatibility.\n- Add \"compact\" signing implementation as replacemen of django based one.\n- Django based generic signing is removed.\n- Update buddy-core version to 0.4.0\n\n\n== Version 0.3.0\n\nDate: 2014-01-18\n\n- First version splitted from monolitic buddy package.\n- No changes from original version.\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Rename CI pipeline to include the version in its name\n\nCloses gh-18844\n" }, { "from": "human", "value": "== Concourse pipeline\n\nEnsure that you've setup the spring-boot target and can login\n\n[source]\n----\n$ fly -t spring-boot login -n spring-boot -c https://ci.spring.io\n----\n\nThe pipeline can be deployed using the following command:\n\n[source]\n----\n$ fly -t spring-boot set-pipeline -p spring-boot -c ci/pipeline.yml -l ci/parameters.yml\n----\n\nNOTE: This assumes that you have credhub integration configured with the appropriate\nsecrets.\n" }, { "from": "gpt", "value": "== Concourse pipeline\n\nEnsure that you've setup the spring-boot target and can login\n\n[source]\n----\n$ fly -t spring-boot login -n spring-boot -c https://ci.spring.io\n----\n\nThe pipeline can be deployed using the following command:\n\n[source]\n----\n$ fly -t spring-boot set-pipeline -p spring-boot-2.2.x -c ci/pipeline.yml -l ci/parameters.yml\n----\n\nNOTE: This assumes that you have credhub integration configured with the appropriate\nsecrets.\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Add note on transaction requirements\n" }, { "from": "human", "value": "==== Dependency Upgrades\n\nGORM 7.0 supports a minimum version of Java 8, Hibernate 5.3.x and Spring 5.2.x.\n\nEach of these underlying components may have changes that require altering your application. These changes are beyond the scope of this documentation.\n\n\n==== Package Restructuring and Deprecations\n\nPreviously deprecated classes have been deleted from this release and in order to support Java 11 modules in the future some package re-structuring has occurred.\n\n==== Changes to Proxy Handling\n\nGORM no longer creates custom proxy factories nor automatically unwraps Hibernate proxies.\n\nThis makes it more consistent to the way regular Hibernate behaves and reduces the complexity required at the framework level.\n\nYou may need to alter `instanceof` checks are manually unwrap proxies in certain cases.\n\n==== Module `grails-validation` Deprecated and Removed\n\nIn GORM 6.x the `grails-validation` module was deprecated and replaced by `grails-datastore-gorm-validation`.\n\nDeprecated interfaces were maintained for backwards compatibility. In GORM 7.0 these deprecated classes have been removed and all dependency on `grails-validation` removed." }, { "from": "gpt", "value": "==== Dependency Upgrades\n\nGORM 7.0 supports a minimum version of Java 8, Hibernate 5.3.x and Spring 5.2.x.\n\nEach of these underlying components may have changes that require altering your application. These changes are beyond the scope of this documentation.\n\n\n==== Package Restructuring and Deprecations\n\nPreviously deprecated classes have been deleted from this release and in order to support Java 11 modules in the future some package re-structuring has occurred.\n\n==== Changes to Proxy Handling\n\nGORM no longer creates custom proxy factories nor automatically unwraps Hibernate proxies.\n\nThis makes it more consistent to the way regular Hibernate behaves and reduces the complexity required at the framework level.\n\nYou may need to alter `instanceof` checks are manually unwrap proxies in certain cases.\n\n==== Module `grails-validation` Deprecated and Removed\n\nIn GORM 6.x the `grails-validation` module was deprecated and replaced by `grails-datastore-gorm-validation`.\n\nDeprecated interfaces were maintained for backwards compatibility. In GORM 7.0 these deprecated classes have been removed and all dependency on `grails-validation` removed.\n\n==== Transactions Now Required for all Operations\n\nPrevious versions of Hibernate allowed read operations to be executed without the presence of a declaration transaction.\n\nHibernate 5.2 and above require the presence of an active transaction. If see a `javax.persistence.TransactionRequiredException` exception it means your method lacks a `@Transactional` annotation around it." } ]
asciidoc
chosen
[ { "from": "system", "value": "Add docs for uploading zip archives.\n" }, { "from": "human", "value": "TinkerPop3\n----------\n\nimage:https://raw.githubusercontent.com/tinkerpop/tinkerpop3/master/docs/static/images/tinkerpop3-splash.png[TinkerPop3]\n\nDocumentation\n~~~~~~~~~~~~~\n\n* link:http://tinkerpop.com/docs/current/[user documentation]\n* link:http://tinkerpop.com/javadocs/current/[javadoc]\n\nBuilding and Testing\n~~~~~~~~~~~~~~~~~~~~\n\n* Build Project: `mvn clean install`\n* Build AsciiDocs: `mvn process-resources -Dasciidoc`\n* Deploy AsciiDocs: `mvn deploy -Dasciidoc`\n* Deploy JavaDocs: `mvn deploy -Djavadoc`\n* Travis Deploy Docs: merge/push to `publish-docs` branch\n* Integration Tests: `mvn clean verify -DskipIntegrationTests=false`\n* Performance Tests: `mvn clean verify -DskipPerformanceTests=false`\n* Gatling Load Tests: `mvn clean test -Pgatling -Dhost={hostname}`\n\nGet Started\n~~~~~~~~~~~\n\n[source,bash]\n----\n$ bin/gremlin.sh\n\n \\,,,/\n (o o)\n-----oOOo-(3)-oOOo-----\ngremlin>\n----\n" }, { "from": "gpt", "value": "TinkerPop3\n----------\n\nimage:https://raw.githubusercontent.com/tinkerpop/tinkerpop3/master/docs/static/images/tinkerpop3-splash.png[TinkerPop3]\n\nDocumentation\n~~~~~~~~~~~~~\n\n* link:http://tinkerpop.com/docs/current/[user documentation]\n* link:http://tinkerpop.com/javadocs/current/[javadoc]\n\nBuilding and Testing\n~~~~~~~~~~~~~~~~~~~~\n\n* Build Project: `mvn clean install`\n* Build AsciiDocs: `mvn process-resources -Dasciidoc`\n* Deploy AsciiDocs: `mvn deploy -Dasciidoc`\n* Deploy JavaDocs: `mvn deploy -Djavadoc`\n* Deploy Zip Archives: `mvn deploy -Darchives`\n* Travis Deploy Docs: merge/push to `publish-docs` branch\n* Integration Tests: `mvn clean verify -DskipIntegrationTests=false`\n* Performance Tests: `mvn clean verify -DskipPerformanceTests=false`\n* Gatling Load Tests: `mvn clean test -Pgatling -Dhost={hostname}`\n\nGet Started\n~~~~~~~~~~~\n\n[source,bash]\n----\n$ bin/gremlin.sh\n\n \\,,,/\n (o o)\n-----oOOo-(3)-oOOo-----\ngremlin>\n----\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Document JVM option MaxFDLimit for macOS ()\n\nThis commit documents a JVM option that is needed on macOS when raising\r\nfile descriptor limits there.\r\n\r\nRelates #26900\r\n" }, { "from": "human", "value": "[[file-descriptors]]\n=== File Descriptors\n\n[NOTE]\nThis is only relevant for Linux and macOS and can be safely ignored if running\nElasticsearch on Windows. On Windows that JVM uses an\nhttps://msdn.microsoft.com/en-us/library/windows/desktop/aa363858(v=vs.85).aspx[API]\nlimited only by available resources.\n\nElasticsearch uses a lot of file descriptors or file handles. Running out of\nfile descriptors can be disastrous and will most probably lead to data loss.\nMake sure to increase the limit on the number of open files descriptors for\nthe user running Elasticsearch to 65,536 or higher.\n\nFor the `.zip` and `.tar.gz` packages, set <<ulimit,`ulimit -n 65536`>> as\nroot before starting Elasticsearch, or set `nofile` to `65536` in\n<<limits.conf,`/etc/security/limits.conf`>>.\n\nRPM and Debian packages already default the maximum number of file\ndescriptors to 65536 and do not require further configuration.\n\nYou can check the `max_file_descriptors` configured for each node\nusing the <<cluster-nodes-stats>> API, with:\n\n[source,js]\n--------------------------------------------------\nGET _nodes/stats/process?filter_path=**.max_file_descriptors\n--------------------------------------------------\n// CONSOLE\n" }, { "from": "gpt", "value": "[[file-descriptors]]\n=== File Descriptors\n\n[NOTE]\nThis is only relevant for Linux and macOS and can be safely ignored if running\nElasticsearch on Windows. On Windows that JVM uses an\nhttps://msdn.microsoft.com/en-us/library/windows/desktop/aa363858(v=vs.85).aspx[API]\nlimited only by available resources.\n\nElasticsearch uses a lot of file descriptors or file handles. Running out of\nfile descriptors can be disastrous and will most probably lead to data loss.\nMake sure to increase the limit on the number of open files descriptors for\nthe user running Elasticsearch to 65,536 or higher.\n\nFor the `.zip` and `.tar.gz` packages, set <<ulimit,`ulimit -n 65536`>> as\nroot before starting Elasticsearch, or set `nofile` to `65536` in\n<<limits.conf,`/etc/security/limits.conf`>>.\n\nOn macOS, you must also pass the JVM option `-XX:-MaxFDLimit`\nto Elasticsearch in order for it to make use of the higher file descriptor limit.\n\nRPM and Debian packages already default the maximum number of file\ndescriptors to 65536 and do not require further configuration.\n\nYou can check the `max_file_descriptors` configured for each node\nusing the <<cluster-nodes-stats>> API, with:\n\n[source,js]\n--------------------------------------------------\nGET _nodes/stats/process?filter_path=**.max_file_descriptors\n--------------------------------------------------\n// CONSOLE\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Remove note that says that no CSI drivers are installed\n" }, { "from": "human", "value": "[id=\"persistent-storage-csi\"]\n= Configuring CSI volumes\ninclude::modules/common-attributes.adoc[]\n:context: persistent-storage-csi\n\ntoc::[]\n\nThe Container Storage Interface (CSI) allows {product-title} to consume\nstorage from storage back ends that implement the\nlink:https://github.com/container-storage-interface/spec[CSI interface]\nas persistent storage.\n\n[IMPORTANT]\n====\n{product-title} does not ship with any CSI drivers. It is recommended\nto use the CSI drivers provided by\nlink:https://kubernetes-csi.github.io/docs/drivers.html[community or storage vendors].\n\nInstallation instructions differ by driver, and are found in each driver's\ndocumentation. Follow the instructions provided by the CSI driver.\n====\n\ninclude::modules/persistent-storage-csi-architecture.adoc[leveloffset=+1]\n\ninclude::modules/persistent-storage-csi-external-controllers.adoc[leveloffset=+2]\n\ninclude::modules/persistent-storage-csi-driver-daemonset.adoc[leveloffset=+2]\n\ninclude::modules/persistent-storage-csi-drivers-supported.adoc[leveloffset=+1]\n\ninclude::modules/persistent-storage-csi-dynamic-provisioning.adoc[leveloffset=+1]\n\ninclude::modules/persistent-storage-csi-mysql-example.adoc[leveloffset=+1]\n" }, { "from": "gpt", "value": "[id=\"persistent-storage-csi\"]\n= Configuring CSI volumes\ninclude::modules/common-attributes.adoc[]\n:context: persistent-storage-csi\n\ntoc::[]\n\nThe Container Storage Interface (CSI) allows {product-title} to consume\nstorage from storage back ends that implement the\nlink:https://github.com/container-storage-interface/spec[CSI interface]\nas persistent storage.\n\ninclude::modules/persistent-storage-csi-architecture.adoc[leveloffset=+1]\n\ninclude::modules/persistent-storage-csi-external-controllers.adoc[leveloffset=+2]\n\ninclude::modules/persistent-storage-csi-driver-daemonset.adoc[leveloffset=+2]\n\ninclude::modules/persistent-storage-csi-drivers-supported.adoc[leveloffset=+1]\n\ninclude::modules/persistent-storage-csi-dynamic-provisioning.adoc[leveloffset=+1]\n\ninclude::modules/persistent-storage-csi-mysql-example.adoc[leveloffset=+1]\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Examples: Add full set of examples for each admonition type\n\nThey may be rendered differently, for example note and tip as aside,\nothers as section.\n" }, { "from": "human", "value": "// .basic\nWARNING: Watch out for dragons!\n\n// .basic_multiline\nNOTE: An admonition paragraph draws the reader's attention to\nauxiliary information.\nIts purpose is determined by the label\nat the beginning of the paragraph.\n\n// .block\n[IMPORTANT]\n====\nWhile werewolves are hardy community members, keep in mind some dietary concerns.\n====\n\n// .block_with_title\n[IMPORTANT]\n.Feeding the Werewolves\n====\nWhile werewolves are hardy community members, keep in mind some dietary concerns.\n====\n\n// .block_with_id_and_role\n[IMPORTANT, id=werewolve, role=member]\n====\nWhile werewolves are hardy community members, keep in mind some dietary concerns.\n====\n" }, { "from": "gpt", "value": "// .note\nNOTE: This is a note.\n\n// .note_with_title\n.Title of note\nNOTE: This is a note with title.\n\n// .note_with_id_and_role\n[#note-1.yellow]\nNOTE: This is a note with id and role.\n\n// .note_block\n[NOTE]\n====\nThis is a note with complex content.\n\n* It contains a list.\n====\n\n// .tip\nTIP: This is a tip.\n\n// .tip_with_title\n.Title of tip\nTIP: This is a tip with title.\n\n// .tip_with_id_and_role\n[#tip-1.blue]\nTIP: This is a tip with id and role.\n\n// .tip_block\n[TIP]\n====\nThis is a tip with complex content.\n\n* It contains a list.\n====\n\n// .important\nIMPORTANT: This is an important notice.\n\n// .important_with_title\n.Title of important notice\nIMPORTANT: This is an important notice with title.\n\n// .important_with_id_and_role\n[#important-1.red]\nIMPORTANT: This is an important notice with id and role.\n\n// .important_block\n[IMPORTANT]\n====\nThis is an important notice with complex content.\n\n* It contains a list.\n====\n\n// .caution\nCAUTION: This is a caution.\n\n// .caution_with_title\n.Title of caution\nCAUTION: This is a caution with title.\n\n// .caution_with_id_and_role\n[#caution-1.red]\nCAUTION: This is a caution with id and role.\n\n// .caution_block\n[CAUTION]\n====\nThis is a caution with complex content.\n\n* It contains a list.\n====\n\n// .warning\nWARNING: This is a warning.\n\n// .warning_with_title\n.Title of warning\nWARNING: This is a warning with title.\n\n// .warning_with_id_and_role\n[#warning-1.red]\nWARNING: This is a warning with id and role.\n\n// .warning_block\n[WARNING]\n====\nThis is a warning with complex content.\n\n* It contains a list.\n====\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Update storage account name in Azure registry storage configuration\n" }, { "from": "human", "value": "// Module included in the following assemblies:\n//\n//* registry/configuring_registry_storage-azure.adoc\n\n[id=\"registry-configuring-storage-azure-user-infra_{context}\"]\n= Configuring registry storage for Azure\n\nDuring installation, your cloud credentials are sufficient to create Azure Blob\nStorage, and the Registry Operator automatically configures storage.\n\n.Prerequisites\n\n* A cluster on Azure with user-provisioned infrastructure.\n* To configure registry storage for Azure, provide Registry Operator\ncloud credentials.\n* For Azure storage the secret is expected to contain one key:\n** `REGISTRY_STORAGE_AZURE_ACCOUNTKEY`\n\n.Procedure\n\n. Create an link:https://docs.microsoft.com/en-us/azure/storage/blobs/storage-quickstart-blobs-portal[Azure storage container].\n\n. Fill in the storage configuration in `configs.imageregistry.operator.openshift.io/cluster`:\n+\n----\n$ oc edit configs.imageregistry.operator.openshift.io/cluster\n\nstorage:\n azure:\n accountName: <account-name>\n container: <container-name>\n----\n" }, { "from": "gpt", "value": "// Module included in the following assemblies:\n//\n//* registry/configuring_registry_storage-azure.adoc\n\n[id=\"registry-configuring-storage-azure-user-infra_{context}\"]\n= Configuring registry storage for Azure\n\nDuring installation, your cloud credentials are sufficient to create Azure Blob\nStorage, and the Registry Operator automatically configures storage.\n\n.Prerequisites\n\n* A cluster on Azure with user-provisioned infrastructure.\n* To configure registry storage for Azure, provide Registry Operator\ncloud credentials.\n* For Azure storage the secret is expected to contain one key:\n** `REGISTRY_STORAGE_AZURE_ACCOUNTKEY`\n\n.Procedure\n\n. Create an link:https://docs.microsoft.com/en-us/azure/storage/blobs/storage-quickstart-blobs-portal[Azure storage container].\n\n. Fill in the storage configuration in `configs.imageregistry.operator.openshift.io/cluster`:\n+\n----\n$ oc edit configs.imageregistry.operator.openshift.io/cluster\n\nstorage:\n azure:\n accountName: <storage-account-name>\n container: <container-name>\n----\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Use `buildCI` target rather than `jenkinsBuild`" }, { "from": "human", "value": "= OmniJ Release Process\n\n== Main Release Process\n\n. Update `CHANGELOG.adoc`\n. Set versions\n.. `gradle.properties`\n.. omnij-dsl `ExtensionModule`\n.. `README.adoc`\n. Commit version bump and changelog.\n. Full build, test\n.. `./gradlew clean jenkinsBuild regTest`\n. Tag: `git tag -a v0.x.y -m \"Release 0.x.y\"`\n. Push: `git push --tags origin master`\n. Publish to Bintray:\n.. `./gradlew bintrayUpload`\n.. Confirm publish of artifacts in Bintray Web UI.\n. Update Github https://github.com/OmniLayer/OmniJ/releases[OmniJ Releases] page.\n. TBD: Update github-pages site (not set up yet)\n\n== Announcements\n\n. Not yet.\n\n== After release\n\n. Set versions back to -SNAPSHOT\n.. `build.gradle`\n.. omnij-dsl `ExtensionModule`\n.. *Not* `README.adoc` -- it should match release version\n. Commit and push to master\n\n\n\n" }, { "from": "gpt", "value": "= OmniJ Release Process\n\n== Main Release Process\n\n. Update `CHANGELOG.adoc`\n. Set versions\n.. `gradle.properties`\n.. omnij-dsl `ExtensionModule`\n.. `README.adoc`\n. Commit version bump and changelog.\n. Full build, test\n.. `./gradlew clean buildCI regTest`\n. Tag: `git tag -a v0.x.y -m \"Release 0.x.y\"`\n. Push: `git push --tags origin master`\n. Publish to Bintray:\n.. `./gradlew bintrayUpload`\n.. Confirm publish of artifacts in Bintray Web UI.\n. Update Github https://github.com/OmniLayer/OmniJ/releases[OmniJ Releases] page.\n. TBD: Update github-pages site (not set up yet)\n\n== Announcements\n\n. Not yet.\n\n== After release\n\n. Set versions back to -SNAPSHOT\n.. `build.gradle`\n.. omnij-dsl `ExtensionModule`\n.. *Not* `README.adoc` -- it should match release version\n. Commit and push to master\n\n\n\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Add myself to the list with authors for the reference documentation\n" }, { "from": "human", "value": "= Reactor Netty\nStephane Maldini;\n:appversion: 0.5.0 (wiki)\nifndef::host-github[:ext-relative: {outfilesuffix}]\n{appversion}\n:doctype: book\n:icons: font\n\n//include::net.adoc[]\ninclude::tcp-server.adoc[]\ninclude::tcp-client.adoc[]\ninclude::http-server.adoc[]\ninclude::http-client.adoc[]\n\n" }, { "from": "gpt", "value": "= Reactor Netty\nStephane Maldini <https://twitter.com/smaldini[@smaldini]> ; Violeta Georgieva <https://twitter.com/violeta_g_g[@violeta_g_g]>\n:appversion: 0.5.0 (wiki)\nifndef::host-github[:ext-relative: {outfilesuffix}]\n{appversion}\n:doctype: book\n:icons: font\n\n//include::net.adoc[]\ninclude::tcp-server.adoc[]\ninclude::tcp-client.adoc[]\ninclude::http-server.adoc[]\ninclude::http-client.adoc[]\n\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "[DOCS] Fix formatting in change password API\n\nOriginal commit: elastic/x-pack-elasticsearch@0c9b1e7484a5359d08f20ed7f122daac337829d1\n" }, { "from": "human", "value": "[role=\"xpack\"]\n[[security-api-change-password]]\n=== Change Password API\n\nThe Change Password API enables you to submit a request to change the password\nof a user.\n\n==== Request\n\n`POST _xpack/security/user/_password` +\n`POST _xpack/security/user/<username>/_password` \n\n\n==== Path Parameters\n\n`username`::\n (string) The user whose password you want to change. If you do not specify\n this parameter, the password is changed for the current user.\n\n\n==== Request Body\n\n`password` (required)::\n (string) The new password value.\n\n\n==== Authorization\n\nEvery user can change their own password. Users with the `manage_security`\nprivilege can change passwords of other users.\n\n\n==== Examples\n\nThe following example updates the password for the `elastic` user:\n\n[source,js]\n--------------------------------------------------\nPOST _xpack/security/user/elastic/_password\n{\n \"password\": \"x-pack-test-password\"\n}\n--------------------------------------------------\n// CONSOLE\n\nA successful call returns an empty JSON structure.\n\n[source,js]\n--------------------------------------------------\n{}\n--------------------------------------------------\n// TESTRESPONSE\n" }, { "from": "gpt", "value": "[role=\"xpack\"]\n[[security-api-change-password]]\n=== Change Password API\n\nThe Change Password API enables you to submit a request to change the password\nof a user.\n\n==== Request\n\n`POST _xpack/security/user/_password` +\n\n`POST _xpack/security/user/<username>/_password`\n\n\n==== Path Parameters\n\n`username`::\n (string) The user whose password you want to change. If you do not specify\n this parameter, the password is changed for the current user.\n\n\n==== Request Body\n\n`password` (required)::\n (string) The new password value.\n\n\n==== Authorization\n\nEvery user can change their own password. Users with the `manage_security`\nprivilege can change passwords of other users.\n\n\n==== Examples\n\nThe following example updates the password for the `elastic` user:\n\n[source,js]\n--------------------------------------------------\nPOST _xpack/security/user/elastic/_password\n{\n \"password\": \"x-pack-test-password\"\n}\n--------------------------------------------------\n// CONSOLE\n\nA successful call returns an empty JSON structure.\n\n[source,js]\n--------------------------------------------------\n{}\n--------------------------------------------------\n// TESTRESPONSE\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Add link to 4th part of JAXenter tutorial\n" }, { "from": "human", "value": "# Janitor image:https://travis-ci.org/techdev-solutions/janitor.svg?branch=master[\"Build Status\",link=\"https://travis-ci.org/techdev-solutions/janitor\"]\n\nAn application to perform cleanup work using the https://getpocket.com[Pocket API].\nReview your existing items via the web interface and have Janitor archive old items once per day.\n\n## API Documentation\nThe documentation for the Kotlin API bindings can be found https://techdev-solutions.github.io/janitor/pocket-api/[here].\n\n## Tutorial (German)\nKotlin: Ein Tutorial für Einsteiger @ JAXenter::\n* https://jaxenter.de/kotlin-tutorial-48156[Part 1]\n* https://jaxenter.de/kotlin-ein-tutorial-fuer-einsteiger-teil-2-48587[Part 2]\n* https://jaxenter.de/kotlin-ein-tutorial-fuer-einsteiger-teil-3-48967[Part 3]\n\n## Screenshots\n### Item Overview\nimage:images/items.png?raw=true[Item Overiew]\n\n### Error View\nimage:images/error.png?raw=true[Error View]" }, { "from": "gpt", "value": "# Janitor image:https://travis-ci.org/techdev-solutions/janitor.svg?branch=master[\"Build Status\",link=\"https://travis-ci.org/techdev-solutions/janitor\"]\n\nAn application to perform cleanup work using the https://getpocket.com[Pocket API].\nReview your existing items via the web interface and have Janitor archive old items once per day.\n\n## API Documentation\nThe documentation for the Kotlin API bindings can be found https://techdev-solutions.github.io/janitor/pocket-api/[here].\n\n## Tutorial (German)\nKotlin: Ein Tutorial für Einsteiger @ JAXenter::\n* https://jaxenter.de/kotlin-tutorial-48156[Part 1]\n* https://jaxenter.de/kotlin-ein-tutorial-fuer-einsteiger-teil-2-48587[Part 2]\n* https://jaxenter.de/kotlin-ein-tutorial-fuer-einsteiger-teil-3-48967[Part 3]\n* https://jaxenter.de/kotlin-ein-tutorial-fuer-einsteiger-teil-4-49160[Part 4]\n\n## Screenshots\n### Item Overview\nimage:images/items.png?raw=true[Item Overiew]\n\n### Error View\nimage:images/error.png?raw=true[Error View]" } ]
asciidoc
chosen
[ { "from": "system", "value": "Fix version constant in Groovy API docs\n\nThis commit fixes the version constant in the Groovy API docs from\n5.0.0-alpha5 to 5.1.0.\n" }, { "from": "human", "value": "= Groovy API\n:ref: http://www.elastic.co/guide/en/elasticsearch/reference/current\n:java: http://www.elastic.co/guide/en/elasticsearch/client/java-api/current\n:version: 5.0.0-alpha5\n\n[preface]\n== Preface\n\nThis section describes the http://groovy-lang.org/[Groovy] API\nelasticsearch provides. All elasticsearch APIs are executed using a\n<<client,GClient>>, and are completely\nasynchronous in nature (they either accept a listener, or return a\nfuture).\n\nThe Groovy API is a wrapper on top of the\n{java}[Java API] exposing it in a groovier\nmanner. The execution options for each API follow a similar manner and\ncovered in <<anatomy>>.\n\n\n[[maven]]\n=== Maven Repository\n\nThe Groovy API is hosted on\nhttp://search.maven.org/#search%7Cga%7C1%7Ca%3A%22elasticsearch-groovy%22[Maven\nCentral].\n\nFor example, you can define the latest version in your `pom.xml` file:\n\n[\"source\",\"xml\",subs=\"attributes\"]\n--------------------------------------------------\n<dependency>\n <groupId>org.elasticsearch</groupId>\n <artifactId>elasticsearch-groovy</artifactId>\n <version>{version}</version>\n</dependency>\n--------------------------------------------------\n\ninclude::anatomy.asciidoc[]\n\ninclude::client.asciidoc[]\n\ninclude::index_.asciidoc[]\n\ninclude::get.asciidoc[]\n\ninclude::delete.asciidoc[]\n\ninclude::search.asciidoc[]\n" }, { "from": "gpt", "value": "= Groovy API\n:ref: http://www.elastic.co/guide/en/elasticsearch/reference/current\n:java: http://www.elastic.co/guide/en/elasticsearch/client/java-api/current\n:version: 5.1.0\n\n[preface]\n== Preface\n\nThis section describes the http://groovy-lang.org/[Groovy] API\nelasticsearch provides. All elasticsearch APIs are executed using a\n<<client,GClient>>, and are completely\nasynchronous in nature (they either accept a listener, or return a\nfuture).\n\nThe Groovy API is a wrapper on top of the\n{java}[Java API] exposing it in a groovier\nmanner. The execution options for each API follow a similar manner and\ncovered in <<anatomy>>.\n\n\n[[maven]]\n=== Maven Repository\n\nThe Groovy API is hosted on\nhttp://search.maven.org/#search%7Cga%7C1%7Ca%3A%22elasticsearch-groovy%22[Maven\nCentral].\n\nFor example, you can define the latest version in your `pom.xml` file:\n\n[\"source\",\"xml\",subs=\"attributes\"]\n--------------------------------------------------\n<dependency>\n <groupId>org.elasticsearch</groupId>\n <artifactId>elasticsearch-groovy</artifactId>\n <version>{version}</version>\n</dependency>\n--------------------------------------------------\n\ninclude::anatomy.asciidoc[]\n\ninclude::client.asciidoc[]\n\ninclude::index_.asciidoc[]\n\ninclude::get.asciidoc[]\n\ninclude::delete.asciidoc[]\n\ninclude::search.asciidoc[]\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Include Hello World Youtube video" }, { "from": "human", "value": "= Minimal-J\n\nJava - but small.\n\nimage::doc/frontends.png[]\n\nMinimal-J applications are\n\n* Responsive to use on every device\n* Straight forward to specify and implement and therefore\n* Easy to plan and manage\n\n=== Idea\n\nBusiness applications tend to get complex and complicated. Minimal-J prevents this by setting clear rules how an application should behave and how it should be implemented.\n\nMinimal applications may not always look the same. But the UI concepts never change. There are no surprises for the user.\n\n== Technical Features\n\n* Independent of the used UI technology. Implementations for Web / Mobile / Desktop.\n* ORM persistence layer for Maria DB or in memory DB. Transactions and Authorization supported.\n* Small: The minimalj.jar is still < 1MB\n* Very few dependencies\n* Applications can run standalone (like SpringBoot)\n\n== Documentation\n\n* link:doc/user_guide/user_guide.adoc[Minimal user guide] User guide for Minimal-J applications.\n* link:doc/topics.adoc[Tutorial and examples] Informations for developers.\n* link:doc/release_notes.adoc[Release Notes]\n\n== Hello World\n\nHow to implement Hello World in Minimal-J:\n\nvideo::0VHz7gv6TpA[youtube]\n\n=== Contact\n\n* Bruno Eberhard, mailto:[email protected][[email protected]] " }, { "from": "gpt", "value": "= Minimal-J\n\nJava - but small.\n\nimage::doc/frontends.png[]\n\nMinimal-J applications are\n\n* Responsive to use on every device\n* Straight forward to specify and implement and therefore\n* Easy to plan and manage\n\n=== Idea\n\nBusiness applications tend to get complex and complicated. Minimal-J prevents this by setting clear rules how an application should behave and how it should be implemented.\n\nMinimal applications may not always look the same. But the UI concepts never change. There are no surprises for the user.\n\n== Technical Features\n\n* Independent of the used UI technology. Implementations for Web / Mobile / Desktop.\n* ORM persistence layer for Maria DB or in memory DB. Transactions and Authorization supported.\n* Small: The minimalj.jar is still < 1MB\n* Very few dependencies\n* Applications can run standalone (like SpringBoot)\n\n== Documentation\n\n* link:doc/user_guide/user_guide.adoc[Minimal user guide] User guide for Minimal-J applications.\n* link:doc/topics.adoc[Tutorial and examples] Informations for developers.\n* link:doc/release_notes.adoc[Release Notes]\n\n== Hello World\n\nHow to implement Hello World in Minimal-J:\n\nlink:_includes/ex-video.adoc[]\n\nvideo::0VHz7gv6TpA[youtube]\n\n=== Contact\n\n* Bruno Eberhard, mailto:[email protected][[email protected]] " } ]
asciidoc
chosen
[ { "from": "system", "value": "Update cayenne version mentioned in the documentation\n" }, { "from": "human", "value": "== Prerequisites\n\n* Java 1.8 or newer\n* A Java web app project that will serve your REST requests.\n* A JAX-RS 2.0 container, such as Jersey 2.x.\n* Cayenne 4.0.M4 or newer. Mapping your database and starting Cayenne\n ServerRuntime is outside the scope of this document. Please refer to the\n http://cayenne.apache.org/docs/4.0/cayenne-guide/index.html[corresponding Cayenne docs]\n\n" }, { "from": "gpt", "value": "== Prerequisites\n\n* Java 1.8 or newer\n* A Java web app project that will serve your REST requests.\n* A JAX-RS 2.0 container, such as Jersey 2.x.\n* Cayenne 4.0 or newer. Mapping your database and starting Cayenne\n ServerRuntime is outside the scope of this document. Please refer to the\n http://cayenne.apache.org/docs/4.0/cayenne-guide/index.html[corresponding Cayenne docs]\n\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Simplify import path for modules" }, { "from": "human", "value": "# AsciiDoc mode for CodeMirror\n\nThis repository contains the AsciiDoc mode for CodeMirror.\n\n## Installation\n\n```\n$ npm install codemirror-asciidoc\n```\n\n## Usage\n\n```js\nvar codemirror = require(\"codemirror/lib/codemirror\"),\n codemirror_asciidoc = require(\"codemirror-asciidoc/lib/asciidoc\");\n\ncodemirror.fromTextArea(document.getElementById(\"editor\"), {\n lineNumbers: true,\n lineWrapping: true,\n mode: \"asciidoc\"\n});\n```\n\n## License\n\nBSD\n\n## Credits\n\nThe AsciiDoc mode for CodeMirror was generated from the AsciiDoc mode for Ace using the https://github.com/espadrine/ace2cm[ace2cm] project by https://github.com/espadrine[Thaddee Tyl].\n" }, { "from": "gpt", "value": "# AsciiDoc mode for CodeMirror\n\nThis repository contains the AsciiDoc mode for CodeMirror.\n\n## Installation\n\n```\n$ npm install codemirror-asciidoc\n```\n\n## Usage\n\n```js\nvar codemirror = require(\"codemirror\"),\n codemirror_asciidoc = require(\"codemirror-asciidoc\");\n\ncodemirror.fromTextArea(document.getElementById(\"editor\"), {\n lineNumbers: true,\n lineWrapping: true,\n mode: \"asciidoc\"\n});\n```\n\n## License\n\nBSD\n\n## Credits\n\nThe AsciiDoc mode for CodeMirror was generated from the AsciiDoc mode for Ace using the https://github.com/espadrine/ace2cm[ace2cm] project by https://github.com/espadrine[Thaddee Tyl].\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Update documentation regarding using Service Catalog on Minishift\n" }, { "from": "human", "value": "[[prerequisites-openshift]]\n\n.Prerequisites\n\n* To install {ProductName}, the OpenShift client tools are required. You can download the OpenShift\nOrigin client from link:https://github.com/openshift/origin/releases[OpenShift Origin^]. {ProductName} has\nbeen tested to work with the latest stable release of the OpenShift Origin Client.\n\n* An OpenShift cluster is required. If you do not have an OpenShift cluster available, see\nlink:https://github.com/minishift/minishift[Minishift^] for an example of how to run a local instance of OpenShift\non your machine.\n\n* A method to generate certificates is required. This guide uses link:https://www.openssl.org/[OpenSSL^].\n" }, { "from": "gpt", "value": "[[prerequisites-openshift]]\n\n.Prerequisites\n\n* To install {ProductName}, the OpenShift client tools are required. You can download the OpenShift\nOrigin client from link:https://github.com/openshift/origin/releases[OpenShift Origin^]. {ProductName} has\nbeen tested to work with the latest stable release of the OpenShift Origin Client.\n\n* An OpenShift cluster is required. If you do not have an OpenShift cluster available, see\nlink:https://github.com/minishift/minishift[Minishift^] for an example of how to run a local instance of OpenShift\non your machine.\n\n* If you want to install {ProductName} on Minishift and want to use Service Catalog, you need to explicitly enable it during the start-up, like\n\n[options=\"nowrap\"]\n----\nMINISHIFT_ENABLE_EXPERIMENTAL=\"y\" minishift start --extra-clusterup-flags \"--service-catalog\"\n----\n\n* A method to generate certificates is required. This guide uses link:https://www.openssl.org/[OpenSSL^].\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Fix type in CI readme\n" }, { "from": "human", "value": "== Concourse pipeline\n\nEnsure that you've have a spring-boot target logged in\n\n[source]\n----\n$ fly -t spring-boot login -n spring-boot -c https://ci.spring-io\n----\n\nThe pipeline can be deployed using the following command:\n\n[source]\n----\n$ fly -t spring-boot set-pipeline -p spring-boot-2.0.x -c ci/pipeline.yml -l ci/parameters.yml\n----\n\nNOTE: This assumes that you have credhub integration configured with the appropriate\nsecrets.\n" }, { "from": "gpt", "value": "== Concourse pipeline\n\nEnsure that you've setup the spring-boot target and can login\n\n[source]\n----\n$ fly -t spring-boot login -n spring-boot -c https://ci.spring.io\n----\n\nThe pipeline can be deployed using the following command:\n\n[source]\n----\n$ fly -t spring-boot set-pipeline -p spring-boot-2.0.x -c ci/pipeline.yml -l ci/parameters.yml\n----\n\nNOTE: This assumes that you have credhub integration configured with the appropriate\nsecrets.\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Update Changelog to describe the list syntax overhaul\n" }, { "from": "human", "value": "= Changelog\n\nThis changelog contains major and/or breaking changes to Kakoune between\nreleased versions.\n\n== Development version\n\n* `%sh{...}` strings are not reparsed automatically anymore, they need\n to go through an explicit `evaluate-commands`\n\n* The `-allow-override` switch from `define-command` has been renamed\n `-override`.\n\n* The search prompt uses buffer word completion so that fuzzy completion\n can be used to quickly search for a buffer word.\n\n* The `wrap` highlighter can accept a new `-marker <marker_text>` switch.\n\n== Kakoune 2018.04.13\n\nFirst official Kakoune release.\n" }, { "from": "gpt", "value": "= Changelog\n\nThis changelog contains major and/or breaking changes to Kakoune between\nreleased versions.\n\n== Development version\n\nThis version contains a significant overhaul of various Kakoune\nfeatures that can break user configuration. This was a necessary\nchange to make Kakoune command model cleaner and more robust.\n\n* `%sh{...}` strings are not reparsed automatically anymore, they need\n to go through an explicit `evaluate-commands`\n\n* The `-allow-override` switch from `define-command` has been renamed\n `-override`.\n\n* The search prompt uses buffer word completion so that fuzzy completion\n can be used to quickly search for a buffer word.\n\n* The `wrap` highlighter can accept a new `-marker <marker_text>` switch.\n\n* The command line syntax has changed to support robust escaping, see\n <<command-parsing#,`:doc command-parsing`>>.\n\n* Various lists (options, registers...) in Kakoune are now written using\n the command line syntax:\n\n - `set-register` now take an arbitrary number of parameters and sets\n the register to multiple strings. `%reg` expands to a list of strings.\n\n - `%opt` expands list options as list of strings.\n\n - selection descs are whitespaces separated instead of `:` separated\n\n== Kakoune 2018.04.13\n\nFirst official Kakoune release.\n" } ]
asciidoc
chosen
[ { "from": "system", "value": "Update What's New for 6.0\n" }, { "from": "human", "value": "[[new]]\n= What's New in Spring Security 6.0\n\nSpring Security 6.0 provides a number of new features.\nBelow are the highlights of the release.\n\n== Breaking Changes\n\n* https://github.com/spring-projects/spring-security/issues/10556[gh-10556] - Remove EOL OpenSaml 3 Support.\nUse the OpenSaml 4 Support instead.\n* https://github.com/spring-projects/spring-security/issues/8980[gh-8980] - Remove unsafe/deprecated `Encryptors.querableText(CharSequence,CharSequence)`.\nInstead use data storage to encrypt values.\n* https://github.com/spring-projects/spring-security/issues/11520[gh-11520] - Remember Me uses SHA256 by default\n* https://github.com/spring-projects/spring-security/issues/8819 - Move filters to web package\nReorganize imports\n* https://github.com/spring-projects/spring-security/issues/7349 - Move filter and token to appropriate packages\nReorganize imports\n* https://github.com/spring-projects/spring-security/issues/11026[gh-11026] - Use `RequestAttributeSecurityContextRepository` instead of `NullSecurityContextRepository`\n* https://github.com/spring-projects/spring-security/pull/11887[gh-11827] - Change default authority for `oauth2Login()`\n" }, { "from": "gpt", "value": "[[new]]\n= What's New in Spring Security 6.0\n\nSpring Security 6.0 provides a number of new features.\nBelow are the highlights of the release.\n\n== Breaking Changes\n\n* https://github.com/spring-projects/spring-security/issues/10556[gh-10556] - Remove EOL OpenSaml 3 Support.\nUse the OpenSaml 4 Support instead.\n* https://github.com/spring-projects/spring-security/issues/8980[gh-8980] - Remove unsafe/deprecated `Encryptors.querableText(CharSequence,CharSequence)`.\nInstead use data storage to encrypt values.\n* https://github.com/spring-projects/spring-security/issues/11520[gh-11520] - Remember Me uses SHA256 by default\n* https://github.com/spring-projects/spring-security/issues/8819 - Move filters to web package\nReorganize imports\n* https://github.com/spring-projects/spring-security/issues/7349 - Move filter and token to appropriate packages\nReorganize imports\n* https://github.com/spring-projects/spring-security/issues/11026[gh-11026] - Use `RequestAttributeSecurityContextRepository` instead of `NullSecurityContextRepository`\n* https://github.com/spring-projects/spring-security/pull/11887[gh-11827] - Change default authority for `oauth2Login()`\n* https://github.com/spring-projects/spring-security/issues/10347[gh-10347] - Remove `UsernamePasswordAuthenticationToken` check in `BasicAuthenticationFilter`" } ]
asciidoc
chosen