commit
stringlengths
40
40
old_file
stringlengths
4
264
new_file
stringlengths
4
264
old_contents
stringlengths
0
4.24k
new_contents
stringlengths
1
5.44k
subject
stringlengths
14
778
message
stringlengths
15
9.92k
lang
stringclasses
277 values
license
stringclasses
13 values
repos
stringlengths
5
127k
1cbc68fa55f0f015e0fc3d2bf9bd0b7c3835d0fc
Arduino-Serial-Sketch/run_serial_test/run_serial_test.ino
Arduino-Serial-Sketch/run_serial_test/run_serial_test.ino
void setup() { Serial.begin(9600); Serial.println("Application Started"); } void loop() { Serial.println("Hello From The Other Side!"); delay(1000); }
Add arduino serial test sketch
Add arduino serial test sketch
Arduino
mit
helmeligi/Arduino-Serial-to-Java-Application
b297c901ebe9418f4c8a2e880717afaae1a8845d
platformio/src_str/adc_string.ino
platformio/src_str/adc_string.ino
/* Prototypical arduino/teensy code. This one sends things as chars, which allows it to be a bit more flexible, at the cost of efficiency. For example, sending the timestamp + two analog channels takes 8 bytes in the more efficient code, but ~15 bytes in this code. Additionally, you would need to parse the string on the other side. The plus is that the output is human-readable. */ #include "ADC.h" #include "IntervalTimer.h" // only the lines below needs to change // first line does which analog channels to read, // second line sets the sampling interval (in microseconds) const unsigned int channel_array[2] = {A0, A7}; const unsigned long period_0 = 100000; const unsigned int array_size = sizeof(channel_array) / sizeof(int); unsigned int value_array[array_size]; unsigned int ii = 0; volatile bool go_flag = false; bool go_flag_copy = false; elapsedMicros current_time; unsigned long current_time_copy = 0; IntervalTimer timer_0; ADC *adc = new ADC(); void setup() { for(ii = 0; ii < array_size; ii++) { pinMode(channel_array[ii], INPUT); } Serial.begin(9600); delay(1000); adc->setReference(ADC_REF_3V3, ADC_0); adc->setAveraging(8); adc->setResolution(12); adc->setConversionSpeed(ADC_HIGH_SPEED); adc->setSamplingSpeed(ADC_HIGH_SPEED); timer_0.priority(10); timer_0.begin(timerCallback, period_0); delay(500); } FASTRUN void timerCallback(void) { go_flag = true; } void loop() { while(!go_flag_copy) { noInterrupts(); go_flag_copy = go_flag; interrupts(); } go_flag_copy = false; go_flag = false; current_time_copy = current_time; for (ii = 0; ii < array_size; ii++) { value_array[ii] = adc->analogRead(channel_array[ii]); } Serial.print(current_time_copy); Serial.print(" "); for (ii = 0; ii < array_size; ii++) { Serial.print(value_array[ii]); Serial.print(" "); } Serial.print("\n"); }
Add string equivalent example for teensy
Add string equivalent example for teensy
Arduino
mit
BLAM-Lab-Projects/finger-6
7567b39b65c4979d9667c71e038722a4f1a68cb4
examples/HIH61xx_Wire_demo/HIH61xx_Wire_demo.ino
examples/HIH61xx_Wire_demo/HIH61xx_Wire_demo.ino
// This example demonstrates how to use the HIH61xx class with the Wire library. The HIH61xx state machine // enables others tasks to run whilst the HIH61xx is powering up etc. #include <Wire.h> #include <HIH61xx.h> #include <AsyncDelay.h> // The "hih" object must be created with a reference to the "Wire" object which represents the I2C bus it is using. // Note that the class for the Wire object is called "TwoWire", and must be included in the templated class name. HIH61xx<TwoWire> hih(Wire); AsyncDelay samplingInterval; void setup(void) { #if F_CPU >= 12000000UL Serial.begin(115200); #else Serial.begin(9600); #endif Wire.begin(); hih.initialise(); samplingInterval.start(3000, AsyncDelay::MILLIS); } bool printed = true; void loop(void) { if (samplingInterval.isExpired() && !hih.isSampling()) { hih.start(); printed = false; samplingInterval.repeat(); Serial.println("Sampling started (using Wire library)"); } hih.process(); if (hih.isFinished() && !printed) { printed = true; // Print saved values Serial.print("RH: "); Serial.print(hih.getRelHumidity() / 100.0); Serial.println(" %"); Serial.print("Ambient: "); Serial.print(hih.getAmbientTemp() / 100.0); Serial.println(" deg C"); Serial.print("Status: "); Serial.println(hih.getStatus()); } }
Add demo which uses the Wire library
Add demo which uses the Wire library
Arduino
lgpl-2.1
stevemarple/HIH61xx
f09501f3914c0f75f03301589bd95fc52c277990
Arduino/LightPlayer/LightPlayer.ino
Arduino/LightPlayer/LightPlayer.ino
#include <SPI.h> #include <SdFat.h> #include <FAB_LED.h> apa106<D, 6> LEDstrip; rgb frame[200]; // Test with reduced SPI speed for breadboards. // Change spiSpeed to SPI_FULL_SPEED for better performance // Use SPI_QUARTER_SPEED for even slower SPI bus speed const uint8_t spiSpeed = SPI_FULL_SPEED; //------------------------------------------------------------------------------ // File system object. SdFat sd; // Serial streams ArduinoOutStream cout(Serial); // SD card chip select const int chipSelect = 4; void setup() { Serial.begin(9600); // Wait for USB Serial while (!Serial) { SysCall::yield(); } cout << F("\nInitializing SD.\n"); if (!sd.begin(chipSelect, spiSpeed)) { if (sd.card()->errorCode()) { cout << F("SD initialization failed.\n"); cout << F("errorCode: ") << hex << showbase; cout << int(sd.card()->errorCode()); cout << F(", errorData: ") << int(sd.card()->errorData()); cout << dec << noshowbase << endl; return; } cout << F("\nCard successfully initialized.\n"); if (sd.vol()->fatType() == 0) { cout << F("Can't find a valid FAT16/FAT32 partition.\n"); return; } if (!sd.vwd()->isOpen()) { cout << F("Can't open root directory.\n"); return; } cout << F("Can't determine error type\n"); return; } cout << F("\nCard successfully initialized.\n"); cout << endl; if (!sd.exists("FRACTAL1.DAT")) { cout << F("FRACTAL1.DAT file not found.\n"); return; } File infile = sd.open("FRACTAL1.DAT"); if (!infile.isOpen()) { cout << F("Failed to open FRACTAL1.DAT\n"); return; } int bytes_read = infile.read(frame, sizeof(frame)); unsigned long prev_millis = millis(); cout << F("\nFrame size in bytes: ") << sizeof(frame); cout << F("\nStarting millis: ") << prev_millis; int i = 0; while (bytes_read == sizeof(frame)) { ++i; while (millis() - prev_millis < 50UL) { // busy loop until its time to paint the lights } prev_millis += 50UL; LEDstrip.sendPixels(sizeof(frame) / sizeof(*frame), frame); bytes_read = infile.read(frame, sizeof(frame)); } cout << F("\nFinal millis: ") << prev_millis; cout << F("\nNum frames: ") << i; } void loop() { }
Add draft of Arduino Read SD/Write Lights loop
Add draft of Arduino Read SD/Write Lights loop Lots of debugging in here. Currently only handles one file on the SD card - no logic yet to switch to the next when the file runs out. Timing is correct, though: every 50 ms, one LED refresh of all 200 lights is sent, then a new frame is read and we busy loop for the remainder of the 50ms time slice.
Arduino
mit
godlygeek/LightRender,MaddAddaM/LightRender
9ffc28c686ed253fd251a1b6854cbcf4fdd4614d
examples/AnalogReadEasyctrl/AnalogReadEasyctrl.ino
examples/AnalogReadEasyctrl/AnalogReadEasyctrl.ino
/* AnalogReadEasyctrl Reads an analog input on pin 0. Attach the center pin of a potentiometer to pin A0, and the outside pins to +5V and ground. This example code is based on the Arduino example AnalogReadSerial */ #include "easyctrl.h" Monitored<int> sensorValue("sensorValue"); // the setup routine runs once when you press reset: void setup() { // initialize serial communication at 115200 bits per second: Serial.begin(115200); Easyctrl.begin("AnalogReadEasyctrl", Serial); } // the loop routine runs over and over again forever: void loop() { // read the input on analog pin 0: sensorValue = analogRead(A0); // Let easyctrl run Easyctrl.update(); delay(1); // delay in between reads for stability }
Add example of outputting an analog read value
Add example of outputting an analog read value
Arduino
mit
arachnidlabs/easyctrl
af42a4926359833c4b3e2076f65c3601b32e523a
examples/TurnOnKelvinatorAC/TurnOnKelvinatorAC.ino
examples/TurnOnKelvinatorAC/TurnOnKelvinatorAC.ino
#include <IRKelvinator.h> IRKelvinatorAC kelvir(D1); // IR led controlled by Pin D1. void printState() { // Display the settings. Serial.println("Kelvinator A/C remote is in the following state:"); Serial.printf(" Basic\n Power: %d, Mode: %d, Temp: %dC, Fan Speed: %d\n", kelvir.getPower(), kelvir.getMode(), kelvir.getTemp(), kelvir.getFan()); Serial.printf(" Options\n X-Fan: %d, Light: %d, Ion Filter: %d\n", kelvir.getXFan(), kelvir.getLight(), kelvir.getIonFilter()); Serial.printf(" Swing (V): %d, Swing (H): %d, Turbo: %d, Quiet: %d\n", kelvir.getSwingVertical(), kelvir.getSwingHorizontal(), kelvir.getTurbo(), kelvir.getQuiet()); // Display the encoded IR sequence. unsigned char* ir_code = kelvir.getRaw(); Serial.print("IR Code: 0x"); for (int i = 0; i < KELVINATOR_STATE_LENGTH; i++) Serial.printf("%02X", ir_code[i]); Serial.println(); } void setup(){ kelvir.begin(); Serial.begin(115200); delay(200); // Set up what we want to send. See IRKelvinator.cpp for all the options. // Most things default to off. Serial.println("Default state of the remote."); printState(); Serial.println("Setting desired state for A/C."); kelvir.on(); kelvir.setFan(1); kelvir.setMode(KELVINATOR_COOL); kelvir.setTemp(26); kelvir.setSwingVertical(false); kelvir.setSwingHorizontal(true); kelvir.setXFan(true); kelvir.setIonFilter(false); kelvir.setLight(true); } void loop() { // Now send the IR signal. Serial.println("Sending IR command to A/C ..."); kelvir.send(); printState(); delay(5000); }
Add example code for Kelvinator A/C control.
Add example code for Kelvinator A/C control.
Arduino
lgpl-2.1
markszabo/IRremoteESP8266,markszabo/IRremoteESP8266,markszabo/IRremoteESP8266,markszabo/IRremoteESP8266
600c765ffe3ab6056c4b5baccc73e910fb098915
libraries/DS18B20/examples/CosaDS18B20calc/CosaDS18B20calc.ino
libraries/DS18B20/examples/CosaDS18B20calc/CosaDS18B20calc.ino
/** * @file CosaDS18B20calc.ino * @version 1.0 * * @section License * Copyright (C) 2015, Mikael Patel * * This library is free software; you can redistribute it and/or * modify it under the terms of the GNU Lesser General Public * License as published by the Free Software Foundation; either * version 2.1 of the License, or (at your option) any later version. * * This library is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * Lesser General Public License for more details. * * @section Description * * This file is part of the Arduino Che Cosa project. */ #include "Cosa/Trace.hh" #include "Cosa/IOStream/Driver/UART.hh" #include "Cosa/Watchdog.hh" void setup() { uart.begin(9600); trace.begin(&uart, PSTR("CosaDS18B20calc: started")); } int32_t iscale(int16_t temp) { bool negative = false; if (temp < 0) { temp = -temp; negative = true; } int32_t res = (temp >> 4) * 1000L + (625 * (temp & 0xf)); return (negative ? -res : res); } float32_t fscale(int16_t temp) { return (temp * 0.0625); } #define CHECK(c,t) trace << PSTR(#c "=") << fscale(t) << endl void loop() { CHECK(+125, 0x07D0); CHECK(+85, 0x0550); CHECK(+25.0625, 0x0191); CHECK(+10.125, 0x00A2); CHECK(+0.5, 0x0008); CHECK(0, 0x0000); CHECK(-0.5, 0xFFF8); CHECK(-10.125, 0xFF5E); CHECK(-25.0625, 0xFE6F); CHECK(-55, 0xFC90); ASSERT(true == false); }
Add validation of DS18B20 temperature calculation.
Add validation of DS18B20 temperature calculation.
Arduino
lgpl-2.1
mikaelpatel/Cosa,mikaelpatel/Cosa,jeditekunum/Cosa,dansut/Cosa,dansut/Cosa,mikaelpatel/Cosa,jeditekunum/Cosa,jeditekunum/Cosa,dansut/Cosa,dansut/Cosa,jeditekunum/Cosa,mikaelpatel/Cosa
35c43899ec3b9d097248e5db6bc8cfd288a7a63a
sketches/clearButtonReceiver.ino
sketches/clearButtonReceiver.ino
/* Kevyn McPhail Deeplocal FOB Receiving module code If Button A is pressed the the arduino returns 1, if button 2 is pressed the arduino returns 2 Button A input is PIN 3, Button B input is PIN 2, and the momentary button press input is PIN 4. On the R02A receiving module, Button A is output D2, Button B is output D3, Momentary button press is output VT. Hardware: Sparkfun Pro Micro 5V/16MHz */ void setup(){ Serial.begin(9600); for (int i = 2; i<5; i++){ pinMode(i, INPUT); } } int firstPin; int secondPin; int thirdPin; void loop(){ firstPin = digitalRead(3); secondPin = digitalRead(2); thirdPin = digitalRead(4); if (firstPin == 1 & secondPin == 0 & thirdPin == 1) { Serial.println(1); delay(200); } if (firstPin == 0 & secondPin == 1 & thirdPin == 1) { Serial.println(2); delay(200); } }
Add Kevyn's original clear button sketch
Add Kevyn's original clear button sketch
Arduino
apache-2.0
EndPointCorp/appctl,EndPointCorp/appctl
d8fcd069f9b3119663da95eb144d2aa811cb6399
src/serial-print.ino
src/serial-print.ino
/* Header file that allows for writing of data from arduino to pi */ #ifndef serial-print_h #define serial-print_h #include "Arduino.h" import processing.serial.*; Serial mySerial; PrintWriter output; void setup() { mySerial = new Serial( this, Serial.list()[0], 9600 ); output = createWriter( "data.txt" ); } void draw() { if (mySerial.available() > 0 ) { String value = mySerial.readString(); if ( value != null ) { output.println( value ); } } } void keyPressed() { output.flush(); // Writes the remaining data to the file output.close(); // Finishes the file exit(); // Stops the program } #endif
Add file to allow saving arduino output from serial
Add file to allow saving arduino output from serial
Arduino
mit
AGilchrist0/balloon
4804830b82330a0788bb4ff5c16622daaecfd55a
examples/simpletest/simpletest.ino
examples/simpletest/simpletest.ino
#include <Arduino.h> #include <SPI.h> #include <ssd1351.h> // use this to do Color c = RGB(...) instead of `RGB c = RGB(...)` or ssd1351::LowColor c = RGB(...) // because it's slightly faster and guarantees you won't be sending wrong colours to the display. // Choose color depth - LowColor and HighColor currently supported // typedef ssd1351::LowColor Color; typedef ssd1351::HighColor Color; // Choose display buffering - NoBuffer or SingleBuffer currently supported // auto display = ssd1351::SSD1351<Color, ssd1351::NoBuffer, 128, 96>(); auto display = ssd1351::SSD1351<Color, ssd1351::SingleBuffer, 128, 96>(); bool up = false; int pos = 127; const int particles = 256; int offsets[particles]; int x_pos[particles]; int y_pos[particles]; Color particle_colors[particles]; void setup() { Serial.begin(9600); Serial.println("Booting..."); display.begin(); Serial.println("Display set up."); for (int i = 0; i < particles; i++) { x_pos[i] = random(0, 128); y_pos[i] = random(0, 96); particle_colors[i] = ssd1351::RGB(0, i + 10, i/2 + 10); } } void loop() { unsigned long before = millis(); display.fillScreen(ssd1351::RGB()); Color circleColor = ssd1351::RGB(0, 128, 255); for (int i = 0; i < particles; i++) { offsets[i] += random(-2, 3); display.drawLine( x_pos[i] + offsets[i], y_pos[i] + offsets[i], pos, 80 + sin(pos / 4.0) * 20, particle_colors[i] ); display.drawCircle( x_pos[i] + offsets[i], y_pos[i] + offsets[i], 1, circleColor ); } display.updateScreen(); Serial.println(millis() - before); if (up) { pos++; if (pos >= 127) { up = false; } } else { pos--; if (pos < 0) { up = true; } } }
Add simple example for the library
Add simple example for the library
Arduino
bsd-2-clause
kirberich/teensy_ssd1351,kirberich/teensy_ssd1351,kirberich/teensy_ssd1351,kirberich/teensy_ssd1351,kirberich/teensy_ssd1351
ff3244494c84f403837590209e3026a5ac24b8c4
examples/EthernetCustom/EthernetCustom.ino
examples/EthernetCustom/EthernetCustom.ino
#include <SPI.h> #include <Ethernet.h> #include "VdlkinoEthernet.h" byte mac[] = { 0xDE, 0xAD, 0xBE, 0xEF, 0xFE, 0xED }; IPAddress ip(192,168,1,177); EthernetServer server(80); VdlkinoEthernet vdlkino(14, 6, &server); void url_get_analog_byte(void *block, char *url) { VdlkinoBlock *vblock = (VdlkinoBlock*) block; char buffer[256]; char *pc = buffer; strcpy(pc, url); if (pc[0] == '/') { pc++; } pc = strtok(pc, "/"); if (strcmp(pc, "analogbyte")) { return; } pc = strtok(NULL, "/"); if (pc == NULL) { return; } vblock->pin = atoi(pc); pc = strtok(NULL, "/"); if (pc != NULL) { return; } vblock->oper = 8; vblock->value = 0; vblock->valid = 1; } uint16_t get_analog_byte(void *block) { VdlkinoBlock *vblock = (VdlkinoBlock*) block; return map(analogRead(vblock->pin), 0, 1023, 0, 255); } void setup() { Ethernet.begin(mac, ip); server.begin(); vdlkino.operations[8] = &get_analog_byte; vdlkino.addUrl(&url_get_analog_byte); } void loop() { vdlkino.run(); }
Add example of a custom function in ethernet
Add example of a custom function in ethernet
Arduino
mit
eduardoklosowski/vdlkino,eduardoklosowski/vdlkino
8e2da67c6a9a1724969e75e071a925718cf0ec87
arduino_atu.ino
arduino_atu.ino
// Copyright 2013 David Turnbull AE9RB // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. // This is an example application for Arduino to read band // data from the Peaberry V2 ATU port. Typical usage is to // change the band on your amplifier or switch antennas. // No support circuitry is needed. // Connect a 1/8" jack directly to Arduino pin. // Leave the ring unconnected, it is not used. #define ATU_0_PIN 2 void setup() { Serial.begin(9600); } // This example prints the band number to the serial // port whenever a change is detected. void loop() { static int band = 0; int i = atu_read(); if (i) { if (band != i) { Serial.println(i); } band = i; } } // Returns a non-zero value when the read is complete. int atu_read() { static int state = 5, data[4], previous; static long t; long m = micros(); int i, ret = 0; if (state < 6) switch(state) { default: i = digitalRead(ATU_0_PIN); if (m - t > 7000) state = 5; if (previous == HIGH && i == LOW) { data[state] = m - t; state++; } if (previous == LOW && i == HIGH) { t = m; } previous = i; break; case 4: for (i=0; i<4; i++) { ret <<= 1; if (data[i] > 2750) ret |= 0x01; } //nobreak; case 5: t = m + 50000; digitalWrite(ATU_0_PIN, LOW); state = 6; break; } else if (t - m < 0) switch(state) { case 6: t = m + 50000; digitalWrite(ATU_0_PIN, HIGH); state = 7; break; case 7: t = m + 5000; digitalWrite(ATU_0_PIN, LOW); state = 8; break; case 8: t = m; previous = LOW; state = 0; break; } return ret; }
Add example for using Arduino to read ATU data
Add example for using Arduino to read ATU data
Arduino
apache-2.0
AE9RB/peaberry,AE9RB/peaberry
f9643d3004928603d431360242875c313b4b171d
examples/BoardWithCustomData/BoardWithCustomData.ino
examples/BoardWithCustomData/BoardWithCustomData.ino
#include <DSPI.h> #include <OpenBCI_32bit_Library.h> #include <OpenBCI_32Bit_Library_Definitions.h> unsigned long timer = 0; byte LEDState = 0; void setup() { // Bring up the OpenBCI Board board.begin(); timer = millis(); LEDState = 1; digitalWrite(OPENBCI_PIN_LED,HIGH); } void loop() { // Downsample if ((millis() - timer) > 20) { // Save new time timer = millis(); sendLEDStatus(); } // Check the serial port for new data if (board.hasDataSerial0()) { // Read one char and process it char c = board.getCharSerial0(); if (c == '0') { // Make the LED turn OFF when a '0' is sent from the PC digitalWrite(OPENBCI_PIN_LED,LOW); LEDState = 0; } else if (c == '1') { // Make the LED turn ON when a '1' is sent from the PC digitalWrite(OPENBCI_PIN_LED,HIGH); LEDState = 1; } } } void sendLEDStatus() { // Must have header byte Serial0.write('A'); // 0x41 1 byte // Write the LED state Serial0.write(LEDState); // 1 byte // Fill the rest with fake data for (int i = 0; i < 30; i++) { Serial0.write(0x00); } // Send a stop byte with an `B` or `1011` in the last nibble to indicate a // different packet type. Serial0.write(0xCB); // 1 byte }
Add example for sending custom data packets
Add example for sending custom data packets
Arduino
mit
OpenBCI/OpenBCI_32bit_Libraries,OpenBCI/OpenBCI_32bit_Libraries
1ba08dd0ae83128c3574bfef3e198e8cac31f9e7
tinkering/sensorboard/light_sensor_check/light_sensor_check.ino
tinkering/sensorboard/light_sensor_check/light_sensor_check.ino
#define LED P1_3 #define Sensor P1_4 float reading; void setup() { // put your setup code here, to run once: Serial.begin(9600); pinMode(LED, OUTPUT); pinMode(Sensor, INPUT); } void loop() { // put your main code here, to run repeatedly: // Turn on LED digitalWrite(LED, HIGH); // Read sensor int i; for (i=1;i<10;i++) { reading = analogRead(Sensor); Serial.print(reading); Serial.print(",\n"); delay(100); } // Turn on LED digitalWrite(LED, LOW); for (i=1;i<10;i++) { reading = analogRead(Sensor); Serial.print(reading); Serial.print(",\n"); delay(100); } }
Test for photoresistor on sensor board made in class.
Test for photoresistor on sensor board made in class.
Arduino
mit
fkmclane/derailleurs,fkmclane/derailleurs,fkmclane/derailleurs
543f1b7c048827d19877dad48c3b552d2d33eb1d
L9110_hbridge/L9110_simple_libtest/L9110_simple_libtest.ino
L9110_hbridge/L9110_simple_libtest/L9110_simple_libtest.ino
// Test L9110 library #include <L9110.h> L9110 L9110; // setup pin for LiPo monitor int LiPoMonitor = 2; void setup() { pinMode(LiPoMonitor, INPUT); } void loop() { if (digitalRead(LiPoMonitor) == LOW) { L9110.forward(); delay(3000); L9110.fullStop(); delay(500); L9110.reverse(); delay(3000); L9110.fullStop(); delay(500); L9110.forwardSlow(); delay(3000); L9110.fullStop(); delay(500); L9110.reverseSlow(); delay(3000); L9110.fullStop(); delay(500); L9110.left(); delay(1500); L9110.fullStop(); delay(500); L9110.right(); delay(1500); L9110.fullStop(); delay(3000); } }
Test of new simple functions in L9110 library
Test of new simple functions in L9110 library
Arduino
mit
AllenRDCo/Arduino
84afd3e29a0afa8d2c8e2b6e78fcf210f32982a6
examples/pcf8574_arduinomicro/pcf8574_arduinomicro.ino
examples/pcf8574_arduinomicro/pcf8574_arduinomicro.ino
/* Example sketch for the PCF8574 for the purposes of showing how to use the interrupt-pin. Attach the positive lead of an LED to PIN7 on the PCF8574 and the negative lead to GND, a wire from Arduino-pin 13 to pin 3 on the PCF8474, a wire from the int-pin on the PCF8574 to Arduino-pin 7 and wires for SDA and SCL to Arduino-pins 2 and 3, respectively. If all goes well you should see the small blue LED on the ESP-module lighting up and the LED connected to the PCF going off, and vice versa. */ #include <pcf8574_esp.h> /* We need to set up the I2C-bus for the library to use */ #include <Wire.h> // Initialize a PCF8574 at I2C-address 0x20 PCF857x pcf8574(0x20, &Wire); //If you had a PCF8575 instead you'd use the below format //PCF857x pcf8575(0x20, &Wire, true); bool PCFInterruptFlag = false; void PCFInterrupt() { PCFInterruptFlag = true; } void setup() { Serial.begin(115200); delay(5000); Serial.println(F("Firing up...")); pinMode(13, OUTPUT); Wire.begin(); //Set to 400KHz Wire.setClock(400000L); pcf8574.begin(); // Most ready-made PCF8574-modules seem to lack an internal pullup-resistor, so you have to use the MCU-internal one. pinMode(7, INPUT_PULLUP); pcf8574.resetInterruptPin(); attachInterrupt(digitalPinToInterrupt(7), PCFInterrupt, FALLING); } void loop() { if(PCFInterruptFlag){ Serial.println(F("Got an interrupt: ")); if(pcf8574.read(3)==1) Serial.println("Pin 3 is HIGH!"); else Serial.println("Pin 3 is LOW!"); // DO NOTE: When you write LOW to a pin on a PCF8574 it becomes an OUTPUT. // It wouldn't generate an interrupt if you were to connect a button to it that pulls it HIGH when you press the button. // Any pin you wish to use as input must be written HIGH and be pulled LOW to generate an interrupt. pcf8574.write(7, pcf8574.read(3)); PCFInterruptFlag=false; } Serial.println(F("Blink.")); if(digitalRead(13)==HIGH) digitalWrite(13, LOW); else digitalWrite(13, HIGH); delay(1000); }
Add an example for Arduino Micro (Atmega32u4)
Add an example for Arduino Micro (Atmega32u4)
Arduino
mit
WereCatf/PCF8574_ESP
6dd824617791b455e89ff75aa02337549539e509
serial/low-speed-transmitter/low-speed-transmitter.ino
serial/low-speed-transmitter/low-speed-transmitter.ino
void setup() { pinMode(LED_BUILTIN, OUTPUT); SerialUSB.begin(2000000); } void loop() { static int counter = 0; SerialUSB.println(counter, DEC); counter = (counter + 1) % (1 << 8); digitalWrite(LED_BUILTIN, counter >> 7 ? HIGH : LOW); delay(20); }
Add low speed serial transmitter example
Add low speed serial transmitter example
Arduino
mit
isaias-b/arduino-examples,isaias-b/arduino-examples,isaias-b/arduino-examples
9aba36ce4d37fb72c446addf5f355cdb2bce17df
examples/On_Chip_Calibration/On_Chip_Calibration.ino
examples/On_Chip_Calibration/On_Chip_Calibration.ino
#include<CS5490.h> #define rx 11 #define tx 12 /* Choose your board */ /* Arduino UNO and ESP8622 */ CS5490 line(MCLK_default,rx,tx); /* ESP and MEGA (Uses Serial2)*/ //CS5490 line(MCLK_default); void setup() { //Initializing communication with CS5490 //600 is the default baud rate velocity. line.begin(600); //Initializing communication arduino/PC to show results in Monitor Serial Serial.begin(115200); // wait for serial port to connect. Needed for Leonardo only while (!Serial); //Set to continous conversion line.contConv(); delay(100); } void loop() { double foo; Serial.println("\n\nWithout calibration"); line.setDcOffsetI(0); //Reset previous calibration foo = line.getInstI(); Serial.print("DC current value: "); Serial.println(foo, 5); foo = line.getDcOffsetI(); Serial.print("DC offset current value: "); Serial.println(foo, 5); /* -------->Types DCoffset ACoffset Gain -------->Channels Current Voltage CurrentAndVoltage -------->How to use? line.calibrate(type,channel) */ line.calibrate(DCoffset,Current); Serial.println("\n\nCalibrated"); foo = line.getInstI(); Serial.print("DC current value: "); Serial.println(foo, 5); foo = line.getDcOffsetI(); Serial.print("DC offset current value: "); Serial.println(foo, 5); Serial.println("\nReset arduino to see it again... "); while(1); }
Add On Chip Calibration example
Add On Chip Calibration example
Arduino
mit
tiagolobao/CS5490
915dba4fde478cfbca640cef0dfb5ea5c60a7ce0
README.adoc
README.adoc
This is a small Java library for parsing the Cloud Foundry environment variables (VCAP_SERVICES and so on). // the first line of this file is used as a description in the POM, so keep it short and sweet! Download from Bintray: image::https://api.bintray.com/packages/pivotal-labs-london/maven/cf-env/images/download.svg[link="https://bintray.com/pivotal-labs-london/maven/cf-env/_latestVersion"] Build with Gradle: -------------------------------------- ./gradlew build -------------------------------------- Release with Gradle: -------------------------------------- # you probably want to make these changes manually rather than like this sed -i -e "s/^version = .*/version = 'x.y.z'/" build.gradle echo -e "bintrayUser=pivotal-labs-london\nbintrayKey=..." >gradle.properties ./gradlew bintrayUpload --------------------------------------
This is a small Java library for parsing the Cloud Foundry environment variables (VCAP_SERVICES and so on). // the first line of this file is used as a description in the POM, so keep it short and sweet! Download from Bintray: image::https://api.bintray.com/packages/pivotal-labs-london/maven/cf-env/images/download.svg[link="https://bintray.com/pivotal-labs-london/maven/cf-env/_latestVersion"] Use as a dependency in your build: -------------------------------------- repositories { jcenter() // or mavenCentral, for transitive dependencies maven { url = 'http://dl.bintray.com/pivotal-labs-london/maven/' } } dependencies { compile group: 'io.pivotal.labs', name: 'cf-env', version: '0.0.1' } -------------------------------------- We hope to have the artifact available via JCenter soon, but until then, please use the repository on Bintray. Build with Gradle: -------------------------------------- ./gradlew build -------------------------------------- Release with Gradle: -------------------------------------- # you probably want to make these changes manually rather than like this sed -i -e "s/^version = .*/version = 'x.y.z'/" build.gradle echo -e "bintrayUser=pivotal-labs-london\nbintrayKey=..." >gradle.properties ./gradlew bintrayUpload --------------------------------------
Document how to use the Bintray repo
Document how to use the Bintray repo
AsciiDoc
bsd-2-clause
pivotal/cf-env
3f83df33d9efe5b4a7f2fe0ad7bcce0f586401d2
README.adoc
README.adoc
= Auxly image:http://img.shields.io/:license-mit-blue.svg["License", link="https://github.com/jeffrimko/Qprompt/blob/master/LICENSE"] image:https://travis-ci.org/jeffrimko/Auxly.svg?branch=master["Build Status"] == Introduction This project provides a Python 2.7/3.x library for common tasks especially when writing shell-like scripts. Some of the functionality overlaps with the standard library but the API is slightly modified. == Status The status of this project is **pre-alpha**. This project is not yet suitable for use other than testing. == Requirements Auxly should run on any Python 2.7/3.x interpreter without additional dependencies. == Usage === The following are basic examples of Auxly (all examples can be found https://github.com/jeffrimko/Auxly/tree/master/examples[here]): - https://github.com/jeffrimko/Auxly/blob/master/examples/delete_1.py[examples/delete_1.py] - Deletes all PYC files in the project. ---- include::examples\delete_1.py[] ----
= Auxly image:http://img.shields.io/:license-mit-blue.svg["License", link="https://github.com/jeffrimko/Qprompt/blob/master/LICENSE"] image:https://travis-ci.org/jeffrimko/Auxly.svg?branch=master["Build Status"] == Introduction This project provides a Python 2.7/3.x library for common tasks especially when writing shell-like scripts. Some of the functionality overlaps with the standard library but the API is slightly modified. == Status The status of this project is **pre-alpha**. This project is not yet suitable for use other than testing. == Requirements Auxly should run on any Python 2.7/3.x interpreter without additional dependencies. == Usage The following are basic examples of Auxly (all examples can be found https://github.com/jeffrimko/Auxly/tree/master/examples[here]): - https://github.com/jeffrimko/Auxly/blob/master/examples/delete_1.py[examples/delete_1.py] - Deletes all PYC files in the project.
Undo include change in readme.
Undo include change in readme.
AsciiDoc
mit
jeffrimko/Auxly
f1b569bfc0f68e487de62febdf6a286d939fa5ae
modules/configuring-scale-bounds-knative.adoc
modules/configuring-scale-bounds-knative.adoc
// Module included in the following assemblies: // // * serverless/configuring-knative-serving-autoscaling.adoc [id="configuring-scale-bounds-knative_{context}"] = Configuring scale bounds Knative Serving autoscaling The `minScale` and `maxScale` annotations can be used to configure the minimum and maximum number of Pods that can serve applications. These annotations can be used to prevent cold starts or to help control computing costs. minScale:: If the `minScale` annotation is not set, Pods will scale to zero (or to 1 if enable-scale-to-zero is false per the `ConfigMap`). maxScale:: If the `maxScale` annotation is not set, there will be no upper limit for the number of Pods created. `minScale` and `maxScale` can be configured as follows in the revision template: [source,yaml] ---- spec: template: metadata: autoscaling.knative.dev/minScale: "2" autoscaling.knative.dev/maxScale: "10" ---- Using these annotations in the revision template will propagate this confguration to `PodAutoscaler` objects. [NOTE] ==== These annotations apply for the full lifetime of a revision. Even when a revision is not referenced by any route, the minimal Pod count specified by `minScale` will still be provided. Keep in mind that non-routeable revisions may be garbage collected, which enables Knative to reclaim the resources. ====
// Module included in the following assemblies: // // * serverless/configuring-knative-serving-autoscaling.adoc [id="configuring-scale-bounds-knative_{context}"] = Configuring scale bounds Knative Serving autoscaling The `minScale` and `maxScale` annotations can be used to configure the minimum and maximum number of Pods that can serve applications. These annotations can be used to prevent cold starts or to help control computing costs. minScale:: If the `minScale` annotation is not set, Pods will scale to zero (or to 1 if enable-scale-to-zero is false per the `ConfigMap`). maxScale:: If the `maxScale` annotation is not set, there will be no upper limit for the number of Pods created. `minScale` and `maxScale` can be configured as follows in the revision template: [source,yaml] ---- spec: template: metadata: annotations: autoscaling.knative.dev/minScale: "2" autoscaling.knative.dev/maxScale: "10" ---- Using these annotations in the revision template will propagate this confguration to `PodAutoscaler` objects. [NOTE] ==== These annotations apply for the full lifetime of a revision. Even when a revision is not referenced by any route, the minimal Pod count specified by `minScale` will still be provided. Keep in mind that non-routeable revisions may be garbage collected, which enables Knative to reclaim the resources. ====
Fix spec.template.metadata.annotations for min and max scale example
Fix spec.template.metadata.annotations for min and max scale example
AsciiDoc
apache-2.0
vikram-redhat/openshift-docs,vikram-redhat/openshift-docs,vikram-redhat/openshift-docs,vikram-redhat/openshift-docs,vikram-redhat/openshift-docs
e0992105f2fbecca2909a19b0cdbbb74aa20a561
docs/index.asciidoc
docs/index.asciidoc
= Packetbeat reference :libbeat: http://www.elastic.co/guide/en/beats/libbeat/1.0.0-rc1 :version: 1.0.0-rc1 include::./overview.asciidoc[] include::./gettingstarted.asciidoc[] include::./configuration.asciidoc[] include::./command-line.asciidoc[] include::./capturing.asciidoc[] include::./https.asciidoc[] include::./fields.asciidoc[] include::./thrift.asciidoc[] include::./windows.asciidoc[] include::./kibana3.asciidoc[] include::./filtering.asciidoc[] include::./troubleshooting.asciidoc[] include::./new_protocol.asciidoc[]
= Packetbeat reference :libbeat: http://www.elastic.co/guide/en/beats/libbeat/master :version: master include::./overview.asciidoc[] include::./gettingstarted.asciidoc[] include::./configuration.asciidoc[] include::./command-line.asciidoc[] include::./capturing.asciidoc[] include::./https.asciidoc[] include::./fields.asciidoc[] include::./thrift.asciidoc[] include::./windows.asciidoc[] include::./kibana3.asciidoc[] include::./filtering.asciidoc[] include::./troubleshooting.asciidoc[] include::./new_protocol.asciidoc[]
Use master version in docs
Use master version in docs
AsciiDoc
mit
yapdns/yapdnsbeat,yapdns/yapdnsbeat
0563338e61781db621d9bfbcf201a4c2c753dae4
README.adoc
README.adoc
= AsciiBinder image:https://badge.fury.io/rb/ascii_binder.svg["Gem Version", link="https://badge.fury.io/rb/ascii_binder"] AsciiBinder is an AsciiDoc-based system for authoring and publishing closely related documentation sets from a single source. == Learn More * See the http://www.asciibinder.org[homepage]. * Have a gander at the http://www.asciibinder.org/latest/welcome/[AsciiBinder documentation]. * Or just take the https://rubygems.org/gems/ascii_binder[ascii_binder Ruby Gem] for a spin. The AsciiBinder system was initially developed for https://github.com/openshift/openshift-docs[OpenShift documentation], but has been revised to work for documenting a wide variety of complex, multi-versioned software projects. == Contributing We are using the https://github.com/redhataccess/ascii_binder/issues[Issues] page to track bugs and feature ideas on the code, so have a look and feel free to ask questions there. You can also chat with us on IRC at FreeNode, http://webchat.freenode.net/?randomnick=1&channels=asciibinder&uio=d4[#asciibinder] channel, or on Twitter - https://twitter.com/AsciiBinder[@AsciiBinder]. == License The gem is available as open source under the terms of the http://opensource.org/licenses/MIT[MIT License].
= AsciiBinder image:https://badge.fury.io/rb/ascii_binder.svg["Gem Version", link="https://badge.fury.io/rb/ascii_binder"] AsciiBinder is an AsciiDoc-based system for authoring and publishing closely related documentation sets from a single source. == Learn More * Have a gander at the https://github.com/redhataccess/ascii_binder-docs/blob/master/welcome/index.adoc[AsciiBinder documentation]. * Or just take the https://rubygems.org/gems/ascii_binder[ascii_binder Ruby Gem] for a spin. The AsciiBinder system was initially developed for https://github.com/openshift/openshift-docs[OpenShift documentation], but has been revised to work for documenting a wide variety of complex, multi-versioned software projects. == Contributing We are using the https://github.com/redhataccess/ascii_binder/issues[Issues] page to track bugs and feature ideas on the code, so have a look and feel free to ask questions there. You can also chat with us on IRC at FreeNode, http://webchat.freenode.net/?randomnick=1&channels=asciibinder&uio=d4[#asciibinder] channel, or on Twitter - https://twitter.com/AsciiBinder[@AsciiBinder]. == License The gem is available as open source under the terms of the http://opensource.org/licenses/MIT[MIT License].
Update docs link to point to docs repo
Update docs link to point to docs repo
AsciiDoc
mit
nhr/ascii_binder,redhataccess/doc_site_builder,redhataccess/ascii_binder,redhataccess/ascii_binder,redhataccess/doc_site_builder,nhr/ascii_binder
3ab42490a909b1b2ba73652ffa172172bea9ad85
README.adoc
README.adoc
:figure-caption!: image::https://travis-ci.org/mmjmanders/ng-iban.svg?branch=master[title="travis status", alt="travis status", link="https://travis-ci.org/mmjmanders/ng-iban"] image::https://app.wercker.com/status/eb4337041c62e162c5dd7af43122647c/m[title="wercker status", alt="wercker status", link="https://app.wercker.com/project/bykey/eb4337041c62e162c5dd7af43122647c"] = ng-iban - validate input fields as IBAN The goal is to provide an easy way to validate an input field as an IBAN number with https://angularjs.org/[AngularJS]. From version `0.4.0` the module uses https://github.com/arhs/iban.js[iban.js] for validation. == Usage First add * `AngularJS` * `ng-iban` to your HTML file. Make sure you require `mm.iban` as a dependency of your AngularJS module. == Installation `bower install ng-iban` === directive [source,html] ---- <input type="text" ng-model="iban" ng-iban/> ---- To use this directive the `ngModel` directive must also be used because this directive depends on it.
:figure-caption!: image::https://travis-ci.org/mmjmanders/ng-iban.svg?branch=master[title="travis status", alt="travis status", link="https://travis-ci.org/mmjmanders/ng-iban"] image::https://app.wercker.com/status/eb4337041c62e162c5dd7af43122647c/m[title="wercker status", alt="wercker status", link="https://app.wercker.com/project/bykey/eb4337041c62e162c5dd7af43122647c"] = ng-iban - validate input fields as IBAN The goal is to provide an easy way to validate an input field as an IBAN number with https://angularjs.org/[AngularJS]. From version `0.4.0` the module uses https://github.com/arhs/iban.js[iban.js] for validation. == Installation === Bower `bower install ng-iban` === NPM `npm install ng-iban` === Other Download file `dist/ng-iban.mni.js`. == Usage Add `mm.iban` as a dependency of your AngularJS module. === directive [source,html] ---- <input type="text" ng-model="iban" ng-iban/> ---- To use this directive the `ngModel` directive must also be used because this directive depends on it.
Update documentation for NPM support
Update documentation for NPM support
AsciiDoc
mit
mmjmanders/ng-iban
768aa1fd4d785b7ab5ccbeca3949f7ed8e73d6d6
README.adoc
README.adoc
= Infinispan Cluster Manager image:https://vertx.ci.cloudbees.com/buildStatus/icon?job=vert.x3-infinispan["Build Status",link="https://vertx.ci.cloudbees.com/view/vert.x-3/job/vert.x3-infinispan/"] This is a cluster manager implementation for Vert.x that uses http://infinispan.org[Infinispan]. Please see the in-source asciidoc documentation or the main documentation on the web-site for a full description of this component: * link:http://vertx.io/docs/vertx-infinispan/java/[web-site docs] * link:src/main/asciidoc/java/index.adoc[in-source docs]
= Infinispan Cluster Manager image:https://vertx.ci.cloudbees.com/buildStatus/icon?job=vert.x3-infinispan["Build Status",link="https://vertx.ci.cloudbees.com/view/vert.x-3/job/vert.x3-infinispan/"] This is a cluster manager implementation for Vert.x that uses http://infinispan.org[Infinispan]. Please see the in-source asciidoc documentation or the main documentation on the web-site for a full description of this component: * link:http://vertx.io/docs/vertx-infinispan/java/[web-site docs] * link:src/main/asciidoc/java/index.adoc[in-source docs] -- will remove --
Revert "Revert "Revert "Revert "Test trigger on push""""
Revert "Revert "Revert "Revert "Test trigger on push"""" This reverts commit 2e835acb457e13f3fc5a49459604488388d80cae.
AsciiDoc
apache-2.0
vert-x3/vertx-infinispan
bea0a3216104837de75798f1cea3ad6202791160
documentation/src/main/asciidoc/ch14.asciidoc
documentation/src/main/asciidoc/ch14.asciidoc
[[validator-further-reading]] == Further reading Last but not least, a few pointers to further information. A great source for examples is the Bean Validation TCK which is available for anonymous access on https://github.com/beanvalidation/beanvalidation-tck/[GitHub]. In particular the TCK's https://github.com/beanvalidation/beanvalidation-tck/tree/master/tests[tests] might be of interest. {bvSpecUrl}[The JSR 380] specification itself is also a great way to deepen your understanding of Bean Validation and Hibernate Validator. If you have any further questions about Hibernate Validator or want to share some of your use cases, have a look at the http://community.jboss.org/en/hibernate/validator[Hibernate Validator Wiki], the https://forum.hibernate.org/viewforum.php?f=9[Hibernate Validator Forum] and the https://stackoverflow.com/questions/tagged/hibernate-validator[Hibernate Validator tag on Stack Overflow]. In case you would like to report a bug use https://hibernate.atlassian.net/projects/HV/[Hibernate's Jira] instance. Feedback is always welcome!
[[validator-further-reading]] == Further reading Last but not least, a few pointers to further information. A great source for examples is the Bean Validation TCK which is available for anonymous access on https://github.com/beanvalidation/beanvalidation-tck/[GitHub]. In particular the TCK's https://github.com/beanvalidation/beanvalidation-tck/tree/master/tests[tests] might be of interest. {bvSpecUrl}[The JSR 380] specification itself is also a great way to deepen your understanding of Bean Validation and Hibernate Validator. If you have any further questions about Hibernate Validator or want to share some of your use cases, have a look at the http://community.jboss.org/en/hibernate/validator[Hibernate Validator Wiki], the https://discourse.hibernate.org/c/hibernate-validator[Hibernate Validator Forum] and the https://stackoverflow.com/questions/tagged/hibernate-validator[Hibernate Validator tag on Stack Overflow]. In case you would like to report a bug use https://hibernate.atlassian.net/projects/HV/[Hibernate's Jira] instance. Feedback is always welcome!
Update the links to the forum
Update the links to the forum
AsciiDoc
apache-2.0
marko-bekhta/hibernate-validator,marko-bekhta/hibernate-validator,hibernate/hibernate-validator,marko-bekhta/hibernate-validator,hibernate/hibernate-validator,hibernate/hibernate-validator
77ccf579cb4a5d8795337d2b85b946c5ecf85475
adoc/regtest-intro.adoc
adoc/regtest-intro.adoc
== Introduction to Regression Test Mode Bitcoin 0.9 and later include support for Regression Test Mode (aka RegTest mode). RegTest mode creates a single node Bitcoin "network" that can confirm blocks upon command. (RegTest mode can also be used to create small, multi-node networks and even to simulate blockchain reorganizations.) For example the following command will generate 101 blocks ./bitcoin-cli -regtest setgenerate true 101 And yes, you get the newly mined coins. They can't be spent anywhere, but they're great for testing. The best documentation of RegTest mode that I've seen so far is https://bitcoinj.github.io/testing[How to test applications] on the new https://bitcoinj.github.io[Bitcoinj website]. Other Links:: * http://geraldkaszuba.com/creating-your-own-experimental-bitcoin-network/[Creating your own experimental Bitcoin network] * https://github.com/gak/docker-bitcoin-regtest[docker-bitcoin-regtest] == Simple Demo of RegTest mode with Bash Scripts These are some really rough Bash scripts that can drive bitcoind in RegTest mode. Procedure:: * Make sure Bitcoin Core 0.9 or later is installed and in your path. * Run the server script ./server.sh & * Run the client setup script to mine some coins to get started: ./setup-client.sh * Run the client script (repeat as desired) ./client.sh * A directory named +regtest-datadir+ is created in the current directory.
== Introduction to Regression Test Mode Bitcoin 0.9 and later include support for Regression Test Mode (aka RegTest mode). RegTest mode creates a single node Bitcoin "network" that can confirm blocks upon command. (RegTest mode can also be used to create small, multi-node networks and even to simulate blockchain reorganizations.) For example the following command will generate 101 blocks ./bitcoin-cli -regtest setgenerate true 101 And yes, you get the newly mined coins. They can't be spent anywhere, but they're great for testing. The best documentation of RegTest mode that I've seen so far is https://bitcoinj.github.io/testing[How to test applications] on the new https://bitcoinj.github.io[Bitcoinj website]. Other Links:: * http://geraldkaszuba.com/creating-your-own-experimental-bitcoin-network/[Creating your own experimental Bitcoin network] * https://github.com/gak/docker-bitcoin-regtest[docker-bitcoin-regtest]
Remove reference to deleted scripts.
Remove reference to deleted scripts.
AsciiDoc
apache-2.0
OmniLayer/OmniJ,OmniLayer/OmniJ,OmniLayer/OmniJ
d0d63e6172efc6e7eb7ca3e8fd29b3a670c18899
subprojects/docs/src/docs/userguide/licenses.adoc
subprojects/docs/src/docs/userguide/licenses.adoc
// Copyright 2017 the original author or authors. // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. = Documentation licenses [[sec:gradle_documentation]] == Gradle Documentation _Copyright © 2007-2018 Gradle, Inc._ Gradle build tool source code is open and licensed under the link:https://github.com/gradle/gradle/blob/master/LICENSE[Apache License 2.0]. Gradle user manual and DSL references are licensed under link:http://creativecommons.org/licenses/by-nc-sa/4.0/[Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License].
// Copyright 2017 the original author or authors. // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. [[licenses]] = License Information [[sec:gradle_documentation]] == Gradle Documentation _Copyright © 2007-2018 Gradle, Inc._ Gradle build tool source code is open-source and licensed under the link:https://github.com/gradle/gradle/blob/master/LICENSE[Apache License 2.0]. Gradle user manual and DSL references are licensed under link:http://creativecommons.org/licenses/by-nc-sa/4.0/[Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License]. [[licenses:build_scan_plugin]] == Gradle Build Scan Plugin Use of the link:https://scans.gradle.com/plugin/[build scan plugin] is subject to link:https://gradle.com/legal/terms-of-service/[Gradle's Terms of Service].
Add license information to docs clarifying build scan plugin license
Add license information to docs clarifying build scan plugin license
AsciiDoc
apache-2.0
blindpirate/gradle,blindpirate/gradle,robinverduijn/gradle,robinverduijn/gradle,robinverduijn/gradle,robinverduijn/gradle,robinverduijn/gradle,gradle/gradle,robinverduijn/gradle,lsmaira/gradle,gradle/gradle,lsmaira/gradle,blindpirate/gradle,lsmaira/gradle,blindpirate/gradle,robinverduijn/gradle,blindpirate/gradle,lsmaira/gradle,gradle/gradle,gradle/gradle,gradle/gradle,blindpirate/gradle,robinverduijn/gradle,lsmaira/gradle,blindpirate/gradle,lsmaira/gradle,blindpirate/gradle,gradle/gradle,lsmaira/gradle,robinverduijn/gradle,gradle/gradle,blindpirate/gradle,gradle/gradle,gradle/gradle,lsmaira/gradle,lsmaira/gradle,robinverduijn/gradle,lsmaira/gradle,gradle/gradle,robinverduijn/gradle,blindpirate/gradle
aba970033b1101b2c1e175a4ba49e513bd4519f7
monitoring/monitoring.adoc
monitoring/monitoring.adoc
[id='monitoring'] = Monitoring include::modules/common-attributes.adoc[] :context: monitoring toc::[] {product-title} uses the Prometheus open source monitoring system. The stack built around Prometheus provides {product-title} cluster monitoring by default. It also provides custom-configured application monitoring as a technology preview. The cluster monitoring stack is only supported for monitoring {product-title} clusters. [id='cluster-monitoring'] == Cluster monitoring include::modules/monitoring-monitoring-overview.adoc[leveloffset=+1] include::monitoring/configuring-monitoring-stack.adoc[leveloffset=+1] include::modules/monitoring-configuring-etcd-monitoring.adoc[leveloffset=+1] include::modules/monitoring-accessing-prometheus-alertmanager-grafana.adoc[leveloffset=+1] [id='application-monitoring'] == Application monitoring You can do custom metrics scraping for your applications. This is done using the Prometheus Operator and a custom Prometheus instance. [IMPORTANT] ==== Application monitoring is a technology preview. Exposing custom metrics will change without the consent of the user of the cluster. ==== include::modules/monitoring-configuring-cluster-for-application-monitoring.adoc[leveloffset=+1] include::modules/monitoring-configuring-monitoring-for-application.adoc[leveloffset=+1] include::modules/monitoring-exposing-application-metrics-for-horizontal-pod-autoscaling.adoc[leveloffset=+1]
[id='monitoring'] = Monitoring include::modules/common-attributes.adoc[] :context: monitoring toc::[] {product-title} uses the Prometheus open source monitoring system. The stack built around Prometheus provides {product-title} cluster monitoring by default. It also provides custom-configured application monitoring as a technology preview. The cluster monitoring stack is only supported for monitoring {product-title} clusters. [id='cluster-monitoring'] == Cluster monitoring include::modules/monitoring-monitoring-overview.adoc[leveloffset=+1] include::monitoring/configuring-monitoring-stack.adoc[leveloffset=+1] include::modules/monitoring-accessing-prometheus-alertmanager-grafana.adoc[leveloffset=+1] [id='application-monitoring'] == Application monitoring You can do custom metrics scraping for your applications. This is done using the Prometheus Operator and a custom Prometheus instance. [IMPORTANT] ==== Application monitoring is a technology preview. Exposing custom metrics will change without the consent of the user of the cluster. ==== include::modules/monitoring-configuring-cluster-for-application-monitoring.adoc[leveloffset=+1] include::modules/monitoring-configuring-monitoring-for-application.adoc[leveloffset=+1] include::modules/monitoring-exposing-application-metrics-for-horizontal-pod-autoscaling.adoc[leveloffset=+1]
Remove the deprecated etcd section
Remove the deprecated etcd section
AsciiDoc
apache-2.0
vikram-redhat/openshift-docs,vikram-redhat/openshift-docs,vikram-redhat/openshift-docs,vikram-redhat/openshift-docs,vikram-redhat/openshift-docs
e71e5cab507a8949b72782b92c3b9f74b33ec729
modules/openshift-developer-cli-installing-odo-on-linux.adoc
modules/openshift-developer-cli-installing-odo-on-linux.adoc
// Module included in the following assemblies: // // * cli_reference/openshift_developer_cli/installing-odo.adoc [id="installing-odo-on-linux"] = Installing {odo-title} on Linux == Binary installation ---- # curl -L https://mirror.openshift.com/pub/openshift-v4/clients/odo/latest/odo-darwin-amd64 -o /usr/local/bin/odo # chmod +x /usr/local/bin/odo ---- == Tarball installation ---- # sh -c 'curl -L https://mirror.openshift.com/pub/openshift-v4/clients/odo/latest/odo-linux-amd64.tar.gz | gzip -d > /usr/local/bin/odo' # chmod +x /usr/local/bin/odo ----
// Module included in the following assemblies: // // * cli_reference/openshift_developer_cli/installing-odo.adoc [id="installing-odo-on-linux"] = Installing {odo-title} on Linux == Binary installation ---- # curl -L https://mirror.openshift.com/pub/openshift-v4/clients/odo/latest/odo-linux-amd64 -o /usr/local/bin/odo # chmod +x /usr/local/bin/odo ---- == Tarball installation ---- # sh -c 'curl -L https://mirror.openshift.com/pub/openshift-v4/clients/odo/latest/odo-linux-amd64.tar.gz | gzip -d > /usr/local/bin/odo' # chmod +x /usr/local/bin/odo ----
Fix url to odo linux binary
Fix url to odo linux binary
AsciiDoc
apache-2.0
vikram-redhat/openshift-docs,vikram-redhat/openshift-docs,vikram-redhat/openshift-docs,vikram-redhat/openshift-docs,vikram-redhat/openshift-docs
4d61fa8fd9011eb6cf4e899c64a8b81fcb2c152c
Documentation/ArchitectureDocumentation.asciidoc
Documentation/ArchitectureDocumentation.asciidoc
ifdef::env-github[] :imagesdir: https://github.com/Moose2Model/Moose2Model/blob/master/Documentation/images/ endif::[] :toc: :toc-placement!: toc::[] This documentation follows the arc42 template for architecture documentation (https://arc42.org/). 1 Introduction and Goals ======================== 1.1 Requirements Overview ------------------------- - Provide diagrams for developers that can be easily kept correct - Reduces the cognitive load of developers who work in complex software systems - Supports only diagram that show the dependencies between components of a software - Supports dependencies between entities of code that are more detailed then a class (method, attribute, ...) - Works in the moment best with models that are extracted by the SAP2Moose project - Shall support all models that are compatible with Moose (http://moosetechnology.org/) 1.2 Quality Goals ----------------- - Shows all dependencies between elements in a diagram - Shows all elements that should be in a diagram 1.3 Stake Holders ----------------- .Stake Holders |=== | Role/Name |Expectations |Developer |Build diagrams where the customiziation is not lost when they are regenerated with new informations. Build diagrams that are sufficiently detailed to support the development. |Software Architect |Have a tool to compare the planned architecture with the realized architecture |===
ifdef::env-github[] :imagesdir: https://github.com/Moose2Model/Moose2Model/blob/master/Documentation/images/ endif::[] :toc: :toc-placement!: toc::[] This documentation follows the arc42 template for architecture documentation (https://arc42.org/). 1 Introduction and Goals ======================== 1.1 Requirements Overview ------------------------- - Provide diagrams for developers that can be easily kept correct - Reduces the cognitive load of developers who work in complex software systems - Supports only diagram that show the dependencies between components of a software - Supports dependencies between entities of code that are more detailed then a class (method, attribute, ...) - Works in the moment best with models that are extracted by the SAP2Moose project - Shall support all models that are compatible with Moose (http://moosetechnology.org/) 1.2 Quality Goals ----------------- - Shows all dependencies between elements in a diagram - Shows all elements that should be in a diagram 1.3 Stake Holders ----------------- .Stake Holders |=== | Role/Name |Expectations |Developer |Build diagrams where the customiziation is not lost when they are regenerated with new informations. Build diagrams that are sufficiently detailed to support the development. |Software Architect |Have a tool to compare the planned architecture with the realized architecture |=== 2 Architecture Constraints ========================== - Easy to install
Add arc 42 Architecture Constraints
Add arc 42 Architecture Constraints
AsciiDoc
mit
RainerWinkler/Moose-Diagram
f31704a0bf79e8a86e03ad938eb4cdd12295d4ff
community/forum.adoc
community/forum.adoc
= Forum (free support) :awestruct-layout: base :showtitle: == Usage questions If you have a question about OptaPlanner, just ask our friendly community: * Ask on http://www.jboss.org/drools/lists[the Drools user mailing list] (recommended). * Or ask on http://stackoverflow.com/questions/tagged/optaplanner[StackOverflow]. Please follow these recommendations when posting a question: * Get to the point. Keep it as short as possible. * Include _relevant_ technical details (stacktrace, short code snippet, ...). * Be polite, friendly and clear. Reread your question before posting it. * Be patient. This isn't link:product.html[paid support]. == Development discussions If you've link:../code/sourceCode.html[build OptaPlanner from source] and you would like to improve it, then come talk with us: * Join us on http://www.jboss.org/drools/lists[the Drools developer mailing list]. * And link:chat.html[chat with us] (recommended).
= Forum :awestruct-layout: base :showtitle: == Usage questions If you have a question about OptaPlanner, just ask our friendly community: * *http://stackoverflow.com/questions/tagged/optaplanner[Ask a usage question on StackOverflow.]* * To start a discussion, use https://groups.google.com/forum/#!forum/optaplanner-dev[the OptaPlanner developer forum]. Please follow these recommendations when posting a question or a discussion: * Get to the point. Keep it as short as possible. * Include _relevant_ technical details (stacktrace, short code snippet, ...). * Be polite, friendly and clear. Reread your question before posting it. * Be patient. This isn't link:product.html[paid support]. == Development discussions If you've link:../code/sourceCode.html[build OptaPlanner from source] and want to improve something, come talk with us: * Join https://groups.google.com/forum/#!forum/optaplanner-dev[the OptaPlanner developer google group]. ** Mail directly to the group via mailto:optaplanner-dev@googlegroups.com[optaplanner-dev@googlegroups.com]. * And link:chat.html[chat with us] (recommended). ** Or ping us on link:socialMedia.html[any social media].
Split optaplanner's dev list away from drools's mailing list
Split optaplanner's dev list away from drools's mailing list
AsciiDoc
apache-2.0
oskopek/optaplanner-website,psiroky/optaplanner-website,psiroky/optaplanner-website,psiroky/optaplanner-website,bibryam/optaplanner-website,bibryam/optaplanner-website,droolsjbpm/optaplanner-website,oskopek/optaplanner-website,oskopek/optaplanner-website,bibryam/optaplanner-website,droolsjbpm/optaplanner-website,droolsjbpm/optaplanner-website
964b9a214abd95783bc74eeb7a91a091a9deb2cc
pages/apim/3.x/kubernetes/apim-kubernetes-overview.adoc
pages/apim/3.x/kubernetes/apim-kubernetes-overview.adoc
[[apim-kubernetes-overview]] = Kubernetes plugin :page-sidebar: apim_3_x_sidebar :page-permalink: apim/3.x/apim_kubernetes_overview.html :page-folder: apim/kubernetes :page-layout: apim3x :page-liquid: [label label-version]#New in version 3.7# == Overview APIM 3.7.0 introduces a Kubernetes plugin for APIM Gateway allowing the deployment of APIs using https://kubernetes.io/docs/concepts/extend-kubernetes/api-extension/custom-resources/[Custom Resource Definitions (CRDs)^]. APIs deployed using CRDs are not visible through APIM Console. WARNING: This plugin is currently in Alpha version; the API key policy isn't available for APIs deployed using this plugin. You can find more detailed information about the plugin in the following sections: * link:/apim/3.x/apim_kubernetes_quick_start.html[Quick Start] * link:/apim/3.x/apim_kubernetes_installation.html[How to install] * link:/apim/3.x/apim_kubernetes_custom_resources.html[Custom Resources] * link:/apim/3.x/apim_kubernetes_admission_hook.html[Admission hook]
[[apim-kubernetes-overview]] = Kubernetes plugin :page-sidebar: apim_3_x_sidebar :page-permalink: apim/3.x/apim_kubernetes_overview.html :page-folder: apim/kubernetes :page-layout: apim3x :page-liquid: [label label-version]#New in version 3.13# == Overview APIM 3.13 introduces a Kubernetes plugin for APIM Gateway allowing the deployment of APIs using https://kubernetes.io/docs/concepts/extend-kubernetes/api-extension/custom-resources/[Custom Resource Definitions (CRDs)^]. APIs deployed using CRDs are not visible through APIM Console. WARNING: This plugin is currently in Alpha version; the API key policy isn't available for APIs deployed using this plugin. You can find more detailed information about the plugin in the following sections: * link:/apim/3.x/apim_kubernetes_quick_start.html[Quick Start] * link:/apim/3.x/apim_kubernetes_installation.html[How to install] * link:/apim/3.x/apim_kubernetes_custom_resources.html[Custom Resources] * link:/apim/3.x/apim_kubernetes_admission_hook.html[Admission hook]
Fix APIM version to 3.13
Fix APIM version to 3.13
AsciiDoc
apache-2.0
gravitee-io/gravitee-docs,gravitee-io/gravitee-docs,gravitee-io/gravitee-docs
fbd8d2ed399230a8e3e1d469e9825eb0602fb981
dsl/camel-cli-connector/src/main/docs/cli-connector.adoc
dsl/camel-cli-connector/src/main/docs/cli-connector.adoc
= CLI Connector Component :doctitle: CLI Connector :shortname: cli-connector :artifactid: camel-cli-connector :description: Runtime adapter connecting with Camel CLI :since: 3.19 :supportlevel: Preview *Since Camel {since}* The camel-cli-connector allows the Camel CLI to be able to manage running Camel integrations. Currently, only a local connector is provided, which means that the Camel CLI can only be managing local running Camel integrations. These integrations can be using different runtimes such as Camel Main, Camel Spring Boot or Camel Quarkus etc. == Auto-detection from classpath To use this implementation all you need to do is to add the `camel-cli-connector` dependency to the classpath, and Camel should auto-detect this on startup and log as follows: [source,text] ---- Local CLI Connector started ----
= CLI Connector Component :doctitle: CLI Connector :shortname: cli-connector :artifactid: camel-cli-connector :description: Runtime adapter connecting with Camel CLI :since: 3.19 :supportlevel: Preview //Manually maintained attributes :camel-spring-boot-name: cli-connector *Since Camel {since}* The camel-cli-connector allows the Camel CLI to be able to manage running Camel integrations. Currently, only a local connector is provided, which means that the Camel CLI can only be managing local running Camel integrations. These integrations can be using different runtimes such as Camel Main, Camel Spring Boot or Camel Quarkus etc. == Auto-detection from classpath To use this implementation all you need to do is to add the `camel-cli-connector` dependency to the classpath, and Camel should auto-detect this on startup and log as follows: [source,text] ---- Local CLI Connector started ---- include::spring-boot:partial$starter.adoc[]
Add spring-boot link in doc
Add spring-boot link in doc
AsciiDoc
apache-2.0
apache/camel,apache/camel,tadayosi/camel,cunningt/camel,tadayosi/camel,cunningt/camel,christophd/camel,cunningt/camel,cunningt/camel,christophd/camel,apache/camel,tadayosi/camel,tadayosi/camel,apache/camel,cunningt/camel,christophd/camel,cunningt/camel,tadayosi/camel,christophd/camel,tadayosi/camel,apache/camel,christophd/camel,apache/camel,christophd/camel
5059698c58cac5335abdf4bb35de4830764ca308
documentation/src/docs/asciidoc/release-notes.adoc
documentation/src/docs/asciidoc/release-notes.adoc
[[release-notes]] == Release Notes :numbered!: include::release-notes-5.0.0-ALPHA.adoc[] include::release-notes-5.0.0-M1.adoc[] include::release-notes-5.0.0-M2.adoc[] include::release-notes-5.0.0-M3.adoc[] include::release-notes-5.0.0-M4.adoc[] include::release-notes-5.0.0-M5.adoc[] include::release-notes-5.0.0-M6.adoc[] include::release-notes-5.0.0-RC1.adoc[] include::release-notes-5.0.0-RC2.adoc[] include::release-notes-5.0.0-RC3.adoc[] include::release-notes-5.0.0.adoc[] include::release-notes-5.1.0-M1.adoc[] :numbered:
[[release-notes]] == Release Notes :numbered!: include::release-notes-5.1.0-M1.adoc[] include::release-notes-5.0.0.adoc[] include::release-notes-5.0.0-RC3.adoc[] include::release-notes-5.0.0-RC2.adoc[] include::release-notes-5.0.0-RC1.adoc[] include::release-notes-5.0.0-M6.adoc[] include::release-notes-5.0.0-M5.adoc[] include::release-notes-5.0.0-M4.adoc[] include::release-notes-5.0.0-M3.adoc[] include::release-notes-5.0.0-M2.adoc[] include::release-notes-5.0.0-M1.adoc[] include::release-notes-5.0.0-ALPHA.adoc[] :numbered:
Order release notes from newest to oldest
Order release notes from newest to oldest Issue: #1066
AsciiDoc
epl-1.0
sbrannen/junit-lambda,junit-team/junit-lambda
6b09164a364abe0e0b7531aca2380a4f33f5c8e4
README.asciidoc
README.asciidoc
== Fusioninventory Plugin https://coveralls.io/r/fusioninventory/fusioninventory-for-glpi[image:https://coveralls.io/repos/fusioninventory/fusioninventory-for-glpi/badge.svg] This plugin makes GLPI to process various types of tasks for Fusioninventory agents: * Computer inventory * Network discovery * Network (SNMP) inventory * Software deployment * VMWare ESX host remote inventory For further information and documentation, please check http://www.fusioninventory.org . If you want to report bugs or check for development status, you can check http://forge.fusioninventory.org . == Third-party code * PluginFusioninventoryFindFiles() is copyright http://rosettacode.org/wiki/Walk_a_directory/Recursively#PHP[rosettacode.org] and made available under GNU Free Documentation License. == Third-party icons and images Some icons used in the project comes from the following set of graphics licensed: * Dortmund is copyright by http://pc.de/icons/[PC.DE] and made available under a http://creativecommons.org/licenses/by/3.0/deed[Creative Commons Attribution 3.0 License]. * Fugue Icons is copyright by http://p.yusukekamiyamane.com/[Yusuke Kamiyamame] and made available under a http://creativecommons.org/licenses/by/3.0/deed[Creative Commons Attribution 3.0 License].
== Fusioninventory Plugin image:https://travis-ci.org/fusioninventory/fusioninventory-for-glpi.svg?branch=master["Build Status", link="https://travis-ci.org/fusioninventory/fusioninventory-for-glpi"] image:https://coveralls.io/repos/fusioninventory/fusioninventory-for-glpi/badge.svg["Coverage Status", link="https://coveralls.io/r/fusioninventory/fusioninventory-for-glpi"] This plugin makes GLPI to process various types of tasks for Fusioninventory agents: * Computer inventory * Network discovery * Network (SNMP) inventory * Software deployment * VMWare ESX host remote inventory For further information and documentation, please check http://www.fusioninventory.org . If you want to report bugs or check for development status, you can check http://forge.fusioninventory.org . == Third-party code * PluginFusioninventoryFindFiles() is copyright http://rosettacode.org/wiki/Walk_a_directory/Recursively#PHP[rosettacode.org] and made available under GNU Free Documentation License. == Third-party icons and images Some icons used in the project comes from the following set of graphics licensed: * Dortmund is copyright by http://pc.de/icons/[PC.DE] and made available under a http://creativecommons.org/licenses/by/3.0/deed[Creative Commons Attribution 3.0 License]. * Fugue Icons is copyright by http://p.yusukekamiyamane.com/[Yusuke Kamiyamame] and made available under a http://creativecommons.org/licenses/by/3.0/deed[Creative Commons Attribution 3.0 License].
Add travis badge + fix coverage badge
Add travis badge + fix coverage badge
AsciiDoc
agpl-3.0
orthagh/fusioninventory-for-glpi,orthagh/fusioninventory-for-glpi,orthagh/fusioninventory-for-glpi,orthagh/fusioninventory-for-glpi,orthagh/fusioninventory-for-glpi
d9579412a0c893779ad2ab5399bb95a667667309
README.asciidoc
README.asciidoc
== Fusioninventory Plugin https://coveralls.io/r/fusioninventory/fusioninventory-for-glpi[image:https://coveralls.io/repos/fusioninventory/fusioninventory-for-glpi/badge.svg] This plugin makes GLPI to process various types of tasks for Fusioninventory agents: * Computer inventory * Network discovery * Network (SNMP) inventory * Software deployment * VMWare ESX host remote inventory For further information and documentation, please check http://www.fusioninventory.org . If you want to report bugs or check for development status, you can check http://forge.fusioninventory.org . == Third-party code * PluginFusioninventoryFindFiles() is copyright http://rosettacode.org/wiki/Walk_a_directory/Recursively#PHP[rosettacode.org] and made available under GNU Free Documentation License. == Third-party icons and images Some icons used in the project comes from the following set of graphics licensed: * Dortmund is copyright by http://pc.de/icons/[PC.DE] and made available under a http://creativecommons.org/licenses/by/3.0/deed[Creative Commons Attribution 3.0 License]. * Fugue Icons is copyright by http://p.yusukekamiyamane.com/[Yusuke Kamiyamame] and made available under a http://creativecommons.org/licenses/by/3.0/deed[Creative Commons Attribution 3.0 License].
== Fusioninventory Plugin image:https://travis-ci.org/fusioninventory/fusioninventory-for-glpi.svg?branch=master["Build Status", link="https://travis-ci.org/fusioninventory/fusioninventory-for-glpi"] image:https://coveralls.io/repos/fusioninventory/fusioninventory-for-glpi/badge.svg["Coverage Status", link="https://coveralls.io/r/fusioninventory/fusioninventory-for-glpi"] This plugin makes GLPI to process various types of tasks for Fusioninventory agents: * Computer inventory * Network discovery * Network (SNMP) inventory * Software deployment * VMWare ESX host remote inventory For further information and documentation, please check http://www.fusioninventory.org . If you want to report bugs or check for development status, you can check http://forge.fusioninventory.org . == Third-party code * PluginFusioninventoryFindFiles() is copyright http://rosettacode.org/wiki/Walk_a_directory/Recursively#PHP[rosettacode.org] and made available under GNU Free Documentation License. == Third-party icons and images Some icons used in the project comes from the following set of graphics licensed: * Dortmund is copyright by http://pc.de/icons/[PC.DE] and made available under a http://creativecommons.org/licenses/by/3.0/deed[Creative Commons Attribution 3.0 License]. * Fugue Icons is copyright by http://p.yusukekamiyamane.com/[Yusuke Kamiyamame] and made available under a http://creativecommons.org/licenses/by/3.0/deed[Creative Commons Attribution 3.0 License].
Add travis badge + fix coverage badge
Add travis badge + fix coverage badge
AsciiDoc
agpl-3.0
TECLIB/fusioninventory-for-glpi,mohierf/fusioninventory-for-glpi,TECLIB/fusioninventory-for-glpi,mohierf/fusioninventory-for-glpi,itinside/fusioninventory-for-glpi,fusioninventory/fusioninventory-for-glpi,fusioninventory/fusioninventory-for-glpi,fusioninventory/fusioninventory-for-glpi,orthagh/fusioninventory-for-glpi,TECLIB/fusioninventory-for-glpi,itinside/fusioninventory-for-glpi,orthagh/fusioninventory-for-glpi,mohierf/fusioninventory-for-glpi,orthagh/fusioninventory-for-glpi,itinside/fusioninventory-for-glpi,TECLIB/fusioninventory-for-glpi,mohierf/fusioninventory-for-glpi,itinside/fusioninventory-for-glpi,fusioninventory/fusioninventory-for-glpi,orthagh/fusioninventory-for-glpi,fusioninventory/fusioninventory-for-glpi,orthagh/fusioninventory-for-glpi,itinside/fusioninventory-for-glpi,mohierf/fusioninventory-for-glpi,fusioninventory/fusioninventory-for-glpi,TECLIB/fusioninventory-for-glpi,TECLIB/fusioninventory-for-glpi,mohierf/fusioninventory-for-glpi
89e7da34b9c4b072e3caab12dbe765b0a33be272
README.adoc
README.adoc
= Hawkular Android Client This repository contains the source code for the Hawkular Android application. == License * http://www.apache.org/licenses/LICENSE-2.0.html[Apache Version 2.0] == Building ifdef::env-github[] [link=https://travis-ci.org/hawkular/hawkular-android-client] image:https://travis-ci.org/hawkular/hawkular-android-client.svg["Build Status", link="https://travis-ci.org/hawkular/hawkular-android-client"] endif::[] You will need JDK 1.7+ installed. Gradle, Android SDK and all dependencies will be downloaded automatically. ---- $ ./gradlew clean assembleDebug ----
= Hawkular Android Client This repository contains the source code for the Hawkular Android application. == License * http://www.apache.org/licenses/LICENSE-2.0.html[Apache Version 2.0] == Building ifdef::env-github[] [link=https://travis-ci.org/hawkular/hawkular-android-client] image:https://travis-ci.org/hawkular/hawkular-android-client.svg["Build Status", link="https://travis-ci.org/hawkular/hawkular-android-client"] endif::[] You will need JDK 1.7+ installed. Gradle, Android SDK and all dependencies will be downloaded automatically. ---- $ ./gradlew clean assembleDebug ---- == Reading There are some documents on the link:../../wiki[Wiki], including API overview, UI mockups and instructions on running necessary servers for using the client in common and push notifications specifically.
Update readme with Wiki reference.
Update readme with Wiki reference.
AsciiDoc
apache-2.0
sauravvishal8797/hawkular-android-client,shubhamvashisht/hawkular-android-client,anuj1708/hawkular-android-client,danielpassos/hawkular-android-client,anuj1708/hawkular-android-buff,pg301/hawkular-android-client,hawkular/hawkular-android-client
81ac2b91b6a6b7cf2422f6aabdd4164434b295dd
README.adoc
README.adoc
= Spring Boot and Two DataSources This project demonstrates how to use two `DataSource` s with Spring Boot 2.0. It utilizes: * Spring Data https://github.com/spring-projects/spring-data-jpa[JPA] / https://github.com/spring-projects/spring-data-rest[REST] * https://github.com/flyway/flyway[Flyway] migrations for the two `DataSource` s * Separate Hibernate properties for each `DataSource` defined in the application.yml * https://github.com/thymeleaf/thymeleaf[Thymeleaf] 3 * https://github.com/DataTables/DataTablesSrc[DataTables] * Unit tests for components Note: It may take a few seconds for the app to start if no one has not accessed it recently
= Spring Boot and Two DataSources This project demonstrates how to use two `DataSource` s with Spring Boot 2.1. It utilizes: * Spring Data https://github.com/spring-projects/spring-data-jpa[JPA] / https://github.com/spring-projects/spring-data-rest[REST] * https://github.com/flyway/flyway[Flyway] migrations for the two `DataSource` s * Separate Hibernate properties for each `DataSource` defined in the application.yml * https://github.com/thymeleaf/thymeleaf[Thymeleaf] 3 * https://github.com/DataTables/DataTablesSrc[DataTables] * Unit tests for components Note: It may take a few seconds for the app to start if no one has not accessed it recently
Adjust readme to Spring Boot 2.1
Adjust readme to Spring Boot 2.1
AsciiDoc
unlicense
drumonii/SpringBootTwoDataSources
ed1b52faf367fc3867a877f0fb63b3cf7034dcab
README.adoc
README.adoc
= Snoop - A Discovery Service for Java EE Snoop is an experimental registration and discovery service for Java EE based microservices. == Getting Started . Start the link:snoop-service.adoc[Snoop Service] . link:service-registration.adoc[Service Registration] . link:service-discovery.adoc[Service Discovery] == Maven . Released artifacts are available in link:http://search.maven.org/#search%7Cga%7C1%7Csnoop[Maven Central] . Snapshots configuration: <repositories> <repository> <id>agilejava-snapshots</id> <url>http://nexus.agilejava.eu/content/groups/public</url> <snapshots> <enabled>true</enabled> </snapshots> </repository> </repositories> == Examples - link:https://github.com/ivargrimstad/snoop-samples[snoop-samples@GitHub] - link:https://github.com/arun-gupta/microservices[https://github.com/arun-gupta/microservices] == FAQ - link:FAQ.adoc[Frequently Asked Questions]
= Snoop - A Discovery Service for Java EE Snoop is an experimental registration and discovery service for Java EE based microservices. == Getting Started . Start the link:snoop-service.adoc[Snoop Service] . link:service-registration.adoc[Service Registration] . link:service-discovery.adoc[Service Discovery] == Maven . Released artifacts are available in link:http://search.maven.org/#search%7Cga%7C1%7Csnoop[Maven Central] . Snapshots configuration: <repositories> <repository> <id>agilejava-snapshots</id> <url>http://nexus.agilejava.eu/content/groups/public</url> <snapshots> <enabled>true</enabled> </snapshots> </repository> </repositories> == Classloader Issue - link:classloader-issue.adoc[Description and Workaround] == Examples - link:https://github.com/ivargrimstad/snoop-samples[snoop-samples@GitHub] - link:https://github.com/arun-gupta/microservices[https://github.com/arun-gupta/microservices] == FAQ - link:FAQ.adoc[Frequently Asked Questions]
Add reference to classloader issue and workaround
Add reference to classloader issue and workaround
AsciiDoc
mit
ivargrimstad/snoopee,ivargrimstad/snoopee,ivargrimstad/snoopee,ivargrimstad/snoop,ivargrimstad/snoop,ivargrimstad/snoop
475366b5f1daeb419f2805ee1e9eefad4df70803
dsl/camel-cli-connector/src/main/docs/cli-connector.adoc
dsl/camel-cli-connector/src/main/docs/cli-connector.adoc
= CLI Connector Component :doctitle: CLI Connector :shortname: cli-connector :artifactid: camel-cli-connector :description: Runtime adapter connecting with Camel CLI :since: 3.19 :supportlevel: Preview //Manually maintained attributes :camel-spring-boot-name: cli-connector *Since Camel {since}* The camel-cli-connector allows the Camel CLI to be able to manage running Camel integrations. Currently, only a local connector is provided, which means that the Camel CLI can only be managing local running Camel integrations. These integrations can be using different runtimes such as Camel Main, Camel Spring Boot or Camel Quarkus etc. == Auto-detection from classpath To use this implementation all you need to do is to add the `camel-cli-connector` dependency to the classpath, and Camel should auto-detect this on startup and log as follows: [source,text] ---- Local CLI Connector started ---- include::spring-boot:partial$starter.adoc[]
= CLI Connector Component :doctitle: CLI Connector :shortname: cli-connector :artifactid: camel-cli-connector :description: Runtime adapter connecting with Camel CLI :since: 3.19 :supportlevel: Preview *Since Camel {since}* The camel-cli-connector allows the Camel CLI to be able to manage running Camel integrations. Currently, only a local connector is provided, which means that the Camel CLI can only be managing local running Camel integrations. These integrations can be using different runtimes such as Camel Main, Camel Spring Boot or Camel Quarkus etc. == Auto-detection from classpath To use this implementation all you need to do is to add the `camel-cli-connector` dependency to the classpath, and Camel should auto-detect this on startup and log as follows: [source,text] ---- Local CLI Connector started ----
Revert "Add spring-boot link in doc"
Revert "Add spring-boot link in doc" This reverts commit fbd8d2ed399230a8e3e1d469e9825eb0602fb981.
AsciiDoc
apache-2.0
cunningt/camel,christophd/camel,tadayosi/camel,cunningt/camel,tadayosi/camel,christophd/camel,apache/camel,apache/camel,tadayosi/camel,christophd/camel,cunningt/camel,tadayosi/camel,christophd/camel,cunningt/camel,cunningt/camel,christophd/camel,cunningt/camel,apache/camel,tadayosi/camel,apache/camel,apache/camel,christophd/camel,apache/camel,tadayosi/camel
c55eb2c72d299f4d15f5f711ab62d13a59b39f28
LICENSE.adoc
LICENSE.adoc
Copyright 2016 higherfrequencytrading.com Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License.
== Copyright 2016 higherfrequencytrading.com Licensed under the *Apache License, Version 2.0* (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License.
Migrate to Apache v2.0 license
Migrate to Apache v2.0 license
AsciiDoc
apache-2.0
OpenHFT/Chronicle-Bytes
009eb4b932af11292fcd86c9427a867dfc4486dc
operators/olm-webhooks.adoc
operators/olm-webhooks.adoc
[id="olm-webhooks"] = Managing admission webhooks in Operator Lifecycle Manager include::modules/common-attributes.adoc[] :context: olm-webhooks toc::[] Validating and mutating admission webhooks allow Operator authors to intercept, modify, and accept or reject resources before they are handled by the Operator controller. Operator Lifecycle Manager (OLM) can manage the lifecycle of these webhooks when they are shipped alongside your Operator. include::modules/olm-defining-csv-webhooks.adoc[leveloffset=+1] include::modules/olm-webhook-considerations.adoc[leveloffset=+1] [id="olm-webhooks-additional-resources"] == Additional resources * xref:../architecture/admission-plug-ins.adoc#admission-webhook-types_admission-plug-ins[Types of webhook admission plug-ins]
[id="olm-webhooks"] = Managing admission webhooks in Operator Lifecycle Manager include::modules/common-attributes.adoc[] :context: olm-webhooks toc::[] Validating and mutating admission webhooks allow Operator authors to intercept, modify, and accept or reject resources before they are saved to the object store and handled by the Operator controller. Operator Lifecycle Manager (OLM) can manage the lifecycle of these webhooks when they are shipped alongside your Operator. include::modules/olm-defining-csv-webhooks.adoc[leveloffset=+1] include::modules/olm-webhook-considerations.adoc[leveloffset=+1] [id="olm-webhooks-additional-resources"] == Additional resources * xref:../architecture/admission-plug-ins.adoc#admission-webhook-types_admission-plug-ins[Types of webhook admission plug-ins]
Edit to OLM webhook workflow
Edit to OLM webhook workflow
AsciiDoc
apache-2.0
vikram-redhat/openshift-docs,vikram-redhat/openshift-docs,vikram-redhat/openshift-docs,vikram-redhat/openshift-docs,vikram-redhat/openshift-docs
e84b61358f35d622c1da53330f5140b011053dbb
docs/reference/migration/index.asciidoc
docs/reference/migration/index.asciidoc
[[breaking-changes]] = Breaking changes [partintro] -- This section discusses the changes that you need to be aware of when migrating your application from one version of Elasticsearch to another. As a general rule: * Migration between major versions -- e.g. `1.x` to `2.x` -- requires a <<restart-upgrade,full cluster restart>>. * Migration between minor versions -- e.g. `1.x` to `1.y` -- can be performed by <<rolling-upgrades,upgrading one node at a time>>. See <<setup-upgrade>> for more info. -- include::migrate_3_0.asciidoc[] include::migrate_2_1.asciidoc[] include::migrate_2_2.asciidoc[] include::migrate_2_0.asciidoc[] include::migrate_1_6.asciidoc[] include::migrate_1_4.asciidoc[] include::migrate_1_0.asciidoc[]
[[breaking-changes]] = Breaking changes [partintro] -- This section discusses the changes that you need to be aware of when migrating your application from one version of Elasticsearch to another. As a general rule: * Migration between major versions -- e.g. `1.x` to `2.x` -- requires a <<restart-upgrade,full cluster restart>>. * Migration between minor versions -- e.g. `1.x` to `1.y` -- can be performed by <<rolling-upgrades,upgrading one node at a time>>. See <<setup-upgrade>> for more info. -- include::migrate_3_0.asciidoc[] include::migrate_2_2.asciidoc[] include::migrate_2_1.asciidoc[] include::migrate_2_0.asciidoc[] include::migrate_1_6.asciidoc[] include::migrate_1_4.asciidoc[] include::migrate_1_0.asciidoc[]
Fix version order for breaking changes docs
Fix version order for breaking changes docs
AsciiDoc
apache-2.0
ESamir/elasticsearch,uschindler/elasticsearch,s1monw/elasticsearch,henakamaMSFT/elasticsearch,StefanGor/elasticsearch,masaruh/elasticsearch,awislowski/elasticsearch,strapdata/elassandra,camilojd/elasticsearch,kalimatas/elasticsearch,brandonkearby/elasticsearch,IanvsPoplicola/elasticsearch,diendt/elasticsearch,gmarz/elasticsearch,JervyShi/elasticsearch,snikch/elasticsearch,jchampion/elasticsearch,bawse/elasticsearch,mapr/elasticsearch,gingerwizard/elasticsearch,jimczi/elasticsearch,fred84/elasticsearch,MisterAndersen/elasticsearch,rajanm/elasticsearch,masaruh/elasticsearch,sreeramjayan/elasticsearch,JSCooke/elasticsearch,jbertouch/elasticsearch,gmarz/elasticsearch,pozhidaevak/elasticsearch,sneivandt/elasticsearch,fforbeck/elasticsearch,avikurapati/elasticsearch,artnowo/elasticsearch,scottsom/elasticsearch,wuranbo/elasticsearch,gingerwizard/elasticsearch,Shepard1212/elasticsearch,glefloch/elasticsearch,shreejay/elasticsearch,vroyer/elasticassandra,jchampion/elasticsearch,pozhidaevak/elasticsearch,davidvgalbraith/elasticsearch,liweinan0423/elasticsearch,ricardocerq/elasticsearch,ivansun1010/elasticsearch,jchampion/elasticsearch,davidvgalbraith/elasticsearch,umeshdangat/elasticsearch,obourgain/elasticsearch,Helen-Zhao/elasticsearch,liweinan0423/elasticsearch,mohit/elasticsearch,episerver/elasticsearch,tebriel/elasticsearch,uschindler/elasticsearch,spiegela/elasticsearch,trangvh/elasticsearch,ivansun1010/elasticsearch,ivansun1010/elasticsearch,strapdata/elassandra,avikurapati/elasticsearch,dpursehouse/elasticsearch,awislowski/elasticsearch,markharwood/elasticsearch,GlenRSmith/elasticsearch,HonzaKral/elasticsearch,coding0011/elasticsearch,nknize/elasticsearch,mortonsykes/elasticsearch,dpursehouse/elasticsearch,HonzaKral/elasticsearch,nezirus/elasticsearch,scottsom/elasticsearch,i-am-Nathan/elasticsearch,mikemccand/elasticsearch,trangvh/elasticsearch,geidies/elasticsearch,geidies/elasticsearch,yanjunh/elasticsearch,yynil/elasticsearch,fforbeck/elasticsearch,yynil/elasticsearch,scorpionvicky/elasticsearch,wenpos/elasticsearch,Stacey-Gammon/elasticsearch,markharwood/elasticsearch,qwerty4030/elasticsearch,coding0011/elasticsearch,avikurapati/elasticsearch,myelin/elasticsearch,nilabhsagar/elasticsearch,maddin2016/elasticsearch,naveenhooda2000/elasticsearch,zkidkid/elasticsearch,s1monw/elasticsearch,JSCooke/elasticsearch,LewayneNaidoo/elasticsearch,snikch/elasticsearch,episerver/elasticsearch,mmaracic/elasticsearch,a2lin/elasticsearch,njlawton/elasticsearch,alexshadow007/elasticsearch,spiegela/elasticsearch,gingerwizard/elasticsearch,jimczi/elasticsearch,s1monw/elasticsearch,jchampion/elasticsearch,elasticdog/elasticsearch,spiegela/elasticsearch,henakamaMSFT/elasticsearch,StefanGor/elasticsearch,fforbeck/elasticsearch,pozhidaevak/elasticsearch,a2lin/elasticsearch,girirajsharma/elasticsearch,wenpos/elasticsearch,sreeramjayan/elasticsearch,glefloch/elasticsearch,vroyer/elasticassandra,mohit/elasticsearch,jpountz/elasticsearch,geidies/elasticsearch,strapdata/elassandra,alexshadow007/elasticsearch,markharwood/elasticsearch,mapr/elasticsearch,qwerty4030/elasticsearch,nilabhsagar/elasticsearch,ThiagoGarciaAlves/elasticsearch,jbertouch/elasticsearch,diendt/elasticsearch,LewayneNaidoo/elasticsearch,xuzha/elasticsearch,sneivandt/elasticsearch,sreeramjayan/elasticsearch,rlugojr/elasticsearch,JervyShi/elasticsearch,qwerty4030/elasticsearch,trangvh/elasticsearch,henakamaMSFT/elasticsearch,tebriel/elasticsearch,jprante/elasticsearch,elasticdog/elasticsearch,cwurm/elasticsearch,markwalkom/elasticsearch,nomoa/elasticsearch,trangvh/elasticsearch,martinstuga/elasticsearch,strapdata/elassandra5-rc,sneivandt/elasticsearch,kalimatas/elasticsearch,episerver/elasticsearch,wuranbo/elasticsearch,tebriel/elasticsearch,shreejay/elasticsearch,fernandozhu/elasticsearch,wuranbo/elasticsearch,rlugojr/elasticsearch,artnowo/elasticsearch,scottsom/elasticsearch,awislowski/elasticsearch,awislowski/elasticsearch,uschindler/elasticsearch,C-Bish/elasticsearch,clintongormley/elasticsearch,mmaracic/elasticsearch,nilabhsagar/elasticsearch,njlawton/elasticsearch,a2lin/elasticsearch,nknize/elasticsearch,dongjoon-hyun/elasticsearch,ThiagoGarciaAlves/elasticsearch,gfyoung/elasticsearch,fforbeck/elasticsearch,JervyShi/elasticsearch,gingerwizard/elasticsearch,JervyShi/elasticsearch,brandonkearby/elasticsearch,xuzha/elasticsearch,a2lin/elasticsearch,Stacey-Gammon/elasticsearch,tebriel/elasticsearch,lks21c/elasticsearch,kalimatas/elasticsearch,jprante/elasticsearch,davidvgalbraith/elasticsearch,sneivandt/elasticsearch,nknize/elasticsearch,HonzaKral/elasticsearch,naveenhooda2000/elasticsearch,JackyMai/elasticsearch,gingerwizard/elasticsearch,mohit/elasticsearch,martinstuga/elasticsearch,umeshdangat/elasticsearch,i-am-Nathan/elasticsearch,nezirus/elasticsearch,scorpionvicky/elasticsearch,markwalkom/elasticsearch,C-Bish/elasticsearch,mmaracic/elasticsearch,JackyMai/elasticsearch,rajanm/elasticsearch,LeoYao/elasticsearch,jimczi/elasticsearch,myelin/elasticsearch,markwalkom/elasticsearch,GlenRSmith/elasticsearch,ESamir/elasticsearch,spiegela/elasticsearch,snikch/elasticsearch,MaineC/elasticsearch,elasticdog/elasticsearch,JSCooke/elasticsearch,nknize/elasticsearch,s1monw/elasticsearch,mmaracic/elasticsearch,snikch/elasticsearch,ivansun1010/elasticsearch,mikemccand/elasticsearch,ESamir/elasticsearch,martinstuga/elasticsearch,davidvgalbraith/elasticsearch,nezirus/elasticsearch,MaineC/elasticsearch,C-Bish/elasticsearch,masaruh/elasticsearch,tebriel/elasticsearch,nknize/elasticsearch,nazarewk/elasticsearch,jpountz/elasticsearch,nazarewk/elasticsearch,cwurm/elasticsearch,scorpionvicky/elasticsearch,IanvsPoplicola/elasticsearch,camilojd/elasticsearch,Shepard1212/elasticsearch,MisterAndersen/elasticsearch,girirajsharma/elasticsearch,trangvh/elasticsearch,avikurapati/elasticsearch,winstonewert/elasticsearch,lks21c/elasticsearch,Helen-Zhao/elasticsearch,C-Bish/elasticsearch,yynil/elasticsearch,palecur/elasticsearch,gfyoung/elasticsearch,gingerwizard/elasticsearch,martinstuga/elasticsearch,xuzha/elasticsearch,tebriel/elasticsearch,nomoa/elasticsearch,henakamaMSFT/elasticsearch,sreeramjayan/elasticsearch,fred84/elasticsearch,i-am-Nathan/elasticsearch,clintongormley/elasticsearch,sreeramjayan/elasticsearch,camilojd/elasticsearch,ThiagoGarciaAlves/elasticsearch,nezirus/elasticsearch,obourgain/elasticsearch,IanvsPoplicola/elasticsearch,StefanGor/elasticsearch,mjason3/elasticsearch,a2lin/elasticsearch,fforbeck/elasticsearch,dongjoon-hyun/elasticsearch,scorpionvicky/elasticsearch,qwerty4030/elasticsearch,umeshdangat/elasticsearch,nilabhsagar/elasticsearch,lks21c/elasticsearch,mjason3/elasticsearch,brandonkearby/elasticsearch,njlawton/elasticsearch,GlenRSmith/elasticsearch,mapr/elasticsearch,markharwood/elasticsearch,wangtuo/elasticsearch,LewayneNaidoo/elasticsearch,njlawton/elasticsearch,winstonewert/elasticsearch,C-Bish/elasticsearch,mikemccand/elasticsearch,palecur/elasticsearch,scottsom/elasticsearch,s1monw/elasticsearch,winstonewert/elasticsearch,ivansun1010/elasticsearch,yynil/elasticsearch,rlugojr/elasticsearch,Helen-Zhao/elasticsearch,xuzha/elasticsearch,jpountz/elasticsearch,snikch/elasticsearch,mortonsykes/elasticsearch,davidvgalbraith/elasticsearch,ZTE-PaaS/elasticsearch,mortonsykes/elasticsearch,winstonewert/elasticsearch,maddin2016/elasticsearch,girirajsharma/elasticsearch,ivansun1010/elasticsearch,yynil/elasticsearch,elasticdog/elasticsearch,GlenRSmith/elasticsearch,vroyer/elassandra,polyfractal/elasticsearch,girirajsharma/elasticsearch,umeshdangat/elasticsearch,nazarewk/elasticsearch,Helen-Zhao/elasticsearch,obourgain/elasticsearch,dongjoon-hyun/elasticsearch,jbertouch/elasticsearch,Shepard1212/elasticsearch,MaineC/elasticsearch,robin13/elasticsearch,wuranbo/elasticsearch,sneivandt/elasticsearch,xuzha/elasticsearch,Shepard1212/elasticsearch,fernandozhu/elasticsearch,ESamir/elasticsearch,LeoYao/elasticsearch,ThiagoGarciaAlves/elasticsearch,markwalkom/elasticsearch,yynil/elasticsearch,shreejay/elasticsearch,ESamir/elasticsearch,vroyer/elassandra,palecur/elasticsearch,umeshdangat/elasticsearch,mmaracic/elasticsearch,artnowo/elasticsearch,jprante/elasticsearch,Shepard1212/elasticsearch,myelin/elasticsearch,kalimatas/elasticsearch,strapdata/elassandra5-rc,markwalkom/elasticsearch,ESamir/elasticsearch,gfyoung/elasticsearch,martinstuga/elasticsearch,scorpionvicky/elasticsearch,awislowski/elasticsearch,maddin2016/elasticsearch,strapdata/elassandra5-rc,rajanm/elasticsearch,clintongormley/elasticsearch,rajanm/elasticsearch,lks21c/elasticsearch,uschindler/elasticsearch,zkidkid/elasticsearch,ThiagoGarciaAlves/elasticsearch,polyfractal/elasticsearch,artnowo/elasticsearch,vroyer/elasticassandra,ricardocerq/elasticsearch,ricardocerq/elasticsearch,JackyMai/elasticsearch,LeoYao/elasticsearch,JackyMai/elasticsearch,LewayneNaidoo/elasticsearch,maddin2016/elasticsearch,shreejay/elasticsearch,robin13/elasticsearch,ZTE-PaaS/elasticsearch,JSCooke/elasticsearch,coding0011/elasticsearch,avikurapati/elasticsearch,LeoYao/elasticsearch,diendt/elasticsearch,JervyShi/elasticsearch,yanjunh/elasticsearch,geidies/elasticsearch,wenpos/elasticsearch,LeoYao/elasticsearch,ZTE-PaaS/elasticsearch,coding0011/elasticsearch,mohit/elasticsearch,jprante/elasticsearch,artnowo/elasticsearch,nazarewk/elasticsearch,mjason3/elasticsearch,jchampion/elasticsearch,kalimatas/elasticsearch,shreejay/elasticsearch,njlawton/elasticsearch,rajanm/elasticsearch,geidies/elasticsearch,cwurm/elasticsearch,clintongormley/elasticsearch,rlugojr/elasticsearch,Helen-Zhao/elasticsearch,elasticdog/elasticsearch,martinstuga/elasticsearch,mjason3/elasticsearch,robin13/elasticsearch,qwerty4030/elasticsearch,myelin/elasticsearch,Stacey-Gammon/elasticsearch,yanjunh/elasticsearch,strapdata/elassandra,glefloch/elasticsearch,naveenhooda2000/elasticsearch,gingerwizard/elasticsearch,obourgain/elasticsearch,nazarewk/elasticsearch,MisterAndersen/elasticsearch,strapdata/elassandra5-rc,JackyMai/elasticsearch,liweinan0423/elasticsearch,nomoa/elasticsearch,MisterAndersen/elasticsearch,jpountz/elasticsearch,bawse/elasticsearch,MaineC/elasticsearch,wangtuo/elasticsearch,mapr/elasticsearch,pozhidaevak/elasticsearch,zkidkid/elasticsearch,brandonkearby/elasticsearch,Stacey-Gammon/elasticsearch,jbertouch/elasticsearch,nezirus/elasticsearch,jimczi/elasticsearch,IanvsPoplicola/elasticsearch,MisterAndersen/elasticsearch,fernandozhu/elasticsearch,pozhidaevak/elasticsearch,LewayneNaidoo/elasticsearch,masaruh/elasticsearch,gfyoung/elasticsearch,fernandozhu/elasticsearch,gmarz/elasticsearch,mikemccand/elasticsearch,GlenRSmith/elasticsearch,cwurm/elasticsearch,mmaracic/elasticsearch,jpountz/elasticsearch,yanjunh/elasticsearch,maddin2016/elasticsearch,strapdata/elassandra,glefloch/elasticsearch,IanvsPoplicola/elasticsearch,fred84/elasticsearch,alexshadow007/elasticsearch,LeoYao/elasticsearch,mortonsykes/elasticsearch,StefanGor/elasticsearch,wenpos/elasticsearch,rajanm/elasticsearch,xuzha/elasticsearch,clintongormley/elasticsearch,ThiagoGarciaAlves/elasticsearch,strapdata/elassandra5-rc,yanjunh/elasticsearch,episerver/elasticsearch,MaineC/elasticsearch,gfyoung/elasticsearch,ZTE-PaaS/elasticsearch,StefanGor/elasticsearch,diendt/elasticsearch,nomoa/elasticsearch,markwalkom/elasticsearch,winstonewert/elasticsearch,polyfractal/elasticsearch,dpursehouse/elasticsearch,dpursehouse/elasticsearch,JervyShi/elasticsearch,fred84/elasticsearch,Stacey-Gammon/elasticsearch,bawse/elasticsearch,geidies/elasticsearch,lks21c/elasticsearch,palecur/elasticsearch,wangtuo/elasticsearch,spiegela/elasticsearch,mapr/elasticsearch,polyfractal/elasticsearch,henakamaMSFT/elasticsearch,jpountz/elasticsearch,palecur/elasticsearch,diendt/elasticsearch,jimczi/elasticsearch,JSCooke/elasticsearch,ricardocerq/elasticsearch,jbertouch/elasticsearch,davidvgalbraith/elasticsearch,uschindler/elasticsearch,robin13/elasticsearch,alexshadow007/elasticsearch,cwurm/elasticsearch,markharwood/elasticsearch,bawse/elasticsearch,LeoYao/elasticsearch,polyfractal/elasticsearch,mohit/elasticsearch,camilojd/elasticsearch,mortonsykes/elasticsearch,jprante/elasticsearch,wangtuo/elasticsearch,scottsom/elasticsearch,wenpos/elasticsearch,nomoa/elasticsearch,coding0011/elasticsearch,jchampion/elasticsearch,myelin/elasticsearch,masaruh/elasticsearch,fernandozhu/elasticsearch,camilojd/elasticsearch,alexshadow007/elasticsearch,girirajsharma/elasticsearch,wuranbo/elasticsearch,sreeramjayan/elasticsearch,gmarz/elasticsearch,jbertouch/elasticsearch,gmarz/elasticsearch,dpursehouse/elasticsearch,i-am-Nathan/elasticsearch,bawse/elasticsearch,mapr/elasticsearch,ZTE-PaaS/elasticsearch,mjason3/elasticsearch,zkidkid/elasticsearch,liweinan0423/elasticsearch,ricardocerq/elasticsearch,clintongormley/elasticsearch,markharwood/elasticsearch,naveenhooda2000/elasticsearch,polyfractal/elasticsearch,mikemccand/elasticsearch,fred84/elasticsearch,obourgain/elasticsearch,dongjoon-hyun/elasticsearch,snikch/elasticsearch,i-am-Nathan/elasticsearch,naveenhooda2000/elasticsearch,girirajsharma/elasticsearch,nilabhsagar/elasticsearch,diendt/elasticsearch,liweinan0423/elasticsearch,HonzaKral/elasticsearch,robin13/elasticsearch,rlugojr/elasticsearch,brandonkearby/elasticsearch,zkidkid/elasticsearch,camilojd/elasticsearch,wangtuo/elasticsearch,episerver/elasticsearch,dongjoon-hyun/elasticsearch,vroyer/elassandra,glefloch/elasticsearch
bb696411acc9f2bed981b1f3c9aa5a13cdfd91c8
doc/resources/doc/sources/index.adoc
doc/resources/doc/sources/index.adoc
= Edge Documentation Edge is a starting point for creating Clojure projects. Not sure if Edge is for you? See <<why-edge.adoc#,Why Edge?>>. == Get Started Are you new to Edge? This is the place to start! . link:https://clojure.org/guides/getting_started[Install clj] (<<windows.adoc#,Additional notes for installing on Windows>>) . <<editor.adoc#,Set up your editor for Clojure>> . <<setup.adoc#,Set up Edge for your project>> . <<dev-guide.adoc#,Developing on Edge>> == Using Edge //. Configuration //. Components * <<dev-guide.adoc#,Developing on Edge>> * <<uberjar.adoc#,Producing an Uberjar>> * <<elastic-beanstalk.adoc#,Using the Elastic Beanstalk Quickstart>> * <<socket-repl.adoc#,Setting up a socket REPL>> == The Edge Project * <<why-edge.adoc#,Why Edge?>> * <<guidelines.adoc#,Contributing Guidelines>> //* Getting help //* How to get involved //* License
= Edge Documentation Edge is a starting point for creating Clojure projects of all sizes. == Get Started Are you new to Edge? This is the place to start! . link:https://clojure.org/guides/getting_started[Install clj] (<<windows.adoc#,Additional notes for installing on Windows>>) . <<editor.adoc#,Set up your editor for Clojure>> . <<setup.adoc#,Set up Edge for your project>> . <<dev-guide.adoc#,Developing on Edge>> == Using Edge //. Configuration //. Components * <<dev-guide.adoc#,Developing on Edge>> * <<uberjar.adoc#,Producing an Uberjar>> * <<elastic-beanstalk.adoc#,Using the Elastic Beanstalk Quickstart>> * <<socket-repl.adoc#,Setting up a socket REPL>> == The Edge Project * <<why-edge.adoc#,Why Edge?>> * <<guidelines.adoc#,Contributing Guidelines>> //* Getting help //* How to get involved //* License
Remove Why Edge? link from preamble
Remove Why Edge? link from preamble It was not useful.
AsciiDoc
mit
juxt/edge,juxt/edge
920d5fe7ab3e56c7a2bc8ebaa186698948ab9b23
documentation/src/docs/asciidoc/overview.adoc
documentation/src/docs/asciidoc/overview.adoc
[[overview]] == Overview The goal of this document is to provide comprehensive reference documentation for both programmers writing tests and extension authors. WARNING: Work in progress! === Supported Java Versions JUnit 5 only supports Java 8 and above. However, you can still test classes compiled with lower versions. == Installation Snapshot artifacts are deployed to Sonatype's {snapshot-repo}[snapshots repository]. [[dependency-metadata]] === Dependency Metadata * *Group ID*: `org.junit` * *Version*: `{junit-version}` * *Artifact IDs*: ** `junit-commons` ** `junit-console` ** `junit-engine-api` ** `junit-gradle` ** `junit-launcher` ** `junit4-engine` ** `junit4-runner` ** `junit5-api` ** `junit5-engine` ** `surefire-junit5` See also: {snapshot-repo}/org/junit/ === JUnit 5 Sample Projects The {junit5-samples-repo}[`junit5-samples`] repository hosts a collection of sample projects based on JUnit 5. You'll find the respective `build.gradle` and `pom.xml` in the projects below. * For Gradle, check out the `{junit5-gradle-consumer}` project. * For Maven, check out the `{junit5-maven-consumer}` project.
[[overview]] == Overview The goal of this document is to provide comprehensive reference documentation for both programmers writing tests and extension authors. WARNING: Work in progress! === Supported Java Versions JUnit 5 only supports Java 8 and above. However, you can still test classes compiled with lower versions. == Installation Snapshot artifacts are deployed to Sonatype's {snapshot-repo}[snapshots repository]. [[dependency-metadata]] === Dependency Metadata * *Group ID*: `org.junit` * *Version*: `{junit-version}` * *Artifact IDs*: ** `junit-commons` ** `junit-console` ** `junit-engine-api` ** `junit-gradle` ** `junit-launcher` ** `junit4-engine` ** `junit4-runner` ** `junit5-api` ** `junit5-engine` ** `surefire-junit5` See also: {snapshot-repo}/org/junit/ === JUnit 5 Sample Projects The {junit5-samples-repo}[`junit5-samples`] repository hosts a collection of sample projects based on JUnit 5. You'll find the respective `build.gradle` and `pom.xml` in the projects below. * For Gradle, check out the `{junit5-gradle-consumer}` project. * For Maven, check out the `{junit5-maven-consumer}` project.
Apply Spotless to User Guide
Apply Spotless to User Guide ------------------------------------------------------------------------ On behalf of the community, the JUnit Lambda Team thanks Klarna AB (http://www.klarna.com) for supporting the JUnit crowdfunding campaign! ------------------------------------------------------------------------
AsciiDoc
epl-1.0
marcphilipp/junit5,marcphilipp/junit-lambda,sbrannen/junit-lambda,junit-team/junit-lambda
babf52d8f8101d018e357bdd47e5ccb9d8af14d5
src/jqassistant/structure.adoc
src/jqassistant/structure.adoc
[[structure:Default]] [role=group,includesConstraints="structure:packagesShouldConformToTheMainBuildingBlocks"] All the blackboxes above should correspond to Java packages. Those packages should have no dependencies to other packages outside themselves but for the support or shared package: [[structure:packagesShouldConformToTheMainBuildingBlocks]] [source,cypher,role=constraint,requiresConcepts="structure:configPackages,structure:supportingPackages"] .Top level packages should conform to the main building blocks. ---- MATCH (a:Artifact {type: 'jar'}) MATCH (a) -[:CONTAINS]-> (p1:Package) -[:DEPENDS_ON]-> (p2:Package) <-[:CONTAINS]- (a) WHERE not p1:Config and not (p1) -[:CONTAINS]-> (p2) and not p2:Support and not p1.fqn = 'ac.simons.biking2.summary' RETURN p1, p2 ----
[[structure:Default]] [role=group,includesConstraints="structure:packagesShouldConformToTheMainBuildingBlocks"] All the blackboxes above should correspond to Java packages. Those packages should have no dependencies to other packages outside themselves but for the support or shared package: [[structure:packagesShouldConformToTheMainBuildingBlocks]] [source,cypher,role=constraint,requiresConcepts="structure:configPackages,structure:supportingPackages"] .Top level packages should conform to the main building blocks. ---- MATCH (a:Main:Artifact) MATCH (a) -[:CONTAINS]-> (p1:Package) -[:DEPENDS_ON]-> (p2:Package) <-[:CONTAINS]- (a) WHERE not p1:Config and not (p1) -[:CONTAINS]-> (p2) and not p2:Support and not p1.fqn = 'ac.simons.biking2.summary' RETURN p1, p2 ----
Use 'Main'-Label instead of type-attribute.
Use 'Main'-Label instead of type-attribute.
AsciiDoc
apache-2.0
tullo/biking2,michael-simons/biking2,tullo/biking2,tullo/biking2,tullo/biking2,michael-simons/biking2,michael-simons/biking2,michael-simons/biking2
8299e439b97f466d83a34db371b265dd265a1267
documentation/src/docs/asciidoc/release-notes/release-notes-5.5.0-M2.adoc
documentation/src/docs/asciidoc/release-notes/release-notes-5.5.0-M2.adoc
[[release-notes-5.5.0-M2]] == 5.5.0-M2️ *Date of Release:* ❓ *Scope:* ❓ For a complete list of all _closed_ issues and pull requests for this release, consult the link:{junit5-repo}+/milestone/37?closed=1+[5.5 M2] milestone page in the JUnit repository on GitHub. [[release-notes-5.5.0-M2-junit-platform]] === JUnit Platform ==== Bug Fixes * ❓ ==== Deprecations and Breaking Changes * ❓ ==== New Features and Improvements * ❓ [[release-notes-5.5.0-M2-junit-jupiter]] === JUnit Jupiter ==== Bug Fixes * Parameterized tests no longer throw an `ArrayStoreException` when creating human-readable test names. ==== Deprecations and Breaking Changes * ❓ ==== New Features and Improvements * New `booleans` property in `ValueSource`. [[release-notes-5.5.0-M2-junit-vintage]] === JUnit Vintage ==== Bug Fixes * ❓ ==== Deprecations and Breaking Changes * ❓ ==== New Features and Improvements * ❓
[[release-notes-5.5.0-M2]] == 5.5.0-M2️ *Date of Release:* ❓ *Scope:* ❓ For a complete list of all _closed_ issues and pull requests for this release, consult the link:{junit5-repo}+/milestone/37?closed=1+[5.5 M2] milestone page in the JUnit repository on GitHub. [[release-notes-5.5.0-M2-junit-platform]] === JUnit Platform ==== Bug Fixes * ❓ ==== Deprecations and Breaking Changes * ❓ ==== New Features and Improvements * ❓ [[release-notes-5.5.0-M2-junit-jupiter]] === JUnit Jupiter ==== Bug Fixes * Parameterized tests no longer throw an `ArrayStoreException` when creating human-readable test names. ==== Deprecations and Breaking Changes * ❓ ==== New Features and Improvements * ❓ [[release-notes-5.5.0-M2-junit-vintage]] === JUnit Vintage ==== Bug Fixes * ❓ ==== Deprecations and Breaking Changes * ❓ ==== New Features and Improvements * ❓
Clean up 5.5 M2 release notes
Clean up 5.5 M2 release notes The following feature enhancement was already shipped with 5.5 M1: * `@ValueSource` now additionally supports literal values of type `boolean` for parameterized tests.
AsciiDoc
epl-1.0
sbrannen/junit-lambda,junit-team/junit-lambda
15d46988dc74d29a3c4567e83de60717b4683561
docs/reference/analysis/tokenfilters/keyword-marker-tokenfilter.asciidoc
docs/reference/analysis/tokenfilters/keyword-marker-tokenfilter.asciidoc
[[analysis-keyword-marker-tokenfilter]] === Keyword Marker Token Filter Protects words from being modified by stemmers. Must be placed before any stemming filters. [cols="<,<",options="header",] |======================================================================= |Setting |Description |`keywords` |A list of words to use. |`keywords_path` |A path (either relative to `config` location, or absolute) to a list of words. |`ignore_case` |Set to `true` to lower case all words first. Defaults to `false`. |======================================================================= Here is an example: [source,js] -------------------------------------------------- index : analysis : analyzer : myAnalyzer : type : custom tokenizer : standard filter : [lowercase, protwods, porter_stem] filter : protwods : type : keyword_marker keywords_path : analysis/protwords.txt --------------------------------------------------
[[analysis-keyword-marker-tokenfilter]] === Keyword Marker Token Filter Protects words from being modified by stemmers. Must be placed before any stemming filters. [cols="<,<",options="header",] |======================================================================= |Setting |Description |`keywords` |A list of words to use. |`keywords_path` |A path (either relative to `config` location, or absolute) to a list of words. |`ignore_case` |Set to `true` to lower case all words first. Defaults to `false`. |======================================================================= Here is an example: [source,js] -------------------------------------------------- index : analysis : analyzer : myAnalyzer : type : custom tokenizer : standard filter : [lowercase, protwords, porter_stem] filter : protwords : type : keyword_marker keywords_path : analysis/protwords.txt --------------------------------------------------
Fix typo in sample json
Fix typo in sample json Fixes #9253
AsciiDoc
apache-2.0
jimczi/elasticsearch,mute/elasticsearch,linglaiyao1314/elasticsearch,schonfeld/elasticsearch,huanzhong/elasticsearch,girirajsharma/elasticsearch,Widen/elasticsearch,jeteve/elasticsearch,ricardocerq/elasticsearch,gingerwizard/elasticsearch,nazarewk/elasticsearch,Clairebi/ElasticsearchClone,liweinan0423/elasticsearch,sauravmondallive/elasticsearch,pritishppai/elasticsearch,strapdata/elassandra-test,vingupta3/elasticsearch,uschindler/elasticsearch,MisterAndersen/elasticsearch,kalburgimanjunath/elasticsearch,karthikjaps/elasticsearch,mapr/elasticsearch,MetSystem/elasticsearch,avikurapati/elasticsearch,jsgao0/elasticsearch,kalimatas/elasticsearch,strapdata/elassandra,mmaracic/elasticsearch,mjhennig/elasticsearch,wimvds/elasticsearch,mute/elasticsearch,mmaracic/elasticsearch,karthikjaps/elasticsearch,ThiagoGarciaAlves/elasticsearch,jeteve/elasticsearch,C-Bish/elasticsearch,dongjoon-hyun/elasticsearch,nrkkalyan/elasticsearch,kenshin233/elasticsearch,kimimj/elasticsearch,avikurapati/elasticsearch,cwurm/elasticsearch,iantruslove/elasticsearch,henakamaMSFT/elasticsearch,masaruh/elasticsearch,kubum/elasticsearch,markllama/elasticsearch,areek/elasticsearch,polyfractal/elasticsearch,ydsakyclguozi/elasticsearch,diendt/elasticsearch,Charlesdong/elasticsearch,jango2015/elasticsearch,SaiprasadKrishnamurthy/elasticsearch,a2lin/elasticsearch,wbowling/elasticsearch,caengcjd/elasticsearch,linglaiyao1314/elasticsearch,mapr/elasticsearch,franklanganke/elasticsearch,davidvgalbraith/elasticsearch,ivansun1010/elasticsearch,overcome/elasticsearch,feiqitian/elasticsearch,yuy168/elasticsearch,maddin2016/elasticsearch,mm0/elasticsearch,amit-shar/elasticsearch,iantruslove/elasticsearch,artnowo/elasticsearch,i-am-Nathan/elasticsearch,trangvh/elasticsearch,JackyMai/elasticsearch,beiske/elasticsearch,winstonewert/elasticsearch,NBSW/elasticsearch,javachengwc/elasticsearch,wimvds/elasticsearch,lydonchandra/elasticsearch,djschny/elasticsearch,Ansh90/elasticsearch,Collaborne/elasticsearch,fforbeck/elasticsearch,Rygbee/elasticsearch,smflorentino/elasticsearch,springning/elasticsearch,rhoml/elasticsearch,caengcjd/elasticsearch,rhoml/elasticsearch,ImpressTV/elasticsearch,scorpionvicky/elasticsearch,myelin/elasticsearch,robin13/elasticsearch,kunallimaye/elasticsearch,kkirsche/elasticsearch,spiegela/elasticsearch,henakamaMSFT/elasticsearch,nezirus/elasticsearch,SergVro/elasticsearch,coding0011/elasticsearch,rlugojr/elasticsearch,tsohil/elasticsearch,JackyMai/elasticsearch,dylan8902/elasticsearch,palecur/elasticsearch,camilojd/elasticsearch,ydsakyclguozi/elasticsearch,pranavraman/elasticsearch,wimvds/elasticsearch,jchampion/elasticsearch,vvcephei/elasticsearch,petabytedata/elasticsearch,ouyangkongtong/elasticsearch,nazarewk/elasticsearch,hydro2k/elasticsearch,onegambler/elasticsearch,mjason3/elasticsearch,alexbrasetvik/elasticsearch,beiske/elasticsearch,vingupta3/elasticsearch,SergVro/elasticsearch,btiernay/elasticsearch,MjAbuz/elasticsearch,Widen/elasticsearch,pritishppai/elasticsearch,hanst/elasticsearch,qwerty4030/elasticsearch,mjhennig/elasticsearch,fooljohnny/elasticsearch,ZTE-PaaS/elasticsearch,EasonYi/elasticsearch,EasonYi/elasticsearch,yynil/elasticsearch,himanshuag/elasticsearch,clintongormley/elasticsearch,Microsoft/elasticsearch,hanst/elasticsearch,gingerwizard/elasticsearch,elasticdog/elasticsearch,mkis-/elasticsearch,tkssharma/elasticsearch,strapdata/elassandra-test,KimTaehee/elasticsearch,mikemccand/elasticsearch,strapdata/elassandra5-rc,bawse/elasticsearch,winstonewert/elasticsearch,onegambler/elasticsearch,franklanganke/elasticsearch,MjAbuz/elasticsearch,mohit/elasticsearch,HarishAtGitHub/elasticsearch,Brijeshrpatel9/elasticsearch,apepper/elasticsearch,beiske/elasticsearch,drewr/elasticsearch,tahaemin/elasticsearch,C-Bish/elasticsearch,pritishppai/elasticsearch,onegambler/elasticsearch,wayeast/elasticsearch,Uiho/elasticsearch,Charlesdong/elasticsearch,StefanGor/elasticsearch,tsohil/elasticsearch,Widen/elasticsearch,ckclark/elasticsearch,linglaiyao1314/elasticsearch,truemped/elasticsearch,andrestc/elasticsearch,onegambler/elasticsearch,wayeast/elasticsearch,franklanganke/elasticsearch,jw0201/elastic,pozhidaevak/elasticsearch,springning/elasticsearch,sc0ttkclark/elasticsearch,vroyer/elassandra,i-am-Nathan/elasticsearch,sposam/elasticsearch,kunallimaye/elasticsearch,ThalaivaStars/OrgRepo1,nknize/elasticsearch,zhiqinghuang/elasticsearch,vietlq/elasticsearch,brandonkearby/elasticsearch,mute/elasticsearch,mcku/elasticsearch,xuzha/elasticsearch,kalburgimanjunath/elasticsearch,nellicus/elasticsearch,kubum/elasticsearch,YosuaMichael/elasticsearch,jaynblue/elasticsearch,sreeramjayan/elasticsearch,huanzhong/elasticsearch,Liziyao/elasticsearch,dataduke/elasticsearch,tsohil/elasticsearch,baishuo/elasticsearch_v2.1.0-baishuo,Widen/elasticsearch,mapr/elasticsearch,rhoml/elasticsearch,amaliujia/elasticsearch,infusionsoft/elasticsearch,sposam/elasticsearch,yynil/elasticsearch,xingguang2013/elasticsearch,rajanm/elasticsearch,polyfractal/elasticsearch,umeshdangat/elasticsearch,ivansun1010/elasticsearch,queirozfcom/elasticsearch,smflorentino/elasticsearch,dpursehouse/elasticsearch,camilojd/elasticsearch,geidies/elasticsearch,sc0ttkclark/elasticsearch,yuy168/elasticsearch,bawse/elasticsearch,strapdata/elassandra,MichaelLiZhou/elasticsearch,tkssharma/elasticsearch,apepper/elasticsearch,strapdata/elassandra-test,ThalaivaStars/OrgRepo1,codebunt/elasticsearch,hanst/elasticsearch,javachengwc/elasticsearch,cwurm/elasticsearch,yuy168/elasticsearch,sc0ttkclark/elasticsearch,AshishThakur/elasticsearch,aglne/elasticsearch,loconsolutions/elasticsearch,kkirsche/elasticsearch,loconsolutions/elasticsearch,girirajsharma/elasticsearch,HarishAtGitHub/elasticsearch,khiraiwa/elasticsearch,ydsakyclguozi/elasticsearch,mnylen/elasticsearch,drewr/elasticsearch,lydonchandra/elasticsearch,chirilo/elasticsearch,thecocce/elasticsearch,iacdingping/elasticsearch,MisterAndersen/elasticsearch,episerver/elasticsearch,Collaborne/elasticsearch,diendt/elasticsearch,jeteve/elasticsearch,fekaputra/elasticsearch,mkis-/elasticsearch,sreeramjayan/elasticsearch,baishuo/elasticsearch_v2.1.0-baishuo,mapr/elasticsearch,zeroctu/elasticsearch,Uiho/elasticsearch,rhoml/elasticsearch,obourgain/elasticsearch,xingguang2013/elasticsearch,fernandozhu/elasticsearch,easonC/elasticsearch,Siddartha07/elasticsearch,Fsero/elasticsearch,awislowski/elasticsearch,StefanGor/elasticsearch,humandb/elasticsearch,sarwarbhuiyan/elasticsearch,fred84/elasticsearch,mrorii/elasticsearch,HonzaKral/elasticsearch,springning/elasticsearch,Collaborne/elasticsearch,franklanganke/elasticsearch,yanjunh/elasticsearch,areek/elasticsearch,wayeast/elasticsearch,kunallimaye/elasticsearch,vroyer/elasticassandra,zeroctu/elasticsearch,huypx1292/elasticsearch,Shekharrajak/elasticsearch,dylan8902/elasticsearch,LewayneNaidoo/elasticsearch,mjason3/elasticsearch,socialrank/elasticsearch,maddin2016/elasticsearch,girirajsharma/elasticsearch,dongjoon-hyun/elasticsearch,shreejay/elasticsearch,iamjakob/elasticsearch,s1monw/elasticsearch,wittyameta/elasticsearch,nomoa/elasticsearch,hanswang/elasticsearch,queirozfcom/elasticsearch,kaneshin/elasticsearch,MichaelLiZhou/elasticsearch,markllama/elasticsearch,milodky/elasticsearch,gmarz/elasticsearch,lchennup/elasticsearch,iantruslove/elasticsearch,C-Bish/elasticsearch,nezirus/elasticsearch,jimczi/elasticsearch,fooljohnny/elasticsearch,Chhunlong/elasticsearch,mgalushka/elasticsearch,Asimov4/elasticsearch,onegambler/elasticsearch,szroland/elasticsearch,mcku/elasticsearch,lzo/elasticsearch-1,robin13/elasticsearch,wangyuxue/elasticsearch,petabytedata/elasticsearch,GlenRSmith/elasticsearch,amaliujia/elasticsearch,loconsolutions/elasticsearch,fekaputra/elasticsearch,yuy168/elasticsearch,chrismwendt/elasticsearch,EasonYi/elasticsearch,tebriel/elasticsearch,dataduke/elasticsearch,nomoa/elasticsearch,mute/elasticsearch,Flipkart/elasticsearch,ouyangkongtong/elasticsearch,s1monw/elasticsearch,palecur/elasticsearch,HonzaKral/elasticsearch,yanjunh/elasticsearch,bestwpw/elasticsearch,wenpos/elasticsearch,fforbeck/elasticsearch,kimimj/elasticsearch,Flipkart/elasticsearch,cnfire/elasticsearch-1,vietlq/elasticsearch,kcompher/elasticsearch,vietlq/elasticsearch,vroyer/elassandra,dataduke/elasticsearch,njlawton/elasticsearch,Ansh90/elasticsearch,wbowling/elasticsearch,drewr/elasticsearch,szroland/elasticsearch,fekaputra/elasticsearch,Kakakakakku/elasticsearch,wittyameta/elasticsearch,mm0/elasticsearch,masaruh/elasticsearch,hanswang/elasticsearch,javachengwc/elasticsearch,anti-social/elasticsearch,Shekharrajak/elasticsearch,Ansh90/elasticsearch,queirozfcom/elasticsearch,kaneshin/elasticsearch,artnowo/elasticsearch,iacdingping/elasticsearch,heng4fun/elasticsearch,henakamaMSFT/elasticsearch,HarishAtGitHub/elasticsearch,javachengwc/elasticsearch,C-Bish/elasticsearch,davidvgalbraith/elasticsearch,martinstuga/elasticsearch,nezirus/elasticsearch,nknize/elasticsearch,mjhennig/elasticsearch,martinstuga/elasticsearch,Asimov4/elasticsearch,truemped/elasticsearch,lightslife/elasticsearch,lchennup/elasticsearch,hechunwen/elasticsearch,tahaemin/elasticsearch,F0lha/elasticsearch,Shepard1212/elasticsearch,MichaelLiZhou/elasticsearch,sreeramjayan/elasticsearch,clintongormley/elasticsearch,mcku/elasticsearch,lchennup/elasticsearch,vvcephei/elasticsearch,elancom/elasticsearch,TonyChai24/ESSource,sauravmondallive/elasticsearch,gingerwizard/elasticsearch,acchen97/elasticsearch,nellicus/elasticsearch,wangyuxue/elasticsearch,jsgao0/elasticsearch,sjohnr/elasticsearch,golubev/elasticsearch,18098924759/elasticsearch,rajanm/elasticsearch,yongminxia/elasticsearch,jimczi/elasticsearch,khiraiwa/elasticsearch,JackyMai/elasticsearch,MichaelLiZhou/elasticsearch,uschindler/elasticsearch,AndreKR/elasticsearch,MetSystem/elasticsearch,jaynblue/elasticsearch,mbrukman/elasticsearch,mjason3/elasticsearch,mcku/elasticsearch,iamjakob/elasticsearch,SergVro/elasticsearch,wangtuo/elasticsearch,markharwood/elasticsearch,MetSystem/elasticsearch,kaneshin/elasticsearch,vvcephei/elasticsearch,areek/elasticsearch,AndreKR/elasticsearch,vroyer/elasticassandra,JSCooke/elasticsearch,zhiqinghuang/elasticsearch,dpursehouse/elasticsearch,koxa29/elasticsearch,weipinghe/elasticsearch,MaineC/elasticsearch,drewr/elasticsearch,masaruh/elasticsearch,truemped/elasticsearch,hechunwen/elasticsearch,nknize/elasticsearch,caengcjd/elasticsearch,apepper/elasticsearch,tsohil/elasticsearch,dpursehouse/elasticsearch,kimimj/elasticsearch,kkirsche/elasticsearch,achow/elasticsearch,jeteve/elasticsearch,IanvsPoplicola/elasticsearch,zeroctu/elasticsearch,StefanGor/elasticsearch,Chhunlong/elasticsearch,hanst/elasticsearch,dylan8902/elasticsearch,kevinkluge/elasticsearch,tebriel/elasticsearch,javachengwc/elasticsearch,Flipkart/elasticsearch,ZTE-PaaS/elasticsearch,Asimov4/elasticsearch,hirdesh2008/elasticsearch,Clairebi/ElasticsearchClone,wangtuo/elasticsearch,markllama/elasticsearch,xuzha/elasticsearch,luiseduardohdbackup/elasticsearch,sarwarbhuiyan/elasticsearch,kaneshin/elasticsearch,robin13/elasticsearch,geidies/elasticsearch,sreeramjayan/elasticsearch,F0lha/elasticsearch,bestwpw/elasticsearch,javachengwc/elasticsearch,thecocce/elasticsearch,kalburgimanjunath/elasticsearch,milodky/elasticsearch,Rygbee/elasticsearch,Uiho/elasticsearch,btiernay/elasticsearch,MjAbuz/elasticsearch,zhiqinghuang/elasticsearch,jimhooker2002/elasticsearch,mortonsykes/elasticsearch,fooljohnny/elasticsearch,gfyoung/elasticsearch,KimTaehee/elasticsearch,sc0ttkclark/elasticsearch,fooljohnny/elasticsearch,ThiagoGarciaAlves/elasticsearch,artnowo/elasticsearch,knight1128/elasticsearch,huanzhong/elasticsearch,ImpressTV/elasticsearch,AshishThakur/elasticsearch,karthikjaps/elasticsearch,smflorentino/elasticsearch,spiegela/elasticsearch,gfyoung/elasticsearch,djschny/elasticsearch,Collaborne/elasticsearch,chrismwendt/elasticsearch,Ansh90/elasticsearch,TonyChai24/ESSource,rento19962/elasticsearch,strapdata/elassandra-test,artnowo/elasticsearch,sc0ttkclark/elasticsearch,i-am-Nathan/elasticsearch,tebriel/elasticsearch,dylan8902/elasticsearch,mnylen/elasticsearch,adrianbk/elasticsearch,petabytedata/elasticsearch,strapdata/elassandra,coding0011/elasticsearch,caengcjd/elasticsearch,dongjoon-hyun/elasticsearch,jaynblue/elasticsearch,Widen/elasticsearch,jw0201/elastic,sarwarbhuiyan/elasticsearch,hydro2k/elasticsearch,abibell/elasticsearch,kcompher/elasticsearch,zeroctu/elasticsearch,Fsero/elasticsearch,sdauletau/elasticsearch,beiske/elasticsearch,hafkensite/elasticsearch,pablocastro/elasticsearch,JSCooke/elasticsearch,strapdata/elassandra,PhaedrusTheGreek/elasticsearch,abibell/elasticsearch,HonzaKral/elasticsearch,trangvh/elasticsearch,hirdesh2008/elasticsearch,achow/elasticsearch,MisterAndersen/elasticsearch,Siddartha07/elasticsearch,markwalkom/elasticsearch,wenpos/elasticsearch,myelin/elasticsearch,kingaj/elasticsearch,mohit/elasticsearch,kingaj/elasticsearch,ouyangkongtong/elasticsearch,markwalkom/elasticsearch,overcome/elasticsearch,apepper/elasticsearch,strapdata/elassandra5-rc,xpandan/elasticsearch,mgalushka/elasticsearch,awislowski/elasticsearch,dylan8902/elasticsearch,hirdesh2008/elasticsearch,jango2015/elasticsearch,kenshin233/elasticsearch,drewr/elasticsearch,hechunwen/elasticsearch,wuranbo/elasticsearch,vietlq/elasticsearch,btiernay/elasticsearch,markharwood/elasticsearch,avikurapati/elasticsearch,Stacey-Gammon/elasticsearch,MaineC/elasticsearch,pablocastro/elasticsearch,vvcephei/elasticsearch,PhaedrusTheGreek/elasticsearch,xuzha/elasticsearch,thecocce/elasticsearch,jw0201/elastic,zhiqinghuang/elasticsearch,chirilo/elasticsearch,socialrank/elasticsearch,ckclark/elasticsearch,phani546/elasticsearch,amit-shar/elasticsearch,martinstuga/elasticsearch,xpandan/elasticsearch,wuranbo/elasticsearch,vrkansagara/elasticsearch,codebunt/elasticsearch,nrkkalyan/elasticsearch,golubev/elasticsearch,Shekharrajak/elasticsearch,StefanGor/elasticsearch,SaiprasadKrishnamurthy/elasticsearch,sjohnr/elasticsearch,himanshuag/elasticsearch,fekaputra/elasticsearch,hanswang/elasticsearch,hafkensite/elasticsearch,camilojd/elasticsearch,mjason3/elasticsearch,rhoml/elasticsearch,Fsero/elasticsearch,weipinghe/elasticsearch,codebunt/elasticsearch,LeoYao/elasticsearch,NBSW/elasticsearch,acchen97/elasticsearch,chrismwendt/elasticsearch,dataduke/elasticsearch,LeoYao/elasticsearch,jbertouch/elasticsearch,AndreKR/elasticsearch,LeoYao/elasticsearch,winstonewert/elasticsearch,skearns64/elasticsearch,Clairebi/ElasticsearchClone,Uiho/elasticsearch,andrestc/elasticsearch,abibell/elasticsearch,petabytedata/elasticsearch,nazarewk/elasticsearch,strapdata/elassandra-test,feiqitian/elasticsearch,sposam/elasticsearch,clintongormley/elasticsearch,MjAbuz/elasticsearch,easonC/elasticsearch,sarwarbhuiyan/elasticsearch,dpursehouse/elasticsearch,ImpressTV/elasticsearch,adrianbk/elasticsearch,golubev/elasticsearch,TonyChai24/ESSource,mgalushka/elasticsearch,masterweb121/elasticsearch,himanshuag/elasticsearch,IanvsPoplicola/elasticsearch,jango2015/elasticsearch,himanshuag/elasticsearch,gmarz/elasticsearch,scorpionvicky/elasticsearch,pranavraman/elasticsearch,NBSW/elasticsearch,huypx1292/elasticsearch,kaneshin/elasticsearch,rhoml/elasticsearch,schonfeld/elasticsearch,sarwarbhuiyan/elasticsearch,franklanganke/elasticsearch,jsgao0/elasticsearch,AndreKR/elasticsearch,avikurapati/elasticsearch,hechunwen/elasticsearch,snikch/elasticsearch,jimhooker2002/elasticsearch,szroland/elasticsearch,dongjoon-hyun/elasticsearch,wenpos/elasticsearch,kkirsche/elasticsearch,bestwpw/elasticsearch,codebunt/elasticsearch,vingupta3/elasticsearch,nilabhsagar/elasticsearch,himanshuag/elasticsearch,markwalkom/elasticsearch,sc0ttkclark/elasticsearch,kimimj/elasticsearch,easonC/elasticsearch,sjohnr/elasticsearch,KimTaehee/elasticsearch,awislowski/elasticsearch,brandonkearby/elasticsearch,elancom/elasticsearch,KimTaehee/elasticsearch,sreeramjayan/elasticsearch,mnylen/elasticsearch,truemped/elasticsearch,tahaemin/elasticsearch,abibell/elasticsearch,cwurm/elasticsearch,qwerty4030/elasticsearch,btiernay/elasticsearch,cnfire/elasticsearch-1,lmtwga/elasticsearch,camilojd/elasticsearch,wangtuo/elasticsearch,lmtwga/elasticsearch,dylan8902/elasticsearch,alexkuk/elasticsearch,karthikjaps/elasticsearch,lzo/elasticsearch-1,fred84/elasticsearch,sauravmondallive/elasticsearch,huanzhong/elasticsearch,Helen-Zhao/elasticsearch,shreejay/elasticsearch,mgalushka/elasticsearch,luiseduardohdbackup/elasticsearch,sreeramjayan/elasticsearch,kalburgimanjunath/elasticsearch,mkis-/elasticsearch,infusionsoft/elasticsearch,szroland/elasticsearch,kcompher/elasticsearch,snikch/elasticsearch,sjohnr/elasticsearch,wangyuxue/elasticsearch,girirajsharma/elasticsearch,nellicus/elasticsearch,phani546/elasticsearch,andrestc/elasticsearch,polyfractal/elasticsearch,kevinkluge/elasticsearch,xingguang2013/elasticsearch,milodky/elasticsearch,likaiwalkman/elasticsearch,hanswang/elasticsearch,lightslife/elasticsearch,jimhooker2002/elasticsearch,jw0201/elastic,hirdesh2008/elasticsearch,queirozfcom/elasticsearch,anti-social/elasticsearch,MisterAndersen/elasticsearch,snikch/elasticsearch,jbertouch/elasticsearch,pritishppai/elasticsearch,gmarz/elasticsearch,EasonYi/elasticsearch,linglaiyao1314/elasticsearch,Clairebi/ElasticsearchClone,tebriel/elasticsearch,acchen97/elasticsearch,cwurm/elasticsearch,dpursehouse/elasticsearch,robin13/elasticsearch,weipinghe/elasticsearch,beiske/elasticsearch,jbertouch/elasticsearch,alexkuk/elasticsearch,pablocastro/elasticsearch,kimimj/elasticsearch,wangtuo/elasticsearch,schonfeld/elasticsearch,Siddartha07/elasticsearch,mkis-/elasticsearch,apepper/elasticsearch,winstonewert/elasticsearch,qwerty4030/elasticsearch,skearns64/elasticsearch,slavau/elasticsearch,luiseduardohdbackup/elasticsearch,queirozfcom/elasticsearch,Liziyao/elasticsearch,huanzhong/elasticsearch,likaiwalkman/elasticsearch,yongminxia/elasticsearch,strapdata/elassandra5-rc,onegambler/elasticsearch,Stacey-Gammon/elasticsearch,ThiagoGarciaAlves/elasticsearch,andrestc/elasticsearch,wittyameta/elasticsearch,socialrank/elasticsearch,kkirsche/elasticsearch,golubev/elasticsearch,MichaelLiZhou/elasticsearch,shreejay/elasticsearch,alexkuk/elasticsearch,zhiqinghuang/elasticsearch,achow/elasticsearch,wuranbo/elasticsearch,feiqitian/elasticsearch,nellicus/elasticsearch,infusionsoft/elasticsearch,polyfractal/elasticsearch,KimTaehee/elasticsearch,brandonkearby/elasticsearch,naveenhooda2000/elasticsearch,masaruh/elasticsearch,andrestc/elasticsearch,IanvsPoplicola/elasticsearch,linglaiyao1314/elasticsearch,Helen-Zhao/elasticsearch,lchennup/elasticsearch,nazarewk/elasticsearch,mgalushka/elasticsearch,MjAbuz/elasticsearch,polyfractal/elasticsearch,szroland/elasticsearch,mcku/elasticsearch,fred84/elasticsearch,yongminxia/elasticsearch,mcku/elasticsearch,slavau/elasticsearch,rajanm/elasticsearch,fforbeck/elasticsearch,strapdata/elassandra-test,jw0201/elastic,apepper/elasticsearch,Clairebi/ElasticsearchClone,HarishAtGitHub/elasticsearch,Shepard1212/elasticsearch,kcompher/elasticsearch,Siddartha07/elasticsearch,mnylen/elasticsearch,spiegela/elasticsearch,NBSW/elasticsearch,cwurm/elasticsearch,nomoa/elasticsearch,MaineC/elasticsearch,nknize/elasticsearch,iamjakob/elasticsearch,Microsoft/elasticsearch,khiraiwa/elasticsearch,wittyameta/elasticsearch,humandb/elasticsearch,umeshdangat/elasticsearch,nezirus/elasticsearch,ricardocerq/elasticsearch,thecocce/elasticsearch,palecur/elasticsearch,MetSystem/elasticsearch,adrianbk/elasticsearch,smflorentino/elasticsearch,infusionsoft/elasticsearch,alexbrasetvik/elasticsearch,ulkas/elasticsearch,ulkas/elasticsearch,SaiprasadKrishnamurthy/elasticsearch,phani546/elasticsearch,yanjunh/elasticsearch,xuzha/elasticsearch,mgalushka/elasticsearch,djschny/elasticsearch,MichaelLiZhou/elasticsearch,Brijeshrpatel9/elasticsearch,springning/elasticsearch,gingerwizard/elasticsearch,andrejserafim/elasticsearch,LewayneNaidoo/elasticsearch,jeteve/elasticsearch,hirdesh2008/elasticsearch,camilojd/elasticsearch,alexshadow007/elasticsearch,mapr/elasticsearch,jprante/elasticsearch,kalimatas/elasticsearch,brandonkearby/elasticsearch,obourgain/elasticsearch,AshishThakur/elasticsearch,lzo/elasticsearch-1,thecocce/elasticsearch,jango2015/elasticsearch,chrismwendt/elasticsearch,JSCooke/elasticsearch,truemped/elasticsearch,diendt/elasticsearch,Fsero/elasticsearch,fernandozhu/elasticsearch,djschny/elasticsearch,elancom/elasticsearch,scorpionvicky/elasticsearch,kunallimaye/elasticsearch,jimhooker2002/elasticsearch,wbowling/elasticsearch,jchampion/elasticsearch,jw0201/elastic,linglaiyao1314/elasticsearch,himanshuag/elasticsearch,Stacey-Gammon/elasticsearch,rmuir/elasticsearch,ivansun1010/elasticsearch,ulkas/elasticsearch,zeroctu/elasticsearch,Shekharrajak/elasticsearch,yuy168/elasticsearch,kenshin233/elasticsearch,hanswang/elasticsearch,tsohil/elasticsearch,episerver/elasticsearch,tebriel/elasticsearch,Microsoft/elasticsearch,lchennup/elasticsearch,lightslife/elasticsearch,koxa29/elasticsearch,ZTE-PaaS/elasticsearch,socialrank/elasticsearch,luiseduardohdbackup/elasticsearch,jimhooker2002/elasticsearch,lks21c/elasticsearch,alexshadow007/elasticsearch,hanst/elasticsearch,rmuir/elasticsearch,weipinghe/elasticsearch,IanvsPoplicola/elasticsearch,alexshadow007/elasticsearch,YosuaMichael/elasticsearch,Uiho/elasticsearch,overcome/elasticsearch,iantruslove/elasticsearch,hydro2k/elasticsearch,wenpos/elasticsearch,jango2015/elasticsearch,overcome/elasticsearch,hydro2k/elasticsearch,mjason3/elasticsearch,jango2015/elasticsearch,jaynblue/elasticsearch,wbowling/elasticsearch,koxa29/elasticsearch,lydonchandra/elasticsearch,yuy168/elasticsearch,milodky/elasticsearch,palecur/elasticsearch,mcku/elasticsearch,Liziyao/elasticsearch,truemped/elasticsearch,koxa29/elasticsearch,episerver/elasticsearch,alexkuk/elasticsearch,liweinan0423/elasticsearch,kingaj/elasticsearch,skearns64/elasticsearch,hafkensite/elasticsearch,xpandan/elasticsearch,jchampion/elasticsearch,jchampion/elasticsearch,njlawton/elasticsearch,Flipkart/elasticsearch,fred84/elasticsearch,mohit/elasticsearch,ulkas/elasticsearch,Collaborne/elasticsearch,vingupta3/elasticsearch,maddin2016/elasticsearch,kaneshin/elasticsearch,abibell/elasticsearch,huanzhong/elasticsearch,SaiprasadKrishnamurthy/elasticsearch,pritishppai/elasticsearch,ulkas/elasticsearch,rmuir/elasticsearch,mmaracic/elasticsearch,ydsakyclguozi/elasticsearch,shreejay/elasticsearch,tkssharma/elasticsearch,socialrank/elasticsearch,aglne/elasticsearch,mm0/elasticsearch,18098924759/elasticsearch,fernandozhu/elasticsearch,xpandan/elasticsearch,mm0/elasticsearch,MetSystem/elasticsearch,elancom/elasticsearch,amaliujia/elasticsearch,strapdata/elassandra-test,hafkensite/elasticsearch,AndreKR/elasticsearch,ThalaivaStars/OrgRepo1,Stacey-Gammon/elasticsearch,cnfire/elasticsearch-1,geidies/elasticsearch,kimimj/elasticsearch,kunallimaye/elasticsearch,pablocastro/elasticsearch,StefanGor/elasticsearch,alexbrasetvik/elasticsearch,huypx1292/elasticsearch,andrestc/elasticsearch,pranavraman/elasticsearch,markwalkom/elasticsearch,coding0011/elasticsearch,fooljohnny/elasticsearch,mnylen/elasticsearch,clintongormley/elasticsearch,GlenRSmith/elasticsearch,EasonYi/elasticsearch,LeoYao/elasticsearch,springning/elasticsearch,pranavraman/elasticsearch,avikurapati/elasticsearch,beiske/elasticsearch,scorpionvicky/elasticsearch,sc0ttkclark/elasticsearch,nknize/elasticsearch,Kakakakakku/elasticsearch,tkssharma/elasticsearch,rento19962/elasticsearch,hirdesh2008/elasticsearch,mortonsykes/elasticsearch,yuy168/elasticsearch,zhiqinghuang/elasticsearch,awislowski/elasticsearch,kcompher/elasticsearch,knight1128/elasticsearch,Charlesdong/elasticsearch,adrianbk/elasticsearch,sdauletau/elasticsearch,sneivandt/elasticsearch,alexshadow007/elasticsearch,iacdingping/elasticsearch,liweinan0423/elasticsearch,KimTaehee/elasticsearch,nrkkalyan/elasticsearch,martinstuga/elasticsearch,njlawton/elasticsearch,codebunt/elasticsearch,mrorii/elasticsearch,nellicus/elasticsearch,sposam/elasticsearch,kevinkluge/elasticsearch,Clairebi/ElasticsearchClone,zhiqinghuang/elasticsearch,mgalushka/elasticsearch,luiseduardohdbackup/elasticsearch,vietlq/elasticsearch,aglne/elasticsearch,andrejserafim/elasticsearch,zeroctu/elasticsearch,hechunwen/elasticsearch,Liziyao/elasticsearch,polyfractal/elasticsearch,fekaputra/elasticsearch,socialrank/elasticsearch,onegambler/elasticsearch,MjAbuz/elasticsearch,fforbeck/elasticsearch,Helen-Zhao/elasticsearch,phani546/elasticsearch,ricardocerq/elasticsearch,mjhennig/elasticsearch,baishuo/elasticsearch_v2.1.0-baishuo,ivansun1010/elasticsearch,alexkuk/elasticsearch,kalimatas/elasticsearch,hydro2k/elasticsearch,clintongormley/elasticsearch,sneivandt/elasticsearch,elasticdog/elasticsearch,glefloch/elasticsearch,iamjakob/elasticsearch,i-am-Nathan/elasticsearch,fred84/elasticsearch,achow/elasticsearch,umeshdangat/elasticsearch,lks21c/elasticsearch,vrkansagara/elasticsearch,a2lin/elasticsearch,scorpionvicky/elasticsearch,Asimov4/elasticsearch,amit-shar/elasticsearch,vingupta3/elasticsearch,uschindler/elasticsearch,yynil/elasticsearch,bawse/elasticsearch,vrkansagara/elasticsearch,kevinkluge/elasticsearch,AshishThakur/elasticsearch,scottsom/elasticsearch,vroyer/elasticassandra,djschny/elasticsearch,njlawton/elasticsearch,rajanm/elasticsearch,mbrukman/elasticsearch,snikch/elasticsearch,mohit/elasticsearch,s1monw/elasticsearch,Rygbee/elasticsearch,Kakakakakku/elasticsearch,iamjakob/elasticsearch,bestwpw/elasticsearch,cnfire/elasticsearch-1,camilojd/elasticsearch,JervyShi/elasticsearch,Kakakakakku/elasticsearch,mikemccand/elasticsearch,mjhennig/elasticsearch,YosuaMichael/elasticsearch,mohit/elasticsearch,glefloch/elasticsearch,markharwood/elasticsearch,mapr/elasticsearch,wuranbo/elasticsearch,LewayneNaidoo/elasticsearch,jimhooker2002/elasticsearch,mute/elasticsearch,sneivandt/elasticsearch,truemped/elasticsearch,diendt/elasticsearch,jpountz/elasticsearch,ckclark/elasticsearch,kenshin233/elasticsearch,knight1128/elasticsearch,ouyangkongtong/elasticsearch,Chhunlong/elasticsearch,shreejay/elasticsearch,sauravmondallive/elasticsearch,cnfire/elasticsearch-1,robin13/elasticsearch,mnylen/elasticsearch,s1monw/elasticsearch,mmaracic/elasticsearch,areek/elasticsearch,ulkas/elasticsearch,martinstuga/elasticsearch,jchampion/elasticsearch,girirajsharma/elasticsearch,yanjunh/elasticsearch,likaiwalkman/elasticsearch,acchen97/elasticsearch,franklanganke/elasticsearch,ESamir/elasticsearch,geidies/elasticsearch,kalimatas/elasticsearch,likaiwalkman/elasticsearch,huypx1292/elasticsearch,davidvgalbraith/elasticsearch,ZTE-PaaS/elasticsearch,golubev/elasticsearch,zkidkid/elasticsearch,hafkensite/elasticsearch,btiernay/elasticsearch,YosuaMichael/elasticsearch,andrejserafim/elasticsearch,Liziyao/elasticsearch,NBSW/elasticsearch,iantruslove/elasticsearch,yongminxia/elasticsearch,pozhidaevak/elasticsearch,JSCooke/elasticsearch,PhaedrusTheGreek/elasticsearch,zkidkid/elasticsearch,pritishppai/elasticsearch,ricardocerq/elasticsearch,Rygbee/elasticsearch,lydonchandra/elasticsearch,hydro2k/elasticsearch,maddin2016/elasticsearch,Chhunlong/elasticsearch,wuranbo/elasticsearch,knight1128/elasticsearch,ImpressTV/elasticsearch,gmarz/elasticsearch,sjohnr/elasticsearch,MichaelLiZhou/elasticsearch,amit-shar/elasticsearch,masterweb121/elasticsearch,iamjakob/elasticsearch,mute/elasticsearch,Collaborne/elasticsearch,gfyoung/elasticsearch,zkidkid/elasticsearch,kubum/elasticsearch,mjhennig/elasticsearch,ThiagoGarciaAlves/elasticsearch,F0lha/elasticsearch,infusionsoft/elasticsearch,bestwpw/elasticsearch,girirajsharma/elasticsearch,mrorii/elasticsearch,jprante/elasticsearch,skearns64/elasticsearch,wittyameta/elasticsearch,rlugojr/elasticsearch,ckclark/elasticsearch,Asimov4/elasticsearch,njlawton/elasticsearch,episerver/elasticsearch,luiseduardohdbackup/elasticsearch,F0lha/elasticsearch,Microsoft/elasticsearch,xingguang2013/elasticsearch,beiske/elasticsearch,sdauletau/elasticsearch,chirilo/elasticsearch,yynil/elasticsearch,alexbrasetvik/elasticsearch,ivansun1010/elasticsearch,sauravmondallive/elasticsearch,naveenhooda2000/elasticsearch,SaiprasadKrishnamurthy/elasticsearch,jaynblue/elasticsearch,jpountz/elasticsearch,a2lin/elasticsearch,umeshdangat/elasticsearch,xuzha/elasticsearch,khiraiwa/elasticsearch,jimhooker2002/elasticsearch,wimvds/elasticsearch,acchen97/elasticsearch,Siddartha07/elasticsearch,hydro2k/elasticsearch,slavau/elasticsearch,dylan8902/elasticsearch,gfyoung/elasticsearch,obourgain/elasticsearch,jpountz/elasticsearch,henakamaMSFT/elasticsearch,ThalaivaStars/OrgRepo1,kkirsche/elasticsearch,sneivandt/elasticsearch,slavau/elasticsearch,kalburgimanjunath/elasticsearch,mikemccand/elasticsearch,kubum/elasticsearch,SergVro/elasticsearch,vvcephei/elasticsearch,glefloch/elasticsearch,gingerwizard/elasticsearch,ouyangkongtong/elasticsearch,lightslife/elasticsearch,jaynblue/elasticsearch,kevinkluge/elasticsearch,jprante/elasticsearch,ESamir/elasticsearch,ckclark/elasticsearch,areek/elasticsearch,Fsero/elasticsearch,gmarz/elasticsearch,pranavraman/elasticsearch,kenshin233/elasticsearch,ouyangkongtong/elasticsearch,Charlesdong/elasticsearch,ydsakyclguozi/elasticsearch,masterweb121/elasticsearch,drewr/elasticsearch,LewayneNaidoo/elasticsearch,thecocce/elasticsearch,strapdata/elassandra5-rc,schonfeld/elasticsearch,henakamaMSFT/elasticsearch,btiernay/elasticsearch,markllama/elasticsearch,yongminxia/elasticsearch,markllama/elasticsearch,hanswang/elasticsearch,knight1128/elasticsearch,F0lha/elasticsearch,qwerty4030/elasticsearch,rento19962/elasticsearch,lmtwga/elasticsearch,wimvds/elasticsearch,MetSystem/elasticsearch,Rygbee/elasticsearch,davidvgalbraith/elasticsearch,uschindler/elasticsearch,wimvds/elasticsearch,mortonsykes/elasticsearch,nrkkalyan/elasticsearch,apepper/elasticsearch,gingerwizard/elasticsearch,ouyangkongtong/elasticsearch,HonzaKral/elasticsearch,feiqitian/elasticsearch,fekaputra/elasticsearch,chirilo/elasticsearch,GlenRSmith/elasticsearch,PhaedrusTheGreek/elasticsearch,lzo/elasticsearch-1,amit-shar/elasticsearch,episerver/elasticsearch,scottsom/elasticsearch,Stacey-Gammon/elasticsearch,slavau/elasticsearch,masterweb121/elasticsearch,zkidkid/elasticsearch,YosuaMichael/elasticsearch,s1monw/elasticsearch,dataduke/elasticsearch,wayeast/elasticsearch,kubum/elasticsearch,kingaj/elasticsearch,rento19962/elasticsearch,lchennup/elasticsearch,likaiwalkman/elasticsearch,nilabhsagar/elasticsearch,vvcephei/elasticsearch,mm0/elasticsearch,jsgao0/elasticsearch,chirilo/elasticsearch,sposam/elasticsearch,ESamir/elasticsearch,mkis-/elasticsearch,cnfire/elasticsearch-1,spiegela/elasticsearch,Brijeshrpatel9/elasticsearch,huypx1292/elasticsearch,Brijeshrpatel9/elasticsearch,zeroctu/elasticsearch,acchen97/elasticsearch,masterweb121/elasticsearch,EasonYi/elasticsearch,anti-social/elasticsearch,18098924759/elasticsearch,kalburgimanjunath/elasticsearch,lchennup/elasticsearch,caengcjd/elasticsearch,ESamir/elasticsearch,clintongormley/elasticsearch,infusionsoft/elasticsearch,maddin2016/elasticsearch,LeoYao/elasticsearch,lzo/elasticsearch-1,gingerwizard/elasticsearch,HarishAtGitHub/elasticsearch,Siddartha07/elasticsearch,GlenRSmith/elasticsearch,kalburgimanjunath/elasticsearch,18098924759/elasticsearch,khiraiwa/elasticsearch,wittyameta/elasticsearch,SergVro/elasticsearch,Collaborne/elasticsearch,overcome/elasticsearch,mm0/elasticsearch,kunallimaye/elasticsearch,lmtwga/elasticsearch,vrkansagara/elasticsearch,knight1128/elasticsearch,ThalaivaStars/OrgRepo1,Flipkart/elasticsearch,achow/elasticsearch,sposam/elasticsearch,Rygbee/elasticsearch,jeteve/elasticsearch,artnowo/elasticsearch,mnylen/elasticsearch,KimTaehee/elasticsearch,phani546/elasticsearch,a2lin/elasticsearch,jbertouch/elasticsearch,tsohil/elasticsearch,mbrukman/elasticsearch,JervyShi/elasticsearch,MaineC/elasticsearch,jprante/elasticsearch,obourgain/elasticsearch,iacdingping/elasticsearch,markharwood/elasticsearch,18098924759/elasticsearch,jpountz/elasticsearch,JervyShi/elasticsearch,wayeast/elasticsearch,liweinan0423/elasticsearch,iacdingping/elasticsearch,bawse/elasticsearch,scottsom/elasticsearch,kimimj/elasticsearch,F0lha/elasticsearch,masaruh/elasticsearch,smflorentino/elasticsearch,PhaedrusTheGreek/elasticsearch,Ansh90/elasticsearch,nrkkalyan/elasticsearch,jbertouch/elasticsearch,qwerty4030/elasticsearch,winstonewert/elasticsearch,rlugojr/elasticsearch,AshishThakur/elasticsearch,jsgao0/elasticsearch,petabytedata/elasticsearch,martinstuga/elasticsearch,Widen/elasticsearch,drewr/elasticsearch,masterweb121/elasticsearch,sarwarbhuiyan/elasticsearch,alexkuk/elasticsearch,nazarewk/elasticsearch,likaiwalkman/elasticsearch,amaliujia/elasticsearch,iamjakob/elasticsearch,szroland/elasticsearch,iantruslove/elasticsearch,karthikjaps/elasticsearch,nomoa/elasticsearch,mortonsykes/elasticsearch,skearns64/elasticsearch,nilabhsagar/elasticsearch,davidvgalbraith/elasticsearch,markwalkom/elasticsearch,pranavraman/elasticsearch,fekaputra/elasticsearch,diendt/elasticsearch,mjhennig/elasticsearch,petabytedata/elasticsearch,jpountz/elasticsearch,caengcjd/elasticsearch,weipinghe/elasticsearch,Brijeshrpatel9/elasticsearch,lydonchandra/elasticsearch,yynil/elasticsearch,dataduke/elasticsearch,mbrukman/elasticsearch,heng4fun/elasticsearch,TonyChai24/ESSource,adrianbk/elasticsearch,mbrukman/elasticsearch,wbowling/elasticsearch,rmuir/elasticsearch,NBSW/elasticsearch,bestwpw/elasticsearch,xingguang2013/elasticsearch,jbertouch/elasticsearch,Shepard1212/elasticsearch,acchen97/elasticsearch,Ansh90/elasticsearch,lightslife/elasticsearch,scottsom/elasticsearch,tkssharma/elasticsearch,ckclark/elasticsearch,LewayneNaidoo/elasticsearch,scottsom/elasticsearch,tahaemin/elasticsearch,YosuaMichael/elasticsearch,ImpressTV/elasticsearch,rento19962/elasticsearch,wayeast/elasticsearch,chirilo/elasticsearch,himanshuag/elasticsearch,adrianbk/elasticsearch,mrorii/elasticsearch,socialrank/elasticsearch,jsgao0/elasticsearch,yynil/elasticsearch,palecur/elasticsearch,djschny/elasticsearch,ImpressTV/elasticsearch,feiqitian/elasticsearch,Microsoft/elasticsearch,schonfeld/elasticsearch,rajanm/elasticsearch,yanjunh/elasticsearch,vietlq/elasticsearch,hafkensite/elasticsearch,milodky/elasticsearch,Kakakakakku/elasticsearch,rmuir/elasticsearch,nrkkalyan/elasticsearch,lzo/elasticsearch-1,lzo/elasticsearch-1,hanst/elasticsearch,wbowling/elasticsearch,snikch/elasticsearch,mute/elasticsearch,rlugojr/elasticsearch,kevinkluge/elasticsearch,SaiprasadKrishnamurthy/elasticsearch,feiqitian/elasticsearch,slavau/elasticsearch,amit-shar/elasticsearch,umeshdangat/elasticsearch,petabytedata/elasticsearch,ricardocerq/elasticsearch,andrestc/elasticsearch,rajanm/elasticsearch,mortonsykes/elasticsearch,mkis-/elasticsearch,springning/elasticsearch,HarishAtGitHub/elasticsearch,schonfeld/elasticsearch,lmtwga/elasticsearch,xingguang2013/elasticsearch,ivansun1010/elasticsearch,lydonchandra/elasticsearch,fernandozhu/elasticsearch,Shepard1212/elasticsearch,geidies/elasticsearch,glefloch/elasticsearch,Chhunlong/elasticsearch,lmtwga/elasticsearch,loconsolutions/elasticsearch,kcompher/elasticsearch,SaiprasadKrishnamurthy/elasticsearch,huypx1292/elasticsearch,pablocastro/elasticsearch,alexshadow007/elasticsearch,iacdingping/elasticsearch,vrkansagara/elasticsearch,golubev/elasticsearch,infusionsoft/elasticsearch,pablocastro/elasticsearch,Brijeshrpatel9/elasticsearch,overcome/elasticsearch,Chhunlong/elasticsearch,GlenRSmith/elasticsearch,AndreKR/elasticsearch,vietlq/elasticsearch,lmtwga/elasticsearch,sauravmondallive/elasticsearch,MjAbuz/elasticsearch,lks21c/elasticsearch,easonC/elasticsearch,markllama/elasticsearch,rento19962/elasticsearch,linglaiyao1314/elasticsearch,codebunt/elasticsearch,markharwood/elasticsearch,andrejserafim/elasticsearch,hechunwen/elasticsearch,loconsolutions/elasticsearch,JackyMai/elasticsearch,iantruslove/elasticsearch,sjohnr/elasticsearch,amaliujia/elasticsearch,PhaedrusTheGreek/elasticsearch,nomoa/elasticsearch,Liziyao/elasticsearch,mbrukman/elasticsearch,naveenhooda2000/elasticsearch,xpandan/elasticsearch,bawse/elasticsearch,nilabhsagar/elasticsearch,Helen-Zhao/elasticsearch,JackyMai/elasticsearch,TonyChai24/ESSource,andrejserafim/elasticsearch,kingaj/elasticsearch,luiseduardohdbackup/elasticsearch,wimvds/elasticsearch,mrorii/elasticsearch,Siddartha07/elasticsearch,elancom/elasticsearch,kubum/elasticsearch,JervyShi/elasticsearch,amaliujia/elasticsearch,xpandan/elasticsearch,tebriel/elasticsearch,snikch/elasticsearch,huanzhong/elasticsearch,easonC/elasticsearch,tkssharma/elasticsearch,Chhunlong/elasticsearch,fforbeck/elasticsearch,geidies/elasticsearch,EasonYi/elasticsearch,achow/elasticsearch,ckclark/elasticsearch,pozhidaevak/elasticsearch,strapdata/elassandra5-rc,sdauletau/elasticsearch,humandb/elasticsearch,Widen/elasticsearch,bestwpw/elasticsearch,jprante/elasticsearch,kenshin233/elasticsearch,ImpressTV/elasticsearch,areek/elasticsearch,vingupta3/elasticsearch,18098924759/elasticsearch,elancom/elasticsearch,humandb/elasticsearch,obourgain/elasticsearch,wayeast/elasticsearch,TonyChai24/ESSource,markwalkom/elasticsearch,ESamir/elasticsearch,markllama/elasticsearch,Uiho/elasticsearch,mrorii/elasticsearch,rmuir/elasticsearch,mikemccand/elasticsearch,Helen-Zhao/elasticsearch,iacdingping/elasticsearch,AshishThakur/elasticsearch,Charlesdong/elasticsearch,Liziyao/elasticsearch,naveenhooda2000/elasticsearch,wangtuo/elasticsearch,aglne/elasticsearch,pozhidaevak/elasticsearch,ulkas/elasticsearch,lightslife/elasticsearch,sneivandt/elasticsearch,tahaemin/elasticsearch,myelin/elasticsearch,adrianbk/elasticsearch,knight1128/elasticsearch,trangvh/elasticsearch,myelin/elasticsearch,franklanganke/elasticsearch,xingguang2013/elasticsearch,ThiagoGarciaAlves/elasticsearch,ZTE-PaaS/elasticsearch,sdauletau/elasticsearch,zkidkid/elasticsearch,dongjoon-hyun/elasticsearch,gfyoung/elasticsearch,davidvgalbraith/elasticsearch,dataduke/elasticsearch,tkssharma/elasticsearch,sposam/elasticsearch,a2lin/elasticsearch,achow/elasticsearch,pozhidaevak/elasticsearch,yongminxia/elasticsearch,uschindler/elasticsearch,khiraiwa/elasticsearch,likaiwalkman/elasticsearch,tahaemin/elasticsearch,alexbrasetvik/elasticsearch,koxa29/elasticsearch,vingupta3/elasticsearch,jango2015/elasticsearch,pablocastro/elasticsearch,jpountz/elasticsearch,baishuo/elasticsearch_v2.1.0-baishuo,Brijeshrpatel9/elasticsearch,C-Bish/elasticsearch,coding0011/elasticsearch,JervyShi/elasticsearch,Ansh90/elasticsearch,trangvh/elasticsearch,anti-social/elasticsearch,trangvh/elasticsearch,djschny/elasticsearch,Fsero/elasticsearch,humandb/elasticsearch,nezirus/elasticsearch,elancom/elasticsearch,weipinghe/elasticsearch,Rygbee/elasticsearch,Shepard1212/elasticsearch,pranavraman/elasticsearch,karthikjaps/elasticsearch,mmaracic/elasticsearch,weipinghe/elasticsearch,LeoYao/elasticsearch,easonC/elasticsearch,andrejserafim/elasticsearch,TonyChai24/ESSource,hanswang/elasticsearch,lks21c/elasticsearch,YosuaMichael/elasticsearch,MisterAndersen/elasticsearch,18098924759/elasticsearch,MaineC/elasticsearch,jimczi/elasticsearch,queirozfcom/elasticsearch,strapdata/elassandra,Asimov4/elasticsearch,kcompher/elasticsearch,hirdesh2008/elasticsearch,brandonkearby/elasticsearch,chrismwendt/elasticsearch,kevinkluge/elasticsearch,abibell/elasticsearch,aglne/elasticsearch,mmaracic/elasticsearch,rlugojr/elasticsearch,kingaj/elasticsearch,kingaj/elasticsearch,loconsolutions/elasticsearch,mbrukman/elasticsearch,vrkansagara/elasticsearch,baishuo/elasticsearch_v2.1.0-baishuo,schonfeld/elasticsearch,coding0011/elasticsearch,Charlesdong/elasticsearch,karthikjaps/elasticsearch,yongminxia/elasticsearch,Shekharrajak/elasticsearch,phani546/elasticsearch,milodky/elasticsearch,jeteve/elasticsearch,fernandozhu/elasticsearch,Kakakakakku/elasticsearch,xuzha/elasticsearch,queirozfcom/elasticsearch,slavau/elasticsearch,baishuo/elasticsearch_v2.1.0-baishuo,nrkkalyan/elasticsearch,SergVro/elasticsearch,wbowling/elasticsearch,markharwood/elasticsearch,tsohil/elasticsearch,Uiho/elasticsearch,lks21c/elasticsearch,MetSystem/elasticsearch,HarishAtGitHub/elasticsearch,i-am-Nathan/elasticsearch,ThalaivaStars/OrgRepo1,nellicus/elasticsearch,Shekharrajak/elasticsearch,IanvsPoplicola/elasticsearch,ThiagoGarciaAlves/elasticsearch,kubum/elasticsearch,vroyer/elassandra,mm0/elasticsearch,Charlesdong/elasticsearch,sdauletau/elasticsearch,kunallimaye/elasticsearch,Fsero/elasticsearch,abibell/elasticsearch,elasticdog/elasticsearch,koxa29/elasticsearch,elasticdog/elasticsearch,spiegela/elasticsearch,kenshin233/elasticsearch,nellicus/elasticsearch,btiernay/elasticsearch,caengcjd/elasticsearch,baishuo/elasticsearch_v2.1.0-baishuo,areek/elasticsearch,sdauletau/elasticsearch,anti-social/elasticsearch,alexbrasetvik/elasticsearch,heng4fun/elasticsearch,heng4fun/elasticsearch,skearns64/elasticsearch,tahaemin/elasticsearch,smflorentino/elasticsearch,PhaedrusTheGreek/elasticsearch,jimczi/elasticsearch,masterweb121/elasticsearch,mikemccand/elasticsearch,lydonchandra/elasticsearch,kalimatas/elasticsearch,sarwarbhuiyan/elasticsearch,LeoYao/elasticsearch,anti-social/elasticsearch,jchampion/elasticsearch,awislowski/elasticsearch,springning/elasticsearch,glefloch/elasticsearch,aglne/elasticsearch,humandb/elasticsearch,Shekharrajak/elasticsearch,fooljohnny/elasticsearch,ESamir/elasticsearch,amit-shar/elasticsearch,diendt/elasticsearch,wittyameta/elasticsearch,heng4fun/elasticsearch,rento19962/elasticsearch,wenpos/elasticsearch,hafkensite/elasticsearch,NBSW/elasticsearch,cnfire/elasticsearch-1,lightslife/elasticsearch,liweinan0423/elasticsearch,ydsakyclguozi/elasticsearch,Flipkart/elasticsearch,JervyShi/elasticsearch,naveenhooda2000/elasticsearch,nilabhsagar/elasticsearch,pritishppai/elasticsearch,elasticdog/elasticsearch,JSCooke/elasticsearch,myelin/elasticsearch,humandb/elasticsearch
b8fb429c0b551c1a34b0fd294dac90a297d774be
AUTHORS.adoc
AUTHORS.adoc
= Authors and contributors - Simon Cruanes (`companion_cube`) - Drup (Gabriel Radanne) - Jacques-Pascal Deplaix - Nicolas Braud-Santoni - Whitequark (Peter Zotov) - hcarty (Hezekiah M. Carty) - struktured (Carmelo Piccione) - Bernardo da Costa - Vincent Bernardoff (vbmithr) - Emmanuel Surleau (emm) - Guillaume Bury (guigui) - JP Rodi - Florian Angeletti (@octachron) - Johannes Kloos - Geoff Gole (@gsg) - Roma Sokolov (@little-arhat) - Malcolm Matalka (`orbitz`) - David Sheets (@dsheets) - Glenn Slotte (glennsl) - @LemonBoy - Leonid Rozenberg (@rleonid) - Bikal Gurung (@bikalgurung) - Fabian Hemmer (copy) - Maciej Woś (@lostman) - Orbifx (Stavros Polymenis) - Rand (@rand00) - Dave Aitken (@actionshrimp) - Etienne Millon (@emillon) - Christopher Zimmermann (@madroach) - Jules Aguillon (@julow)
= Authors and contributors - Simon Cruanes (`companion_cube`) - Drup (Gabriel Radanne) - Jacques-Pascal Deplaix - Nicolas Braud-Santoni - Whitequark (Peter Zotov) - hcarty (Hezekiah M. Carty) - struktured (Carmelo Piccione) - Bernardo da Costa - Vincent Bernardoff (vbmithr) - Emmanuel Surleau (emm) - Guillaume Bury (guigui) - JP Rodi - Florian Angeletti (@octachron) - Johannes Kloos - Geoff Gole (@gsg) - Roma Sokolov (@little-arhat) - Malcolm Matalka (`orbitz`) - David Sheets (@dsheets) - Glenn Slotte (glennsl) - @LemonBoy - Leonid Rozenberg (@rleonid) - Bikal Gurung (@bikalgurung) - Fabian Hemmer (copy) - Maciej Woś (@lostman) - Orbifx (Stavros Polymenis) - Rand (@rand00) - Dave Aitken (@actionshrimp) - Etienne Millon (@emillon) - Christopher Zimmermann (@madroach) - Jules Aguillon (@julow) - Metin Akat (@loxs)
Add myself to the authors file
Add myself to the authors file
AsciiDoc
bsd-2-clause
c-cube/ocaml-containers
f832929a9f279dc9d3bfa6c440179b3453f00a58
adoc/omnij-devguide.adoc
adoc/omnij-devguide.adoc
= OmniJ Developer's Guide Sean Gilligan v0.1, July 30, 2015: Early draft :numbered: :toc: :toclevels: 3 :linkattrs: Paragraph TBD. == Introduction to OmniJ This section is TBD. For now the project http://github.com/OmniLayer/OmniJ/README.adoc[README] is the best place to get started. == JSON-RPC Clients [plantuml, diagram-classes, svg] .... skinparam packageStyle Rect skinparam shadowing false hide empty members namespace com.msgilligan.bitcoin.rpc { class RPCClient RPCClient <|-- class DynamicRPCClient << Groovy >> RPCClient <|-- BitcoinClient BitcoinClient <|-- class BitcoinCLIClient << Groovy >> } namespace foundation.omni.rpc { com.msgilligan.bitcoin.rpc.BitcoinClient <|-- OmniClient OmniClient <|-- OmniExtendedClient OmniExtendedClient <|-- class OmniCLIClient << Groovy >> } ....
= OmniJ Developer's Guide Sean Gilligan v0.1, July 30, 2015: Early draft :numbered: :toc: :toclevels: 3 :linkattrs: :imagesdir: images Paragraph TBD. == Introduction to OmniJ This section is TBD. For now the project http://github.com/OmniLayer/OmniJ/README.adoc[README] is the best place to get started. == JSON-RPC Clients [plantuml, diagram-classes, svg] .... skinparam packageStyle Rect skinparam shadowing false hide empty members namespace com.msgilligan.bitcoin.rpc { class RPCClient RPCClient <|-- class DynamicRPCClient << Groovy >> RPCClient <|-- BitcoinClient BitcoinClient <|-- class BitcoinCLIClient << Groovy >> } namespace foundation.omni.rpc { com.msgilligan.bitcoin.rpc.BitcoinClient <|-- OmniClient OmniClient <|-- OmniExtendedClient OmniExtendedClient <|-- class OmniCLIClient << Groovy >> } ....
Add images directory attribute to devguide
Add images directory attribute to devguide
AsciiDoc
apache-2.0
OmniLayer/OmniJ,OmniLayer/OmniJ,OmniLayer/OmniJ
1873528b34edfa2fc400ead1cfb4d95625e18b79
docs/groundwork.adoc
docs/groundwork.adoc
= Groundwork :toc: :source-highlighter: pygments link:index.html[back to index page] == Laying the Groundwork To redeploy RecordTrac, you need support from key stakeholders _within_ government. The administrator or elected official in charge of overseeing public records request must agree to use this system, and instruct their colleagues to do so. RecordTrac assumes there is a contact for a given municipality or department within the municipality to handle public records requests. If a government agency has no process at all in place, but is interested in using the system, they could start with one ‘champion’ that is knowledgeable about who has access to different records. The champion can then route requests to the proper parties within government who may have the documents or information a requester needs. == Best Practices to Consider RecordTrac is flexible and could complement almost any governmental agency's process for fulfilling records requests. There are however, best practices a governmental agency should adopt to really leverage the power of RecordTrac. Below is an example lifted from the City of Oakland: * Track all public records requests through RecordTrac, even if you originally received it over the phone, by email, fax, or mail. * Don't reveal sensitive information in your message or upload documents that haven't been thoroughly redacted. Everything you do on the site is immediately viewable to the public. * Upload scanned copies of the [redacted] records online instead of sending the document only to the requester. This prevents you from answering the same public records request multiple times. It also provides proof you responded to the request and exactly what you provided. * Communicate with everyone through RecordTrac. Only take conversations offline if it involves confidential or sensitive information. * Review requests no later than two business days after you receive them. This ensures the person responsible for fulfilling a records request gets it in time if it needs to be re-routed to him or her.
= Groundwork :toc: :source-highlighter: pygments link:index.html[back to index page] == Laying the Groundwork To redeploy RecordTrac, you need support from key stakeholders _within_ government. The administrator or elected official in charge of overseeing public records request must agree to use this system, and instruct their colleagues to do so. RecordTrac assumes there is a contact for a given municipality or department within the municipality to handle public records requests. If a government agency has no process at all in place, but is interested in using the system, they could start with one ‘champion’ that is knowledgeable about who has access to different records. The champion can then route requests to the proper parties within government who may have the documents or information a requester needs.
Put Best practices in a separate section.
Put Best practices in a separate section.
AsciiDoc
apache-2.0
CityOfNewYork/NYCOpenRecords,CityOfNewYork/NYCOpenRecords,CityOfNewYork/NYCOpenRecords,CityOfNewYork/NYCOpenRecords,CityOfNewYork/NYCOpenRecords
fbd313be9d59492957f3cac1437a6304fb06b21d
README.adoc
README.adoc
= Qprompt == Introduction This project provides a Python 2.x library that allows the user to quickly create CLI prompts for user input. The main features are the following: - Simple multi-entry menus. - Prompt for yes/no response. - Prompt for integer response. - Prompt for float response. - Optional default value. - Optional validity check. - Should work on any platform without additional dependencies. == Status Currently, this project is **under active development**. The contents of the repository should be considered unstable during active development. == Requirements Qprompt should run on any Python 2.x interpreter without additional dependencies. == Installation Qprompt can be installed with pip using the following command: `pip install qprompt` Additional, Qprompt can be installed from source by running: `python setup.py install` == Examples Example of basic information prompting: -------- include::examples/ask_1.py[] -------- Example of menu usage: -------- include::examples/menu_1.py[] -------- == Similar The following projects are similar and may be worth checking out: - https://github.com/Sleft/cliask[cliask] - https://github.com/aventurella/promptly[Promptly] - https://github.com/tylerdave/prompter[prompter] - https://github.com/magmax/python-inquirer[python-inquirer]
= Qprompt == Introduction This project provides a Python 2.x library that allows the user to quickly create CLI prompts for user input. The main features are the following: - Simple multi-entry menus. - Prompt for yes/no response. - Prompt for integer response. - Prompt for float response. - Optional default value. - Optional validity check. - Should work on any platform without additional dependencies. == Status Currently, this project is **under active development**. The contents of the repository should be considered unstable during active development. == Requirements Qprompt should run on any Python 2.x interpreter without additional dependencies. == Installation Qprompt can be installed with pip using the following command: `pip install qprompt` Additional, Qprompt can be installed from source by running: `python setup.py install` == Examples The following are basic examples of Qprompt (all examples can be found https://github.com/jeffrimko/Qprompt/tree/master/examples[here]): - https://github.com/jeffrimko/Qprompt/blob/master/examples/ask_1.py[`examples/ask_1.py`] - Basic info prompting. - https://github.com/jeffrimko/Qprompt/blob/master/examples/menu_1.py[`examples/menu_1.py`] - Basic menu usage. == Similar The following projects are similar and may be worth checking out: - https://github.com/Sleft/cliask[cliask] - https://github.com/aventurella/promptly[Promptly] - https://github.com/tylerdave/prompter[prompter] - https://github.com/magmax/python-inquirer[python-inquirer]
Update to the example section.
Update to the example section.
AsciiDoc
mit
jeffrimko/Qprompt
e938ebfa19e8742a255592dfa11ac879700dbade
adoc/include/common_options.adoc
adoc/include/common_options.adoc
*--map-http-status* 'TEXT':: Map non success HTTP response codes to exit codes other than 1. e.g. "--map-http-satus 403=0,404=0" would exit with 0 even if a 403 or 404 http error code was received. Valid exit codes are 0,1,50-99. include::format_option.adoc[] include::jmespath_option.adoc[] include::help_option.adoc[] include::verbose_option.adoc[]
*--map-http-status* 'TEXT':: Map non success HTTP response codes to exit codes other than 1. e.g. "--map-http-satus 403=0,404=0" would exit with 0 even if a 403 or 404 http error code was received. Valid exit codes are 0,1,50-99. *-F, --format* '[json|text]':: Set the output format for stdout. Defaults to "text". *--jq, --jmespath* 'EXPR':: Supply a JMESPath expression to apply to json output. Takes precedence over any specified '--format' and forces the format to be json processed by this expression. + A full specification of the JMESPath language for querying JSON structures may be found at https://jmespath.org/ *-h, --help*:: Show help text for this command. *-v, --verbose*:: Control the level of output. + Use -v or --verbose to show warnings and any additional text output. + Use -vv to add informative logging. + Use -vvv to add debug logging and full stack on any errors. (equivalent to -v --debug)
Remove nested includes from adoc
Remove nested includes from adoc
AsciiDoc
apache-2.0
globus/globus-cli,globus/globus-cli
52b6608eace07c90c3dbe9e3a87e5510675406f1
modules/dedicated-managing-dedicated-administrators.adoc
modules/dedicated-managing-dedicated-administrators.adoc
// Module included in the following assemblies: // // administering_a_cluster/dedicated-admin-role.adoc [id="dedicated-managing-dedicated-administrators_{context}"] = Managing {product-title} administrators Administrator roles are managed using a `dedicated-admins` group on the cluster. Existing members of this group can edit membership via the link:https://cloud.redhat.com/openshift[{cloud-redhat-com}] site. [id="dedicated-administrators-adding-user_{context}"] == Adding a user . Navigate to the *Cluster Details* page and *Users* tab. . Click the *Add user* button. (first user only) . Enter the user name and select the group (*dedicated-admins*) . Click the *Add* button. [id="dedicated-administrators-removing-user_{context}"] == Removing a user . Navigate to the *Cluster Details* page and *Users* tab. . Click the *X* to the right of the user / group combination to be deleted..
// Module included in the following assemblies: // // administering_a_cluster/dedicated-admin-role.adoc [id="dedicated-managing-dedicated-administrators_{context}"] = Managing {product-title} administrators Administrator roles are managed using a `dedicated-admins` group on the cluster. Existing members of this group can edit membership via the link:https://cloud.redhat.com/openshift[{cloud-redhat-com}] site. [id="dedicated-administrators-adding-user_{context}"] == Adding a user . Navigate to the *Cluster Details* page and *Access Control* tab. . Click the *Add user* button. (first user only) . Enter the user name and select the group (*dedicated-admins*) . Click the *Add* button. [id="dedicated-administrators-removing-user_{context}"] == Removing a user . Navigate to the *Cluster Details* page and *Access Control* tab. . Click the 3 vertical dots to the right of the user / group combination to show a menu, then click on *Delete*.
Update documentation to current interface on cloud.rh.c
Update documentation to current interface on cloud.rh.c
AsciiDoc
apache-2.0
vikram-redhat/openshift-docs,vikram-redhat/openshift-docs,vikram-redhat/openshift-docs,vikram-redhat/openshift-docs,vikram-redhat/openshift-docs
c5002afeefa5636f66cc673a80f130175f931a60
CHANGELOG.asciidoc
CHANGELOG.asciidoc
2020/06/18: Concuerror integration has been added. It is currently minimal but usable. Experimentation and feedback is welcome. 2020/11/30: Support for publishing Hex releases and docs has been added. It is currently experimental. Feedback is more than welcome. 2022/03/25: The -Wrace_conditions Dialyzer flag was removed as it is no longer available starting from OTP 25. 2022/??/??: Relx has been updated to v4. Relx v4 is no longer an escript, therefore breaking changes were introduced. The `RELX`, `RELX_URL` and `RELX_OPTS` variables were removed. The `relx` project must be added as a `DEPS`, `BUILD_DEPS` or `REL_DEPS` dependency to enable building releases. For example: `REL_DEPS = relx`. Relx itself has had some additional changes: the `start` command has been replaced by `daemon`, and configuration defaults have changed so that you may need to add the following to your relx.config file: ``` erlang {dev_mode, false}. {include_erts, true}. ```
2020/06/18: Concuerror integration has been added. It is currently minimal but usable. Experimentation and feedback is welcome. 2020/11/30: Support for publishing Hex releases and docs has been added. It is currently experimental. Feedback is more than welcome. 2022/03/25: The -Wrace_conditions Dialyzer flag was removed as it is no longer available starting from OTP 25. 2022/05/20: Relx has been updated to v4. Relx v4 is no longer an escript, therefore breaking changes were introduced. The `RELX`, `RELX_URL` and `RELX_OPTS` variables were removed. The `relx` project must be added as a `DEPS`, `BUILD_DEPS` or `REL_DEPS` dependency to enable building releases. For example: `REL_DEPS = relx`. Relx itself has had some additional changes: the `start` command has been replaced by `daemon`, and configuration defaults have changed so that you may need to add the following to your relx.config file: ``` erlang {dev_mode, false}. {include_erts, true}. ```
Set date for breaking Relx 4 change
Set date for breaking Relx 4 change
AsciiDoc
isc
ninenines/erlang.mk,rabbitmq/erlang.mk
436b3063271078581d376774c2a6d534742dd777
doc/cookbook/zone.adoc
doc/cookbook/zone.adoc
== Time Zones & Offset Extract a zone from a `java.time.ZonedDateTime`: ==== [source.code,clojure] ---- (t/zone (t/zoned-date-time "2000-01-01T00:00:00Z[Europe/Paris]")) ---- [source.code,clojure] ---- (t/zone) ---- ==== Create a `java.time.ZonedDateTime` in a particular time zone: ==== [source.code,clojure] ---- (t/in (t/instant "2000-01-01T00:00") "Australia/Darwin") ---- ==== === TBD : offsets
== Time Zones & Offset Extract a zone from a `java.time.ZonedDateTime`: ==== [source.code,clojure] ---- (t/zone (t/zoned-date-time "2000-01-01T00:00:00Z[Europe/Paris]")) ---- [source.code,clojure] ---- (t/zone) ---- ==== Create a `java.time.ZonedDateTime` in a particular time zone: ==== [source.code,clojure] ---- (t/in (t/instant "2000-01-01T00:00") "Australia/Darwin") ---- ==== Give the `OffsetDateTime` instead of `ZonedDateTime`: ==== [source.code,clojure] ---- (t/offset-date-time (t/zoned-date-time "2000-01-01T00:00:00Z[Australia/Darwin]")) ---- ==== Specify the offset for a `LocalDateTime`: ==== [source.code,clojure] ---- (t/offset-by (t/date-time "2018-01-01T00:00") 9) ---- ====
Add offset examples to Zones
Add offset examples to Zones
AsciiDoc
mit
juxt/tick,juxt/tick
7ce834e70d12fb870f79af99daaf22e9f5c6aee8
src/main/docs/index.adoc
src/main/docs/index.adoc
= geo-shell Jared Erickson v0.7-SNAPSHOT ifndef::imagesdir[:imagesdir: images] include::intro.adoc[] include::workspace.adoc[] include::layer.adoc[] include::format.adoc[] include::raster.adoc[] include::tile.adoc[] include::style.adoc[] include::map.adoc[] include::builtin.adoc[]
= Geo Shell Jared Erickson v0.7-SNAPSHOT :title-logo-image: image:geoshell.png[pdfwidth=5.5in,align=center] ifndef::imagesdir[:imagesdir: images] include::intro.adoc[] include::workspace.adoc[] include::layer.adoc[] include::format.adoc[] include::raster.adoc[] include::tile.adoc[] include::style.adoc[] include::map.adoc[] include::builtin.adoc[]
Add title image to pdf
Add title image to pdf
AsciiDoc
mit
jericks/geo-shell,jericks/geo-shell,jericks/geo-shell
6b23a4a24353bad169fa4340a30d8a357a512fce
impl/src/docs/asciidoc/index.adoc
impl/src/docs/asciidoc/index.adoc
:generated: ../../../target/generated-docs/asciidoc include::{generated}/overview.adoc[] include::manual_rest_doc.adoc[] include::{generated}/paths.adoc[]
:generated: ../../../target/generated-docs/asciidoc include::{generated}/overview.adoc[] include::manual_rest_doc.adoc[] include::{generated}/paths.adoc[] include::{generated}/definitions.adoc[]
Add generated data type definitions to Swagger documentation
Add generated data type definitions to Swagger documentation
AsciiDoc
mit
jugda/dukecon_server,dukecon/dukecon_server,jugda/dukecon_server,dukecon/dukecon_server,dukecon/dukecon_server,jugda/dukecon_server
b582f7574430e1946ffaac7ac7365c48fc3ac1b4
README.adoc
README.adoc
= Spring Boot and Two DataSources This project demonstrates how to use two `DataSource` s with Spring Boot 2.1. It utilizes: * Spring Data https://github.com/spring-projects/spring-data-jpa[JPA] * https://github.com/flyway/flyway[Flyway] migrations for the two `DataSource` s * Separate Hibernate properties for each `DataSource` defined in the application.yml * Tests for components Note: It may take a few seconds for the app to start if no one has not accessed it recently
= Spring Boot and Two DataSources This project demonstrates how to use two `DataSource` s with Spring Boot 2.1. It utilizes: * Spring Data https://github.com/spring-projects/spring-data-jpa[JPA] * https://github.com/flyway/flyway[Flyway] migrations for the two `DataSource` s * Separate Hibernate properties for each `DataSource` defined in the application.yml * Tests for components
Remove note about app starting up
Remove note about app starting up It's no longer applicable to the branch that is purely backend components without any frontend.
AsciiDoc
unlicense
drumonii/SpringBootTwoDataSources
b978bf8d377f09c132050297e45b3c8fbb652a4a
src/main/asciidoc/development.adoc
src/main/asciidoc/development.adoc
[[development]] == Development Github repository: {datasource-proxy} === Build Documentation ```sh > ./mvnw asciidoctor:process-asciidoc@output-html ```
[[development]] == Development Github repository: {datasource-proxy} === Build Documentation Generate `index.html` ```sh > ./mvnw asciidoctor:process-asciidoc@output-html ``` Http preview ```sh > ./mvnw asciidoctor:http@output-html ```
Add how to use asciidoctor plugin in dev
Add how to use asciidoctor plugin in dev
AsciiDoc
mit
ttddyy/datasource-proxy,ttddyy/datasource-proxy
2e835acb457e13f3fc5a49459604488388d80cae
README.adoc
README.adoc
= Infinispan Cluster Manager image:https://vertx.ci.cloudbees.com/buildStatus/icon?job=vert.x3-infinispan["Build Status",link="https://vertx.ci.cloudbees.com/view/vert.x-3/job/vert.x3-infinispan/"] This is a cluster manager implementation for Vert.x that uses http://infinispan.org[Infinispan]. Please see the in-source asciidoc documentation or the main documentation on the web-site for a full description of this component: * link:http://vertx.io/docs/vertx-infinispan/java/[web-site docs] * link:src/main/asciidoc/java/index.adoc[in-source docs] -- will remove --
= Infinispan Cluster Manager image:https://vertx.ci.cloudbees.com/buildStatus/icon?job=vert.x3-infinispan["Build Status",link="https://vertx.ci.cloudbees.com/view/vert.x-3/job/vert.x3-infinispan/"] This is a cluster manager implementation for Vert.x that uses http://infinispan.org[Infinispan]. Please see the in-source asciidoc documentation or the main documentation on the web-site for a full description of this component: * link:http://vertx.io/docs/vertx-infinispan/java/[web-site docs] * link:src/main/asciidoc/java/index.adoc[in-source docs]
Revert "Revert "Revert "Test trigger on push"""
Revert "Revert "Revert "Test trigger on push""" This reverts commit f235543226884c6293a70b2540441e5b1528ff6c.
AsciiDoc
apache-2.0
vert-x3/vertx-infinispan
242825445a9b0adf2b5ef2b3de4d8c612ade0c22
docs/src/docs/asciidoc/multiTenancy/tenantTransforms.adoc
docs/src/docs/asciidoc/multiTenancy/tenantTransforms.adoc
The next transformations can be applied to any class to simplify greatly the development of Multi-Tenant applications. These include: - `@CurrentTenant` - Resolve the current tenant for the context of a class or method - `@Tenant` - Use a specific tenant for the context of a class or method - `@WithoutTenant` - Execute logic without a specific tenant (using the default connection) For example: [source,groovy] ---- import grails.gorm.multitenancy.* // resolve the current tenant for every method @CurrentTenant class TeamService { // execute the countPlayers method without a tenant id @WithoutTenant int countPlayers() { Player.count() } // use the tenant id "another" for all GORM logic within the method @Tenant({"another"}) List<Team> allTwoTeams() { Team.list() } List<Team> listTeams() { Team.list(max:10) } @Transactional void addTeam(String name) { new Team(name:name).save(flush:true) } } ----
The following transformations can be applied to any class to simplify greatly the development of Multi-Tenant applications. These include: - `@CurrentTenant` - Resolve the current tenant for the context of a class or method - `@Tenant` - Use a specific tenant for the context of a class or method - `@WithoutTenant` - Execute logic without a specific tenant (using the default connection) For example: [source,groovy] ---- import grails.gorm.multitenancy.* // resolve the current tenant for every method @CurrentTenant class TeamService { // execute the countPlayers method without a tenant id @WithoutTenant int countPlayers() { Player.count() } // use the tenant id "another" for all GORM logic within the method @Tenant({"another"}) List<Team> allTwoTeams() { Team.list() } List<Team> listTeams() { Team.list(max:10) } @Transactional void addTeam(String name) { new Team(name:name).save(flush:true) } } ----
Replace "the next" with "the following"
Replace "the next" with "the following"
AsciiDoc
apache-2.0
grails/gorm-hibernate5
a78ed8d4a2590eb9759cf1283971056af637d94b
libbeat/docs/communitybeats.asciidoc
libbeat/docs/communitybeats.asciidoc
[[community-beats]] == Community Beats The open source community has been hard at work developing new Beats. You can check out a few of them here: [horizontal] https://github.com/Ingensi/dockerbeat[dockerbeat]:: Reads docker container statistics and indexes them in Elasticsearch https://github.com/christiangalsterer/httpbeat[httpbeat]:: Polls multiple HTTP(S) endpoints and sends the data to Logstash, Elasticsearch. Supports all HTTP methods and proxies. https://github.com/mrkschan/nginxbeat[nginxbeat]:: Reads status from Nginx https://github.com/joshuar/pingbeat[pingbeat]:: Sends ICMP pings to a list of targets and stores the round trip time (RTT) in Elasticsearch https://github.com/mrkschan/uwsgibeat[uwsgibeat]:: Reads stats from uWSGI https://github.com/kozlice/phpfpmbeat[phpfpmbeat]:: Reads status from PHP-FPM Have you created a Beat that's not listed? Open a pull request to add your link here: https://github.com/elastic/libbeat/blob/master/docs/communitybeats.asciidoc NOTE: Elastic provides no warranty or support for community-sourced Beats. [[contributing-beats]] === Contributing to Beats Remember, you can be a Beats developer, too. <<new-beat, Learn how>>
[[community-beats]] == Community Beats The open source community has been hard at work developing new Beats. You can check out a few of them here: [horizontal] https://github.com/Ingensi/dockerbeat[dockerbeat]:: Reads docker container statistics and indexes them in Elasticsearch https://github.com/christiangalsterer/httpbeat[httpbeat]:: Polls multiple HTTP(S) endpoints and sends the data to Logstash, Elasticsearch. Supports all HTTP methods and proxies. https://github.com/mrkschan/nginxbeat[nginxbeat]:: Reads status from Nginx https://github.com/joshuar/pingbeat[pingbeat]:: Sends ICMP pings to a list of targets and stores the round trip time (RTT) in Elasticsearch https://github.com/mrkschan/uwsgibeat[uwsgibeat]:: Reads stats from uWSGI https://github.com/kozlice/phpfpmbeat[phpfpmbeat]:: Reads status from PHP-FPM https://github.com/radoondas/apachebeat[apachebeat]:: Reads status from Apache HTTPD server-status Have you created a Beat that's not listed? Open a pull request to add your link here: https://github.com/elastic/libbeat/blob/master/docs/communitybeats.asciidoc NOTE: Elastic provides no warranty or support for community-sourced Beats. [[contributing-beats]] === Contributing to Beats Remember, you can be a Beats developer, too. <<new-beat, Learn how>>
Add apachebeat to the list of beats from opensource
Add apachebeat to the list of beats from opensource
AsciiDoc
mit
yapdns/yapdnsbeat,yapdns/yapdnsbeat
47212b4db9302d1b8f194bc8c43247772c54afbd
doc/src/main/asciidoc/rest-spec.adoc
doc/src/main/asciidoc/rest-spec.adoc
= RESTful API Endpoint specification == Nodes === Idea for accessing fields directly * RUD: /nodes/:uuid/relatedProducts/:uuid -> Pageable list of nodes * R: /nodes/:uuid/name TODO: Do we want to restrict the primitiv types to read only? The user can update the node via PUT /nodes/:uuid anyway. == Webroot == Tags TODO: Define how tag familes are setup and tags are assigned to those families. == Users / Groups / Roles == Projects TODO: Define how languages are assigned to projects
= RESTful API Endpoint specification == Nodes === Idea for accessing fields directly * RUD: /nodes/:uuid/relatedProducts/:uuid -> Pageable list of nodes * R: /nodes/:uuid/name TODO: Do we want to restrict the primitiv types to read only? The user can update the node via PUT /nodes/:uuid anyway. == Breadcrumbs `/breadcrumb/:uuid` -> object containing breadcrumb to the node specified by uuid in the following format: [source,json] ---- [ { "uuid": "e0c64dsgasdgasdgdgasdgasdgasdg33", "name": "products" }, { "uuid": "e0c64ad00a9343cc864ad00a9373cc23", "name": "aeroplane.en.html" } ] ---- TODO: Where does the "name" property come from? It should be the field specified by "segmentField" in the schema. Two issues: 1. Should we normalize the property name to "name", or retain the name of the field specified by "segmentField"? 2. If the schema of a node in the breadcrumb path does not specify "segmentField", what do we do? Just display the uuid? == Webroot == Tags TODO: Define how tag familes are setup and tags are assigned to those families. == Users / Groups / Roles == Projects TODO: Define how languages are assigned to projects
Add section on breadcrumbs endpoint
Add section on breadcrumbs endpoint
AsciiDoc
apache-2.0
gentics/mesh,gentics/mesh,gentics/mesh,gentics/mesh
ccb2545c35c7a88014f2c0dd06a1582b029ab298
conoha/dokku-apps/README.adoc
conoha/dokku-apps/README.adoc
= conoha/dokku-apps .Add pytohn-getting-started app ---- alias dokku="ssh -t dokku@conoha" cd python-getting-started dokku apps:create python-getting-started git remote add dokku dokku@conoha:python-getting-started git push dokku master ---- And this app can be available at http://python-getting-started.d.10sr.f5.si .
= conoha/dokku-apps First you have to run: ---- cat .ssh/id_rsa.pub | ssh conoha 'sudo sshcommand acl-add dokku dokkudeploy' ---- .Add pytohn-getting-started app ---- alias dokku="ssh -t dokku@conoha" cd python-getting-started dokku apps:create python-getting-started git remote add dokku dokku@conoha:python-getting-started git push dokku master ---- And this app can be available at http://python-getting-started.d.10sr.f5.si .
Add note to add keys for dokku
Add note to add keys for dokku
AsciiDoc
unlicense
10sr/server-provisions,10sr/machine-setups,10sr/machine-setups,10sr/machine-setups,10sr/server-provisions,10sr/machine-setups
d61a7352efcf33d8bcb1bf1fd9052480e2ee15ba
code/continuousIntegration.adoc
code/continuousIntegration.adoc
= Continuous integration :awestruct-description: Check if the latest nightly build passes all automated tests. :awestruct-layout: normalBase :showtitle: == OptaPlanner We use Jenkins for continuous integration. *Show https://hudson.jboss.org/hudson/job/optaplanner/[the public Jenkins job].* This is a mirror of a Red Hat internal Jenkins job. Keep the build blue! == Project website (optaplanner.org) We use Travis to build this project website, see https://travis-ci.org/droolsjbpm/optaplanner-website[the travis job].
= Continuous integration :awestruct-description: Check if the latest nightly build passes all automated tests. :awestruct-layout: normalBase :showtitle: == OptaPlanner We use Jenkins for continuous integration. *Show https://jenkins-kieci.rhcloud.com/job/optaplanner/[the public Jenkins job].* This is a mirror of a Red Hat internal Jenkins job. Keep the build green! == Project website (optaplanner.org) We use Travis to build this project website, see https://travis-ci.org/droolsjbpm/optaplanner-website[the travis job].
Update public Jenkins job URL
Update public Jenkins job URL
AsciiDoc
apache-2.0
bibryam/optaplanner-website,psiroky/optaplanner-website,oskopek/optaplanner-website,droolsjbpm/optaplanner-website,bibryam/optaplanner-website,oskopek/optaplanner-website,psiroky/optaplanner-website,bibryam/optaplanner-website,psiroky/optaplanner-website,oskopek/optaplanner-website,droolsjbpm/optaplanner-website,droolsjbpm/optaplanner-website
c9fd39c5785d21b75d32348f8643409668e985bd
modules/nw-dns-operator-logs.adoc
modules/nw-dns-operator-logs.adoc
// Module included in the following assemblies: // // * dns/dns-operator.adoc [id="nw-dns-operator-logs_{context}"] = DNS Operator logs You can view DNS Operator logs by using the `oc logs` command. .Procedure View the logs of the DNS Operator: ---- $ oc logs --namespace=openshift-dns-operator deployment/dns-operator ----
// Module included in the following assemblies: // // * dns/dns-operator.adoc [id="nw-dns-operator-logs_{context}"] = DNS Operator logs You can view DNS Operator logs by using the `oc logs` command. .Procedure View the logs of the DNS Operator: ---- $ oc logs -n openshift-dns-operator deployment/dns-operator -c dns-operator ----
Fix command to get DNS logs
Fix command to get DNS logs - https://bugzilla.redhat.com/show_bug.cgi?id=1834702
AsciiDoc
apache-2.0
vikram-redhat/openshift-docs,vikram-redhat/openshift-docs,vikram-redhat/openshift-docs,vikram-redhat/openshift-docs,vikram-redhat/openshift-docs
04d5d44e744c96aa641b534a22a546ae7262579e
src/docs/manual/03_task_exportEA.adoc
src/docs/manual/03_task_exportEA.adoc
:filename: manual/03_task_exportEA.adoc ifndef::imagesdir[:imagesdir: ../images] = exportEA IMPORTANT: Currently this feature is WINDOWS-only. https://github.com/docToolchain/docToolchain/issues/231[See related issue] include::feedback.adoc[] image::ea/Manual/exportEA.png[] TIP: Blog-Posts: https://rdmueller.github.io/jria2eac/[JIRA to Sparx EA], https://rdmueller.github.io/sparx-ea/[Did you ever wish you had better Diagrams?] == Source .build.gradle [source,groovy] ---- include::../../../scripts/exportEA.gradle[tags=exportEA] ---- .scripts/exportEAP.vbs [source] ---- include::../../../scripts/exportEAP.vbs[] ----
:filename: manual/03_task_exportEA.adoc ifndef::imagesdir[:imagesdir: ../images] = exportEA IMPORTANT: Currently this feature is WINDOWS-only. https://github.com/docToolchain/docToolchain/issues/231[See related issue] include::feedback.adoc[] image::ea/Manual/exportEA.png[] TIP: Blog-Posts: https://rdmueller.github.io/jria2eac/[JIRA to Sparx EA], https://rdmueller.github.io/sparx-ea/[Did you ever wish you had better Diagrams?] == Configuration By default no special configuration is necessary. But, to be more specific on the project and its packages to be used for export, two optional parameter configurations are available. The parameters can be used independently from each other. A sample how to edit your projects Config.groovy is given in the 'Config.groovy' of the docToolchain project itself. connection:: Set the connection to a certain project or comment it out to use all project files inside the src folder or its child folder. packageFilter:: Add one or multiple packageGUIDs to be used for export. All packages are analysed, if no packageFilter is set. == Source .build.gradle [source,groovy] ---- include::../../../scripts/exportEA.gradle[tags=exportEA] ---- .scripts/exportEAP.vbs [source] ---- include::../../../scripts/exportEAP.vbs[] ----
Add documentation for the parameters offered by the exportEA configuration.
Add documentation for the parameters offered by the exportEA configuration.
AsciiDoc
mit
docToolchain/docToolchain,docToolchain/docToolchain,docToolchain/docToolchain,docToolchain/docToolchain
573e2c24a90c6422ec873124f93fc1418c1ea00e
python/template/docs/problem.adoc
python/template/docs/problem.adoc
:doctitle: :author: Jerod Gawne :email: jerodgawne@gmail.com :docdate: June 07, 2018 :revdate: {docdatetime} :src-uri: https://github.com/jerodg/hackerrank :difficulty: :time-complexity: :required-knowledge: :advanced-knowledge: :solution-variability: :score: :keywords: python, {required-knowledge}, {advanced-knowledge} :summary: :doctype: article :sectanchors: :sectlinks: :sectnums: :toc: {summary} == Learning == Tutorial == Improving the Template === Convention .Missing * shebang * encoding * doc-comments === Extraneous N/A === Pep8 * No new-line at end of file === Syntax N/A == Reference
:doctitle: :author: Jerod Gawne :email: jerodgawne@gmail.com :docdate: June 07, 2018 :revdate: {docdatetime} :src-uri: https://github.com/jerodg/hackerrank :difficulty: :time-complexity: :required-knowledge: :advanced-knowledge: :solution-variability: :score: :keywords: python, {required-knowledge}, {advanced-knowledge} :summary: :doctype: article :sectanchors: :sectlinks: :sectnums: :toc: {summary} == Learning == Tutorial == Improving the Problem == Improving the Template === Convention .Missing * shebang * encoding * doc-comments === Extraneous N/A === Pep8 * No new-line at end of file === Syntax N/A == Reference
Add Improving the Problem section.
Add Improving the Problem section.
AsciiDoc
mit
jerodg/hackerrank-python
5ba5e9426c278aba34de90e183f8e7e8a74556e5
examples/camel-example-spring-boot-supervising-route-controller/readme.adoc
examples/camel-example-spring-boot-supervising-route-controller/readme.adoc
# Camel Supervising Route Controller Example Spring Boot This example shows how to work with a simple Apache Camel application using Spring Boot and a Supervising Route Controller. ## How to run You can run this example using mvn spring-boot:run Beside JMX you can use Spring Boot Endpoints to interact with the routes: * To get info about the routes + [source] ---- curl -XGET -s http://localhost:8080/actuator/camelroutes ---- + +* To get details about a route ++ +[source] +---- +curl -XGET -s http://localhost:8080/actuator/camelroutes/{id}/detail +---- * To get info about a route + [source] ---- curl -XGET -s http://localhost:8080/actuator/camelroutes/{id}/info ---- * To stop a route + [source] ---- curl -XPOST -H "Content-Type: application/json" -s http://localhost:8080/actuator/camelroutes/{id}/stop ---- * To start a route + [source] ---- curl -XPOST -H "Content-Type: application/json" -s http://localhost:8080/actuator/camelroutes/{id}/start ---- ## More information You can find more information about Apache Camel at the website: http://camel.apache.org/
# Camel Supervising Route Controller Example Spring Boot This example shows how to work with a simple Apache Camel application using Spring Boot and a Supervising Route Controller. ## How to run You can run this example using mvn spring-boot:run Beside JMX you can use Spring Boot Endpoints to interact with the routes: * To get info about the routes + [source] ---- curl -XGET -s http://localhost:8080/actuator/camelroutes ---- * To get details about a route + [source] ---- curl -XGET -s http://localhost:8080/actuator/camelroutes/{id}/detail ---- * To get info about a route + [source] ---- curl -XGET -s http://localhost:8080/actuator/camelroutes/{id}/info ---- * To stop a route + [source] ---- curl -XPOST -H "Content-Type: application/json" -s http://localhost:8080/actuator/camelroutes/{id}/stop ---- * To start a route + [source] ---- curl -XPOST -H "Content-Type: application/json" -s http://localhost:8080/actuator/camelroutes/{id}/start ---- ## More information You can find more information about Apache Camel at the website: http://camel.apache.org/
Fix the example document error
Fix the example document error Fixed the document error of camel-example-spring-boot-supervising-route-controller
AsciiDoc
apache-2.0
gnodet/camel,pax95/camel,mcollovati/camel,alvinkwekel/camel,apache/camel,tdiesler/camel,Fabryprog/camel,Fabryprog/camel,adessaigne/camel,gnodet/camel,cunningt/camel,tadayosi/camel,tdiesler/camel,DariusX/camel,pax95/camel,pmoerenhout/camel,pmoerenhout/camel,mcollovati/camel,zregvart/camel,pax95/camel,nikhilvibhav/camel,adessaigne/camel,pmoerenhout/camel,adessaigne/camel,objectiser/camel,DariusX/camel,punkhorn/camel-upstream,CodeSmell/camel,objectiser/camel,nicolaferraro/camel,nikhilvibhav/camel,tdiesler/camel,apache/camel,tadayosi/camel,christophd/camel,pmoerenhout/camel,apache/camel,alvinkwekel/camel,adessaigne/camel,apache/camel,gnodet/camel,cunningt/camel,adessaigne/camel,cunningt/camel,cunningt/camel,alvinkwekel/camel,pmoerenhout/camel,tadayosi/camel,punkhorn/camel-upstream,pax95/camel,ullgren/camel,christophd/camel,zregvart/camel,mcollovati/camel,tdiesler/camel,objectiser/camel,ullgren/camel,tdiesler/camel,DariusX/camel,nicolaferraro/camel,CodeSmell/camel,davidkarlsen/camel,gnodet/camel,nicolaferraro/camel,apache/camel,davidkarlsen/camel,pmoerenhout/camel,gnodet/camel,Fabryprog/camel,christophd/camel,ullgren/camel,objectiser/camel,pax95/camel,CodeSmell/camel,nicolaferraro/camel,punkhorn/camel-upstream,cunningt/camel,alvinkwekel/camel,nikhilvibhav/camel,pax95/camel,davidkarlsen/camel,punkhorn/camel-upstream,tadayosi/camel,davidkarlsen/camel,tadayosi/camel,zregvart/camel,apache/camel,christophd/camel,nikhilvibhav/camel,mcollovati/camel,Fabryprog/camel,ullgren/camel,adessaigne/camel,zregvart/camel,DariusX/camel,tadayosi/camel,christophd/camel,CodeSmell/camel,christophd/camel,cunningt/camel,tdiesler/camel
324d6806d57251e53f54e9baf541aa06a1a372a5
README.adoc
README.adoc
= The Ehcache 3.x line is currently the development line. Status of the build: image:https://ehcache.ci.cloudbees.com/buildStatus/icon?job=ehcache3 For more information, you might want to go check the https://github.com/ehcache/ehcache3/wiki[wiki]. image:http://cloudbees.prod.acquia-sites.com/sites/default/files/styles/large/public/Button-Built-on-CB-1.png?itok=3Tnkun-C
= The Ehcache 3.x line is currently the development line. Status of the build: image:https://ehcache.ci.cloudbees.com/buildStatus/icon?job=ehcache3[Ehcache@Cloudbees, link="https://ehcache.ci.cloudbees.com/job/ehcache3/"] For more information, you might want to go check the https://github.com/ehcache/ehcache3/wiki[wiki]. image:http://cloudbees.prod.acquia-sites.com/sites/default/files/styles/large/public/Button-Powered-by-CB.png?itok=uMDWINfY[Cloudbees, link="http://www.cloudbees.com/resources/foss"]
Fix image tags in updated readme
Fix image tags in updated readme
AsciiDoc
apache-2.0
rkavanap/ehcache3,palmanojkumar/ehcache3,mingyaaaa/ehcache3,lorban/ehcache3,sreekanth-r/ehcache3,aurbroszniowski/ehcache3,lorban/ehcache3,kedar031/ehcache3,GaryWKeim/ehcache3,wantstudy/ehcache3,jhouserizer/ehcache3,AbfrmBlr/ehcache3,GaryWKeim/ehcache3,rishabhmonga/ehcache3,ehcache/ehcache3,ljacomet/ehcache3,albinsuresh/ehcache3,292388900/ehcache3,henri-tremblay/ehcache3,ehcache/ehcache3,albinsuresh/ehcache3,AbfrmBlr/ehcache3,aurbroszniowski/ehcache3,chrisdennis/ehcache3,ljacomet/ehcache3,CapeSepias/ehcache3,akomakom/ehcache3,jhouserizer/ehcache3,cljohnso/ehcache3,cljohnso/ehcache3,rkavanap/ehcache3,chenrui2014/ehcache3,chrisdennis/ehcache3,cschanck/ehcache3,alexsnaps/ehcache3,anthonydahanne/ehcache3,cschanck/ehcache3
b1b639d60403f3fc22318bbc013bfaca8acd470f
README.adoc
README.adoc
= clublist - Club Membership List Track members for a small non-profit club. This shows off some basic functionality of JPA and DeltaSpike Data in a JSF environment. == Deployment . Copy config-sample.properties to config.properties, change the name in the orgName property in this file from 'Sample Club' to your organization's name (keep it short). . Setup the datasource in your app server . Setup the deployment as needed (e.g., edit jboss-deployment.xml) . Create a user in the role of club_exec in your app server. . Deploy (e.g., mvn wildfly:deploy if you use JBoss WildFly). . Enjoy! == ToDo Security: Maybe allow members to update their own record (only!) Use redirect after editing to really go back to the List page. "Position" should be a relationship to another Entity, with a dropdown chooser. "Membership Type" should be a relationship to another Entity, with a dropdown. Search with 'Like' method in DS Data A confirmation (p:dialog?) on the Edit->Delete button would be a good idea. Implement the Mailing List page. Implement the Print Member Badge/Label page. Even though people should not use spreadsheets for database work, you will probably be pressured to impelement the "Export" capability. You will need Apache POI for this. Refactor Home object to merge w/ darwinsys-ee EntityHome
= clublist - Club Membership List Track members for a small non-profit club. This shows off some basic functionality of JPA and DeltaSpike Data in a JSF environment. == Deployment . Copy config-sample.properties to config.properties, change the name in the orgName property in this file from 'Sample Club' to your organization's name (keep it short). . Setup the datasource in your app server . Setup the deployment as needed (e.g., edit jboss-deployment.xml) . Create a user in the role of club_exec in your app server. . Deploy (e.g., mvn wildfly:deploy if you use JBoss WildFly). . Enjoy! == ToDo Here are some things that should be added. https://github.com/IanDarwin/clublist[Fork this project on GitHub] and send pull requests when you get one working! . Search with 'Like' method in DS Data . Use redirect after editing to really go back to the List page. . Maybe allow members to update their own record (only!) Probably requires moving to app-managed security since you already have a record for each person. . "Position" should be a relationship to another Entity, with a dropdown chooser. . "Membership Type" should be a relationship to another Entity, with a dropdown. . A confirmation (p:dialog?) on the Edit->Delete button would be a good idea. . Implement the Mailing List page. . Implement the Print Member Badge/Label page. . Even though people should not use spreadsheets for database work, you will probably be pressured to impelement the "Export" capability. You will need Apache POI for this. . Refactor Home object to merge w/ darwinsys-ee EntityHome
Format ToDo as a list
Format ToDo as a list
AsciiDoc
bsd-2-clause
IanDarwin/clublist,IanDarwin/clublist
463ffdf5dc24117a2d245ca293ddb3bf1ad19ee8
README.adoc
README.adoc
proxy ===== [quote] A development proxy with logging and redirect-rewriting Installation ------------ [source,bash] ---- go get -u github.com/ciarand/proxy ---- Usage ----- [source,bash] ---- # start the proxy in one shell: proxy -from=https://www.google.com -to=http://0.0.0.0:8080 # and in another, run curl: curl -s http://localhost:8080/ | head -c 15 <!doctype html> # result from proxy shell (shortened for width): INFO[0022] request started client_address=[::1]:58988 method=GET uri=/ INFO[0023] request complete elapsed=624.644152ms status=200 ---- License ------- See the link:LICENSE[LICENSE] file.
proxy ===== [quote] A development proxy with logging and redirect-rewriting Installation ------------ Download a prebuilt binary for your platform and architecture from the link:https://github.com/ciarand/proxy/releases[release page]. Or, build from source: [source,bash] ---- # from source go get -u github.com/ciarand/proxy ---- Usage ----- [source,bash] ---- # start the proxy in one shell: proxy -from=https://www.google.com -to=http://0.0.0.0:8080 # and in another, run curl: curl -s http://localhost:8080/ | head -c 15 <!doctype html> # result from proxy shell (shortened for width): INFO[0022] request started client_address=[::1]:58988 method=GET uri=/ INFO[0023] request complete elapsed=624.644152ms status=200 ---- License ------- See the link:LICENSE[LICENSE] file.
Add prebuilt binary installation instructions
Add prebuilt binary installation instructions
AsciiDoc
isc
ciarand/proxy
99db481af16fdd49197e96ce8e67ac31b38577e3
README.adoc
README.adoc
# android-images Yet another repo with docker images for Android developers [source,planzuml] ------ node "java jdk-8" as jdk8 node "java jdk-7" as jdk7 artifact "Android" { node "gradle" as gradle node "sdk" as sdk node "ndk 11" as ndk11 node "ndk 13" as ndk13 node "vlc" as vlc } artifact "Tools" { node "ruby" as ruby node "Asciidoctor" as asciidoctor } gradle -up-> jdk8 sdk -up-> gradle ndk11 -up-> sdk ndk13 -up-> sdk vlc -up-> ndk11 ruby -up-> jdk8 asciidoctor -up-> ruby ------
# android-images Yet another repo with docker images for Android developers [source,plantuml] ------ node "java jdk-8" as jdk8 node "java jdk-7" as jdk7 artifact "Android" { node "gradle" as gradle node "sdk" as sdk node "ndk 11" as ndk11 node "ndk 13" as ndk13 node "vlc" as vlc } artifact "Tools" { node "ruby" as ruby node "Asciidoctor" as asciidoctor } gradle -up-> jdk8 sdk -up-> gradle ndk11 -up-> sdk ndk13 -up-> sdk vlc -up-> ndk11 ruby -up-> jdk8 asciidoctor -up-> ruby ------
Fix typo in plantuml definition
Fix typo in plantuml definition
AsciiDoc
apache-2.0
michalharakal/android-images
8379c9433703f7a107d5618528ec7066dcfc4f34
README.adoc
README.adoc
= Blueprint :author: Hafid Haddouti image:https://travis-ci.org/haf-tech/blueprint.svg?branch=master["Build Status", link="https://travis-ci.org/haf-tech/blueprint"] image:https://img.shields.io/badge/License-Apache%202.0-blue.svg["License", link="https://opensource.org/licenses/Apache-2.0"] .... Blueprint is a playground for illustrating different paradigms. In the meantime the following concepts are integrated or planned: - Spring Boot - AsciiDoctor integration, with UML and different outputs - Onion architecture - Docker build/push Next: - reactive programming
= Blueprint :author: Hafid Haddouti image:https://travis-ci.org/haf-tech/blueprint.svg?branch=master["Build Status", link="https://travis-ci.org/haf-tech/blueprint"] image:https://img.shields.io/badge/License-Apache%202.0-blue.svg["License", link="https://opensource.org/licenses/Apache-2.0"] Blueprint is a playground for illustrating different paradigms. In the meantime the following concepts are integrated or planned: - Spring Boot - AsciiDoctor integration, with UML and different outputs - Onion architecture - Docker build/push Next: - reactive programming An up to date documentation is https://haf-tech.github.io/blueprint/[online] available.
Add link to GitHup page
Add link to GitHup page
AsciiDoc
apache-2.0
haf-tech/blueprint
0a6537d3ad2914b49e7a104ddf6bb9fb236b96b7
community/users.asciidoc
community/users.asciidoc
= Who's Using Debezium? :awestruct-layout: doc :linkattrs: :icons: font :source-highlighter: highlight.js Debezium is used in production by a wide range of companies and organizations. This list contains users of Debezium who agreed to serve as public reference; where available, further resources with more details are linked. If your organization would like to be added to (or removed from) this list, please send a pull request for updating the https://github.com/debezium/debezium.github.io/blob/develop/docs/users.asciidoc[source of this page]. * Convoy (https://medium.com/convoy-tech/logs-offsets-near-real-time-elt-with-apache-kafka-snowflake-473da1e4d776[details]) * JW Player (https://www.slideshare.net/jwplayer/polylog-a-logbased-architecture-for-distributed-systems-124997666[details]) * OYO * Usabilla by Surveymonkey * WePay, Inc. (https://wecode.wepay.com/posts/streaming-databases-in-realtime-with-mysql-debezium-kafka[details], https://wecode.wepay.com/posts/streaming-cassandra-at-wepay-part-1[more details]) * ... and you? Then let us know and get added to the list, too. Thanks!
= Who's Using Debezium? :awestruct-layout: doc :linkattrs: :icons: font :source-highlighter: highlight.js Debezium is used in production by a wide range of companies and organizations. This list contains users of Debezium who agreed to serve as public reference; where available, further resources with more details are linked. If your organization would like to be added to (or removed from) this list, please send a pull request for updating the https://github.com/debezium/debezium.github.io/blob/develop/community/users.asciidoc[source of this page]. * Convoy (https://medium.com/convoy-tech/logs-offsets-near-real-time-elt-with-apache-kafka-snowflake-473da1e4d776[details]) * JW Player (https://www.slideshare.net/jwplayer/polylog-a-logbased-architecture-for-distributed-systems-124997666[details]) * OYO * Usabilla by Surveymonkey * WePay, Inc. (https://wecode.wepay.com/posts/streaming-databases-in-realtime-with-mysql-debezium-kafka[details], https://wecode.wepay.com/posts/streaming-cassandra-at-wepay-part-1[more details]) * ... and you? Then let us know and get added to the list, too. Thanks!
Fix broken link to source of page
Fix broken link to source of page
AsciiDoc
apache-2.0
debezium/debezium.github.io,debezium/debezium.github.io,debezium/debezium.github.io
3ffbb20f3476795d59a9f9e97864531e5ed075aa
master.adoc
master.adoc
= Sample Book Author Name <author@example.com> v1.0, October 4, 2015: First Draft :doctype: book :docinfo: :toc: left :toclevels: 2 :sectnums: :linkcss: An sample book to show case AsciiDoctor folder structure. include::book/chapter-1/chapter-1.adoc[leveloffset=+1] include::book/chapter-2/chapter-2.adoc[leveloffset=+1]
= Sample Book Author Name <author@example.com> v1.0, October 4, 2015: First Draft :doctype: book :docinfo: :toc: left :toclevels: 2 :sectnums: :linkcss: An sample book to show case AsciiDoctor folder structure. include::book/chapter-1/chapter-1.adoc[leveloffset=+1] include::book/chapter-2/chapter-2.adoc[leveloffset=+1]
Add extra line between include.
Add extra line between include.
AsciiDoc
mit
makzan/asciidoc-book-starter
eadf5c1973037cd3ff3a3fcb1fb48df7eb29be68
doc/release-process.adoc
doc/release-process.adoc
= bitcoinj-addons Release Process == Main Release Process . Update `CHANGELOG.adoc` . Set versions .. `README.adoc` .. bitcoinj-groovy `ExtensionModule` .. `gradle.properties` . Commit version bump and changelog. . Tag: `git tag -a v0.x.y -m "Release 0.x.y"` . Push: `git push --tags origin master` . Full build, test .. `./gradlew clean jenkinsBuild regTest` .. Recommended: test with *OmniJ* regTests. . Publish to Bintray: .. `./gradlew bintrayUpload` .. Confirm publish of artifacts in Bintray Web UI. . Update github-pages site (including JavaDoc): `./gradlew publishSite` == Announcements . Not yet. == After release . Set versions back to -SNAPSHOT .. `gradle.properties` .. bitcoinj-groovy `ExtensionModule` .. *Not* `README.adoc` -- it should match release version . Commit and push to master
= bitcoinj-addons Release Process == Main Release Process . Update `CHANGELOG.adoc` . Set versions .. `README.adoc` .. bitcoinj-groovy `ExtensionModule` .. `build.gradle` (should move to `gradle.properties`) . Commit version bump and changelog. . Tag: `git tag -a v0.x.y -m "Release 0.x.y"` . Push: `git push --tags origin master` . Full build, test .. `./gradlew clean jenkinsBuild regTest` .. Recommended: test with *OmniJ* regTests. . Publish to Bintray: .. `./gradlew bintrayUpload` .. Confirm publish of artifacts in Bintray Web UI. . Update github-pages site (including JavaDoc): `./gradlew publishSite` == Announcements . Not yet. == After release . Set versions back to -SNAPSHOT .. `gradle.properties` .. bitcoinj-groovy `ExtensionModule` .. *Not* `README.adoc` -- it should match release version . Commit and push to master
Fix minor error in release process doc.
Fix minor error in release process doc.
AsciiDoc
apache-2.0
msgilligan/bitcoinj-addons,msgilligan/bitcoinj-addons,msgilligan/bitcoinj-addons,msgilligan/bitcoinj-addons
c07974784ac6d323a1fb8820b609d2a4c3b717e0
README.adoc
README.adoc
|=== |image:http://goreportcard.com/badge/spohnan/ci-bot-01["Go Report Card",link="http://goreportcard.com/report/spohnan/ci-bot-01", window="_blank"]|image:https://travis-ci.org/spohnan/ci-bot-01.svg?branch=master["Build Status", link="https://travis-ci.org/spohnan/ci-bot-01", window="_blank"] |=== === Automated GitHub Interactions
[options="header"] |=== |CI Build and Tests|Static Analysis |image:https://travis-ci.org/spohnan/ci-bot-01.svg?branch=master["Build Status", link="https://travis-ci.org/spohnan/ci-bot-01", window="_blank"]|image:http://goreportcard.com/badge/spohnan/ci-bot-01["Go Report Card",link="http://goreportcard.com/report/spohnan/ci-bot-01", window="_blank"] |=== === Automated GitHub Interactions
Add travis badge to readme
Add travis badge to readme
AsciiDoc
mit
spohnan/ci-bot-01,spohnan/ci-bot-01
76e3eeb08f5964513bcd73f57a94e02f019c0301
components/camel-spring-cloud-netflix/src/main/docs/spring-cloud-netflix.adoc
components/camel-spring-cloud-netflix/src/main/docs/spring-cloud-netflix.adoc
[[SpringCloud-SpringCloud]] Spring Cloud ~~~~~~~~~~~ *Available as of Camel 2.19* Spring Cloud component Maven users will need to add the following dependency to their `pom.xml` in order to use this component: [source,xml] ------------------------------------------------------------------------------------------------ <dependency> <groupId>org.apache.camel</groupId> <artifactId>camel-spring-cloud</artifactId> <version>${camel.version}</version> <!-- use the same version as your Camel core version --> </dependency> ------------------------------------------------------------------------------------------------ `camel-spring-cloud` jar comes with the `spring.factories` file, so as soon as you add that dependency into your classpath, Spring Boot will automatically auto-configure Camel for you. [[SpringCloud-CamelSpringCloudStarter]] Camel Spring Cloud Starter ^^^^^^^^^^^^^^^^^^^^^^^^^ *Available as of Camel 2.19* To use the starter, add the following to your spring boot pom.xml file: [source,xml] ------------------------------------------------------ <dependency> <groupId>org.apache.camel</groupId> <artifactId>camel-spring-cloud-starter</artifactId> <version>${camel.version}</version> <!-- use the same version as your Camel core version --> </dependency> ------------------------------------------------------
=== Spring Cloud Netflix *Available as of Camel 2.19* The Spring Cloud Netflix component bridges Camel Cloud and Spring Cloud Netflix so you can leverage Spring Cloud Netflix service discovery and load balance features in Camel and/or you can use Camel Service Discovery implementations as ServerList source for Spring Cloud Netflix's Ribbon load balabncer. Maven users will need to add the following dependency to their `pom.xml` in order to use this component: [source,xml] ---- <dependency> <groupId>org.apache.camel</groupId> <artifactId>camel-spring-cloud-netflix</artifactId> <version>${camel.version}</version> <!-- use the same version as your Camel core version --> </dependency> ---- `camel-spring-cloud-netflix` jar comes with the `spring.factories` file, so as soon as you add that dependency into your classpath, Spring Boot will automatically auto-configure Camel for you. You can disable Camel Spring Cloud Netflix with the following properties: [source,properties] ---- # Enable/Disable the whole integration, default true camel.cloud.netflix = true # Enable/Disable the integration with Ribbon, default true camel.cloud.netflix.ribbon = true ---- === Spring Cloud Netflix Starter *Available as of Camel 2.19* To use the starter, add the following to your spring boot pom.xml file: [source,xml] ---- <dependency> <groupId>org.apache.camel</groupId> <artifactId>camel-spring-cloud-netflix-starter</artifactId> <version>${camel.version}</version> <!-- use the same version as your Camel core version --> </dependency> ----
Fix copy and paste doc
Fix copy and paste doc
AsciiDoc
apache-2.0
akhettar/camel,gautric/camel,akhettar/camel,punkhorn/camel-upstream,scranton/camel,objectiser/camel,dmvolod/camel,onders86/camel,jonmcewen/camel,objectiser/camel,nboukhed/camel,nboukhed/camel,acartapanis/camel,gnodet/camel,jamesnetherton/camel,dmvolod/camel,jonmcewen/camel,dmvolod/camel,pmoerenhout/camel,tdiesler/camel,Thopap/camel,christophd/camel,rmarting/camel,christophd/camel,isavin/camel,sverkera/camel,yuruki/camel,DariusX/camel,CodeSmell/camel,akhettar/camel,pmoerenhout/camel,tlehoux/camel,yuruki/camel,mcollovati/camel,anoordover/camel,alvinkwekel/camel,christophd/camel,CodeSmell/camel,christophd/camel,apache/camel,objectiser/camel,rmarting/camel,gnodet/camel,jamesnetherton/camel,gautric/camel,pkletsko/camel,gautric/camel,jonmcewen/camel,davidkarlsen/camel,ullgren/camel,nikhilvibhav/camel,acartapanis/camel,anton-k11/camel,punkhorn/camel-upstream,onders86/camel,jonmcewen/camel,mgyongyosi/camel,acartapanis/camel,cunningt/camel,pkletsko/camel,DariusX/camel,snurmine/camel,nikhilvibhav/camel,Fabryprog/camel,jamesnetherton/camel,kevinearls/camel,salikjan/camel,CodeSmell/camel,tdiesler/camel,mgyongyosi/camel,CodeSmell/camel,sverkera/camel,pax95/camel,christophd/camel,prashant2402/camel,rmarting/camel,pkletsko/camel,nboukhed/camel,Thopap/camel,nicolaferraro/camel,tdiesler/camel,dmvolod/camel,gautric/camel,onders86/camel,drsquidop/camel,prashant2402/camel,kevinearls/camel,davidkarlsen/camel,anoordover/camel,adessaigne/camel,curso007/camel,snurmine/camel,drsquidop/camel,akhettar/camel,tadayosi/camel,Thopap/camel,gnodet/camel,anton-k11/camel,anoordover/camel,apache/camel,alvinkwekel/camel,scranton/camel,yuruki/camel,jamesnetherton/camel,Thopap/camel,rmarting/camel,kevinearls/camel,curso007/camel,tadayosi/camel,acartapanis/camel,isavin/camel,tlehoux/camel,nikhilvibhav/camel,scranton/camel,drsquidop/camel,alvinkwekel/camel,mgyongyosi/camel,tdiesler/camel,snurmine/camel,mgyongyosi/camel,prashant2402/camel,mcollovati/camel,pkletsko/camel,anton-k11/camel,anton-k11/camel,punkhorn/camel-upstream,DariusX/camel,pmoerenhout/camel,pmoerenhout/camel,nboukhed/camel,tdiesler/camel,drsquidop/camel,onders86/camel,pax95/camel,rmarting/camel,tadayosi/camel,isavin/camel,jamesnetherton/camel,ullgren/camel,tlehoux/camel,pkletsko/camel,snurmine/camel,prashant2402/camel,acartapanis/camel,tadayosi/camel,cunningt/camel,zregvart/camel,acartapanis/camel,pax95/camel,gnodet/camel,drsquidop/camel,zregvart/camel,nboukhed/camel,kevinearls/camel,apache/camel,adessaigne/camel,sverkera/camel,tlehoux/camel,alvinkwekel/camel,isavin/camel,sverkera/camel,adessaigne/camel,mcollovati/camel,yuruki/camel,punkhorn/camel-upstream,mgyongyosi/camel,scranton/camel,pax95/camel,scranton/camel,Fabryprog/camel,snurmine/camel,yuruki/camel,davidkarlsen/camel,gnodet/camel,isavin/camel,tdiesler/camel,sverkera/camel,cunningt/camel,drsquidop/camel,davidkarlsen/camel,cunningt/camel,nicolaferraro/camel,nboukhed/camel,jonmcewen/camel,dmvolod/camel,curso007/camel,Fabryprog/camel,dmvolod/camel,onders86/camel,Thopap/camel,sverkera/camel,cunningt/camel,kevinearls/camel,anoordover/camel,nikhilvibhav/camel,salikjan/camel,isavin/camel,apache/camel,tlehoux/camel,mgyongyosi/camel,onders86/camel,nicolaferraro/camel,ullgren/camel,prashant2402/camel,nicolaferraro/camel,adessaigne/camel,Fabryprog/camel,yuruki/camel,rmarting/camel,anoordover/camel,anton-k11/camel,jamesnetherton/camel,pmoerenhout/camel,adessaigne/camel,DariusX/camel,pax95/camel,pax95/camel,anton-k11/camel,anoordover/camel,ullgren/camel,mcollovati/camel,zregvart/camel,tadayosi/camel,christophd/camel,gautric/camel,akhettar/camel,Thopap/camel,akhettar/camel,apache/camel,curso007/camel,apache/camel,jonmcewen/camel,kevinearls/camel,tadayosi/camel,curso007/camel,scranton/camel,prashant2402/camel,tlehoux/camel,objectiser/camel,gautric/camel,pmoerenhout/camel,snurmine/camel,adessaigne/camel,curso007/camel,pkletsko/camel,cunningt/camel,zregvart/camel
aa949c664a89ab9d6c2c21fa9b36af585b721db9
docs/index.asciidoc
docs/index.asciidoc
= Filebeat :libbeat: http://www.elastic.co/guide/en/beats/libbeat/1.0.0-rc1 :version: 1.0.0-rc1 include::./overview.asciidoc[] include::./getting-started.asciidoc[] include::./fields.asciidoc[] include::./configuration.asciidoc[] include::./command-line.asciidoc[] include::./migration.asciidoc[] include::./support.asciidoc[]
= Filebeat :libbeat: http://www.elastic.co/guide/en/beats/libbeat/master :version: master include::./overview.asciidoc[] include::./getting-started.asciidoc[] include::./fields.asciidoc[] include::./configuration.asciidoc[] include::./command-line.asciidoc[] include::./migration.asciidoc[] include::./support.asciidoc[]
Use master version in docs
Use master version in docs
AsciiDoc
mit
yapdns/yapdnsbeat,yapdns/yapdnsbeat
8f10c771e62016bc156a01310ae35089b605f93d
docs/reference/migration/migrate_7_0/scripting.asciidoc
docs/reference/migration/migrate_7_0/scripting.asciidoc
[float] [[breaking_70_scripting_changes]] === Scripting changes [float] ==== getDate() and getDates() removed Fields of type `long` and `date` had `getDate()` and `getDates()` methods (for multi valued fields) to get an object with date specific helper methods for the current doc value. In 5.3.0, `date` fields were changed to expose this same date object directly when calling `doc["myfield"].value`, and the getter methods for date objects were deprecated. These methods have now been removed. Instead, use `.value` on `date` fields, or explicitly parse `long` fields into a date object using `Instance.ofEpochMillis(doc["myfield"].value)`. [float] ==== Script errors will return as `400` error codes Malformed scripts, either in search templates, ingest pipelines or search requests, return `400 - Bad request` while they would previously return `500 - Internal Server Error`. This also applies for stored scripts.
[float] [[breaking_70_scripting_changes]] === Scripting changes [float] ==== getDate() and getDates() removed Fields of type `long` and `date` had `getDate()` and `getDates()` methods (for multi valued fields) to get an object with date specific helper methods for the current doc value. In 5.3.0, `date` fields were changed to expose this same date object directly when calling `doc["myfield"].value`, and the getter methods for date objects were deprecated. These methods have now been removed. Instead, use `.value` on `date` fields, or explicitly parse `long` fields into a date object using `Instance.ofEpochMillis(doc["myfield"].value)`. [float] ==== Accessing missing document values will throw an error `doc['field'].value` will throw an exception if the document is missing a value for the field `field`. To check if a document is missing a value, you can use `doc['field'].size() == 0`. [float] ==== Script errors will return as `400` error codes Malformed scripts, either in search templates, ingest pipelines or search requests, return `400 - Bad request` while they would previously return `500 - Internal Server Error`. This also applies for stored scripts.
Add migration info for missing values in script
Add migration info for missing values in script Relates to #30975
AsciiDoc
apache-2.0
scorpionvicky/elasticsearch,uschindler/elasticsearch,HonzaKral/elasticsearch,robin13/elasticsearch,robin13/elasticsearch,uschindler/elasticsearch,gfyoung/elasticsearch,gfyoung/elasticsearch,gingerwizard/elasticsearch,uschindler/elasticsearch,nknize/elasticsearch,coding0011/elasticsearch,nknize/elasticsearch,coding0011/elasticsearch,HonzaKral/elasticsearch,scorpionvicky/elasticsearch,scorpionvicky/elasticsearch,gingerwizard/elasticsearch,scorpionvicky/elasticsearch,robin13/elasticsearch,GlenRSmith/elasticsearch,nknize/elasticsearch,gingerwizard/elasticsearch,GlenRSmith/elasticsearch,gfyoung/elasticsearch,gfyoung/elasticsearch,gingerwizard/elasticsearch,gfyoung/elasticsearch,GlenRSmith/elasticsearch,HonzaKral/elasticsearch,scorpionvicky/elasticsearch,nknize/elasticsearch,GlenRSmith/elasticsearch,gingerwizard/elasticsearch,uschindler/elasticsearch,HonzaKral/elasticsearch,robin13/elasticsearch,coding0011/elasticsearch,nknize/elasticsearch,coding0011/elasticsearch,gingerwizard/elasticsearch,GlenRSmith/elasticsearch,robin13/elasticsearch,gingerwizard/elasticsearch,coding0011/elasticsearch,uschindler/elasticsearch
d9e3a3fc36a4d9770f53cb105d60ae7d01965aed
src/main/asciidoc/inc/_goals.adoc
src/main/asciidoc/inc/_goals.adoc
= Maven Goals This plugin supports the following goals which are explained in detail in the next sections. .Plugin Goals [cols="1,3"] |=== |Goal | Description |**<<{plugin}:build>>** |Build images |**<<{plugin}:start>>** or **<<{plugin}:start,{plugin}:run>>** |Create and start containers |**<<{plugin}:stop>>** |Stop and destroy containers |**<<{plugin}:push>>** |Push images to a registry |**<<{plugin}:watch>>** |Watch for doing rebuilds and restarts |**<<{plugin}:remove>>** |Remove images from local docker host |**<<{plugin}:logs>>** |Show container logs |**<<{plugin}:source>>** |Attach docker build archive to Maven project |**<<{plugin}:save>>** |Save images to a file |**<<{plugin}:volume-create>>** |Create a volume for containers to share data |**<<{plugin}:volume-remove>>** |Remove a volume |=== Note that all goals are orthogonal to each other. For example in order to start a container for your application you typically have to build its image before. `{plugin}:start` does *not* imply building the image so you should use it then in combination with `{plugin}:build`.
= Maven Goals This plugin supports the following goals which are explained in detail in the next sections. .Plugin Goals [cols="1,1,2"] |=== |Goal | Default Lifecycle Phase | Description |**<<{plugin}:build>>** |install |Build images |**<<{plugin}:start>>** or **<<{plugin}:start,{plugin}:run>>** |pre-integration-test |Create and start containers |**<<{plugin}:stop>>** |post-integration-test |Stop and destroy containers |**<<{plugin}:push>>** |deploy |Push images to a registry |**<<{plugin}:watch>>** | |Watch for doing rebuilds and restarts |**<<{plugin}:remove>>** |post-integration-test |Remove images from local docker host |**<<{plugin}:logs>>** | |Show container logs |**<<{plugin}:source>>** | |Attach docker build archive to Maven project |**<<{plugin}:save>>** | |Save images to a file |**<<{plugin}:volume-create>>** |pre-integration-test |Create a volume for containers to share data |**<<{plugin}:volume-remove>>** |post-integration-test |Remove a volume |=== Note that all goals are orthogonal to each other. For example in order to start a container for your application you typically have to build its image before. `{plugin}:start` does *not* imply building the image so you should use it then in combination with `{plugin}:build`.
Add default Maven lifecycle bindings to the manual
Add default Maven lifecycle bindings to the manual
AsciiDoc
apache-2.0
rhuss/docker-maven-plugin,vjuranek/docker-maven-plugin,thomasvandoren/docker-maven-plugin,scoplin/docker-maven-plugin,vjuranek/docker-maven-plugin,fabric8io/docker-maven-plugin,fabric8io/docker-maven-plugin,fabric8io/docker-maven-plugin,vjuranek/docker-maven-plugin,rhuss/docker-maven-plugin
886e5e7db96a6c1a021dcc9d08268303d8402227
docs/hacky-implicit-provisioning-version2.adoc
docs/hacky-implicit-provisioning-version2.adoc
# Hacky Implicit Provisioning version 2 This document describes a quick hack to test the implicit provisioning flow as it currently works in aktualizr before full support is available in meta-updater and on the server. ## Goals * end-to-end installation of updates using OSTree ## Steps 1. Edit `recipes-sota/aktualizr/files/sota_autoprov.toml` in meta-updater: * Remove the `provision_path` line. * Add a line in the `[tls]` section with `server = "https://..."`. Use the URL from your account's credentials.zip file. 1. Edit `recipes-sota/aktualizr/aktualizr_git.bb` in meta-updater: * Change aktualizr `SRC_URI` to `git://github.com/advancedtelematic/aktualizr;branch=bugfix/cert_prov_install`. * Change `SRCREV` to `13828774baa5a7369e6f5ca552b879ad1ce773d5`. * Increment `PR`. 1. Build a standard image using bitbake. Make sure to set `SOTA_PACKED_CREDENTIALS` like normal. 1. Boot the image. 1. Optionally, verify that aktualizr is not provisioning. Make sure the device is not visible in the Garage. 1. Run `cert_provider` from the aktualizr repo: `cert_provider -c credentials.zip -t <device>` 1. Verify that aktualizr provisions correctly with the server using the device_id generated by `cert_provider`.
# Hacky Implicit Provisioning version 2 This document describes a quick hack to test the implicit provisioning flow as it currently works in aktualizr before full support is available in meta-updater and on the server. ## Goals * end-to-end installation of updates using OSTree ## Steps 1. Edit `recipes-sota/aktualizr/files/sota_autoprov.toml` in meta-updater: * Remove the `provision_path` line. * Add a line in the `[tls]` section with `server = "https://..."`. Use the URL from your account's credentials.zip file. 1. Edit `recipes-sota/aktualizr/aktualizr_git.bb` in meta-updater: * Change `SRCREV` to `1c635c67d70cf38d2c841d449e750f08855e41a4`. * Increment `PR`. 1. Build a standard image using bitbake. Make sure to set `SOTA_PACKED_CREDENTIALS` like normal. 1. Boot the image. 1. Optionally, verify that aktualizr is not provisioning. Make sure the device is not visible in the Garage. 1. Run `cert_provider` from the aktualizr repo: `cert_provider -c credentials.zip -t <device>` 1. Verify that aktualizr provisions correctly with the server using the device_id generated by `cert_provider`. ## Known Issues At this time, although the device will be shown in the Garage, its installed package will not be.
Use latest version of aktualizr. Add known issue.
Use latest version of aktualizr. Add known issue.
AsciiDoc
mpl-2.0
advancedtelematic/aktualizr,advancedtelematic/aktualizr,advancedtelematic/aktualizr,advancedtelematic/sota_client_cpp,advancedtelematic/aktualizr,advancedtelematic/sota_client_cpp
fae66814a50c60e47772ad6feec9ee222920fd9c
doc/videos.adoc
doc/videos.adoc
= Videos James Elliott <james@deepsymmetry.org> :icons: font // Set up support for relative links on GitHub; add more conditions // if you need to support other environments and extensions. ifdef::env-github[:outfilesuffix: .adoc] This page collects performance videos that highlight Afterglow in action. If you have any to share, please post them to the https://gitter.im/brunchboy/afterglow[Afterglow room on Gitter]! https://vimeo.com/153492480[image:assets/Deepower-2015.png[Deepower Live Report,align="right",float="right"]] Since I have been too busy working on the software to get out and perform with it, I am deeply grateful to https://github.com/dandaka[Vlad Rafeev] for getting this page started by sharing a video of a https://www.facebook.com/deepowerband/[Deepower audiovisual] show he lit using Afterglow in December, 2015, in Kaliningrad, Russia last month. He modestly says, “My first experience with lights, still a lot to accomplish.” And I say, there is a ton left to implement in the software too. But it is already fun to see https://vimeo.com/153492480[great video like this]!
= Videos James Elliott <james@deepsymmetry.org> :icons: font // Set up support for relative links on GitHub; add more conditions // if you need to support other environments and extensions. ifdef::env-github[:outfilesuffix: .adoc] This page collects performance videos that highlight Afterglow in action. If you have any to share, please post them to the https://gitter.im/brunchboy/afterglow[Afterglow room on Gitter]! +++<a href="https://vimeo.com/153492480"><img src="assets/Deepower-2015.png" align="right" alt="Deepower Live Report"></a>+++ Since I have been too busy working on the software to get out and perform with it, I am deeply grateful to https://github.com/dandaka[Vlad Rafeev] for getting this page started by sharing a video of a https://www.facebook.com/deepowerband/[Deepower audiovisual] show he lit using Afterglow in December, 2015, in Kaliningrad, Russia last month. He modestly says, “My first experience with lights, still a lot to accomplish.” And I say, there is a ton left to implement in the software too. But it is already fun to see https://vimeo.com/153492480[great video like this]!
Work around Github image float issue.
Work around Github image float issue.
AsciiDoc
epl-1.0
brunchboy/afterglow,brunchboy/afterglow,brunchboy/afterglow
0ba8df5939f33a3d6ee7a4eaab0c993e35a839d0
migration/migrating_3_4/about-migration.adoc
migration/migrating_3_4/about-migration.adoc
[id="about-migration"] = About migrating {product-title} 3 to 4 include::modules/common-attributes.adoc[] :context: about-migration toc::[] {product-title} 4 includes new technologies and functionality that results in a cluster that is self-managing, flexible, and automated. The way that {product-title} 4 clusters are deployed and managed drastically differs from {product-title} 3. To successfully transition from {product-title} 3 to {product-title} 4, it is important that you review the following information: xref:../../migration/migrating_3_4/planning-migration-3-to-4.adoc#planning-migration-3-to-4[Planning your transition]:: Learn about the differences between {product-title} versions 3 and 4. Prior to transitioning, be sure that you have reviewed and prepared for storage, networking, logging, security, and monitoring considerations. xref:../../migration/migrating_3_4/migrating-application-workloads-3-4.adoc#migrating-application-workloads-3-4[Performing your migration]:: Learn about and use the tools to perform your migration: * {mtc-full} ({mtc-short}) to migrate your application workloads
[id="about-migration"] = About migrating {product-title} 3 to 4 include::modules/common-attributes.adoc[] :context: about-migration toc::[] {product-title} 4 includes new technologies and functionality that results in a cluster that is self-managing, flexible, and automated. The way that {product-title} 4 clusters are deployed and managed drastically differs from {product-title} 3. To successfully transition from {product-title} 3 to {product-title} 4, it is important that you review the following information: xref:../../migration/migrating_3_4/planning-migration-3-to-4.adoc#planning-migration-3-to-4[Planning your transition]:: Learn about the differences between {product-title} versions 3 and 4. Prior to transitioning, be sure that you have reviewed and prepared for storage, networking, logging, security, and monitoring considerations. xref:../../migration/migrating_3_4/migrating-application-workloads-3-4.adoc#migrating-application-workloads-3-4[Performing your migration]:: Learn about and use the {mtc-full} ({mtc-short}) to migrate your application workloads.
Update tools sentence after removing CPMA
Update tools sentence after removing CPMA
AsciiDoc
apache-2.0
vikram-redhat/openshift-docs,vikram-redhat/openshift-docs,vikram-redhat/openshift-docs,vikram-redhat/openshift-docs,vikram-redhat/openshift-docs
38adcbe7aeee0cdaa5600b91cc4297c399bdc33b
README.adoc
README.adoc
= Jira Release Notes Generator (JiraRnGen) :Author: David Thompson :Email: <dthompsn1@gmail.com> :Revision: 0.1.0 2016-08-03 == Description I was looking for a way with hosted Jira to be able to send out release notes with our own template. Since the hosted version doesn't allow you to edit the templates, I decided to build an application that worked with the Jira api to build an html page that could be emailed to our team. This NodeJS app is the result of this work. == ReleaseNotes 0.1.0 - First iteration at being able to send Jira Release Notes == Usage First time to use, run within the JiraRnGen directory. ---- $ npm install ---- Then to run each time ---- $ node release-notes.js -u username -p password --email -n jira-hostname -f release-name -s email-address ----
= Jira Release Notes Generator (JiraRnGen) :Author: David Thompson :Email: <dthompsn1@gmail.com> :Revision: 0.1.0 2016-08-03 == Description I was looking for a way with hosted Jira to be able to send out release notes with our own template. Since the hosted version doesn't allow you to edit the templates, I decided to build an application that worked with the Jira api to build an html page that could be emailed to our team. This NodeJS app is the result of this work. == ReleaseNotes 0.1.0 - First iteration at being able to send Jira Release Notes == Usage First time to use, run within the JiraRnGen directory. ---- $ npm install ---- You will need to create a image/logo.json file; you can copy the example to logo.json for testing. You will need to set up a sendgrid account and set up an API key to use with sendgrid. Copy conf/config-example.json to conf/config.json. Then to run each time ---- $ node release-notes.js -u username -p password --email -n jira-hostname -f release-name -s email-address ----
Update docs for running at start.
Update docs for running at start.
AsciiDoc
mit
applitect/JiraRnGen,applitect/JiraRnGen,applitect/JiraRnGen
b15bdde83f43d3b6c0564a359ae3a8bf43b94071
opennms-doc/guide-user/src/asciidoc/index.adoc
opennms-doc/guide-user/src/asciidoc/index.adoc
// Global settings :ascii-ids: :encoding: UTF-8 :lang: en :icons: font :toc: left :toclevels: 3 :numbered: [[gu]] = Users Guide :author: Copyright (c) 2014-2016 The OpenNMS Group, Inc. :revnumber: {opennms-product-name} {opennms-version} :revdate: {last-update-label} {docdatetime} :version-label!: [[gu-service-assurance]] == Service Assurance include::text/service-assurance/introduction.adoc[] include::text/service-assurance/critical-service.adoc[] include::text/service-assurance/path-outage.adoc[] include::text/surveillance-view.adoc[] include::text/dashboard.adoc[] [[gu-bsm]] == Busines Service Monitoring include::text/bsm/introduction.adoc[] include::text/bsm/business-service-hierarchy.adoc[] include::text/bsm/operational-status.adoc[] include::text/bsm/root-cause-impact-analysis.adoc[] include::text/bsm/simulation-mode.adoc[] include::text/bsm/share-bsm-view.adoc[] include::text/bsm/change-icons.adoc[] [[gu-alarms]] == Alarms include::text/alarms/alarm-notes.adoc[]
// Global settings :ascii-ids: :encoding: UTF-8 :lang: en :icons: font :toc: left :toclevels: 3 :numbered: [[gu]] = Users Guide :author: Copyright (c) 2014-2016 The OpenNMS Group, Inc. :revnumber: {opennms-product-name} {opennms-version} :revdate: {last-update-label} {docdatetime} :version-label!: [[gu-service-assurance]] == Service Assurance include::text/service-assurance/introduction.adoc[] include::text/service-assurance/critical-service.adoc[] include::text/service-assurance/path-outage.adoc[] include::text/surveillance-view.adoc[] include::text/dashboard.adoc[] [[gu-bsm]] == Business Service Monitoring include::text/bsm/introduction.adoc[] include::text/bsm/business-service-hierarchy.adoc[] include::text/bsm/operational-status.adoc[] include::text/bsm/root-cause-impact-analysis.adoc[] include::text/bsm/simulation-mode.adoc[] include::text/bsm/share-bsm-view.adoc[] include::text/bsm/change-icons.adoc[] [[gu-alarms]] == Alarms include::text/alarms/alarm-notes.adoc[]
Fix typo in Busines -> Business
Fix typo in Busines -> Business
AsciiDoc
agpl-3.0
aihua/opennms,aihua/opennms,aihua/opennms,aihua/opennms,aihua/opennms,aihua/opennms,aihua/opennms,aihua/opennms,aihua/opennms
f663decf9b6ee767d81a7b9e9bd028eada319548
README.adoc
README.adoc
image:https://jenkins-kieci.rhcloud.com/buildStatus/icon?job=optaplanner["Build Status", link="https://jenkins-kieci.rhcloud.com/job/optaplanner"] == Developing Drools, OptaPlanner and jBPM *If you want to build or contribute to a droolsjbpm project, https://github.com/droolsjbpm/droolsjbpm-build-bootstrap/blob/master/README.md[read this document].* *It will save you and us a lot of time by setting up your development environment correctly.* It solves all known pitfalls that can disrupt your development. It also describes all guidelines, tips and tricks. If you want your pull requests (or patches) to be merged into master, please respect those guidelines.
image:https://jenkins-kieci.rhcloud.com/buildStatus/icon?job=optaplanner["Build Status", link="https://jenkins-kieci.rhcloud.com/job/optaplanner"] == Quick development start To build and run from source: [source,sh] ---- $ mvn clean install $ cd optaplanner-examples $ mvn exec:java ---- To develop with IntelliJ IDEA, Eclipse or NetBeans, open the root `pom.xml` as a new project and configure a _Run/Debug configuration_ like this: * Main class: `org.optaplanner.examples.app.OptaPlannerExamplesApp` * VM options: `-Xmx2G -server` (Memory only needed when using the big datasets in the examples) * Program arguments: `` * Working directory: `$MODULE_DIR$` (must resolve to optaplanner-examples directory) * Use classpath of module: `optaplanner-examples` == Developing Drools, OptaPlanner and jBPM *If you want to build or contribute to a droolsjbpm project, https://github.com/droolsjbpm/droolsjbpm-build-bootstrap/blob/master/README.md[read this document].* *It will save you and us a lot of time by setting up your development environment correctly.* It solves all known pitfalls that can disrupt your development. It also describes all guidelines, tips and tricks. If you want your pull requests (or patches) to be merged into master, please respect those guidelines.
Add build instructions to readme
Add build instructions to readme
AsciiDoc
apache-2.0
baldimir/optaplanner,droolsjbpm/optaplanner,oskopek/optaplanner,droolsjbpm/optaplanner,oskopek/optaplanner,gsheldon/optaplanner,baldimir/optaplanner,baldimir/optaplanner,oskopek/optaplanner,oskopek/optaplanner,tkobayas/optaplanner,gsheldon/optaplanner,droolsjbpm/optaplanner,baldimir/optaplanner,tkobayas/optaplanner,tkobayas/optaplanner,gsheldon/optaplanner,gsheldon/optaplanner,tkobayas/optaplanner,droolsjbpm/optaplanner
8274fc56d531bd3b5bc8b47fb699dfae15da84a0
README.adoc
README.adoc
= Explorer for Hawkular http://hawkular.org/[Hawkular] is a set of components for Monitoring. This explorer connects to a Hawkular server and allows to browse trough inventory and view entities, graph metrics. ifndef::env-github[] image::docs/screenshot.png[] endif::[] ifdef::env-github[] image::https://github.com/pilhuhn/hawkfx/blob/master/docs/screenshot.png[] endif::[] == Running The explorer requires JRuby in version 9+ If you are using `rvm` you can select it via `rvm use jruby-9.0.5.0` then use `bundler` to install the required gems `bundle install` then run `jruby hawkfx.rb`
= Explorer for Hawkular http://hawkular.org/[Hawkular] is a set of components for Monitoring. This explorer connects to a Hawkular server and allows to browse trough inventory and view entities, graph metrics. ifndef::env-github[] image::docs/screenshot.png[] endif::[] ifdef::env-github[] image::https://github.com/pilhuhn/hawkfx/blob/master/docs/screenshot.png[] endif::[] == Running The explorer requires JRuby in version 9+ If you are using `rvm` you can select it via `rvm use jruby-9.0.5.0` install and use `bundler` to install the required gems `gem install bundler` `bundle install` then run `jruby hawkfx.rb`
Make build step more clear.
Make build step more clear.
AsciiDoc
apache-2.0
pilhuhn/hawkfx
6c960898d801f348d640e17ab5d496d9f7a8e767
docs/reference/mapping/fields/id-field.asciidoc
docs/reference/mapping/fields/id-field.asciidoc
[[mapping-id-field]] === `_id` field Each document indexed is associated with a <<mapping-type-field,`_type`>> (see <<mapping-type>>) and an <<mapping-id-field,`_id`>>. The `_id` field is not indexed as its value can be derived automatically from the <<mapping-uid-field,`_uid`>> field. The value of the `_id` field is accessible in certain queries (`term`, `terms`, `match`, `query_string`, `simple_query_string`) and scripts, but _not_ in aggregations or when sorting, where the <<mapping-uid-field,`_uid`>> field should be used instead: [source,js] -------------------------- # Example documents PUT my_index/my_type/1 { "text": "Document with ID 1" } PUT my_index/my_type/2 { "text": "Document with ID 2" } GET my_index/_search { "query": { "terms": { "_id": [ "1", "2" ] <1> } }, "script_fields": { "UID": { "script": "doc['_id']" <2> } } } -------------------------- // AUTOSENSE <1> Querying on the `_id` field (also see the <<query-dsl-ids-query,`ids` query>>) <2> Accessing the `_id` field in scripts (inline scripts must be <<enable-dynamic-scripting,enabled>> for this example to work)
[[mapping-id-field]] === `_id` field Each document indexed is associated with a <<mapping-type-field,`_type`>> (see <<mapping-type>>) and an <<mapping-id-field,`_id`>>. The `_id` field is not indexed as its value can be derived automatically from the <<mapping-uid-field,`_uid`>> field. The value of the `_id` field is accessible in certain queries (`term`, `terms`, `match`, `query_string`, `simple_query_string`), but _not_ in aggregations, scripts or when sorting, where the <<mapping-uid-field,`_uid`>> field should be used instead: [source,js] -------------------------- # Example documents PUT my_index/my_type/1 { "text": "Document with ID 1" } PUT my_index/my_type/2 { "text": "Document with ID 2" } GET my_index/_search { "query": { "terms": { "_id": [ "1", "2" ] <1> } } } -------------------------- // AUTOSENSE <1> Querying on the `_id` field (also see the <<query-dsl-ids-query,`ids` query>>)
Fix docs example for the _id field, the field is not accessible in scripts
Fix docs example for the _id field, the field is not accessible in scripts Closes #19274
AsciiDoc
apache-2.0
strapdata/elassandra-test,strapdata/elassandra-test,strapdata/elassandra-test,strapdata/elassandra-test,strapdata/elassandra-test,strapdata/elassandra-test,strapdata/elassandra-test
1e6fc7b9e405c2c9e05868db8534dadbb80437f7
doc/src/guide/external_plugins_list.asciidoc
doc/src/guide/external_plugins_list.asciidoc
[[plugins_list]] == List of plugins This is a non-exhaustive list of Erlang.mk plugins, sorted alphabetically. === elvis.mk An https://github.com/inaka/elvis.mk[Elvis plugin] for Erlang.mk. Elvis is an https://github.com/inaka/elvis[Erlang style reviewer]. === geas https://github.com/crownedgrouse/geas[Geas] gives aggregated information on a project and its dependencies, and is available as an Erlang.mk plugin. === hexer.mk An https://github.com/inaka/hexer.mk[Hex plugin] for Erlang.mk. Hex is a https://hex.pm/[package manager for the Elixir ecosystem]. === reload.mk A https://github.com/bullno1/reload.mk[live reload plugin] for Erlang.mk.
[[plugins_list]] == List of plugins This is a non-exhaustive list of Erlang.mk plugins, sorted alphabetically. === elixir.mk An https://github.com/botsunit/elixir.mk[Elixir plugin] for Erlang.mk. http://elixir-lang.org/[Elixir] is an alternative language for the BEAM. === elvis.mk An https://github.com/inaka/elvis.mk[Elvis plugin] for Erlang.mk. Elvis is an https://github.com/inaka/elvis[Erlang style reviewer]. === geas https://github.com/crownedgrouse/geas[Geas] gives aggregated information on a project and its dependencies, and is available as an Erlang.mk plugin. === hexer.mk An https://github.com/inaka/hexer.mk[Hex plugin] for Erlang.mk. Hex is a https://hex.pm/[package manager for the Elixir ecosystem]. === lfe.mk An https://github.com/ninenines/lfe.mk[LFE plugin] for Erlang.mk. LFE, or http://lfe.io/[Lisp Flavoured Erlang], is an alternative language for the BEAM. === reload.mk A https://github.com/bullno1/reload.mk[live reload plugin] for Erlang.mk.
Add elixir.mk and lfe.mk to the plugins list
Add elixir.mk and lfe.mk to the plugins list
AsciiDoc
isc
crownedgrouse/erlang.mk,hairyhum/erlang.mk,ingwinlu/erlang.mk,nevar/erlang.mk,jj1bdx/erlang.mk,rabbitmq/erlang.mk,ninenines/erlang.mk
290377a0fa42b7da5bce6583834bc525fdfe354c
README.adoc
README.adoc
= Json-lib :version: 3.0.0.SNAPSHOT :linkattrs: :project-name: json-lib image:http://img.shields.io/travis/aalmiray/{project-name}/development.svg["Build Status", link="https://travis-ci.org/aalmiray/{project-name}"] image:http://img.shields.io/coveralls/aalmiray/{project-name}/development.svg["Coverage Status", link="https://coveralls.io/r/aalmiray/{project-name}"] image:http://img.shields.io/:semver-{version}-blue.svg["Semantic Versioning", link="http://semver.org"] image:http://img.shields.io/badge/license-ASF2-blue.svg["Apache License 2", link="http://www.apache.org/licenses/LICENSE-2.0.txt"] image:http://img.shields.io/badge/download-latest-bb00bb.svg[link="https://bintray.com/aalmiray/kordamp/{project-name}/_latestVersion"] --- JSON-lib is a java library for transforming beans, maps, collections, java arrays and XML to JSON and back again to beans and DynaBeans. Refer to the link:http://aalmiray.github.io/{project-name}/[project guide, window="_blank"] for further information on configuration and usage.
= Json-lib :version: 3.0.0.SNAPSHOT :linkattrs: image:http://img.shields.io/travis/aalmiray/Json-lib/development.svg["Build Status", link="https://travis-ci.org/aalmiray/Json-lib"] image:http://img.shields.io/coveralls/aalmiray/Json-lib/development.svg["Coverage Status", link="https://coveralls.io/r/aalmiray/Json-lib"] image:http://img.shields.io/:semver-{version}-blue.svg["Semantic Versioning", link="http://semver.org"] image:http://img.shields.io/badge/license-ASF2-blue.svg["Apache License 2", link="http://www.apache.org/licenses/LICENSE-2.0.txt"] image:http://img.shields.io/badge/download-latest-bb00bb.svg[link="https://bintray.com/aalmiray/kordamp/json-lib/_latestVersion"] --- JSON-lib is a java library for transforming beans, maps, collections, java arrays and XML to JSON and back again to beans and DynaBeans. Refer to the link:http://aalmiray.github.io/json-lib/[project guide, window="_blank"] for further information on configuration and usage.
Fix project name in readme
Fix project name in readme [skip ci]
AsciiDoc
apache-2.0
aalmiray/Json-lib,aalmiray/Json-lib
3e4b6c5ea4716a2d5ddc753bcc10260b75eaa762
README.asciidoc
README.asciidoc
== ANDROID MAVEN PLUGIN A plugin for Android application development with http://maven.apache.org[Apache Maven 3.0.3+] and the http://tools.android.com[Android SDK]. === Links * http://code.google.com/p/maven-android-plugin[Project site] with wiki and more * http://code.google.com/p/maven-android-plugin/issues/list[Issue tracker] * http://maven-android-plugin-m2site.googlecode.com/svn/index.html[Maven generated plugin documentation site] * http://www.sonatype.com/books/mvnref-book/reference/android-dev.html[Maven: Complete Reference - Chapter - Android Application Development with Maven] * https://groups.google.com/forum/?fromgroups#!forum/maven-android-developers[Mailinglist] * http://code.google.com/p/maven-android-plugin/wiki/Changelog[Changelog] * http://jenkins.josefson.org/[Continuous Integration Server Builds] === Contributions We welcome your feature enhancements and bug fixes in pull requests!
== ANDROID MAVEN PLUGIN A plugin for Android application development with http://maven.apache.org[Apache Maven 3.0.3+] and the http://tools.android.com[Android SDK]. === Links * http://code.google.com/p/maven-android-plugin[Project site] with wiki and more * http://code.google.com/p/maven-android-plugin/issues/list[Issue tracker] * http://maven-android-plugin-m2site.googlecode.com/svn/index.html[Maven generated plugin documentation site] * http://www.sonatype.com/books/mvnref-book/reference/android-dev.html[Maven: Complete Reference - Chapter - Android Application Development with Maven] * https://groups.google.com/forum/?fromgroups#!forum/maven-android-developers[Mailinglist] * http://code.google.com/p/maven-android-plugin/wiki/Changelog[Changelog] * image:https://travis-ci.org/jayway/maven-android-plugin.png["Build Status", link="https://travis-ci.org/jayway/maven-android-plugin"] === Contributions We welcome your feature enhancements and bug fixes in pull requests!
Change link to CI server.
Change link to CI server.
AsciiDoc
apache-2.0
xiaojiaqiao/android-maven-plugin,Stuey86/android-maven-plugin,secondsun/maven-android-plugin,psorobka/android-maven-plugin,xiaojiaqiao/android-maven-plugin,greek1979/maven-android-plugin,repanda/android-maven-plugin,mitchhentges/android-maven-plugin,Cha0sX/android-maven-plugin,hgl888/android-maven-plugin,b-cuts/android-maven-plugin,CJstar/android-maven-plugin,b-cuts/android-maven-plugin,secondsun/maven-android-plugin,kedzie/maven-android-plugin,Stuey86/android-maven-plugin,WonderCsabo/maven-android-plugin,jdegroot/android-maven-plugin,wskplho/android-maven-plugin,WonderCsabo/maven-android-plugin,repanda/android-maven-plugin,xiaojiaqiao/android-maven-plugin,kedzie/maven-android-plugin,Cha0sX/android-maven-plugin,Cha0sX/android-maven-plugin,wskplho/android-maven-plugin,hgl888/android-maven-plugin,greek1979/maven-android-plugin,xieningtao/android-maven-plugin,ashutoshbhide/android-maven-plugin,mitchhentges/android-maven-plugin,CJstar/android-maven-plugin,psorobka/android-maven-plugin,simpligility/android-maven-plugin,secondsun/maven-android-plugin,xieningtao/android-maven-plugin,jdegroot/android-maven-plugin,ashutoshbhide/android-maven-plugin
f2de1064b10e4876a36c3c2d3cbc2685445323cc
attendees/cicd/git-server/README.adoc
attendees/cicd/git-server/README.adoc
= Install a Git Server == Start Services . All services can be started, in detached mode, by giving the command: + docker-compose up -d + And this shows the output as: + Creating git_serverdata... Creating gitserver_git_dbdata_1... Creating gitserver_git_db_1... Creating git... + . Configure the installation .. execute the following script + ./install-gogs.sh <DOCKER_HOST_IP> <git_server_PORT> + _example: ./install.sh 192.168.99.100 3000_ == Sign Up . Access to this URL: http://dockerhost:3000/user/sign_up . Create an account and enjoy! . Enjoy! :)
= Install a Git Server == Start Services . All services can be started, in detached mode, by giving the command: + docker-compose up -d + And this shows the output as: + Creating git_serverdata... Creating gitserver_git_dbdata_1... Creating gitserver_git_db_1... Creating git... + . Configure the installation .. execute the following script + ./install-gogs.sh <DOCKER_HOST_IP> <git_server_PORT> + _example: ./install-gogs.sh 192.168.99.100 3000_ == Sign Up . Access to this URL: http://dockerhost:3000/user/sign_up . Create an account and enjoy! . Enjoy! :)
Fix usage command with install-gogs.sh
Fix usage command with install-gogs.sh
AsciiDoc
apache-2.0
redhat-developer-demos/docker-java,redhat-developer-demos/docker-java
57583364532eed5c3a603307a24ab7857cf27b22
doc/src/manual/cowboy_router.compile.asciidoc
doc/src/manual/cowboy_router.compile.asciidoc
= cowboy_router:compile(3) == Name cowboy_router:compile - Compile routes to the resources == Description [source,erlang] ---- compile(cowboy_router:routes()) -> cowboy_router:dispatch_rules() ---- Compile routes to the resources. Takes a human readable list of routes and transforms it into a form more efficient to process. == Arguments Routes:: Human readable list of routes. == Return value An opaque dispatch rules value is returned. This value must be given to Cowboy as a middleware environment value. == Changelog * *1.0*: Function introduced. == Examples .Compile routes and start a listener [source,erlang] ---- Dispatch = cowboy_router:compile([ {'_', [ {"/", toppage_h, []}, {"/[...], cowboy_static, {priv_dir, my_example_app, ""}} ]} ]), {ok, _} = cowboy:start_clear(example, [{port, 8080}], #{ env => #{dispatch => Dispatch} }). ---- == See also link:man:cowboy_router(3)[cowboy_router(3)]
= cowboy_router:compile(3) == Name cowboy_router:compile - Compile routes to the resources == Description [source,erlang] ---- compile(cowboy_router:routes()) -> cowboy_router:dispatch_rules() ---- Compile routes to the resources. Takes a human readable list of routes and transforms it into a form more efficient to process. == Arguments Routes:: Human readable list of routes. == Return value An opaque dispatch rules value is returned. This value must be given to Cowboy as a middleware environment value. == Changelog * *1.0*: Function introduced. == Examples .Compile routes and start a listener [source,erlang] ---- Dispatch = cowboy_router:compile([ {'_', [ {"/", toppage_h, []}, {"/[...]", cowboy_static, {priv_dir, my_example_app, ""}} ]} ]), {ok, _} = cowboy:start_clear(example, [{port, 8080}], #{ env => #{dispatch => Dispatch} }). ---- == See also link:man:cowboy_router(3)[cowboy_router(3)]
Fix an example missing a " in the manual
Fix an example missing a " in the manual
AsciiDoc
isc
kivra/cowboy,hairyhum/cowboy,turtleDeng/cowboy,K2InformaticsGmbH/cowboy,CrankWheel/cowboy,ninenines/cowboy
506f06bb577587936a0c3efb10080fde8d62dce6
spring-content-solr/src/main/asciidoc/solr-rest.adoc
spring-content-solr/src/main/asciidoc/solr-rest.adoc
[[search]] = Search == The SearchContent Resource When a Store extending `Searchable` is exported, a `searchContent` endpoint will be available at the `/{store}/searchContent` URI. ==== [source, sh] ---- curl -H 'Accept: application/hal+json' http://localhost:8080/searchContent?queryString=foo ---- ==== === Supported HTTP Methods As the SearchContent resource is read-only it supports `GET` only. All other HTTP methods will cause a `405 Method Not Allowed`. ==== Supported media types - `application/hal+json` - `application/json`.
[[search]] = Search == The SearchContent Resource When a Store extending `Searchable` is exported, a `searchContent` endpoint will be available at the `/{store}/searchContent` URI. ==== [source, sh] ---- curl -H 'Accept: application/hal+json' http://localhost:8080/searchContent?queryString=foo ---- ==== === Supported HTTP Methods As the SearchContent resource is read-only it supports `GET` only. All other HTTP methods will cause a `405 Method Not Allowed`. ==== Supported media types - `application/hal+json` - `application/json`. === Format of the response payload This resource can return entities, or a custom search result type, depending on how the Searchable interface is specified and the type of Store it decorates. When Searchable decorates an AssociativeStore this resource will lookup and return representations of the content's associated entities. These lookups can be made more efficient by specifying a `@FulltextEntityLookupQuery` query method. This is a custom `findAll` method that accepts a single `Collection` parameter annotated with the name `contentIds`, as follows: ``` public interface MyRepository extends CrudRepository<MyEntity, Long> { @FulltextEntityLookupQuery List<MyEntity> findAllByContentIdIn(@Param("contentIds") List<UUID> contentIds); } public interface MyStore extends AssociativeStore<MyEntity, UUID>, Searchable<UUID> {} ``` When Searchable is typed to your own search return type the resource will return a representation of this type instead. See `Search Return Types` in the respective documentation for your chosen Spring Content fulltext module; solr or elasticsearch, for more information on specifying a custom search return types.
Add documentation explaining the format of the searchContent response payload
Add documentation explaining the format of the searchContent response payload
AsciiDoc
apache-2.0
paulcwarren/spring-content,paulcwarren/spring-content,paulcwarren/spring-content
7c9da1be4bb85002515de4a46fb0f48d11c2410c
modules/adding-custom-notification-banners.adoc
modules/adding-custom-notification-banners.adoc
// Module included in the following assemblies: // // * web_console/customizing-the-web-console.adoc [id="creating-custom-notification-banners_{context}"] = Creating custom notification banners .Prerequisites * You must have administrator privileges. .Procedure . From *Administration* -> *Custom Resource Definitions*, click on *ConsoleNotification*. . Click *YAML* and edit the file: + ---- apiVersion: console.openshift.io/v1 kind: ConsoleNotification metadata: name: example spec: backgroundColor: '#0088ce' color: '#fff' link: href: 'https://www.example.com' text: Optional link text location: BannerTop <1> text: This is an example notification message with an optional link. ---- <1> Valid location settings are `BannerTop`, `BannerBottom`, and `BannerTopBottom`. . Click the *Save* button to apply your changes.
// Module included in the following assemblies: // // * web_console/customizing-the-web-console.adoc [id="creating-custom-notification-banners_{context}"] = Creating custom notification banners .Prerequisites * You must have administrator privileges. .Procedure . From *Administration* -> *Custom Resource Definitions*, click on *ConsoleNotification*. . Select *Instances* tab . Click *Create Console Notification* and edit the file: + ---- apiVersion: console.openshift.io/v1 kind: ConsoleNotification metadata: name: example spec: text: This is an example notification message with an optional link. location: BannerTop <1> link: href: 'https://www.example.com' text: Optional link text color: '#fff' backgroundColor: '#0088ce' ---- <1> Valid location settings are `BannerTop`, `BannerBottom`, and `BannerTopBottom`. . Click the *Create* button to apply your changes.
Fix navigation path and CRD format in creating-custom-notification-banners
Fix navigation path and CRD format in creating-custom-notification-banners
AsciiDoc
apache-2.0
vikram-redhat/openshift-docs,vikram-redhat/openshift-docs,vikram-redhat/openshift-docs,vikram-redhat/openshift-docs,vikram-redhat/openshift-docs
b491eccf9891ea00adbfc25ee2ba8d818c94aab0
storage/persistent_storage/persistent-storage-fibre.adoc
storage/persistent_storage/persistent-storage-fibre.adoc
[id="persistent-storage-using-fibre"] = Persistent storage using Fibre Channel include::modules/common-attributes.adoc[] :context: persistent-storage-fibre toc::[] {product-title} supports Fibre Channel, allowing you to provision your {product-title} cluster with persistent storage using Fibre channel volumes. Some familiarity with Kubernetes and Fibre Channel is assumed. The Kubernetes persistent volume framework allows administrators to provision a cluster with persistent storage and gives users a way to request those resources without having any knowledge of the underlying infrastructure. Persistent volumes are not bound to a single project or namespace; they can be shared across the {product-title} cluster. Persistent volume claims are specific to a project or namespace and can be requested by users. [IMPORTANT] ==== High availability of storage in the infrastructure is left to the underlying storage provider. ==== .Additional resources * link:https://access.redhat.com/documentation/en-us/red_hat_enterprise_linux/7/html/storage_administration_guide/ch-fibrechanel[Fibre Channel] include::modules/persistent-storage-fibre-provisioning.adoc[leveloffset=+1] include::modules/persistent-storage-fibre-disk-quotas.adoc[leveloffset=+2] include::modules/persistent-storage-fibre-volume-security.adoc[leveloffset=+2]
[id="persistent-storage-using-fibre"] = Persistent storage using Fibre Channel include::modules/common-attributes.adoc[] :context: persistent-storage-fibre toc::[] {product-title} supports Fibre Channel, allowing you to provision your {product-title} cluster with persistent storage using Fibre channel volumes. Some familiarity with Kubernetes and Fibre Channel is assumed. The Kubernetes persistent volume framework allows administrators to provision a cluster with persistent storage and gives users a way to request those resources without having any knowledge of the underlying infrastructure. Persistent volumes are not bound to a single project or namespace; they can be shared across the {product-title} cluster. Persistent volume claims are specific to a project or namespace and can be requested by users. [IMPORTANT] ==== High availability of storage in the infrastructure is left to the underlying storage provider. ==== .Additional resources * link:https://access.redhat.com/documentation/en-us/red_hat_enterprise_linux/8/html/managing_storage_devices/using-fibre-channel-devices_managing-storage-devices[Using Fibre Channel devices] include::modules/persistent-storage-fibre-provisioning.adoc[leveloffset=+1] include::modules/persistent-storage-fibre-disk-quotas.adoc[leveloffset=+2] include::modules/persistent-storage-fibre-volume-security.adoc[leveloffset=+2]
Update FC link to RHEL8 doc
Update FC link to RHEL8 doc
AsciiDoc
apache-2.0
vikram-redhat/openshift-docs,vikram-redhat/openshift-docs,vikram-redhat/openshift-docs,vikram-redhat/openshift-docs,vikram-redhat/openshift-docs