prompt stringlengths 77 6.9k | completion stringlengths 1 3.38k | text stringlengths 163 9.03k | commit stringlengths 40 40 | old_file stringlengths 4 264 | new_file stringlengths 4 264 | lang stringclasses 277
values |
|---|---|---|---|---|---|---|
<|file_sep|>original/lib/SimpleSAML/Error/UnserializableException.php
<|file_sep|>current/lib/SimpleSAML/Error/UnserializableException.php
<|file_sep|>updated/lib/SimpleSAML/Error/UnserializableException.php | <?php
/**
* Class for saving normal exceptions for serialization.
*
* This class is used by the SimpleSAML_Auth_State class when it needs
* to serialize an exception which doesn't subclass the
* SimpleSAML_Error_Exception class.
*
* It creates a new exception which contains the backtrace and message
* of the original exception.
*
* @package simpleSAMLphp
* @version $Id$
*/
class SimpleSAML_Error_UnserializableException extends SimpleSAML_Error_Exception {
public function __construct(Exception $original) {
$msg = get_class($original) . ': ' . $original->getMessage();
$code = $original->getCode(); | <|file_sep|>original/lib/SimpleSAML/Error/UnserializableException.php
<|file_sep|>current/lib/SimpleSAML/Error/UnserializableException.php
<|file_sep|>updated/lib/SimpleSAML/Error/UnserializableException.php
<?php
/**
* Class for saving normal exceptions for serialization.
*
* This class is used by the SimpleSAML_Auth_State class when it needs
* to serialize an exception which doesn't subclass the
* SimpleSAML_Error_Exception class.
*
* It creates a new exception which contains the backtrace and message
* of the original exception.
*
* @package simpleSAMLphp
* @version $Id$
*/
class SimpleSAML_Error_UnserializableException extends SimpleSAML_Error_Exception {
public function __construct(Exception $original) {
$msg = get_class($original) . ': ' . $original->getMessage();
$code = $original->getCode(); | c2d23ab92293d8059943b5cbaa3b7a2b6d49266b | lib/SimpleSAML/Error/UnserializableException.php | lib/SimpleSAML/Error/UnserializableException.php | PHP |
<|file_sep|>pkgs/development/interpreters/starlark/default.nix.diff
original:
{ stdenv, lib, fetchFromGitHub, buildGoModule }:
updated:
{ stdenv, lib, fetchFromGitHub, buildGoModule, fetchpatch }:
<|file_sep|>pkgs/development/interpreters/starlark/default.nix.diff
original:
updated:
patches = [
# Fix floating point imprecision issue in the test suite.
# https://github.com/google/starlark-go/pull/409
(fetchpatch {
url = "https://github.com/google/starlark-go/commit/be6ed3bfcc376e5bf6fe2257ae89ddfb00d14e2c.patch";
sha256 = "sha256-A0tHPso6SfFn73kICcA9/5n3JHd7hMdQMGty+4L6T4k=";
})
];
<|file_sep|>original/pkgs/development/interpreters/starlark/default.nix
version = "unstable-2022-03-02";
src = fetchFromGitHub {
owner = "google";
repo = "starlark-go";
rev = "5411bad688d12781515a91cc032645331b4fc302";
sha256 = "sha256-JNsGyGlIVMS5w0W4jHVsrPqqNms3Xfpa4n/XcEWqt6I=";
};
vendorSha256 = "sha256-lgL5o3MQfZekZ++BNESwV0LeoTxwEZfziQAe99zm4RY=";
ldflags = [ "-s" "-w" ];
meta = with lib; {
broken = (stdenv.isLinux && stdenv.isAarch64);
homepage = "https://github.com/google/starlark-go";
description = "An interpreter for Starlark, implemented in Go";
license = licenses.bsd3;
maintainers = with maintainers; [ aaronjheng ];
};
}
<|file_sep|>current/pkgs/development/interpreters/starlark/default.nix
vendorSha256 = "sha256-lgL5o3MQfZekZ++BNESwV0LeoTxwEZfziQAe99zm4RY=";
patches = [
# Fix floating point imprecision issue in the test suite.
# https://github.com/google/starlark-go/pull/409
(fetchpatch {
url = "https://github.com/google/starlark-go/commit/be6ed3bfcc376e5bf6fe2257ae89ddfb00d14e2c.patch";
sha256 = "sha256-A0tHPso6SfFn73kICcA9/5n3JHd7hMdQMGty+4L6T4k=";
})
];
ldflags = [ "-s" "-w" ];
meta = with lib; {
broken = (stdenv.isLinux && stdenv.isAarch64);
homepage = "https://github.com/google/starlark-go";
description = "An interpreter for Starlark, implemented in Go";
license = licenses.bsd3;
maintainers = with maintainers; [ aaronjheng ];
};
}
<|file_sep|>updated/pkgs/development/interpreters/starlark/default.nix |
vendorSha256 = "sha256-lgL5o3MQfZekZ++BNESwV0LeoTxwEZfziQAe99zm4RY=";
patches = [
# Fix floating point imprecision issue in the test suite.
# https://github.com/google/starlark-go/pull/409
(fetchpatch {
url = "https://github.com/google/starlark-go/commit/be6ed3bfcc376e5bf6fe2257ae89ddfb00d14e2c.patch";
sha256 = "sha256-A0tHPso6SfFn73kICcA9/5n3JHd7hMdQMGty+4L6T4k=";
})
];
ldflags = [ "-s" "-w" ];
meta = with lib; {
homepage = "https://github.com/google/starlark-go";
description = "An interpreter for Starlark, implemented in Go";
license = licenses.bsd3;
maintainers = with maintainers; [ aaronjheng ];
};
} | <|file_sep|>pkgs/development/interpreters/starlark/default.nix.diff
original:
{ stdenv, lib, fetchFromGitHub, buildGoModule }:
updated:
{ stdenv, lib, fetchFromGitHub, buildGoModule, fetchpatch }:
<|file_sep|>pkgs/development/interpreters/starlark/default.nix.diff
original:
updated:
patches = [
# Fix floating point imprecision issue in the test suite.
# https://github.com/google/starlark-go/pull/409
(fetchpatch {
url = "https://github.com/google/starlark-go/commit/be6ed3bfcc376e5bf6fe2257ae89ddfb00d14e2c.patch";
sha256 = "sha256-A0tHPso6SfFn73kICcA9/5n3JHd7hMdQMGty+4L6T4k=";
})
];
<|file_sep|>original/pkgs/development/interpreters/starlark/default.nix
version = "unstable-2022-03-02";
src = fetchFromGitHub {
owner = "google";
repo = "starlark-go";
rev = "5411bad688d12781515a91cc032645331b4fc302";
sha256 = "sha256-JNsGyGlIVMS5w0W4jHVsrPqqNms3Xfpa4n/XcEWqt6I=";
};
vendorSha256 = "sha256-lgL5o3MQfZekZ++BNESwV0LeoTxwEZfziQAe99zm4RY=";
ldflags = [ "-s" "-w" ];
meta = with lib; {
broken = (stdenv.isLinux && stdenv.isAarch64);
homepage = "https://github.com/google/starlark-go";
description = "An interpreter for Starlark, implemented in Go";
license = licenses.bsd3;
maintainers = with maintainers; [ aaronjheng ];
};
}
<|file_sep|>current/pkgs/development/interpreters/starlark/default.nix
vendorSha256 = "sha256-lgL5o3MQfZekZ++BNESwV0LeoTxwEZfziQAe99zm4RY=";
patches = [
# Fix floating point imprecision issue in the test suite.
# https://github.com/google/starlark-go/pull/409
(fetchpatch {
url = "https://github.com/google/starlark-go/commit/be6ed3bfcc376e5bf6fe2257ae89ddfb00d14e2c.patch";
sha256 = "sha256-A0tHPso6SfFn73kICcA9/5n3JHd7hMdQMGty+4L6T4k=";
})
];
ldflags = [ "-s" "-w" ];
meta = with lib; {
broken = (stdenv.isLinux && stdenv.isAarch64);
homepage = "https://github.com/google/starlark-go";
description = "An interpreter for Starlark, implemented in Go";
license = licenses.bsd3;
maintainers = with maintainers; [ aaronjheng ];
};
}
<|file_sep|>updated/pkgs/development/interpreters/starlark/default.nix
vendorSha256 = "sha256-lgL5o3MQfZekZ++BNESwV0LeoTxwEZfziQAe99zm4RY=";
patches = [
# Fix floating point imprecision issue in the test suite.
# https://github.com/google/starlark-go/pull/409
(fetchpatch {
url = "https://github.com/google/starlark-go/commit/be6ed3bfcc376e5bf6fe2257ae89ddfb00d14e2c.patch";
sha256 = "sha256-A0tHPso6SfFn73kICcA9/5n3JHd7hMdQMGty+4L6T4k=";
})
];
ldflags = [ "-s" "-w" ];
meta = with lib; {
homepage = "https://github.com/google/starlark-go";
description = "An interpreter for Starlark, implemented in Go";
license = licenses.bsd3;
maintainers = with maintainers; [ aaronjheng ];
};
} | 699a374e8a24edbd3296240f6da5d454e7fb3c21 | pkgs/development/interpreters/starlark/default.nix | pkgs/development/interpreters/starlark/default.nix | Nix |
<|file_sep|>original/src/main/groovy/com/bmuschko/gradle/docker/utils/IOUtils.groovy
}
} catch (IOException ignored) {
// ignore
}
}
/**
* Create a progress logger for an arbitrary project and class.
*
* @param project the project to create a ProgressLogger for.
* @param clazz optional class to pair the ProgressLogger to. Defaults to _this_ class if null.
* @return instance of ProgressLogger.
*/
static ProgressLogger getProgressLogger(final Project project, final Class clazz) {
ServiceRegistry registry = (project.gradle as GradleInternal).getServices()
ProgressLoggerFactory factory = registry.get(ProgressLoggerFactory)
ProgressLogger progressLogger = factory.newOperation(Objects.requireNonNull(clazz))
progressLogger.setDescription("ProgressLogger for ${clazz.getSimpleName()}")
progressLogger.setLoggingHeader(null)
}
}
<|file_sep|>current/src/main/groovy/com/bmuschko/gradle/docker/utils/IOUtils.groovy
}
} catch (IOException ignored) {
// ignore
}
}
/**
* Create a progress logger for an arbitrary project and class.
*
* @param project the project to create a ProgressLogger for.
* @param clazz optional class to pair the ProgressLogger to. Defaults to _this_ class if null.
* @return instance of ProgressLogger.
*/
static ProgressLogger getProgressLogger(final Project project, final Class clazz) {
ServiceRegistry registry = (project.gradle as GradleInternal).getServices()
ProgressLoggerFactory factory = registry.get(ProgressLoggerFactory)
ProgressLogger progressLogger = factory.newOperation(Objects.requireNonNull(clazz))
progressLogger.setDescription("ProgressLogger for ${clazz.getSimpleName()}")
progressLogger.setLoggingHeader(null)
}
}
<|file_sep|>updated/src/main/groovy/com/bmuschko/gradle/docker/utils/IOUtils.groovy | toClose.close()
}
} catch (IOException ignored) {
// ignore
}
}
/**
* Create a progress logger for an arbitrary project and class.
*
* @param project the project to create a ProgressLogger for.
* @param clazz optional class to pair the ProgressLogger to. Defaults to _this_ class if null.
* @return instance of ProgressLogger.
*/
static ProgressLogger getProgressLogger(final Project project, final Class clazz) {
ServiceRegistry registry = (project.gradle as GradleInternal).getServices()
ProgressLoggerFactory factory = registry.get(ProgressLoggerFactory)
ProgressLogger progressLogger = factory.newOperation(Objects.requireNonNull(clazz))
progressLogger.setDescription("ProgressLogger for ${clazz.getSimpleName()}")
}
} | <|file_sep|>original/src/main/groovy/com/bmuschko/gradle/docker/utils/IOUtils.groovy
}
} catch (IOException ignored) {
// ignore
}
}
/**
* Create a progress logger for an arbitrary project and class.
*
* @param project the project to create a ProgressLogger for.
* @param clazz optional class to pair the ProgressLogger to. Defaults to _this_ class if null.
* @return instance of ProgressLogger.
*/
static ProgressLogger getProgressLogger(final Project project, final Class clazz) {
ServiceRegistry registry = (project.gradle as GradleInternal).getServices()
ProgressLoggerFactory factory = registry.get(ProgressLoggerFactory)
ProgressLogger progressLogger = factory.newOperation(Objects.requireNonNull(clazz))
progressLogger.setDescription("ProgressLogger for ${clazz.getSimpleName()}")
progressLogger.setLoggingHeader(null)
}
}
<|file_sep|>current/src/main/groovy/com/bmuschko/gradle/docker/utils/IOUtils.groovy
}
} catch (IOException ignored) {
// ignore
}
}
/**
* Create a progress logger for an arbitrary project and class.
*
* @param project the project to create a ProgressLogger for.
* @param clazz optional class to pair the ProgressLogger to. Defaults to _this_ class if null.
* @return instance of ProgressLogger.
*/
static ProgressLogger getProgressLogger(final Project project, final Class clazz) {
ServiceRegistry registry = (project.gradle as GradleInternal).getServices()
ProgressLoggerFactory factory = registry.get(ProgressLoggerFactory)
ProgressLogger progressLogger = factory.newOperation(Objects.requireNonNull(clazz))
progressLogger.setDescription("ProgressLogger for ${clazz.getSimpleName()}")
progressLogger.setLoggingHeader(null)
}
}
<|file_sep|>updated/src/main/groovy/com/bmuschko/gradle/docker/utils/IOUtils.groovy
toClose.close()
}
} catch (IOException ignored) {
// ignore
}
}
/**
* Create a progress logger for an arbitrary project and class.
*
* @param project the project to create a ProgressLogger for.
* @param clazz optional class to pair the ProgressLogger to. Defaults to _this_ class if null.
* @return instance of ProgressLogger.
*/
static ProgressLogger getProgressLogger(final Project project, final Class clazz) {
ServiceRegistry registry = (project.gradle as GradleInternal).getServices()
ProgressLoggerFactory factory = registry.get(ProgressLoggerFactory)
ProgressLogger progressLogger = factory.newOperation(Objects.requireNonNull(clazz))
progressLogger.setDescription("ProgressLogger for ${clazz.getSimpleName()}")
}
} | c0ae71b52281cdd107aacf56286af97d2b4b9d6b | src/main/groovy/com/bmuschko/gradle/docker/utils/IOUtils.groovy | src/main/groovy/com/bmuschko/gradle/docker/utils/IOUtils.groovy | Groovy |
<|file_sep|>original/app.json
"description":"Send and receive messages from a Habitica Party Chat to a Slack Channel",
"repository":"https://github.com/niteshpatel/habitica-slack",
"keywords":[
"habitica",
"slack"
],
"env":{
"DJANGO_SECRETKEY":{
"description":"Secret key for your Django instance (autogenerated).",
"generator":"secret"
},
"HABITICA_APIUSER":{
"description":"Your Habitica User ID (found here https://habitica.com/#/options/settings/api)."
},
"HABITICA_APIKEY":{
"description":"Your Habitica API Token (found here https://habitica.com/#/options/settings/api)."
},
"HABITICA_GROUPID":{
"description":"Your Habitica Group ID (found here https://habitica.com/#/options/groups/party)."
},
"SLACK_TOKEN":{
<|file_sep|>current/app.json
"description":"Send and receive messages from a Habitica Party Chat to a Slack Channel",
"repository":"https://github.com/niteshpatel/habitica-slack",
"keywords":[
"habitica",
"slack"
],
"env":{
"DJANGO_SECRETKEY":{
"description":"Secret key for your Django instance (autogenerated).",
"generator":"secret"
},
"HABITICA_APIUSER":{
"description":"Your Habitica User ID (found here https://habitica.com/#/options/settings/api)."
},
"HABITICA_APIKEY":{
"description":"Your Habitica API Token (found here https://habitica.com/#/options/settings/api)."
},
"HABITICA_GROUPID":{
"description":"Your Habitica Group ID (found here https://habitica.com/#/options/groups/party)."
},
"SLACK_TOKEN":{
<|file_sep|>updated/app.json | "description": "Send and receive messages from a Habitica Party Chat to a Slack Channel",
"repository": "https://github.com/niteshpatel/habitica-slack",
"keywords": [
"habitica",
"slack"
],
"env": {
"DJANGO_SECRETKEY": {
"description": "Secret key for your Django instance (autogenerated).",
"generator": "secret"
},
"HABITICA_APIUSER": {
"description": "Your Habitica User ID (found here https://habitica.com/#/options/settings/api)."
},
"HABITICA_APIKEY": {
"description": "Your Habitica API Token (found here https://habitica.com/#/options/settings/api)."
},
"HABITICA_GROUPID": {
"description": "Your Habitica Group ID (found here https://habitica.com/#/options/groups/party)."
},
"SLACK_TOKEN": { | <|file_sep|>original/app.json
"description":"Send and receive messages from a Habitica Party Chat to a Slack Channel",
"repository":"https://github.com/niteshpatel/habitica-slack",
"keywords":[
"habitica",
"slack"
],
"env":{
"DJANGO_SECRETKEY":{
"description":"Secret key for your Django instance (autogenerated).",
"generator":"secret"
},
"HABITICA_APIUSER":{
"description":"Your Habitica User ID (found here https://habitica.com/#/options/settings/api)."
},
"HABITICA_APIKEY":{
"description":"Your Habitica API Token (found here https://habitica.com/#/options/settings/api)."
},
"HABITICA_GROUPID":{
"description":"Your Habitica Group ID (found here https://habitica.com/#/options/groups/party)."
},
"SLACK_TOKEN":{
<|file_sep|>current/app.json
"description":"Send and receive messages from a Habitica Party Chat to a Slack Channel",
"repository":"https://github.com/niteshpatel/habitica-slack",
"keywords":[
"habitica",
"slack"
],
"env":{
"DJANGO_SECRETKEY":{
"description":"Secret key for your Django instance (autogenerated).",
"generator":"secret"
},
"HABITICA_APIUSER":{
"description":"Your Habitica User ID (found here https://habitica.com/#/options/settings/api)."
},
"HABITICA_APIKEY":{
"description":"Your Habitica API Token (found here https://habitica.com/#/options/settings/api)."
},
"HABITICA_GROUPID":{
"description":"Your Habitica Group ID (found here https://habitica.com/#/options/groups/party)."
},
"SLACK_TOKEN":{
<|file_sep|>updated/app.json
"description": "Send and receive messages from a Habitica Party Chat to a Slack Channel",
"repository": "https://github.com/niteshpatel/habitica-slack",
"keywords": [
"habitica",
"slack"
],
"env": {
"DJANGO_SECRETKEY": {
"description": "Secret key for your Django instance (autogenerated).",
"generator": "secret"
},
"HABITICA_APIUSER": {
"description": "Your Habitica User ID (found here https://habitica.com/#/options/settings/api)."
},
"HABITICA_APIKEY": {
"description": "Your Habitica API Token (found here https://habitica.com/#/options/settings/api)."
},
"HABITICA_GROUPID": {
"description": "Your Habitica Group ID (found here https://habitica.com/#/options/groups/party)."
},
"SLACK_TOKEN": { | 9728be88389c1ff45fd79413bb3751f2b74e9c50 | app.json | app.json | JSON |
<|file_sep|>original/admin/app/views/transactions/forms/_delete.html.erb
<%= form_tag("/generate_delete_transaction", remote: true) do %>
<%= text_field_tag 'public_key', creator_address, hidden: true %>
<%= text_field_tag 'payload', payload, hidden: true %>
<%= text_field_tag 'priv_key', private_key, hidden: true %>
<%= submit_tag submit_name, class: "btn btn-default", data: { disable_with: "#{submit_name.eql?('Delete') ? 'Deleting...' : 'Accepting...'}" } %>
<% end %>
<|file_sep|>current/admin/app/views/transactions/forms/_delete.html.erb
<%= form_tag("/generate_delete_transaction", remote: true) do %>
<%= text_field_tag 'public_key', creator_address, hidden: true %>
<%= text_field_tag 'payload', payload, hidden: true %>
<%= text_field_tag 'priv_key', private_key, hidden: true %>
<%= submit_tag submit_name, class: "btn btn-default", data: { disable_with: "#{submit_name.eql?('Delete') ? 'Deleting...' : 'Accepting...'}" } %>
<% end %>
<|file_sep|>updated/admin/app/views/transactions/forms/_delete.html.erb | <%= form_tag("/generate_delete_transaction", remote: true) do %>
<%= text_field_tag 'public_key', creator_address, hidden: true %>
<%= text_field_tag 'payload', payload, hidden: true %>
<%= text_field_tag 'priv_key', private_key, hidden: true %>
<%= submit_tag submit_name, class: "btn btn-default", data: { disable_with: "#{submit_name.eql?('Accept') ? 'Accepting...' : 'Cancelling...'}" } %>
<% end %> | <|file_sep|>original/admin/app/views/transactions/forms/_delete.html.erb
<%= form_tag("/generate_delete_transaction", remote: true) do %>
<%= text_field_tag 'public_key', creator_address, hidden: true %>
<%= text_field_tag 'payload', payload, hidden: true %>
<%= text_field_tag 'priv_key', private_key, hidden: true %>
<%= submit_tag submit_name, class: "btn btn-default", data: { disable_with: "#{submit_name.eql?('Delete') ? 'Deleting...' : 'Accepting...'}" } %>
<% end %>
<|file_sep|>current/admin/app/views/transactions/forms/_delete.html.erb
<%= form_tag("/generate_delete_transaction", remote: true) do %>
<%= text_field_tag 'public_key', creator_address, hidden: true %>
<%= text_field_tag 'payload', payload, hidden: true %>
<%= text_field_tag 'priv_key', private_key, hidden: true %>
<%= submit_tag submit_name, class: "btn btn-default", data: { disable_with: "#{submit_name.eql?('Delete') ? 'Deleting...' : 'Accepting...'}" } %>
<% end %>
<|file_sep|>updated/admin/app/views/transactions/forms/_delete.html.erb
<%= form_tag("/generate_delete_transaction", remote: true) do %>
<%= text_field_tag 'public_key', creator_address, hidden: true %>
<%= text_field_tag 'payload', payload, hidden: true %>
<%= text_field_tag 'priv_key', private_key, hidden: true %>
<%= submit_tag submit_name, class: "btn btn-default", data: { disable_with: "#{submit_name.eql?('Accept') ? 'Accepting...' : 'Cancelling...'}" } %>
<% end %> | fe87adc4d4567f2c162a5bffb97f21c7d819550c | admin/app/views/transactions/forms/_delete.html.erb | admin/app/views/transactions/forms/_delete.html.erb | HTML+ERB |
<|file_sep|>original/common_audio/signal_processing/cross_correlation.c
/* C version of WebRtcSpl_CrossCorrelation() for generic platforms. */
void WebRtcSpl_CrossCorrelationC(int32_t* cross_correlation,
const int16_t* seq1,
const int16_t* seq2,
int16_t dim_seq,
int16_t dim_cross_correlation,
int right_shifts,
int step_seq2) {
int i = 0, j = 0;
for (i = 0; i < dim_cross_correlation; i++) {
int32_t corr = 0;
/* Unrolling doesn't seem to improve performance. */
for (j = 0; j < dim_seq; j++) {
// It's not clear why casting |right_shifts| here helps performance.
corr += (seq1[j] * seq2[j]) >> (int16_t)right_shifts;
}
seq2 += step_seq2;
*cross_correlation++ = corr;
}
}
<|file_sep|>current/common_audio/signal_processing/cross_correlation.c
/* C version of WebRtcSpl_CrossCorrelation() for generic platforms. */
void WebRtcSpl_CrossCorrelationC(int32_t* cross_correlation,
const int16_t* seq1,
const int16_t* seq2,
int16_t dim_seq,
int16_t dim_cross_correlation,
int right_shifts,
int step_seq2) {
int i = 0, j = 0;
for (i = 0; i < dim_cross_correlation; i++) {
int32_t corr = 0;
/* Unrolling doesn't seem to improve performance. */
for (j = 0; j < dim_seq; j++) {
// It's not clear why casting |right_shifts| here helps performance.
corr += (seq1[j] * seq2[j]) >> (int16_t)right_shifts;
}
seq2 += step_seq2;
*cross_correlation++ = corr;
}
}
<|file_sep|>updated/common_audio/signal_processing/cross_correlation.c | #include "webrtc/common_audio/signal_processing/include/signal_processing_library.h"
/* C version of WebRtcSpl_CrossCorrelation() for generic platforms. */
void WebRtcSpl_CrossCorrelationC(int32_t* cross_correlation,
const int16_t* seq1,
const int16_t* seq2,
int16_t dim_seq,
int16_t dim_cross_correlation,
int right_shifts,
int step_seq2) {
int i = 0, j = 0;
for (i = 0; i < dim_cross_correlation; i++) {
int32_t corr = 0;
/* Unrolling doesn't seem to improve performance. */
for (j = 0; j < dim_seq; j++)
corr += (seq1[j] * seq2[j]) >> right_shifts;
seq2 += step_seq2;
*cross_correlation++ = corr;
}
} | <|file_sep|>original/common_audio/signal_processing/cross_correlation.c
/* C version of WebRtcSpl_CrossCorrelation() for generic platforms. */
void WebRtcSpl_CrossCorrelationC(int32_t* cross_correlation,
const int16_t* seq1,
const int16_t* seq2,
int16_t dim_seq,
int16_t dim_cross_correlation,
int right_shifts,
int step_seq2) {
int i = 0, j = 0;
for (i = 0; i < dim_cross_correlation; i++) {
int32_t corr = 0;
/* Unrolling doesn't seem to improve performance. */
for (j = 0; j < dim_seq; j++) {
// It's not clear why casting |right_shifts| here helps performance.
corr += (seq1[j] * seq2[j]) >> (int16_t)right_shifts;
}
seq2 += step_seq2;
*cross_correlation++ = corr;
}
}
<|file_sep|>current/common_audio/signal_processing/cross_correlation.c
/* C version of WebRtcSpl_CrossCorrelation() for generic platforms. */
void WebRtcSpl_CrossCorrelationC(int32_t* cross_correlation,
const int16_t* seq1,
const int16_t* seq2,
int16_t dim_seq,
int16_t dim_cross_correlation,
int right_shifts,
int step_seq2) {
int i = 0, j = 0;
for (i = 0; i < dim_cross_correlation; i++) {
int32_t corr = 0;
/* Unrolling doesn't seem to improve performance. */
for (j = 0; j < dim_seq; j++) {
// It's not clear why casting |right_shifts| here helps performance.
corr += (seq1[j] * seq2[j]) >> (int16_t)right_shifts;
}
seq2 += step_seq2;
*cross_correlation++ = corr;
}
}
<|file_sep|>updated/common_audio/signal_processing/cross_correlation.c
#include "webrtc/common_audio/signal_processing/include/signal_processing_library.h"
/* C version of WebRtcSpl_CrossCorrelation() for generic platforms. */
void WebRtcSpl_CrossCorrelationC(int32_t* cross_correlation,
const int16_t* seq1,
const int16_t* seq2,
int16_t dim_seq,
int16_t dim_cross_correlation,
int right_shifts,
int step_seq2) {
int i = 0, j = 0;
for (i = 0; i < dim_cross_correlation; i++) {
int32_t corr = 0;
/* Unrolling doesn't seem to improve performance. */
for (j = 0; j < dim_seq; j++)
corr += (seq1[j] * seq2[j]) >> right_shifts;
seq2 += step_seq2;
*cross_correlation++ = corr;
}
} | 8504c7c838219029c7e92818f663c777b8bbe5b6 | common_audio/signal_processing/cross_correlation.c | common_audio/signal_processing/cross_correlation.c | C |
<|file_sep|>original/roles/simple/graphviz/tasks/main.yml
- name: Install graphviz
homebrew:
name: graphviz
install_options: with-app
state: present
when: ansible_distribution == "MacOSX"
- name: Install graphviz
become: yes
apt:
name: graphviz
state: present
when: ansible_os_family == "Debian"
- name: Install graphviz
become: yes
pkgng:
name: graphviz
state: present
when: ansible_distribution == "FreeBSD"
<|file_sep|>current/roles/simple/graphviz/tasks/main.yml
- name: Install graphviz
homebrew:
name: graphviz
install_options: with-app
state: present
when: ansible_distribution == "MacOSX"
- name: Install graphviz
become: yes
apt:
name: graphviz
state: present
when: ansible_os_family == "Debian"
- name: Install graphviz
become: yes
pkgng:
name: graphviz
state: present
when: ansible_distribution == "FreeBSD"
<|file_sep|>updated/roles/simple/graphviz/tasks/main.yml | - name: Install graphviz
homebrew:
name: graphviz
install_options: with-app, with-bindings
state: present
when: ansible_distribution == "MacOSX"
- name: Install graphviz
become: yes
apt:
name: graphviz
state: present
when: ansible_os_family == "Debian"
- name: Install graphviz
become: yes
pkgng:
name: graphviz
state: present
when: ansible_distribution == "FreeBSD" | <|file_sep|>original/roles/simple/graphviz/tasks/main.yml
- name: Install graphviz
homebrew:
name: graphviz
install_options: with-app
state: present
when: ansible_distribution == "MacOSX"
- name: Install graphviz
become: yes
apt:
name: graphviz
state: present
when: ansible_os_family == "Debian"
- name: Install graphviz
become: yes
pkgng:
name: graphviz
state: present
when: ansible_distribution == "FreeBSD"
<|file_sep|>current/roles/simple/graphviz/tasks/main.yml
- name: Install graphviz
homebrew:
name: graphviz
install_options: with-app
state: present
when: ansible_distribution == "MacOSX"
- name: Install graphviz
become: yes
apt:
name: graphviz
state: present
when: ansible_os_family == "Debian"
- name: Install graphviz
become: yes
pkgng:
name: graphviz
state: present
when: ansible_distribution == "FreeBSD"
<|file_sep|>updated/roles/simple/graphviz/tasks/main.yml
- name: Install graphviz
homebrew:
name: graphviz
install_options: with-app, with-bindings
state: present
when: ansible_distribution == "MacOSX"
- name: Install graphviz
become: yes
apt:
name: graphviz
state: present
when: ansible_os_family == "Debian"
- name: Install graphviz
become: yes
pkgng:
name: graphviz
state: present
when: ansible_distribution == "FreeBSD" | d9f054c1a36da903820e6ebbd444ad926408fb8c | roles/simple/graphviz/tasks/main.yml | roles/simple/graphviz/tasks/main.yml | YAML |
<|file_sep|>tests/parser/parserTests.txt.diff
original:
!! Version 2
updated:
!! options
version=2
parsoid-compatible
requirements=extension:Disambiguator
!! end
<|file_sep|>original/tests/parser/parserTests.txt
!! article
Disambig
!! text
This is a disambig page.
__DISAMBIG__
!! endarticle
!! article
Normal
!! text
This is a normal page.
!! endarticle
!! test
Link to disambig and non-disambig page
!! wikitext
I am linking to [[Disambig]] and [[Normal]].
!! html
<p>I am linking to <a href="/wiki/Disambig" class="mw-disambig" title="Disambig">Disambig</a> and <a href="/wiki/Normal" title="Normal">Normal</a>.
</p>
!! end
<|file_sep|>current/tests/parser/parserTests.txt
!! article
Disambig
!! text
This is a disambig page.
__DISAMBIG__
!! endarticle
!! article
Normal
!! text
This is a normal page.
!! endarticle
!! test
Link to disambig and non-disambig page
!! wikitext
I am linking to [[Disambig]] and [[Normal]].
!! html
<p>I am linking to <a href="/wiki/Disambig" class="mw-disambig" title="Disambig">Disambig</a> and <a href="/wiki/Normal" title="Normal">Normal</a>.
</p>
!! end
<|file_sep|>updated/tests/parser/parserTests.txt | This is a normal page.
!! endarticle
!! test
Link to disambig and non-disambig page
!! wikitext
I am linking to [[Disambig]] and [[Normal]].
!! html
<p>I am linking to <a href="/wiki/Disambig" class="mw-disambig" title="Disambig">Disambig</a> and <a href="/wiki/Normal" title="Normal">Normal</a>.
</p>
!! end
!! test
Ensure disambiguation links are marked properly in Parsoid
!! options
parsoid=wt2html
!! wikitext
[[Disambig]]
!! html/parsoid
<p><a rel="mw:WikiLink" href="./Disambig" title="Disambig" class="mw-disambig">Disambig</a></p>
!! end | <|file_sep|>tests/parser/parserTests.txt.diff
original:
!! Version 2
updated:
!! options
version=2
parsoid-compatible
requirements=extension:Disambiguator
!! end
<|file_sep|>original/tests/parser/parserTests.txt
!! article
Disambig
!! text
This is a disambig page.
__DISAMBIG__
!! endarticle
!! article
Normal
!! text
This is a normal page.
!! endarticle
!! test
Link to disambig and non-disambig page
!! wikitext
I am linking to [[Disambig]] and [[Normal]].
!! html
<p>I am linking to <a href="/wiki/Disambig" class="mw-disambig" title="Disambig">Disambig</a> and <a href="/wiki/Normal" title="Normal">Normal</a>.
</p>
!! end
<|file_sep|>current/tests/parser/parserTests.txt
!! article
Disambig
!! text
This is a disambig page.
__DISAMBIG__
!! endarticle
!! article
Normal
!! text
This is a normal page.
!! endarticle
!! test
Link to disambig and non-disambig page
!! wikitext
I am linking to [[Disambig]] and [[Normal]].
!! html
<p>I am linking to <a href="/wiki/Disambig" class="mw-disambig" title="Disambig">Disambig</a> and <a href="/wiki/Normal" title="Normal">Normal</a>.
</p>
!! end
<|file_sep|>updated/tests/parser/parserTests.txt
This is a normal page.
!! endarticle
!! test
Link to disambig and non-disambig page
!! wikitext
I am linking to [[Disambig]] and [[Normal]].
!! html
<p>I am linking to <a href="/wiki/Disambig" class="mw-disambig" title="Disambig">Disambig</a> and <a href="/wiki/Normal" title="Normal">Normal</a>.
</p>
!! end
!! test
Ensure disambiguation links are marked properly in Parsoid
!! options
parsoid=wt2html
!! wikitext
[[Disambig]]
!! html/parsoid
<p><a rel="mw:WikiLink" href="./Disambig" title="Disambig" class="mw-disambig">Disambig</a></p>
!! end | f9e5a56bf8d41d90b3dd220e7c8497a56c8f62ce | tests/parser/parserTests.txt | tests/parser/parserTests.txt | Text |
<|file_sep|>original/assets/be/views/elements/dataTableForm.html
<div class="detail" ng-controller="FormEditController as fec">
<form ng-submit="fec.onSubmit()" novalidate class="form-horizontal" name="fec.edit" ng-init='fec.init(row);'>
<formly-form model="fec.model" fields="fec.fields" options="fec.options" form="fec.form" editable-form onaftersave="fec.onSubmit()" root-el="form">
<div class="form-submit-bouton col-xs-6 col-xs-push-2" ng-if="fec.conditionEditDetail(row)">
<!-- button to show form -->
<button type="button" class="btn btn-default" ng-click="fec.editForm()" ng-show="!fec.form.$visible" ng-bind="'UI.FORM_EDIT' | translate"></button>
<!-- buttons to submit / cancel form -->
<span ng-show="fec.form.$visible">
<button type="submit" class="btn btn-primary" ng-disabled="fec.form.$invalid || fec.form.$waiting" ng-bind="'UI.FORM_SAVE' | translate"></button>
<button type="button" class="btn btn-default" ng-disabled="fec.form.$waiting" ng-click="fec.cancelForm()" ng-bind="'UI.FORM_CANCEL' | translate"></button>
</span>
</div>
</formly-form>
</form>
</div>
<|file_sep|>current/assets/be/views/elements/dataTableForm.html
<div class="detail" ng-controller="FormEditController as fec">
<form ng-submit="fec.onSubmit()" novalidate class="form-horizontal" name="fec.edit" ng-init='fec.init(row);'>
<formly-form model="fec.model" fields="fec.fields" options="fec.options" form="fec.form" editable-form onaftersave="fec.onSubmit()" root-el="form">
<div class="form-submit-bouton col-xs-6 col-xs-push-2" ng-if="fec.conditionEditDetail(row)">
<!-- button to show form -->
<button type="button" class="btn btn-default" ng-click="fec.editForm()" ng-show="!fec.form.$visible" ng-bind="'UI.FORM_EDIT' | translate"></button>
<!-- buttons to submit / cancel form -->
<span ng-show="fec.form.$visible">
<button type="submit" class="btn btn-primary" ng-disabled="fec.form.$invalid || fec.form.$waiting" ng-bind="'UI.FORM_SAVE' | translate"></button>
<button type="button" class="btn btn-default" ng-disabled="fec.form.$waiting" ng-click="fec.cancelForm()" ng-bind="'UI.FORM_CANCEL' | translate"></button>
</span>
</div>
</formly-form>
</form>
</div>
<|file_sep|>updated/assets/be/views/elements/dataTableForm.html | <div class="detail" ng-controller="FormEditController as fec">
<form novalidate class="form-horizontal" name="fec.edit" ng-init='fec.init(row);'>
<formly-form model="fec.model" fields="fec.fields" options="fec.options" form="fec.form" editable-form onaftersave="fec.onSubmit()" root-el="form">
<div class="form-submit-bouton col-xs-6 col-xs-push-2" ng-if="fec.conditionEditDetail(row)">
<!-- button to show form -->
<button type="button" class="btn btn-default" ng-click="fec.editForm()" ng-show="!fec.form.$visible" ng-bind="'UI.FORM_EDIT' | translate"></button>
<!-- buttons to submit / cancel form -->
<span ng-show="fec.form.$visible">
<button type="submit" class="btn btn-primary" ng-disabled="fec.form.$invalid || fec.form.$waiting" ng-bind="'UI.FORM_SAVE' | translate"></button>
<button type="button" class="btn btn-default" ng-disabled="fec.form.$waiting" ng-click="fec.cancelForm()" ng-bind="'UI.FORM_CANCEL' | translate"></button>
</span>
</div>
</formly-form>
</form>
</div> | <|file_sep|>original/assets/be/views/elements/dataTableForm.html
<div class="detail" ng-controller="FormEditController as fec">
<form ng-submit="fec.onSubmit()" novalidate class="form-horizontal" name="fec.edit" ng-init='fec.init(row);'>
<formly-form model="fec.model" fields="fec.fields" options="fec.options" form="fec.form" editable-form onaftersave="fec.onSubmit()" root-el="form">
<div class="form-submit-bouton col-xs-6 col-xs-push-2" ng-if="fec.conditionEditDetail(row)">
<!-- button to show form -->
<button type="button" class="btn btn-default" ng-click="fec.editForm()" ng-show="!fec.form.$visible" ng-bind="'UI.FORM_EDIT' | translate"></button>
<!-- buttons to submit / cancel form -->
<span ng-show="fec.form.$visible">
<button type="submit" class="btn btn-primary" ng-disabled="fec.form.$invalid || fec.form.$waiting" ng-bind="'UI.FORM_SAVE' | translate"></button>
<button type="button" class="btn btn-default" ng-disabled="fec.form.$waiting" ng-click="fec.cancelForm()" ng-bind="'UI.FORM_CANCEL' | translate"></button>
</span>
</div>
</formly-form>
</form>
</div>
<|file_sep|>current/assets/be/views/elements/dataTableForm.html
<div class="detail" ng-controller="FormEditController as fec">
<form ng-submit="fec.onSubmit()" novalidate class="form-horizontal" name="fec.edit" ng-init='fec.init(row);'>
<formly-form model="fec.model" fields="fec.fields" options="fec.options" form="fec.form" editable-form onaftersave="fec.onSubmit()" root-el="form">
<div class="form-submit-bouton col-xs-6 col-xs-push-2" ng-if="fec.conditionEditDetail(row)">
<!-- button to show form -->
<button type="button" class="btn btn-default" ng-click="fec.editForm()" ng-show="!fec.form.$visible" ng-bind="'UI.FORM_EDIT' | translate"></button>
<!-- buttons to submit / cancel form -->
<span ng-show="fec.form.$visible">
<button type="submit" class="btn btn-primary" ng-disabled="fec.form.$invalid || fec.form.$waiting" ng-bind="'UI.FORM_SAVE' | translate"></button>
<button type="button" class="btn btn-default" ng-disabled="fec.form.$waiting" ng-click="fec.cancelForm()" ng-bind="'UI.FORM_CANCEL' | translate"></button>
</span>
</div>
</formly-form>
</form>
</div>
<|file_sep|>updated/assets/be/views/elements/dataTableForm.html
<div class="detail" ng-controller="FormEditController as fec">
<form novalidate class="form-horizontal" name="fec.edit" ng-init='fec.init(row);'>
<formly-form model="fec.model" fields="fec.fields" options="fec.options" form="fec.form" editable-form onaftersave="fec.onSubmit()" root-el="form">
<div class="form-submit-bouton col-xs-6 col-xs-push-2" ng-if="fec.conditionEditDetail(row)">
<!-- button to show form -->
<button type="button" class="btn btn-default" ng-click="fec.editForm()" ng-show="!fec.form.$visible" ng-bind="'UI.FORM_EDIT' | translate"></button>
<!-- buttons to submit / cancel form -->
<span ng-show="fec.form.$visible">
<button type="submit" class="btn btn-primary" ng-disabled="fec.form.$invalid || fec.form.$waiting" ng-bind="'UI.FORM_SAVE' | translate"></button>
<button type="button" class="btn btn-default" ng-disabled="fec.form.$waiting" ng-click="fec.cancelForm()" ng-bind="'UI.FORM_CANCEL' | translate"></button>
</span>
</div>
</formly-form>
</form>
</div> | de4202964111fc6aa792122052c19a78d8902e6b | assets/be/views/elements/dataTableForm.html | assets/be/views/elements/dataTableForm.html | HTML |
<|file_sep|>original/bower.json
"dist/neue.css",
"dist/neue.min.js"
],
"dependencies": {
"jquery": "^1.7",
"susy": "2.2.0.rc.2"
},
"devDependencies": {
"jquery": "1.8.3",
"almond": "~0.2.9",
"html5shiv": "~3.7.2",
"respond": "~1.4.2",
"highlightjs": "~8.0.0",
"filament-sticky": "~0.1.4",
"dosomething-modal": "DoSomething/modal#~0.1.0",
"dosomething-validation": "DoSomething/validation#~0.0.3"
},
"resolutions": {
"jquery": "1.8.3"
}
}
<|file_sep|>current/bower.json
"dist/neue.css",
"dist/neue.min.js"
],
"dependencies": {
"jquery": "^1.7",
"susy": "2.2.0.rc.2"
},
"devDependencies": {
"jquery": "1.8.3",
"almond": "~0.2.9",
"html5shiv": "~3.7.2",
"respond": "~1.4.2",
"highlightjs": "~8.0.0",
"filament-sticky": "~0.1.4",
"dosomething-modal": "DoSomething/modal#~0.1.0",
"dosomething-validation": "DoSomething/validation#~0.0.3"
},
"resolutions": {
"jquery": "1.8.3"
}
}
<|file_sep|>updated/bower.json | "dist/neue.css",
"dist/neue.min.js"
],
"dependencies": {
"jquery": "^1.7",
"susy": "2.2.0.rc.2"
},
"devDependencies": {
"jquery": "1.8.3",
"almond": "~0.2.9",
"html5shiv": "~3.7.2",
"respond": "~1.4.2",
"highlightjs": "~8.0.0",
"filament-sticky": "~0.1.4",
"dosomething-modal": "DoSomething/modal#~0.1.0",
"dosomething-validation": "DoSomething/validation#~0.1.0"
},
"resolutions": {
"jquery": "1.8.3"
}
} | <|file_sep|>original/bower.json
"dist/neue.css",
"dist/neue.min.js"
],
"dependencies": {
"jquery": "^1.7",
"susy": "2.2.0.rc.2"
},
"devDependencies": {
"jquery": "1.8.3",
"almond": "~0.2.9",
"html5shiv": "~3.7.2",
"respond": "~1.4.2",
"highlightjs": "~8.0.0",
"filament-sticky": "~0.1.4",
"dosomething-modal": "DoSomething/modal#~0.1.0",
"dosomething-validation": "DoSomething/validation#~0.0.3"
},
"resolutions": {
"jquery": "1.8.3"
}
}
<|file_sep|>current/bower.json
"dist/neue.css",
"dist/neue.min.js"
],
"dependencies": {
"jquery": "^1.7",
"susy": "2.2.0.rc.2"
},
"devDependencies": {
"jquery": "1.8.3",
"almond": "~0.2.9",
"html5shiv": "~3.7.2",
"respond": "~1.4.2",
"highlightjs": "~8.0.0",
"filament-sticky": "~0.1.4",
"dosomething-modal": "DoSomething/modal#~0.1.0",
"dosomething-validation": "DoSomething/validation#~0.0.3"
},
"resolutions": {
"jquery": "1.8.3"
}
}
<|file_sep|>updated/bower.json
"dist/neue.css",
"dist/neue.min.js"
],
"dependencies": {
"jquery": "^1.7",
"susy": "2.2.0.rc.2"
},
"devDependencies": {
"jquery": "1.8.3",
"almond": "~0.2.9",
"html5shiv": "~3.7.2",
"respond": "~1.4.2",
"highlightjs": "~8.0.0",
"filament-sticky": "~0.1.4",
"dosomething-modal": "DoSomething/modal#~0.1.0",
"dosomething-validation": "DoSomething/validation#~0.1.0"
},
"resolutions": {
"jquery": "1.8.3"
}
} | 08f1a3d5d15ab79c6c0ce7467e570acbee30b568 | bower.json | bower.json | JSON |
<|file_sep|>Tools/idle/ZoomHeight.py.diff
original:
height = top.winfo_screenheight()
updated:
newheight = top.winfo_screenheight()
<|file_sep|>Tools/idle/ZoomHeight.py.diff
original:
y = 0
height = height - 72
updated:
newy = 0
newheight = newheight - 72
<|file_sep|>Tools/idle/ZoomHeight.py.diff
original:
y = 24
height = height - 96
newgeom = "%dx%d+%d+%d" % (width, height, x, y)
if geom == newgeom:
updated:
newy = 24
newheight = newheight - 96
if height >= newheight:
<|file_sep|>original/Tools/idle/ZoomHeight.py
def __init__(self, editwin):
self.editwin = editwin
def zoom_height_event(self, event):
top = self.editwin.top
geom = top.wm_geometry()
m = re.match(r"(\d+)x(\d+)\+(-?\d+)\+(-?\d+)", geom)
if not m:
top.bell()
return
width, height, x, y = map(int, m.groups())
height = top.winfo_screenheight()
if sys.platform == 'win32':
y = 0
height = height - 72
else:
y = 24
height = height - 96
newgeom = "%dx%d+%d+%d" % (width, height, x, y)
if geom == newgeom:
newgeom = ""
<|file_sep|>current/Tools/idle/ZoomHeight.py
def __init__(self, editwin):
self.editwin = editwin
def zoom_height_event(self, event):
top = self.editwin.top
geom = top.wm_geometry()
m = re.match(r"(\d+)x(\d+)\+(-?\d+)\+(-?\d+)", geom)
if not m:
top.bell()
return
width, height, x, y = map(int, m.groups())
newheight = top.winfo_screenheight()
if sys.platform == 'win32':
newy = 0
newheight = newheight - 72
else:
newy = 24
newheight = newheight - 96
if height >= newheight:
newgeom = ""
top.wm_geometry(newgeom)
<|file_sep|>updated/Tools/idle/ZoomHeight.py |
def zoom_height_event(self, event):
top = self.editwin.top
geom = top.wm_geometry()
m = re.match(r"(\d+)x(\d+)\+(-?\d+)\+(-?\d+)", geom)
if not m:
top.bell()
return
width, height, x, y = map(int, m.groups())
newheight = top.winfo_screenheight()
if sys.platform == 'win32':
newy = 0
newheight = newheight - 72
else:
newy = 24
newheight = newheight - 96
if height >= newheight:
newgeom = ""
else:
newgeom = "%dx%d+%d+%d" % (width, newheight, x, newy)
top.wm_geometry(newgeom) | <|file_sep|>Tools/idle/ZoomHeight.py.diff
original:
height = top.winfo_screenheight()
updated:
newheight = top.winfo_screenheight()
<|file_sep|>Tools/idle/ZoomHeight.py.diff
original:
y = 0
height = height - 72
updated:
newy = 0
newheight = newheight - 72
<|file_sep|>Tools/idle/ZoomHeight.py.diff
original:
y = 24
height = height - 96
newgeom = "%dx%d+%d+%d" % (width, height, x, y)
if geom == newgeom:
updated:
newy = 24
newheight = newheight - 96
if height >= newheight:
<|file_sep|>original/Tools/idle/ZoomHeight.py
def __init__(self, editwin):
self.editwin = editwin
def zoom_height_event(self, event):
top = self.editwin.top
geom = top.wm_geometry()
m = re.match(r"(\d+)x(\d+)\+(-?\d+)\+(-?\d+)", geom)
if not m:
top.bell()
return
width, height, x, y = map(int, m.groups())
height = top.winfo_screenheight()
if sys.platform == 'win32':
y = 0
height = height - 72
else:
y = 24
height = height - 96
newgeom = "%dx%d+%d+%d" % (width, height, x, y)
if geom == newgeom:
newgeom = ""
<|file_sep|>current/Tools/idle/ZoomHeight.py
def __init__(self, editwin):
self.editwin = editwin
def zoom_height_event(self, event):
top = self.editwin.top
geom = top.wm_geometry()
m = re.match(r"(\d+)x(\d+)\+(-?\d+)\+(-?\d+)", geom)
if not m:
top.bell()
return
width, height, x, y = map(int, m.groups())
newheight = top.winfo_screenheight()
if sys.platform == 'win32':
newy = 0
newheight = newheight - 72
else:
newy = 24
newheight = newheight - 96
if height >= newheight:
newgeom = ""
top.wm_geometry(newgeom)
<|file_sep|>updated/Tools/idle/ZoomHeight.py
def zoom_height_event(self, event):
top = self.editwin.top
geom = top.wm_geometry()
m = re.match(r"(\d+)x(\d+)\+(-?\d+)\+(-?\d+)", geom)
if not m:
top.bell()
return
width, height, x, y = map(int, m.groups())
newheight = top.winfo_screenheight()
if sys.platform == 'win32':
newy = 0
newheight = newheight - 72
else:
newy = 24
newheight = newheight - 96
if height >= newheight:
newgeom = ""
else:
newgeom = "%dx%d+%d+%d" % (width, newheight, x, newy)
top.wm_geometry(newgeom) | ffd1ba9eee804fdd55e86908b158a4ad94f2f366 | Tools/idle/ZoomHeight.py | Tools/idle/ZoomHeight.py | Python |
<|file_sep|>original/config/fish/env.fish
append-to-path ~/projects/git-toolbelt
set -gx WORKON_HOME ~/.virtualenvs
set -gx PYTHONPATH ~/.local/bin
append-to-path ~/.local/bin
append-to-path /Applications/Postgres.app/Contents/Versions/9.5/bin
<|file_sep|>current/config/fish/env.fish
append-to-path ~/projects/git-toolbelt
set -gx WORKON_HOME ~/.virtualenvs
set -gx PYTHONPATH ~/.local/bin
append-to-path ~/.local/bin
append-to-path /Applications/Postgres.app/Contents/Versions/9.5/bin
<|file_sep|>updated/config/fish/env.fish | append-to-path ~/projects/git-toolbelt
set -gx WORKON_HOME ~/.virtualenvs
set -gx PYTHONPATH ~/.local/bin
append-to-path ~/.local/bin
append-to-path /Applications/Postgres.app/Contents/Versions/9.6/bin | <|file_sep|>original/config/fish/env.fish
append-to-path ~/projects/git-toolbelt
set -gx WORKON_HOME ~/.virtualenvs
set -gx PYTHONPATH ~/.local/bin
append-to-path ~/.local/bin
append-to-path /Applications/Postgres.app/Contents/Versions/9.5/bin
<|file_sep|>current/config/fish/env.fish
append-to-path ~/projects/git-toolbelt
set -gx WORKON_HOME ~/.virtualenvs
set -gx PYTHONPATH ~/.local/bin
append-to-path ~/.local/bin
append-to-path /Applications/Postgres.app/Contents/Versions/9.5/bin
<|file_sep|>updated/config/fish/env.fish
append-to-path ~/projects/git-toolbelt
set -gx WORKON_HOME ~/.virtualenvs
set -gx PYTHONPATH ~/.local/bin
append-to-path ~/.local/bin
append-to-path /Applications/Postgres.app/Contents/Versions/9.6/bin | af1f95bde61eb9553d26ca6ee3adf8e95ca8e35b | config/fish/env.fish | config/fish/env.fish | fish |
<|file_sep|>index.js.diff
original:
updated:
var confluenceUrl = process.env.CONFLUENCE_URL
<|file_sep|>original/index.js
#!/usr/bin/env node
var program = require('commander');
var userArgs = process.argv.splice(2);
var message = userArgs.join(' ');
if (message.length > 140) {
console.log('Message was too long. Can only be 140 characters. It was: ', message.length);
process.exit(1);
}
console.log(message);
<|file_sep|>current/index.js
#!/usr/bin/env node
var program = require('commander');
var userArgs = process.argv.splice(2);
var message = userArgs.join(' ');
var confluenceUrl = process.env.CONFLUENCE_URL
if (message.length > 140) {
console.log('Message was too long. Can only be 140 characters. It was: ', message.length);
process.exit(1);
}
console.log(message);
<|file_sep|>updated/index.js | #!/usr/bin/env node
var program = require('commander');
var userArgs = process.argv.splice(2);
var message = userArgs.join(' ');
var confluenceUrl = process.env.CONFLUENCE_URL
if (message.length > 140) {
console.log('Message was too long. Can only be 140 characters. It was: ', message.length);
process.exit(1);
}
if (!confluenceUrl) {
console.log('Please set the environment variable CONFLUENCE_URL.')
process.exit(2);
}
console.log(message); | <|file_sep|>index.js.diff
original:
updated:
var confluenceUrl = process.env.CONFLUENCE_URL
<|file_sep|>original/index.js
#!/usr/bin/env node
var program = require('commander');
var userArgs = process.argv.splice(2);
var message = userArgs.join(' ');
if (message.length > 140) {
console.log('Message was too long. Can only be 140 characters. It was: ', message.length);
process.exit(1);
}
console.log(message);
<|file_sep|>current/index.js
#!/usr/bin/env node
var program = require('commander');
var userArgs = process.argv.splice(2);
var message = userArgs.join(' ');
var confluenceUrl = process.env.CONFLUENCE_URL
if (message.length > 140) {
console.log('Message was too long. Can only be 140 characters. It was: ', message.length);
process.exit(1);
}
console.log(message);
<|file_sep|>updated/index.js
#!/usr/bin/env node
var program = require('commander');
var userArgs = process.argv.splice(2);
var message = userArgs.join(' ');
var confluenceUrl = process.env.CONFLUENCE_URL
if (message.length > 140) {
console.log('Message was too long. Can only be 140 characters. It was: ', message.length);
process.exit(1);
}
if (!confluenceUrl) {
console.log('Please set the environment variable CONFLUENCE_URL.')
process.exit(2);
}
console.log(message); | 65bb0ff4fce04e58febe70936adf503d432f9d41 | index.js | index.js | JavaScript |
<|file_sep|>original/server/websocket.js
var clients = {}
wss.on('connection', (ws) => {
ws.on('message', (data) => {
var req = JSON.parse(data)
if (req.message === 'ping') {
ws.id = req.id
clients[ws.id] = ws
clients[ws.id].send(JSON.stringify({message: 'pong'}))
}
else if (req.message === 'gems') {
ds.run(req.gem,
(gem) => {
if (!clients[req.id]) {
return true
}
clients[req.id].send(JSON.stringify({message: 'node', gem: gem}))
},
(err, gems) => {
<|file_sep|>current/server/websocket.js
var clients = {}
wss.on('connection', (ws) => {
ws.on('message', (data) => {
var req = JSON.parse(data)
if (req.message === 'ping') {
ws.id = req.id
clients[ws.id] = ws
clients[ws.id].send(JSON.stringify({message: 'pong'}))
}
else if (req.message === 'gems') {
ds.run(req.gem,
(gem) => {
if (!clients[req.id]) {
return true
}
clients[req.id].send(JSON.stringify({message: 'node', gem: gem}))
},
(err, gems) => {
<|file_sep|>updated/server/websocket.js | var clients = {}
wss.on('connection', (ws) => {
ws.on('message', (data) => {
var req = JSON.parse(data)
if (req.message === 'ping') {
ws.id = req.id
clients[ws.id] = ws
clients[ws.id].send(JSON.stringify({message: 'pong', id: ws.id}))
}
else if (req.message === 'gems') {
ds.run(req.gem,
(gem) => {
if (!clients[req.id]) {
return true
}
clients[req.id].send(JSON.stringify({message: 'node', gem: gem}))
},
(err, gems) => { | <|file_sep|>original/server/websocket.js
var clients = {}
wss.on('connection', (ws) => {
ws.on('message', (data) => {
var req = JSON.parse(data)
if (req.message === 'ping') {
ws.id = req.id
clients[ws.id] = ws
clients[ws.id].send(JSON.stringify({message: 'pong'}))
}
else if (req.message === 'gems') {
ds.run(req.gem,
(gem) => {
if (!clients[req.id]) {
return true
}
clients[req.id].send(JSON.stringify({message: 'node', gem: gem}))
},
(err, gems) => {
<|file_sep|>current/server/websocket.js
var clients = {}
wss.on('connection', (ws) => {
ws.on('message', (data) => {
var req = JSON.parse(data)
if (req.message === 'ping') {
ws.id = req.id
clients[ws.id] = ws
clients[ws.id].send(JSON.stringify({message: 'pong'}))
}
else if (req.message === 'gems') {
ds.run(req.gem,
(gem) => {
if (!clients[req.id]) {
return true
}
clients[req.id].send(JSON.stringify({message: 'node', gem: gem}))
},
(err, gems) => {
<|file_sep|>updated/server/websocket.js
var clients = {}
wss.on('connection', (ws) => {
ws.on('message', (data) => {
var req = JSON.parse(data)
if (req.message === 'ping') {
ws.id = req.id
clients[ws.id] = ws
clients[ws.id].send(JSON.stringify({message: 'pong', id: ws.id}))
}
else if (req.message === 'gems') {
ds.run(req.gem,
(gem) => {
if (!clients[req.id]) {
return true
}
clients[req.id].send(JSON.stringify({message: 'node', gem: gem}))
},
(err, gems) => { | af6314d64914d4450ceabb197b2c451e97ac2a62 | server/websocket.js | server/websocket.js | JavaScript |
<|file_sep|>oh-my-zsh.sh.diff
original:
# Load all of your custom configurations from custom/
for config_file ($ZSH/custom/*.zsh) source $config_file
updated:
<|file_sep|>original/oh-my-zsh.sh
# Load and run compinit
autoload -U compinit
compinit -i
# Load all of your custom configurations from custom/
for config_file ($ZSH/custom/*.zsh) source $config_file
# Load all of the plugins that were defined in ~/.zshrc
for plugin ($plugins); do
if [ -f $ZSH/plugins/$plugin/$plugin.plugin.zsh ]; then
source $ZSH/plugins/$plugin/$plugin.plugin.zsh
fi
done
# Load the theme
source "$ZSH/themes/$ZSH_THEME.zsh-theme"
# Check for updates on initial load...
if [ "$DISABLE_AUTO_UPDATE" = "true" ]
then
return
<|file_sep|>current/oh-my-zsh.sh
# Load and run compinit
autoload -U compinit
compinit -i
# Load all of the plugins that were defined in ~/.zshrc
for plugin ($plugins); do
if [ -f $ZSH/plugins/$plugin/$plugin.plugin.zsh ]; then
source $ZSH/plugins/$plugin/$plugin.plugin.zsh
fi
done
# Load the theme
source "$ZSH/themes/$ZSH_THEME.zsh-theme"
# Check for updates on initial load...
if [ "$DISABLE_AUTO_UPDATE" = "true" ]
then
return
else
/usr/bin/env zsh $ZSH/tools/check_for_upgrade.sh
fi
<|file_sep|>updated/oh-my-zsh.sh | autoload -U compinit
compinit -i
# Load all of the plugins that were defined in ~/.zshrc
for plugin ($plugins); do
if [ -f $ZSH/plugins/$plugin/$plugin.plugin.zsh ]; then
source $ZSH/plugins/$plugin/$plugin.plugin.zsh
fi
done
# Load all of your custom configurations from custom/
for config_file ($ZSH/custom/*.zsh) source $config_file
# Load the theme
source "$ZSH/themes/$ZSH_THEME.zsh-theme"
# Check for updates on initial load...
if [ "$DISABLE_AUTO_UPDATE" = "true" ]
then
return
else | <|file_sep|>oh-my-zsh.sh.diff
original:
# Load all of your custom configurations from custom/
for config_file ($ZSH/custom/*.zsh) source $config_file
updated:
<|file_sep|>original/oh-my-zsh.sh
# Load and run compinit
autoload -U compinit
compinit -i
# Load all of your custom configurations from custom/
for config_file ($ZSH/custom/*.zsh) source $config_file
# Load all of the plugins that were defined in ~/.zshrc
for plugin ($plugins); do
if [ -f $ZSH/plugins/$plugin/$plugin.plugin.zsh ]; then
source $ZSH/plugins/$plugin/$plugin.plugin.zsh
fi
done
# Load the theme
source "$ZSH/themes/$ZSH_THEME.zsh-theme"
# Check for updates on initial load...
if [ "$DISABLE_AUTO_UPDATE" = "true" ]
then
return
<|file_sep|>current/oh-my-zsh.sh
# Load and run compinit
autoload -U compinit
compinit -i
# Load all of the plugins that were defined in ~/.zshrc
for plugin ($plugins); do
if [ -f $ZSH/plugins/$plugin/$plugin.plugin.zsh ]; then
source $ZSH/plugins/$plugin/$plugin.plugin.zsh
fi
done
# Load the theme
source "$ZSH/themes/$ZSH_THEME.zsh-theme"
# Check for updates on initial load...
if [ "$DISABLE_AUTO_UPDATE" = "true" ]
then
return
else
/usr/bin/env zsh $ZSH/tools/check_for_upgrade.sh
fi
<|file_sep|>updated/oh-my-zsh.sh
autoload -U compinit
compinit -i
# Load all of the plugins that were defined in ~/.zshrc
for plugin ($plugins); do
if [ -f $ZSH/plugins/$plugin/$plugin.plugin.zsh ]; then
source $ZSH/plugins/$plugin/$plugin.plugin.zsh
fi
done
# Load all of your custom configurations from custom/
for config_file ($ZSH/custom/*.zsh) source $config_file
# Load the theme
source "$ZSH/themes/$ZSH_THEME.zsh-theme"
# Check for updates on initial load...
if [ "$DISABLE_AUTO_UPDATE" = "true" ]
then
return
else | 58a7cf19b7af136c5f59ee706ffe4772871f00fa | oh-my-zsh.sh | oh-my-zsh.sh | Shell |
<|file_sep|>kitchen.yml.diff
original:
- name: ubuntu-14.04
updated:
- name: centos-8
<|file_sep|>original/kitchen.yml
environments_path: test/fixtures/environments
data_bags_path: test/fixtures/data_bags
client_rb:
environment: test
platforms:
- name: centos-6
- name: centos-7
- name: ubuntu-14.04
- name: ubuntu-16.04
suites:
- name: default
run_list:
- recipe[chef-vault]
- recipe[test]
- name: secret_resource
run_list:
- recipe[chef-vault]
- recipe[test::chef_vault_secret]
- name: with-sugar
<|file_sep|>current/kitchen.yml
environments_path: test/fixtures/environments
data_bags_path: test/fixtures/data_bags
client_rb:
environment: test
platforms:
- name: centos-6
- name: centos-7
- name: centos-8
- name: ubuntu-16.04
suites:
- name: default
run_list:
- recipe[chef-vault]
- recipe[test]
- name: secret_resource
run_list:
- recipe[chef-vault]
- recipe[test::chef_vault_secret]
- name: with-sugar
<|file_sep|>updated/kitchen.yml | environments_path: test/fixtures/environments
data_bags_path: test/fixtures/data_bags
client_rb:
environment: test
platforms:
- name: centos-6
- name: centos-7
- name: centos-8
- name: ubuntu-16.04
- name: ubuntu-18.04
suites:
- name: default
run_list:
- recipe[chef-vault]
- recipe[test]
- name: secret_resource
run_list:
- recipe[chef-vault]
- recipe[test::chef_vault_secret] | <|file_sep|>kitchen.yml.diff
original:
- name: ubuntu-14.04
updated:
- name: centos-8
<|file_sep|>original/kitchen.yml
environments_path: test/fixtures/environments
data_bags_path: test/fixtures/data_bags
client_rb:
environment: test
platforms:
- name: centos-6
- name: centos-7
- name: ubuntu-14.04
- name: ubuntu-16.04
suites:
- name: default
run_list:
- recipe[chef-vault]
- recipe[test]
- name: secret_resource
run_list:
- recipe[chef-vault]
- recipe[test::chef_vault_secret]
- name: with-sugar
<|file_sep|>current/kitchen.yml
environments_path: test/fixtures/environments
data_bags_path: test/fixtures/data_bags
client_rb:
environment: test
platforms:
- name: centos-6
- name: centos-7
- name: centos-8
- name: ubuntu-16.04
suites:
- name: default
run_list:
- recipe[chef-vault]
- recipe[test]
- name: secret_resource
run_list:
- recipe[chef-vault]
- recipe[test::chef_vault_secret]
- name: with-sugar
<|file_sep|>updated/kitchen.yml
environments_path: test/fixtures/environments
data_bags_path: test/fixtures/data_bags
client_rb:
environment: test
platforms:
- name: centos-6
- name: centos-7
- name: centos-8
- name: ubuntu-16.04
- name: ubuntu-18.04
suites:
- name: default
run_list:
- recipe[chef-vault]
- recipe[test]
- name: secret_resource
run_list:
- recipe[chef-vault]
- recipe[test::chef_vault_secret] | cbd6ab722422548f0ce8a4436831f9c6a23310cb | kitchen.yml | kitchen.yml | YAML |
<|file_sep|>original/setup/shells.sh
#!/bin/sh
. ./utils.sh || exit 1
# add shells to /etc/shells
shells="zsh bash dash fish ksh pdksh xonsh"
for shell in $shells; do
_test_executable "$shell" || return 1
shell_path="$(command -v "$shell")"
if grep -Fqx "$shell_path" /etc/shells; then
printf "Shell %s is already installed\n" "$shell_path"
else
printf "Adding shell %s to /etc/shells\n" "$shell_path"
printf "%s\n" "$shell_path" \
| sudo tee -a /etc/shells >/dev/null
fi
done
# switch default shell to zsh
# _test_executable "zsh" && chsh -s "$(command -v zsh)"
<|file_sep|>current/setup/shells.sh
#!/bin/sh
. ./utils.sh || exit 1
# add shells to /etc/shells
shells="zsh bash dash fish ksh pdksh xonsh"
for shell in $shells; do
_test_executable "$shell" || return 1
shell_path="$(command -v "$shell")"
if grep -Fqx "$shell_path" /etc/shells; then
printf "Shell %s is already installed\n" "$shell_path"
else
printf "Adding shell %s to /etc/shells\n" "$shell_path"
printf "%s\n" "$shell_path" \
| sudo tee -a /etc/shells >/dev/null
fi
done
# switch default shell to zsh
# _test_executable "zsh" && chsh -s "$(command -v zsh)"
<|file_sep|>updated/setup/shells.sh | #!/bin/sh
. ./utils.sh || exit 1
# add shells to /etc/shells
shells="zsh bash dash fish ksh pdksh xonsh"
for shell in $shells; do
_test_executable "$shell" || return 1
shell_path="$(command -v "$shell")"
if grep -Fqx "$shell_path" /etc/shells; then
printf "Shell %s is already installed\n" "$shell_path"
else
printf "Adding shell %s to /etc/shells\n" "$shell_path"
printf "%s\n" "$shell_path" \
| sudo tee -a /etc/shells >/dev/null
fi
done
# switch default shell to zsh
_test_executable "zsh" 2>/dev/null && chsh -s "$(command -v zsh)" | <|file_sep|>original/setup/shells.sh
#!/bin/sh
. ./utils.sh || exit 1
# add shells to /etc/shells
shells="zsh bash dash fish ksh pdksh xonsh"
for shell in $shells; do
_test_executable "$shell" || return 1
shell_path="$(command -v "$shell")"
if grep -Fqx "$shell_path" /etc/shells; then
printf "Shell %s is already installed\n" "$shell_path"
else
printf "Adding shell %s to /etc/shells\n" "$shell_path"
printf "%s\n" "$shell_path" \
| sudo tee -a /etc/shells >/dev/null
fi
done
# switch default shell to zsh
# _test_executable "zsh" && chsh -s "$(command -v zsh)"
<|file_sep|>current/setup/shells.sh
#!/bin/sh
. ./utils.sh || exit 1
# add shells to /etc/shells
shells="zsh bash dash fish ksh pdksh xonsh"
for shell in $shells; do
_test_executable "$shell" || return 1
shell_path="$(command -v "$shell")"
if grep -Fqx "$shell_path" /etc/shells; then
printf "Shell %s is already installed\n" "$shell_path"
else
printf "Adding shell %s to /etc/shells\n" "$shell_path"
printf "%s\n" "$shell_path" \
| sudo tee -a /etc/shells >/dev/null
fi
done
# switch default shell to zsh
# _test_executable "zsh" && chsh -s "$(command -v zsh)"
<|file_sep|>updated/setup/shells.sh
#!/bin/sh
. ./utils.sh || exit 1
# add shells to /etc/shells
shells="zsh bash dash fish ksh pdksh xonsh"
for shell in $shells; do
_test_executable "$shell" || return 1
shell_path="$(command -v "$shell")"
if grep -Fqx "$shell_path" /etc/shells; then
printf "Shell %s is already installed\n" "$shell_path"
else
printf "Adding shell %s to /etc/shells\n" "$shell_path"
printf "%s\n" "$shell_path" \
| sudo tee -a /etc/shells >/dev/null
fi
done
# switch default shell to zsh
_test_executable "zsh" 2>/dev/null && chsh -s "$(command -v zsh)" | 796ca0bf40baa76320a07ca6525fc60d0aff53c0 | setup/shells.sh | setup/shells.sh | Shell |
<|file_sep|>original/requirements/base.txt
dj-database-url==0.5.0
django-apiblueprint-view==1.1.2
django-basicauth==0.5.1
django-extensions==2.1.3
django-localflavor==2.1
django-markdown-deux==1.0.5
django==1.11.16 # pyup: >=1.11,<2.0
djangorestframework==3.7.7
djangorestframework-gis==0.12.0
django-cors-headers==2.4.0
fastkml==0.11
lxml==4.2.5
psycopg2-binary==2.7.5
pyshp==1.2.12
requests==2.20.0
boto==2.49.0
uk-geo-utils==0.6.0
git+git://github.com/DemocracyClub/dc_base_theme.git@0.3.7
git+https://github.com/DemocracyClub/dc_signup_form.git@2.0.1
<|file_sep|>current/requirements/base.txt
dj-database-url==0.5.0
django-apiblueprint-view==1.1.2
django-basicauth==0.5.1
django-extensions==2.1.3
django-localflavor==2.1
django-markdown-deux==1.0.5
django==1.11.16 # pyup: >=1.11,<2.0
djangorestframework==3.7.7
djangorestframework-gis==0.12.0
django-cors-headers==2.4.0
fastkml==0.11
lxml==4.2.5
psycopg2-binary==2.7.5
pyshp==1.2.12
requests==2.20.0
boto==2.49.0
uk-geo-utils==0.6.0
git+git://github.com/DemocracyClub/dc_base_theme.git@0.3.7
git+https://github.com/DemocracyClub/dc_signup_form.git@2.0.1
<|file_sep|>updated/requirements/base.txt | dj-database-url==0.5.0
django-apiblueprint-view==2.0.0
django-basicauth==0.5.1
django-extensions==2.1.3
django-localflavor==2.1
django-markdown-deux==1.0.5
django==1.11.16 # pyup: >=1.11,<2.0
djangorestframework==3.7.7
djangorestframework-gis==0.12.0
django-cors-headers==2.4.0
fastkml==0.11
lxml==4.2.5
psycopg2-binary==2.7.5
pyshp==1.2.12
requests==2.20.0
boto==2.49.0
uk-geo-utils==0.6.0
git+git://github.com/DemocracyClub/dc_base_theme.git@0.3.7
git+https://github.com/DemocracyClub/dc_signup_form.git@2.0.1 | <|file_sep|>original/requirements/base.txt
dj-database-url==0.5.0
django-apiblueprint-view==1.1.2
django-basicauth==0.5.1
django-extensions==2.1.3
django-localflavor==2.1
django-markdown-deux==1.0.5
django==1.11.16 # pyup: >=1.11,<2.0
djangorestframework==3.7.7
djangorestframework-gis==0.12.0
django-cors-headers==2.4.0
fastkml==0.11
lxml==4.2.5
psycopg2-binary==2.7.5
pyshp==1.2.12
requests==2.20.0
boto==2.49.0
uk-geo-utils==0.6.0
git+git://github.com/DemocracyClub/dc_base_theme.git@0.3.7
git+https://github.com/DemocracyClub/dc_signup_form.git@2.0.1
<|file_sep|>current/requirements/base.txt
dj-database-url==0.5.0
django-apiblueprint-view==1.1.2
django-basicauth==0.5.1
django-extensions==2.1.3
django-localflavor==2.1
django-markdown-deux==1.0.5
django==1.11.16 # pyup: >=1.11,<2.0
djangorestframework==3.7.7
djangorestframework-gis==0.12.0
django-cors-headers==2.4.0
fastkml==0.11
lxml==4.2.5
psycopg2-binary==2.7.5
pyshp==1.2.12
requests==2.20.0
boto==2.49.0
uk-geo-utils==0.6.0
git+git://github.com/DemocracyClub/dc_base_theme.git@0.3.7
git+https://github.com/DemocracyClub/dc_signup_form.git@2.0.1
<|file_sep|>updated/requirements/base.txt
dj-database-url==0.5.0
django-apiblueprint-view==2.0.0
django-basicauth==0.5.1
django-extensions==2.1.3
django-localflavor==2.1
django-markdown-deux==1.0.5
django==1.11.16 # pyup: >=1.11,<2.0
djangorestframework==3.7.7
djangorestframework-gis==0.12.0
django-cors-headers==2.4.0
fastkml==0.11
lxml==4.2.5
psycopg2-binary==2.7.5
pyshp==1.2.12
requests==2.20.0
boto==2.49.0
uk-geo-utils==0.6.0
git+git://github.com/DemocracyClub/dc_base_theme.git@0.3.7
git+https://github.com/DemocracyClub/dc_signup_form.git@2.0.1 | afdc86a98581935edd45b94a9d754d2dddac030d | requirements/base.txt | requirements/base.txt | Text |
<|file_sep|>original/requirements-dev.txt
coverage
cython>=0.18
mock>=1.0
nose>=1.2
numpy
https://bitbucket.org/birkenfeld/sphinx/get/default.tar.gz
<|file_sep|>current/requirements-dev.txt
coverage
cython>=0.18
mock>=1.0
nose>=1.2
numpy
https://bitbucket.org/birkenfeld/sphinx/get/default.tar.gz
<|file_sep|>updated/requirements-dev.txt | coverage
cython>=0.18
mock>=1.0
nose==1.2.1
numpy
https://bitbucket.org/birkenfeld/sphinx/get/default.tar.gz | <|file_sep|>original/requirements-dev.txt
coverage
cython>=0.18
mock>=1.0
nose>=1.2
numpy
https://bitbucket.org/birkenfeld/sphinx/get/default.tar.gz
<|file_sep|>current/requirements-dev.txt
coverage
cython>=0.18
mock>=1.0
nose>=1.2
numpy
https://bitbucket.org/birkenfeld/sphinx/get/default.tar.gz
<|file_sep|>updated/requirements-dev.txt
coverage
cython>=0.18
mock>=1.0
nose==1.2.1
numpy
https://bitbucket.org/birkenfeld/sphinx/get/default.tar.gz | 05a4cc53dad9360bfa643e34f113697c52e3ad54 | requirements-dev.txt | requirements-dev.txt | Text |
<|file_sep|>src/Allegro/C.hsc.diff
original:
foreign import ccall "al_install_system"
updated:
foreign import ccall unsafe "al_install_system"
<|file_sep|>original/src/Allegro/C.hsc
{-# LANGUAGE ForeignFunctionInterface #-}
module Allegro.C (al_init) where
import Foreign
import Foreign.C.Types
#include <allegro5/allegro.h>
al_init :: IO Bool
al_init = al_install_system #{const ALLEGRO_VERSION_INT} fp_atexit
foreign import ccall "al_install_system"
al_install_system :: CInt -> FunPtr (Ptr () -> IO CInt) -> IO Bool
foreign import ccall "&atexit"
fp_atexit :: FunPtr (Ptr () -> IO CInt)
<|file_sep|>current/src/Allegro/C.hsc
{-# LANGUAGE ForeignFunctionInterface #-}
module Allegro.C (al_init) where
import Foreign
import Foreign.C.Types
#include <allegro5/allegro.h>
al_init :: IO Bool
al_init = al_install_system #{const ALLEGRO_VERSION_INT} fp_atexit
foreign import ccall unsafe "al_install_system"
al_install_system :: CInt -> FunPtr (Ptr () -> IO CInt) -> IO Bool
foreign import ccall "&atexit"
fp_atexit :: FunPtr (Ptr () -> IO CInt)
<|file_sep|>updated/src/Allegro/C.hsc | {-# LANGUAGE ForeignFunctionInterface #-}
module Allegro.C (al_init) where
import Foreign
import Foreign.C.Types
#include <allegro5/allegro.h>
al_init :: IO Bool
al_init = al_install_system #{const ALLEGRO_VERSION_INT} fp_atexit
foreign import ccall unsafe "al_install_system"
al_install_system :: CInt -> FunPtr (Ptr () -> IO CInt) -> IO Bool
foreign import ccall unsafe "&atexit"
fp_atexit :: FunPtr (Ptr () -> IO CInt)
| <|file_sep|>src/Allegro/C.hsc.diff
original:
foreign import ccall "al_install_system"
updated:
foreign import ccall unsafe "al_install_system"
<|file_sep|>original/src/Allegro/C.hsc
{-# LANGUAGE ForeignFunctionInterface #-}
module Allegro.C (al_init) where
import Foreign
import Foreign.C.Types
#include <allegro5/allegro.h>
al_init :: IO Bool
al_init = al_install_system #{const ALLEGRO_VERSION_INT} fp_atexit
foreign import ccall "al_install_system"
al_install_system :: CInt -> FunPtr (Ptr () -> IO CInt) -> IO Bool
foreign import ccall "&atexit"
fp_atexit :: FunPtr (Ptr () -> IO CInt)
<|file_sep|>current/src/Allegro/C.hsc
{-# LANGUAGE ForeignFunctionInterface #-}
module Allegro.C (al_init) where
import Foreign
import Foreign.C.Types
#include <allegro5/allegro.h>
al_init :: IO Bool
al_init = al_install_system #{const ALLEGRO_VERSION_INT} fp_atexit
foreign import ccall unsafe "al_install_system"
al_install_system :: CInt -> FunPtr (Ptr () -> IO CInt) -> IO Bool
foreign import ccall "&atexit"
fp_atexit :: FunPtr (Ptr () -> IO CInt)
<|file_sep|>updated/src/Allegro/C.hsc
{-# LANGUAGE ForeignFunctionInterface #-}
module Allegro.C (al_init) where
import Foreign
import Foreign.C.Types
#include <allegro5/allegro.h>
al_init :: IO Bool
al_init = al_install_system #{const ALLEGRO_VERSION_INT} fp_atexit
foreign import ccall unsafe "al_install_system"
al_install_system :: CInt -> FunPtr (Ptr () -> IO CInt) -> IO Bool
foreign import ccall unsafe "&atexit"
fp_atexit :: FunPtr (Ptr () -> IO CInt)
| 87adce0d38a35398ee200484d02d2975a6b4126e | src/Allegro/C.hsc | src/Allegro/C.hsc | Haskell |
<|file_sep|>metadata/mf.asciitext.lite.yml.diff
original:
updated:
- versionName: 2.4.0
versionCode: 13
commit: 2.4.0
subdir: base/
gradle:
- yes
<|file_sep|>original/metadata/mf.asciitext.lite.yml
AuthorWebSite: https://oss.mobilefirst.me/
SourceCode: https://github.com/MobileFirstLLC/irregular-expressions
IssueTracker: https://github.com/MobileFirstLLC/irregular-expressions/issues
AutoName: Irregular Expressions
RepoType: git
Repo: https://github.com/MobileFirstLLC/irregular-expressions
Builds:
- versionName: 2.3.0
versionCode: 12
commit: 2.3.0
subdir: base/
gradle:
- yes
AutoUpdateMode: Version %v
UpdateCheckMode: Tags
CurrentVersion: 2.3.0
CurrentVersionCode: 12
<|file_sep|>current/metadata/mf.asciitext.lite.yml
Repo: https://github.com/MobileFirstLLC/irregular-expressions
Builds:
- versionName: 2.3.0
versionCode: 12
commit: 2.3.0
subdir: base/
gradle:
- yes
- versionName: 2.4.0
versionCode: 13
commit: 2.4.0
subdir: base/
gradle:
- yes
AutoUpdateMode: Version %v
UpdateCheckMode: Tags
CurrentVersion: 2.3.0
CurrentVersionCode: 12
<|file_sep|>updated/metadata/mf.asciitext.lite.yml | Repo: https://github.com/MobileFirstLLC/irregular-expressions
Builds:
- versionName: 2.3.0
versionCode: 12
commit: 2.3.0
subdir: base/
gradle:
- yes
- versionName: 2.4.0
versionCode: 13
commit: 2.4.0
subdir: base/
gradle:
- yes
AutoUpdateMode: Version %v
UpdateCheckMode: Tags
CurrentVersion: 2.4.0
CurrentVersionCode: 13 | <|file_sep|>metadata/mf.asciitext.lite.yml.diff
original:
updated:
- versionName: 2.4.0
versionCode: 13
commit: 2.4.0
subdir: base/
gradle:
- yes
<|file_sep|>original/metadata/mf.asciitext.lite.yml
AuthorWebSite: https://oss.mobilefirst.me/
SourceCode: https://github.com/MobileFirstLLC/irregular-expressions
IssueTracker: https://github.com/MobileFirstLLC/irregular-expressions/issues
AutoName: Irregular Expressions
RepoType: git
Repo: https://github.com/MobileFirstLLC/irregular-expressions
Builds:
- versionName: 2.3.0
versionCode: 12
commit: 2.3.0
subdir: base/
gradle:
- yes
AutoUpdateMode: Version %v
UpdateCheckMode: Tags
CurrentVersion: 2.3.0
CurrentVersionCode: 12
<|file_sep|>current/metadata/mf.asciitext.lite.yml
Repo: https://github.com/MobileFirstLLC/irregular-expressions
Builds:
- versionName: 2.3.0
versionCode: 12
commit: 2.3.0
subdir: base/
gradle:
- yes
- versionName: 2.4.0
versionCode: 13
commit: 2.4.0
subdir: base/
gradle:
- yes
AutoUpdateMode: Version %v
UpdateCheckMode: Tags
CurrentVersion: 2.3.0
CurrentVersionCode: 12
<|file_sep|>updated/metadata/mf.asciitext.lite.yml
Repo: https://github.com/MobileFirstLLC/irregular-expressions
Builds:
- versionName: 2.3.0
versionCode: 12
commit: 2.3.0
subdir: base/
gradle:
- yes
- versionName: 2.4.0
versionCode: 13
commit: 2.4.0
subdir: base/
gradle:
- yes
AutoUpdateMode: Version %v
UpdateCheckMode: Tags
CurrentVersion: 2.4.0
CurrentVersionCode: 13 | e60cec9826471920ab393bfe386f2b3dc36206a8 | metadata/mf.asciitext.lite.yml | metadata/mf.asciitext.lite.yml | YAML |
<|file_sep|>spec/api-system-preferences-spec.js.diff
original:
if (process.platform !== 'darwin') {
return
}
updated:
describe('systemPreferences.getAccentColor', function () {
if (process.platform !== 'win32') {
return
}
it('should return a non-empty string', function () {
let accentColor = systemPreferences.getAccentColor();
assert.notEqual(accentColor, null);
assert(accentColor.length > 0);
})
})
<|file_sep|>original/spec/api-system-preferences-spec.js
const assert = require('assert')
const {remote} = require('electron')
const {systemPreferences} = remote
describe('systemPreferences module', function () {
if (process.platform !== 'darwin') {
return
}
describe('systemPreferences.getUserDefault(key, type)', function () {
it('returns values for known user defaults', function () {
let locale = systemPreferences.getUserDefault('AppleLocale', 'string')
assert.notEqual(locale, null)
assert(locale.length > 0)
let languages = systemPreferences.getUserDefault('AppleLanguages', 'array')
assert.notEqual(languages, null)
assert(languages.length > 0)
})
})
})
<|file_sep|>current/spec/api-system-preferences-spec.js
}
it('should return a non-empty string', function () {
let accentColor = systemPreferences.getAccentColor();
assert.notEqual(accentColor, null);
assert(accentColor.length > 0);
})
})
describe('systemPreferences.getUserDefault(key, type)', function () {
it('returns values for known user defaults', function () {
let locale = systemPreferences.getUserDefault('AppleLocale', 'string')
assert.notEqual(locale, null)
assert(locale.length > 0)
let languages = systemPreferences.getUserDefault('AppleLanguages', 'array')
assert.notEqual(languages, null)
assert(languages.length > 0)
})
})
})
<|file_sep|>updated/spec/api-system-preferences-spec.js | }
it('should return a non-empty string', function () {
let accentColor = systemPreferences.getAccentColor();
assert.notEqual(accentColor, null);
assert(accentColor.length > 0);
})
})
describe('systemPreferences.getUserDefault(key, type)', function () {
if (process.platform !== 'darwin') {
return
}
it('returns values for known user defaults', function () {
let locale = systemPreferences.getUserDefault('AppleLocale', 'string')
assert.notEqual(locale, null)
assert(locale.length > 0)
let languages = systemPreferences.getUserDefault('AppleLanguages', 'array')
assert.notEqual(languages, null) | <|file_sep|>spec/api-system-preferences-spec.js.diff
original:
if (process.platform !== 'darwin') {
return
}
updated:
describe('systemPreferences.getAccentColor', function () {
if (process.platform !== 'win32') {
return
}
it('should return a non-empty string', function () {
let accentColor = systemPreferences.getAccentColor();
assert.notEqual(accentColor, null);
assert(accentColor.length > 0);
})
})
<|file_sep|>original/spec/api-system-preferences-spec.js
const assert = require('assert')
const {remote} = require('electron')
const {systemPreferences} = remote
describe('systemPreferences module', function () {
if (process.platform !== 'darwin') {
return
}
describe('systemPreferences.getUserDefault(key, type)', function () {
it('returns values for known user defaults', function () {
let locale = systemPreferences.getUserDefault('AppleLocale', 'string')
assert.notEqual(locale, null)
assert(locale.length > 0)
let languages = systemPreferences.getUserDefault('AppleLanguages', 'array')
assert.notEqual(languages, null)
assert(languages.length > 0)
})
})
})
<|file_sep|>current/spec/api-system-preferences-spec.js
}
it('should return a non-empty string', function () {
let accentColor = systemPreferences.getAccentColor();
assert.notEqual(accentColor, null);
assert(accentColor.length > 0);
})
})
describe('systemPreferences.getUserDefault(key, type)', function () {
it('returns values for known user defaults', function () {
let locale = systemPreferences.getUserDefault('AppleLocale', 'string')
assert.notEqual(locale, null)
assert(locale.length > 0)
let languages = systemPreferences.getUserDefault('AppleLanguages', 'array')
assert.notEqual(languages, null)
assert(languages.length > 0)
})
})
})
<|file_sep|>updated/spec/api-system-preferences-spec.js
}
it('should return a non-empty string', function () {
let accentColor = systemPreferences.getAccentColor();
assert.notEqual(accentColor, null);
assert(accentColor.length > 0);
})
})
describe('systemPreferences.getUserDefault(key, type)', function () {
if (process.platform !== 'darwin') {
return
}
it('returns values for known user defaults', function () {
let locale = systemPreferences.getUserDefault('AppleLocale', 'string')
assert.notEqual(locale, null)
assert(locale.length > 0)
let languages = systemPreferences.getUserDefault('AppleLanguages', 'array')
assert.notEqual(languages, null) | 9f946527929d65607c0d42d33733d370fee27f83 | spec/api-system-preferences-spec.js | spec/api-system-preferences-spec.js | JavaScript |
<|file_sep|>original/config/routes.rb
end
scope :path => Refinery::Products.products_categories_path do
resources :categories, :path => '', :as => :categories, :controller => 'categories'
end
end
# Admin routes
namespace :products, :path => '' do
namespace :admin, :path => Refinery::Core.backend_route do
scope :path => Refinery::Products.products_path do
root :to => "products#index"
resources :products, :except => :show do
collection do
post :update_positions
get :uncategorized
end
end
resources :categories
<|file_sep|>current/config/routes.rb
end
scope :path => Refinery::Products.products_categories_path do
resources :categories, :path => '', :as => :categories, :controller => 'categories'
end
end
# Admin routes
namespace :products, :path => '' do
namespace :admin, :path => Refinery::Core.backend_route do
scope :path => Refinery::Products.products_path do
root :to => "products#index"
resources :products, :except => :show do
collection do
post :update_positions
get :uncategorized
end
end
resources :categories
<|file_sep|>updated/config/routes.rb | end
scope :path => Refinery::Products.products_categories_path do
resources :categories, :path => '', :as => :categories, :controller => 'categories'
end
end
# Admin routes
namespace :products, :path => '' do
namespace :admin, :path => Refinery::Core.backend_route do
scope :path => Refinery::Products.shop_path do
root :to => "products#index"
resources :products, :except => :show do
collection do
post :update_positions
get :uncategorized
end
end
resources :categories | <|file_sep|>original/config/routes.rb
end
scope :path => Refinery::Products.products_categories_path do
resources :categories, :path => '', :as => :categories, :controller => 'categories'
end
end
# Admin routes
namespace :products, :path => '' do
namespace :admin, :path => Refinery::Core.backend_route do
scope :path => Refinery::Products.products_path do
root :to => "products#index"
resources :products, :except => :show do
collection do
post :update_positions
get :uncategorized
end
end
resources :categories
<|file_sep|>current/config/routes.rb
end
scope :path => Refinery::Products.products_categories_path do
resources :categories, :path => '', :as => :categories, :controller => 'categories'
end
end
# Admin routes
namespace :products, :path => '' do
namespace :admin, :path => Refinery::Core.backend_route do
scope :path => Refinery::Products.products_path do
root :to => "products#index"
resources :products, :except => :show do
collection do
post :update_positions
get :uncategorized
end
end
resources :categories
<|file_sep|>updated/config/routes.rb
end
scope :path => Refinery::Products.products_categories_path do
resources :categories, :path => '', :as => :categories, :controller => 'categories'
end
end
# Admin routes
namespace :products, :path => '' do
namespace :admin, :path => Refinery::Core.backend_route do
scope :path => Refinery::Products.shop_path do
root :to => "products#index"
resources :products, :except => :show do
collection do
post :update_positions
get :uncategorized
end
end
resources :categories | bfdfd0b50ad78a31a93485b19e74b1a83e11e51a | config/routes.rb | config/routes.rb | Ruby |
<|file_sep|>original/{{cookiecutter.project_slug}}/requirements/base.txt
Django==1.11.1
wagtail==1.10.1
wagtailfontawesome==1.0.6
celery==4.0.2
django-cachalot==1.4.1
django-compressor==2.1.1
django-environ==0.4.3
django-foundation-formtags==0.0.8
django-wagtail-feeds==0.0.7
django-libsass==0.7
django-redis==4.8.0
libsass==0.12.3
psycopg2==2.7.1
wand==0.4.4
{% if cookiecutter.use_wagalytics_app == 'y' %}wagalytics==0.6 {% endif %}
{% if cookiecutter.use_django_cachalot == 'y' %}django-cachalot==1.2.1 {% endif %}
<|file_sep|>current/{{cookiecutter.project_slug}}/requirements/base.txt
Django==1.11.1
wagtail==1.10.1
wagtailfontawesome==1.0.6
celery==4.0.2
django-cachalot==1.4.1
django-compressor==2.1.1
django-environ==0.4.3
django-foundation-formtags==0.0.8
django-wagtail-feeds==0.0.7
django-libsass==0.7
django-redis==4.8.0
libsass==0.12.3
psycopg2==2.7.1
wand==0.4.4
{% if cookiecutter.use_wagalytics_app == 'y' %}wagalytics==0.6 {% endif %}
{% if cookiecutter.use_django_cachalot == 'y' %}django-cachalot==1.2.1 {% endif %}
<|file_sep|>updated/{{cookiecutter.project_slug}}/requirements/base.txt | Django==1.11.1
wagtail==1.10.1
wagtailfontawesome==1.0.6
celery==4.0.2
django-cachalot==1.5.0
django-compressor==2.1.1
django-environ==0.4.3
django-foundation-formtags==0.0.8
django-wagtail-feeds==0.0.7
django-libsass==0.7
django-redis==4.8.0
libsass==0.12.3
psycopg2==2.7.1
wand==0.4.4
{% if cookiecutter.use_wagalytics_app == 'y' %}wagalytics==0.6 {% endif %}
{% if cookiecutter.use_django_cachalot == 'y' %}django-cachalot==1.2.1 {% endif %} | <|file_sep|>original/{{cookiecutter.project_slug}}/requirements/base.txt
Django==1.11.1
wagtail==1.10.1
wagtailfontawesome==1.0.6
celery==4.0.2
django-cachalot==1.4.1
django-compressor==2.1.1
django-environ==0.4.3
django-foundation-formtags==0.0.8
django-wagtail-feeds==0.0.7
django-libsass==0.7
django-redis==4.8.0
libsass==0.12.3
psycopg2==2.7.1
wand==0.4.4
{% if cookiecutter.use_wagalytics_app == 'y' %}wagalytics==0.6 {% endif %}
{% if cookiecutter.use_django_cachalot == 'y' %}django-cachalot==1.2.1 {% endif %}
<|file_sep|>current/{{cookiecutter.project_slug}}/requirements/base.txt
Django==1.11.1
wagtail==1.10.1
wagtailfontawesome==1.0.6
celery==4.0.2
django-cachalot==1.4.1
django-compressor==2.1.1
django-environ==0.4.3
django-foundation-formtags==0.0.8
django-wagtail-feeds==0.0.7
django-libsass==0.7
django-redis==4.8.0
libsass==0.12.3
psycopg2==2.7.1
wand==0.4.4
{% if cookiecutter.use_wagalytics_app == 'y' %}wagalytics==0.6 {% endif %}
{% if cookiecutter.use_django_cachalot == 'y' %}django-cachalot==1.2.1 {% endif %}
<|file_sep|>updated/{{cookiecutter.project_slug}}/requirements/base.txt
Django==1.11.1
wagtail==1.10.1
wagtailfontawesome==1.0.6
celery==4.0.2
django-cachalot==1.5.0
django-compressor==2.1.1
django-environ==0.4.3
django-foundation-formtags==0.0.8
django-wagtail-feeds==0.0.7
django-libsass==0.7
django-redis==4.8.0
libsass==0.12.3
psycopg2==2.7.1
wand==0.4.4
{% if cookiecutter.use_wagalytics_app == 'y' %}wagalytics==0.6 {% endif %}
{% if cookiecutter.use_django_cachalot == 'y' %}django-cachalot==1.2.1 {% endif %} | cdeb40250439322fb444633545abef3c0e82eb61 | {{cookiecutter.project_slug}}/requirements/base.txt | {{cookiecutter.project_slug}}/requirements/base.txt | Text |
<|file_sep|>original/package.json
{
"name": "cloudcp",
"version": "0.1.0",
"description": "Copy local files to any cloud storage provider supported by pkgcloud",
"main": "index.js",
"scripts": {
"test": "tap test/index.js",
"test-ci": "standard && istanbul cover test/index.js",
"preversion": "npm test",
"postversion": "git push && git push --tags"
},
"keywords": [
"copy",
"fs",
"cp",
"cloud"
],
"author": "Pier-Luc Gendreau",
"license": "MIT",
"devDependencies": {
"glob": "^5.0.15",
<|file_sep|>current/package.json
{
"name": "cloudcp",
"version": "0.1.0",
"description": "Copy local files to any cloud storage provider supported by pkgcloud",
"main": "index.js",
"scripts": {
"test": "tap test/index.js",
"test-ci": "standard && istanbul cover test/index.js",
"preversion": "npm test",
"postversion": "git push && git push --tags"
},
"keywords": [
"copy",
"fs",
"cp",
"cloud"
],
"author": "Pier-Luc Gendreau",
"license": "MIT",
"devDependencies": {
"glob": "^5.0.15",
<|file_sep|>updated/package.json | {
"name": "cloudcp",
"version": "0.1.0",
"description": "Copy local files to any cloud storage provider supported by pkgcloud",
"main": "index.js",
"scripts": {
"test": "tap test/index.js",
"test-ci": "standard && istanbul cover test/index.js"
},
"keywords": [
"copy",
"fs",
"cp",
"cloud"
],
"author": "Pier-Luc Gendreau",
"license": "MIT",
"devDependencies": {
"glob": "^5.0.15",
"istanbul": "^0.4.0",
"sinon": "^1.17.2", | <|file_sep|>original/package.json
{
"name": "cloudcp",
"version": "0.1.0",
"description": "Copy local files to any cloud storage provider supported by pkgcloud",
"main": "index.js",
"scripts": {
"test": "tap test/index.js",
"test-ci": "standard && istanbul cover test/index.js",
"preversion": "npm test",
"postversion": "git push && git push --tags"
},
"keywords": [
"copy",
"fs",
"cp",
"cloud"
],
"author": "Pier-Luc Gendreau",
"license": "MIT",
"devDependencies": {
"glob": "^5.0.15",
<|file_sep|>current/package.json
{
"name": "cloudcp",
"version": "0.1.0",
"description": "Copy local files to any cloud storage provider supported by pkgcloud",
"main": "index.js",
"scripts": {
"test": "tap test/index.js",
"test-ci": "standard && istanbul cover test/index.js",
"preversion": "npm test",
"postversion": "git push && git push --tags"
},
"keywords": [
"copy",
"fs",
"cp",
"cloud"
],
"author": "Pier-Luc Gendreau",
"license": "MIT",
"devDependencies": {
"glob": "^5.0.15",
<|file_sep|>updated/package.json
{
"name": "cloudcp",
"version": "0.1.0",
"description": "Copy local files to any cloud storage provider supported by pkgcloud",
"main": "index.js",
"scripts": {
"test": "tap test/index.js",
"test-ci": "standard && istanbul cover test/index.js"
},
"keywords": [
"copy",
"fs",
"cp",
"cloud"
],
"author": "Pier-Luc Gendreau",
"license": "MIT",
"devDependencies": {
"glob": "^5.0.15",
"istanbul": "^0.4.0",
"sinon": "^1.17.2", | 9fce8b4849a712e559929feac34cc5c1fa8b6c76 | package.json | package.json | JSON |
<|file_sep|>plugins/cache_reload.js.diff
original:
url: "",
initialized: false
updated:
url: ""
<|file_sep|>plugins/cache_reload.js.diff
original:
if(this.initalized)
updated:
if(!impl.url)
<|file_sep|>original/plugins/cache_reload.js
if(this.initalized)
return this;
BOOMR.subscribe(
"page_ready",
function() {
if(!impl.url)
return;
// we use document and not BOOMR.window.document since
// we can run inside the boomerang iframe if any
var i=document.createElement('iframe');
i.style.display="none";
i.src=impl.url;
document.body.appendChild(i);
},
null,
null
);
this.initialized = true;
<|file_sep|>current/plugins/cache_reload.js
if(!impl.url)
return this;
BOOMR.subscribe(
"page_ready",
function() {
if(!impl.url)
return;
// we use document and not BOOMR.window.document since
// we can run inside the boomerang iframe if any
var i=document.createElement('iframe');
i.style.display="none";
i.src=impl.url;
document.body.appendChild(i);
},
null,
null
);
this.initialized = true;
return this;
<|file_sep|>updated/plugins/cache_reload.js | if(!impl.url)
return this;
// we use document and not BOOMR.window.document since
// we can run inside the boomerang iframe if any
var i=document.createElement('iframe');
i.style.display="none";
i.src=impl.url;
document.body.appendChild(i);
return this;
},
is_complete: function() {
// we always return true since this plugin never adds anything to the beacon
return true;
}
};
}());
| <|file_sep|>plugins/cache_reload.js.diff
original:
url: "",
initialized: false
updated:
url: ""
<|file_sep|>plugins/cache_reload.js.diff
original:
if(this.initalized)
updated:
if(!impl.url)
<|file_sep|>original/plugins/cache_reload.js
if(this.initalized)
return this;
BOOMR.subscribe(
"page_ready",
function() {
if(!impl.url)
return;
// we use document and not BOOMR.window.document since
// we can run inside the boomerang iframe if any
var i=document.createElement('iframe');
i.style.display="none";
i.src=impl.url;
document.body.appendChild(i);
},
null,
null
);
this.initialized = true;
<|file_sep|>current/plugins/cache_reload.js
if(!impl.url)
return this;
BOOMR.subscribe(
"page_ready",
function() {
if(!impl.url)
return;
// we use document and not BOOMR.window.document since
// we can run inside the boomerang iframe if any
var i=document.createElement('iframe');
i.style.display="none";
i.src=impl.url;
document.body.appendChild(i);
},
null,
null
);
this.initialized = true;
return this;
<|file_sep|>updated/plugins/cache_reload.js
if(!impl.url)
return this;
// we use document and not BOOMR.window.document since
// we can run inside the boomerang iframe if any
var i=document.createElement('iframe');
i.style.display="none";
i.src=impl.url;
document.body.appendChild(i);
return this;
},
is_complete: function() {
// we always return true since this plugin never adds anything to the beacon
return true;
}
};
}());
| 60a2b50f5c49c3c109f3f90e53897a1601bdd256 | plugins/cache_reload.js | plugins/cache_reload.js | JavaScript |
<|file_sep|>original/.travis.yml
before_install:
- node -v
- npm -v
install:
- npm install
- npm install -g grunt-cli@0.1.13
- grunt build
- mvn install -DskipTests=true
script:
- mvn test
after_success:
- git config --global user.email "simon@simondean.org"
- git config --global user.name "Simon Dean"
- mvn --batch-mode release:prepare release:perform --settings settings.xml
- git log -5
env:
secure: aKqIEWlkoAdsMALaxp6lvdIQjvMgBw9a27Wzssp2imRMWqryu36TJkoiAFLPCAZhYBC9avYY/Af6ViQFHY+i8yR959VsOR852X7IguItXnMRio4ucy8WNWpiFoY6bFnLubMKEGiok9d9t8hoNrR5YwnmBCUxh3YyYych2BUjX+s=
<|file_sep|>current/.travis.yml
before_install:
- node -v
- npm -v
install:
- npm install
- npm install -g grunt-cli@0.1.13
- grunt build
- mvn install -DskipTests=true
script:
- mvn test
after_success:
- git config --global user.email "simon@simondean.org"
- git config --global user.name "Simon Dean"
- mvn --batch-mode release:prepare release:perform --settings settings.xml
- git log -5
env:
secure: aKqIEWlkoAdsMALaxp6lvdIQjvMgBw9a27Wzssp2imRMWqryu36TJkoiAFLPCAZhYBC9avYY/Af6ViQFHY+i8yR959VsOR852X7IguItXnMRio4ucy8WNWpiFoY6bFnLubMKEGiok9d9t8hoNrR5YwnmBCUxh3YyYych2BUjX+s=
<|file_sep|>updated/.travis.yml | language: java
jdk:
- oraclejdk8
before_install:
- node -v
- npm -v
install:
- npm install
- npm install -g grunt-cli@0.1.13
- grunt build
- mvn install -DskipTests=true
script:
- mvn test
- mvn --batch-mode deploy -DskipTests=true --settings settings.xml
env:
secure: aKqIEWlkoAdsMALaxp6lvdIQjvMgBw9a27Wzssp2imRMWqryu36TJkoiAFLPCAZhYBC9avYY/Af6ViQFHY+i8yR959VsOR852X7IguItXnMRio4ucy8WNWpiFoY6bFnLubMKEGiok9d9t8hoNrR5YwnmBCUxh3YyYych2BUjX+s= | <|file_sep|>original/.travis.yml
before_install:
- node -v
- npm -v
install:
- npm install
- npm install -g grunt-cli@0.1.13
- grunt build
- mvn install -DskipTests=true
script:
- mvn test
after_success:
- git config --global user.email "simon@simondean.org"
- git config --global user.name "Simon Dean"
- mvn --batch-mode release:prepare release:perform --settings settings.xml
- git log -5
env:
secure: aKqIEWlkoAdsMALaxp6lvdIQjvMgBw9a27Wzssp2imRMWqryu36TJkoiAFLPCAZhYBC9avYY/Af6ViQFHY+i8yR959VsOR852X7IguItXnMRio4ucy8WNWpiFoY6bFnLubMKEGiok9d9t8hoNrR5YwnmBCUxh3YyYych2BUjX+s=
<|file_sep|>current/.travis.yml
before_install:
- node -v
- npm -v
install:
- npm install
- npm install -g grunt-cli@0.1.13
- grunt build
- mvn install -DskipTests=true
script:
- mvn test
after_success:
- git config --global user.email "simon@simondean.org"
- git config --global user.name "Simon Dean"
- mvn --batch-mode release:prepare release:perform --settings settings.xml
- git log -5
env:
secure: aKqIEWlkoAdsMALaxp6lvdIQjvMgBw9a27Wzssp2imRMWqryu36TJkoiAFLPCAZhYBC9avYY/Af6ViQFHY+i8yR959VsOR852X7IguItXnMRio4ucy8WNWpiFoY6bFnLubMKEGiok9d9t8hoNrR5YwnmBCUxh3YyYych2BUjX+s=
<|file_sep|>updated/.travis.yml
language: java
jdk:
- oraclejdk8
before_install:
- node -v
- npm -v
install:
- npm install
- npm install -g grunt-cli@0.1.13
- grunt build
- mvn install -DskipTests=true
script:
- mvn test
- mvn --batch-mode deploy -DskipTests=true --settings settings.xml
env:
secure: aKqIEWlkoAdsMALaxp6lvdIQjvMgBw9a27Wzssp2imRMWqryu36TJkoiAFLPCAZhYBC9avYY/Af6ViQFHY+i8yR959VsOR852X7IguItXnMRio4ucy8WNWpiFoY6bFnLubMKEGiok9d9t8hoNrR5YwnmBCUxh3YyYych2BUjX+s= | ce5fd980f05a32a21b1a1747c3864c4ede8e4a97 | .travis.yml | .travis.yml | YAML |
<|file_sep|>original/manifest.json
{
"manifest_version": 2,
"name": "AutoFlix",
"version": "1.1.3",
"description": "Simplifies the access to Netflix for all users.",
"icons": {
"16": "icon-16.png",
"32": "icon-32.png",
"48": "icon-48.png",
"64": "icon-64.png" ,
"128": "icon-128.png"
},
"browser_action": {
"default_icon": "icon-128.png"
},
"background": {
"scripts": ["background.js"]
<|file_sep|>current/manifest.json
{
"manifest_version": 2,
"name": "AutoFlix",
"version": "1.1.3",
"description": "Simplifies the access to Netflix for all users.",
"icons": {
"16": "icon-16.png",
"32": "icon-32.png",
"48": "icon-48.png",
"64": "icon-64.png" ,
"128": "icon-128.png"
},
"browser_action": {
"default_icon": "icon-128.png"
},
"background": {
"scripts": ["background.js"]
<|file_sep|>updated/manifest.json | {
"manifest_version": 2,
"name": "AutoFlix",
"version": "1.1.4",
"description": "Simplifies the access to Netflix for all users.",
"icons": {
"16": "icon-16.png",
"32": "icon-32.png",
"48": "icon-48.png",
"64": "icon-64.png" ,
"128": "icon-128.png"
},
"browser_action": {
"default_icon": "icon-128.png"
},
"background": {
"scripts": ["background.js"] | <|file_sep|>original/manifest.json
{
"manifest_version": 2,
"name": "AutoFlix",
"version": "1.1.3",
"description": "Simplifies the access to Netflix for all users.",
"icons": {
"16": "icon-16.png",
"32": "icon-32.png",
"48": "icon-48.png",
"64": "icon-64.png" ,
"128": "icon-128.png"
},
"browser_action": {
"default_icon": "icon-128.png"
},
"background": {
"scripts": ["background.js"]
<|file_sep|>current/manifest.json
{
"manifest_version": 2,
"name": "AutoFlix",
"version": "1.1.3",
"description": "Simplifies the access to Netflix for all users.",
"icons": {
"16": "icon-16.png",
"32": "icon-32.png",
"48": "icon-48.png",
"64": "icon-64.png" ,
"128": "icon-128.png"
},
"browser_action": {
"default_icon": "icon-128.png"
},
"background": {
"scripts": ["background.js"]
<|file_sep|>updated/manifest.json
{
"manifest_version": 2,
"name": "AutoFlix",
"version": "1.1.4",
"description": "Simplifies the access to Netflix for all users.",
"icons": {
"16": "icon-16.png",
"32": "icon-32.png",
"48": "icon-48.png",
"64": "icon-64.png" ,
"128": "icon-128.png"
},
"browser_action": {
"default_icon": "icon-128.png"
},
"background": {
"scripts": ["background.js"] | dc06a0292108738f657197352142ce459c9e14df | manifest.json | manifest.json | JSON |
<|file_sep|>original/Casks/tomtom-mysports-connect.rb
<|file_sep|>current/Casks/tomtom-mysports-connect.rb
<|file_sep|>updated/Casks/tomtom-mysports-connect.rb | cask 'tomtom-mysports-connect' do
version :latest
sha256 :no_check
url 'http://download.tomtom.com/sweet/fitness/MySportsConnectDesktop/current/TomTomMySportsConnectInstaller.pkg'
name 'TomTom MySports Connect'
homepage 'http://www.tomtom.com/de_ch/mysports/getstarted/'
license :gpl
pkg 'TomTomMySportsConnectInstaller.pkg'
uninstall :quit => 'com.tomtom.mysportsconnect',
:pkgutil => [
'com.tomtom.tomtomfa.pkg',
'com.tomtom.tomtomfa.temp.pkg',
],
:delete => '/Applications/TomTom MySports Connect.app'
end | <|file_sep|>original/Casks/tomtom-mysports-connect.rb
<|file_sep|>current/Casks/tomtom-mysports-connect.rb
<|file_sep|>updated/Casks/tomtom-mysports-connect.rb
cask 'tomtom-mysports-connect' do
version :latest
sha256 :no_check
url 'http://download.tomtom.com/sweet/fitness/MySportsConnectDesktop/current/TomTomMySportsConnectInstaller.pkg'
name 'TomTom MySports Connect'
homepage 'http://www.tomtom.com/de_ch/mysports/getstarted/'
license :gpl
pkg 'TomTomMySportsConnectInstaller.pkg'
uninstall :quit => 'com.tomtom.mysportsconnect',
:pkgutil => [
'com.tomtom.tomtomfa.pkg',
'com.tomtom.tomtomfa.temp.pkg',
],
:delete => '/Applications/TomTom MySports Connect.app'
end | b513cbe1839215d107b620e897e279e26dcbcb70 | Casks/tomtom-mysports-connect.rb | Casks/tomtom-mysports-connect.rb | Ruby |
<|file_sep|>original/.travis.yml
r_packages:
# - gWidgets
# - gWidgetstcltk
- stringdist
- igraph
# - covr
# r packages directly installed from github
r_github_pages:
- jimhester/covr
# before_install:
# - sudo apt-get install tk
after_success:
- Rscript -e 'covr::codecov()'
# or coveralls
- Rscript -e 'covr::coveralls()'
notifications:
email: false
<|file_sep|>current/.travis.yml
r_packages:
# - gWidgets
# - gWidgetstcltk
- stringdist
- igraph
# - covr
# r packages directly installed from github
r_github_pages:
- jimhester/covr
# before_install:
# - sudo apt-get install tk
after_success:
- Rscript -e 'covr::codecov()'
# or coveralls
- Rscript -e 'covr::coveralls()'
notifications:
email: false
<|file_sep|>updated/.travis.yml | # - gWidgetstcltk
- stringdist
- igraph
# - covr
# r packages directly installed from github
r_github_pages:
- jimhester/covr
# before_install:
# - sudo apt-get install tk
after_success:
- Rscript -e 'covr::codecov()'
# or coveralls
- Rscript -e 'covr::coveralls()'
notifications:
email:
on_success: never # default: change
on_failure: always # default: always | <|file_sep|>original/.travis.yml
r_packages:
# - gWidgets
# - gWidgetstcltk
- stringdist
- igraph
# - covr
# r packages directly installed from github
r_github_pages:
- jimhester/covr
# before_install:
# - sudo apt-get install tk
after_success:
- Rscript -e 'covr::codecov()'
# or coveralls
- Rscript -e 'covr::coveralls()'
notifications:
email: false
<|file_sep|>current/.travis.yml
r_packages:
# - gWidgets
# - gWidgetstcltk
- stringdist
- igraph
# - covr
# r packages directly installed from github
r_github_pages:
- jimhester/covr
# before_install:
# - sudo apt-get install tk
after_success:
- Rscript -e 'covr::codecov()'
# or coveralls
- Rscript -e 'covr::coveralls()'
notifications:
email: false
<|file_sep|>updated/.travis.yml
# - gWidgetstcltk
- stringdist
- igraph
# - covr
# r packages directly installed from github
r_github_pages:
- jimhester/covr
# before_install:
# - sudo apt-get install tk
after_success:
- Rscript -e 'covr::codecov()'
# or coveralls
- Rscript -e 'covr::coveralls()'
notifications:
email:
on_success: never # default: change
on_failure: always # default: always | c8f8edb7d88164336319247459b3b11a8002e1e4 | .travis.yml | .travis.yml | YAML |
<|file_sep|>modules/hdo/templates/deployer.json.diff
original:
"command": ["cap", "staging", "deploy:web:disable", "deploy:migrations", "cache:precompute", "deploy:web:enable"],
updated:
<|file_sep|>original/modules/hdo/templates/deployer.json
{
"logdir": "<%= logdir %>",
"github_client_id": "<%= github_client_id %>",
"github_client_secret": "<%= github_client_secret %>",
"statsd": "ops1.holderdeord.no:8125",
"projects": {
"holderdeord/hdo-site#master": {
"directory": "<%= appdir %>",
"command": ["cap", "staging", "deploy:web:disable", "deploy:migrations", "cache:precompute", "deploy:web:enable"],
"token": "<%= token %>",
"timeout": 180
}
}
}
<|file_sep|>current/modules/hdo/templates/deployer.json
{
"logdir": "<%= logdir %>",
"github_client_id": "<%= github_client_id %>",
"github_client_secret": "<%= github_client_secret %>",
"statsd": "ops1.holderdeord.no:8125",
"projects": {
"holderdeord/hdo-site#master": {
"directory": "<%= appdir %>",
"token": "<%= token %>",
"timeout": 180
}
}
}
<|file_sep|>updated/modules/hdo/templates/deployer.json | {
"logdir": "<%= logdir %>",
"github_client_id": "<%= github_client_id %>",
"github_client_secret": "<%= github_client_secret %>",
"statsd": "ops1.holderdeord.no:8125",
"projects": {
"holderdeord/hdo-site#master": {
"directory": "<%= appdir %>",
"token": "<%= token %>",
"timeout": 180,
"command": [
"cap",
"staging",
"deploy:web:disable",
"deploy:migrations",
"cache:precompute",
"deploy:web:enable",
"deploy:cleanup"
]
}
} | <|file_sep|>modules/hdo/templates/deployer.json.diff
original:
"command": ["cap", "staging", "deploy:web:disable", "deploy:migrations", "cache:precompute", "deploy:web:enable"],
updated:
<|file_sep|>original/modules/hdo/templates/deployer.json
{
"logdir": "<%= logdir %>",
"github_client_id": "<%= github_client_id %>",
"github_client_secret": "<%= github_client_secret %>",
"statsd": "ops1.holderdeord.no:8125",
"projects": {
"holderdeord/hdo-site#master": {
"directory": "<%= appdir %>",
"command": ["cap", "staging", "deploy:web:disable", "deploy:migrations", "cache:precompute", "deploy:web:enable"],
"token": "<%= token %>",
"timeout": 180
}
}
}
<|file_sep|>current/modules/hdo/templates/deployer.json
{
"logdir": "<%= logdir %>",
"github_client_id": "<%= github_client_id %>",
"github_client_secret": "<%= github_client_secret %>",
"statsd": "ops1.holderdeord.no:8125",
"projects": {
"holderdeord/hdo-site#master": {
"directory": "<%= appdir %>",
"token": "<%= token %>",
"timeout": 180
}
}
}
<|file_sep|>updated/modules/hdo/templates/deployer.json
{
"logdir": "<%= logdir %>",
"github_client_id": "<%= github_client_id %>",
"github_client_secret": "<%= github_client_secret %>",
"statsd": "ops1.holderdeord.no:8125",
"projects": {
"holderdeord/hdo-site#master": {
"directory": "<%= appdir %>",
"token": "<%= token %>",
"timeout": 180,
"command": [
"cap",
"staging",
"deploy:web:disable",
"deploy:migrations",
"cache:precompute",
"deploy:web:enable",
"deploy:cleanup"
]
}
} | 350ff106b2d356700eb24e8b9ceb819af263b7f4 | modules/hdo/templates/deployer.json | modules/hdo/templates/deployer.json | JSON |
<|file_sep|>lib/mix/lib/mix/tasks/do.ex.diff
original:
updated:
<|file_sep|>lib/mix/lib/mix/tasks/do.ex.diff
original:
{current, commands} =
case String.split(arg, ",") do
[arg] -> {[arg | current], commands}
# special care if the argument contains a comma
args -> update_commands(args, current, commands)
end
gather_commands(rest, current, commands)
updated:
case String.split(arg, ",", parts: 2) do
[arg] ->
gather_commands(rest, [arg | current], commands)
[left, right] ->
rest = append_unless_empty(right, rest)
current = append_unless_empty(left, current)
gather_commands(rest, [], [current | commands])
end
<|file_sep|>original/lib/mix/lib/mix/tasks/do.ex
case String.split(arg, ",") do
[arg] -> {[arg | current], commands}
# special care if the argument contains a comma
args -> update_commands(args, current, commands)
end
gather_commands(rest, current, commands)
end
defp update_commands([], current, commands) do
{current, commands}
end
defp update_commands([arg], current, commands) when arg != "" do
{[arg], [current | commands]}
end
defp update_commands([arg | args], current, commands) do
# if the argument is empty, we had a leading or trailing comma
# so we simply terminate the current command
command = if arg == "", do: current, else: [arg | current]
update_commands(args, [], [command | commands])
end
end
<|file_sep|>current/lib/mix/lib/mix/tasks/do.ex
gather_commands(rest, [arg | current], commands)
[left, right] ->
rest = append_unless_empty(right, rest)
current = append_unless_empty(left, current)
gather_commands(rest, [], [current | commands])
end
end
defp update_commands([], current, commands) do
{current, commands}
end
defp update_commands([arg], current, commands) when arg != "" do
{[arg], [current | commands]}
end
defp update_commands([arg | args], current, commands) do
# if the argument is empty, we had a leading or trailing comma
# so we simply terminate the current command
command = if arg == "", do: current, else: [arg | current]
update_commands(args, [], [command | commands])
end
end
<|file_sep|>updated/lib/mix/lib/mix/tasks/do.ex | defp gather_commands([], current, commands) do
[current | commands]
|> Enum.reject(&(&1 == []))
|> Enum.map(&Enum.reverse(&1))
|> Enum.reverse
end
defp gather_commands([arg | rest], current, commands) do
case String.split(arg, ",", parts: 2) do
[arg] ->
gather_commands(rest, [arg | current], commands)
[left, right] ->
rest = append_unless_empty(right, rest)
current = append_unless_empty(left, current)
gather_commands(rest, [], [current | commands])
end
end
defp append_unless_empty("", list), do: list
defp append_unless_empty(h, list), do: [h | list]
end | <|file_sep|>lib/mix/lib/mix/tasks/do.ex.diff
original:
updated:
<|file_sep|>lib/mix/lib/mix/tasks/do.ex.diff
original:
{current, commands} =
case String.split(arg, ",") do
[arg] -> {[arg | current], commands}
# special care if the argument contains a comma
args -> update_commands(args, current, commands)
end
gather_commands(rest, current, commands)
updated:
case String.split(arg, ",", parts: 2) do
[arg] ->
gather_commands(rest, [arg | current], commands)
[left, right] ->
rest = append_unless_empty(right, rest)
current = append_unless_empty(left, current)
gather_commands(rest, [], [current | commands])
end
<|file_sep|>original/lib/mix/lib/mix/tasks/do.ex
case String.split(arg, ",") do
[arg] -> {[arg | current], commands}
# special care if the argument contains a comma
args -> update_commands(args, current, commands)
end
gather_commands(rest, current, commands)
end
defp update_commands([], current, commands) do
{current, commands}
end
defp update_commands([arg], current, commands) when arg != "" do
{[arg], [current | commands]}
end
defp update_commands([arg | args], current, commands) do
# if the argument is empty, we had a leading or trailing comma
# so we simply terminate the current command
command = if arg == "", do: current, else: [arg | current]
update_commands(args, [], [command | commands])
end
end
<|file_sep|>current/lib/mix/lib/mix/tasks/do.ex
gather_commands(rest, [arg | current], commands)
[left, right] ->
rest = append_unless_empty(right, rest)
current = append_unless_empty(left, current)
gather_commands(rest, [], [current | commands])
end
end
defp update_commands([], current, commands) do
{current, commands}
end
defp update_commands([arg], current, commands) when arg != "" do
{[arg], [current | commands]}
end
defp update_commands([arg | args], current, commands) do
# if the argument is empty, we had a leading or trailing comma
# so we simply terminate the current command
command = if arg == "", do: current, else: [arg | current]
update_commands(args, [], [command | commands])
end
end
<|file_sep|>updated/lib/mix/lib/mix/tasks/do.ex
defp gather_commands([], current, commands) do
[current | commands]
|> Enum.reject(&(&1 == []))
|> Enum.map(&Enum.reverse(&1))
|> Enum.reverse
end
defp gather_commands([arg | rest], current, commands) do
case String.split(arg, ",", parts: 2) do
[arg] ->
gather_commands(rest, [arg | current], commands)
[left, right] ->
rest = append_unless_empty(right, rest)
current = append_unless_empty(left, current)
gather_commands(rest, [], [current | commands])
end
end
defp append_unless_empty("", list), do: list
defp append_unless_empty(h, list), do: [h | list]
end | 70ba8464c9fb2745aaeb0f53783c1361b8db982e | lib/mix/lib/mix/tasks/do.ex | lib/mix/lib/mix/tasks/do.ex | Elixir |
<|file_sep|>original/examples/wfs.html
height: 475px;
border: 1px solid black;
}
</style>
<script src="../lib/OpenLayers.js"></script>
<script type="text/javascript">
<!--
var map, layer;
function init(){
map = new OpenLayers.Map('map');
layer = new OpenLayers.Layer.WMS( "OpenLayers WMS",
"http://labs.metacarta.com/wms/vmap0", {layers: 'basic'} );
map.addLayer(layer);
layer = new OpenLayers.Layer.WFS( "OpenLayers WFS",
"http://octo.metacarta.com/services/search/WFS",
{q: 'dolphins', maxfeatures: 10} );
map.addLayer(layer);
<|file_sep|>current/examples/wfs.html
height: 475px;
border: 1px solid black;
}
</style>
<script src="../lib/OpenLayers.js"></script>
<script type="text/javascript">
<!--
var map, layer;
function init(){
map = new OpenLayers.Map('map');
layer = new OpenLayers.Layer.WMS( "OpenLayers WMS",
"http://labs.metacarta.com/wms/vmap0", {layers: 'basic'} );
map.addLayer(layer);
layer = new OpenLayers.Layer.WFS( "OpenLayers WFS",
"http://octo.metacarta.com/services/search/WFS",
{q: 'dolphins', maxfeatures: 10} );
map.addLayer(layer);
<|file_sep|>updated/examples/wfs.html | height: 475px;
border: 1px solid black;
}
</style>
<script src="../lib/OpenLayers.js"></script>
<script type="text/javascript">
<!--
var map, layer;
function init(){
OpenLayers.ProxyHost = '/dev/examples/proxy.cgi?url=';
map = new OpenLayers.Map('map');
layer = new OpenLayers.Layer.WMS( "OpenLayers WMS",
"http://labs.metacarta.com/wms/vmap0", {layers: 'basic'} );
map.addLayer(layer);
layer = new OpenLayers.Layer.WFS( "OpenLayers WFS",
"http://octo.metacarta.com/services/search/WFS",
{q: 'dolphins', maxfeatures: 10} );
map.addLayer(layer); | <|file_sep|>original/examples/wfs.html
height: 475px;
border: 1px solid black;
}
</style>
<script src="../lib/OpenLayers.js"></script>
<script type="text/javascript">
<!--
var map, layer;
function init(){
map = new OpenLayers.Map('map');
layer = new OpenLayers.Layer.WMS( "OpenLayers WMS",
"http://labs.metacarta.com/wms/vmap0", {layers: 'basic'} );
map.addLayer(layer);
layer = new OpenLayers.Layer.WFS( "OpenLayers WFS",
"http://octo.metacarta.com/services/search/WFS",
{q: 'dolphins', maxfeatures: 10} );
map.addLayer(layer);
<|file_sep|>current/examples/wfs.html
height: 475px;
border: 1px solid black;
}
</style>
<script src="../lib/OpenLayers.js"></script>
<script type="text/javascript">
<!--
var map, layer;
function init(){
map = new OpenLayers.Map('map');
layer = new OpenLayers.Layer.WMS( "OpenLayers WMS",
"http://labs.metacarta.com/wms/vmap0", {layers: 'basic'} );
map.addLayer(layer);
layer = new OpenLayers.Layer.WFS( "OpenLayers WFS",
"http://octo.metacarta.com/services/search/WFS",
{q: 'dolphins', maxfeatures: 10} );
map.addLayer(layer);
<|file_sep|>updated/examples/wfs.html
height: 475px;
border: 1px solid black;
}
</style>
<script src="../lib/OpenLayers.js"></script>
<script type="text/javascript">
<!--
var map, layer;
function init(){
OpenLayers.ProxyHost = '/dev/examples/proxy.cgi?url=';
map = new OpenLayers.Map('map');
layer = new OpenLayers.Layer.WMS( "OpenLayers WMS",
"http://labs.metacarta.com/wms/vmap0", {layers: 'basic'} );
map.addLayer(layer);
layer = new OpenLayers.Layer.WFS( "OpenLayers WFS",
"http://octo.metacarta.com/services/search/WFS",
{q: 'dolphins', maxfeatures: 10} );
map.addLayer(layer); | 1a3d466f2a681cd5f3f9bff74a6eddf15d4a7c6b | examples/wfs.html | examples/wfs.html | HTML |
<|file_sep|>original/requirements_dev.txt
# Database
psycopg2-binary==2.7.7
pip==19.0.1
bumpversion==0.5.3
wheel==0.32.3
watchdog==0.9.0
flake8==3.7.5
tox==3.7.0
cryptography==2.5
Sphinx==1.8.4
coverage==4.5.2
pytest==4.2.0
pytest-runner==4.2
pytest-cov==2.6.1
# miscellaneous
sphinx_rtd_theme==0.4.2
pre-commit==1.14.3
<|file_sep|>current/requirements_dev.txt
# Database
psycopg2-binary==2.7.7
pip==19.0.1
bumpversion==0.5.3
wheel==0.32.3
watchdog==0.9.0
flake8==3.7.5
tox==3.7.0
cryptography==2.5
Sphinx==1.8.4
coverage==4.5.2
pytest==4.2.0
pytest-runner==4.2
pytest-cov==2.6.1
# miscellaneous
sphinx_rtd_theme==0.4.2
pre-commit==1.14.3
<|file_sep|>updated/requirements_dev.txt |
# Database
psycopg2-binary==2.7.7
pip==19.0.1
bumpversion==0.5.3
wheel==0.33.0
watchdog==0.9.0
flake8==3.7.5
tox==3.7.0
cryptography==2.5
Sphinx==1.8.4
coverage==4.5.2
pytest==4.2.0
pytest-runner==4.2
pytest-cov==2.6.1
# miscellaneous
sphinx_rtd_theme==0.4.2
pre-commit==1.14.3
| <|file_sep|>original/requirements_dev.txt
# Database
psycopg2-binary==2.7.7
pip==19.0.1
bumpversion==0.5.3
wheel==0.32.3
watchdog==0.9.0
flake8==3.7.5
tox==3.7.0
cryptography==2.5
Sphinx==1.8.4
coverage==4.5.2
pytest==4.2.0
pytest-runner==4.2
pytest-cov==2.6.1
# miscellaneous
sphinx_rtd_theme==0.4.2
pre-commit==1.14.3
<|file_sep|>current/requirements_dev.txt
# Database
psycopg2-binary==2.7.7
pip==19.0.1
bumpversion==0.5.3
wheel==0.32.3
watchdog==0.9.0
flake8==3.7.5
tox==3.7.0
cryptography==2.5
Sphinx==1.8.4
coverage==4.5.2
pytest==4.2.0
pytest-runner==4.2
pytest-cov==2.6.1
# miscellaneous
sphinx_rtd_theme==0.4.2
pre-commit==1.14.3
<|file_sep|>updated/requirements_dev.txt
# Database
psycopg2-binary==2.7.7
pip==19.0.1
bumpversion==0.5.3
wheel==0.33.0
watchdog==0.9.0
flake8==3.7.5
tox==3.7.0
cryptography==2.5
Sphinx==1.8.4
coverage==4.5.2
pytest==4.2.0
pytest-runner==4.2
pytest-cov==2.6.1
# miscellaneous
sphinx_rtd_theme==0.4.2
pre-commit==1.14.3
| 0f41d6a1639841697d16fe0ee84664f8c41b3a78 | requirements_dev.txt | requirements_dev.txt | Text |
<|file_sep|>original/LICENSE.md
The MIT License (MIT)
Copyright (c) 2015 Patrick M. Jensen, patmjen@gmail.com
Permission is hereby granted, free of charge, to any person obtaining a copy
of this software and associated documentation files (the "Software"), to deal
in the Software without restriction, including without limitation the rights
to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
copies of the Software, and to permit persons to whom the Software is
furnished to do so, subject to the following conditions:
The above copyright notice and this permission notice shall be included in
all copies or substantial portions of the Software.
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
THE SOFTWARE.
<|file_sep|>current/LICENSE.md
The MIT License (MIT)
Copyright (c) 2015 Patrick M. Jensen, patmjen@gmail.com
Permission is hereby granted, free of charge, to any person obtaining a copy
of this software and associated documentation files (the "Software"), to deal
in the Software without restriction, including without limitation the rights
to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
copies of the Software, and to permit persons to whom the Software is
furnished to do so, subject to the following conditions:
The above copyright notice and this permission notice shall be included in
all copies or substantial portions of the Software.
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
THE SOFTWARE.
<|file_sep|>updated/LICENSE.md | The MIT License (MIT)
Copyright (c) 2016 Patrick M. Jensen, patmjen@gmail.com
Permission is hereby granted, free of charge, to any person obtaining a copy
of this software and associated documentation files (the "Software"), to deal
in the Software without restriction, including without limitation the rights
to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
copies of the Software, and to permit persons to whom the Software is
furnished to do so, subject to the following conditions:
The above copyright notice and this permission notice shall be included in
all copies or substantial portions of the Software.
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
THE SOFTWARE. | <|file_sep|>original/LICENSE.md
The MIT License (MIT)
Copyright (c) 2015 Patrick M. Jensen, patmjen@gmail.com
Permission is hereby granted, free of charge, to any person obtaining a copy
of this software and associated documentation files (the "Software"), to deal
in the Software without restriction, including without limitation the rights
to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
copies of the Software, and to permit persons to whom the Software is
furnished to do so, subject to the following conditions:
The above copyright notice and this permission notice shall be included in
all copies or substantial portions of the Software.
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
THE SOFTWARE.
<|file_sep|>current/LICENSE.md
The MIT License (MIT)
Copyright (c) 2015 Patrick M. Jensen, patmjen@gmail.com
Permission is hereby granted, free of charge, to any person obtaining a copy
of this software and associated documentation files (the "Software"), to deal
in the Software without restriction, including without limitation the rights
to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
copies of the Software, and to permit persons to whom the Software is
furnished to do so, subject to the following conditions:
The above copyright notice and this permission notice shall be included in
all copies or substantial portions of the Software.
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
THE SOFTWARE.
<|file_sep|>updated/LICENSE.md
The MIT License (MIT)
Copyright (c) 2016 Patrick M. Jensen, patmjen@gmail.com
Permission is hereby granted, free of charge, to any person obtaining a copy
of this software and associated documentation files (the "Software"), to deal
in the Software without restriction, including without limitation the rights
to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
copies of the Software, and to permit persons to whom the Software is
furnished to do so, subject to the following conditions:
The above copyright notice and this permission notice shall be included in
all copies or substantial portions of the Software.
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
THE SOFTWARE. | 148f46220e7c3028936032307706d4a7ddc85330 | LICENSE.md | LICENSE.md | Markdown |
<|file_sep|>original/app/widgets/ChatOmemo/locales.ini
[omemo]
fingerprints_title = Fingerprints
fingerprints = OMEMO Fingerprints
generating_bundle = "Generating the encryption keys, please wait…"
generated_bundle = Encryption keys generated
resolving_devices = Resolving OMEMO devices
building_sessions = Building encryption sessions
building_own_sessions = Building own encryption sessions
encrypted = Encrypted message
encrypted_disabled = Encryption disabled
encrypted_loading = Send a message to build the encrypted session
enable_contact = Enabling OMEMO for the contact
disable_contact = Disabling OMEMO for the contact
last_activity = Last activity
sessions_built = %s sessions built
<|file_sep|>current/app/widgets/ChatOmemo/locales.ini
[omemo]
fingerprints_title = Fingerprints
fingerprints = OMEMO Fingerprints
generating_bundle = "Generating the encryption keys, please wait…"
generated_bundle = Encryption keys generated
resolving_devices = Resolving OMEMO devices
building_sessions = Building encryption sessions
building_own_sessions = Building own encryption sessions
encrypted = Encrypted message
encrypted_disabled = Encryption disabled
encrypted_loading = Send a message to build the encrypted session
enable_contact = Enabling OMEMO for the contact
disable_contact = Disabling OMEMO for the contact
last_activity = Last activity
sessions_built = %s sessions built
<|file_sep|>updated/app/widgets/ChatOmemo/locales.ini | [omemo]
fingerprints_title = Fingerprints
fingerprints = OMEMO Fingerprints
generating_bundle = "Generating the encryption keys, please wait…"
generated_bundle = Encryption keys generated
resolving_devices = Resolving OMEMO devices
building_sessions = Building encryption sessions
building_own_sessions = Building own encryption sessions
encrypted = Encrypted message
encrypted_disabled = Encryption disabled
encrypted_loading = Send a message to build the encrypted session
enable_contact = Enabling encryption for the contact
disable_contact = Disabling encryption for the contact
last_activity = Last activity
sessions_built = %s sessions built | <|file_sep|>original/app/widgets/ChatOmemo/locales.ini
[omemo]
fingerprints_title = Fingerprints
fingerprints = OMEMO Fingerprints
generating_bundle = "Generating the encryption keys, please wait…"
generated_bundle = Encryption keys generated
resolving_devices = Resolving OMEMO devices
building_sessions = Building encryption sessions
building_own_sessions = Building own encryption sessions
encrypted = Encrypted message
encrypted_disabled = Encryption disabled
encrypted_loading = Send a message to build the encrypted session
enable_contact = Enabling OMEMO for the contact
disable_contact = Disabling OMEMO for the contact
last_activity = Last activity
sessions_built = %s sessions built
<|file_sep|>current/app/widgets/ChatOmemo/locales.ini
[omemo]
fingerprints_title = Fingerprints
fingerprints = OMEMO Fingerprints
generating_bundle = "Generating the encryption keys, please wait…"
generated_bundle = Encryption keys generated
resolving_devices = Resolving OMEMO devices
building_sessions = Building encryption sessions
building_own_sessions = Building own encryption sessions
encrypted = Encrypted message
encrypted_disabled = Encryption disabled
encrypted_loading = Send a message to build the encrypted session
enable_contact = Enabling OMEMO for the contact
disable_contact = Disabling OMEMO for the contact
last_activity = Last activity
sessions_built = %s sessions built
<|file_sep|>updated/app/widgets/ChatOmemo/locales.ini
[omemo]
fingerprints_title = Fingerprints
fingerprints = OMEMO Fingerprints
generating_bundle = "Generating the encryption keys, please wait…"
generated_bundle = Encryption keys generated
resolving_devices = Resolving OMEMO devices
building_sessions = Building encryption sessions
building_own_sessions = Building own encryption sessions
encrypted = Encrypted message
encrypted_disabled = Encryption disabled
encrypted_loading = Send a message to build the encrypted session
enable_contact = Enabling encryption for the contact
disable_contact = Disabling encryption for the contact
last_activity = Last activity
sessions_built = %s sessions built | 3a6212ea81b8689041e898df1fa458c6e1c7e806 | app/widgets/ChatOmemo/locales.ini | app/widgets/ChatOmemo/locales.ini | INI |
<|file_sep|>original/app/jobs/geocode_firm_job.rb
require 'geocoder'
class GeocodeFirmJob < ActiveJob::Base
def perform(firm)
coordinates = Geocoder.coordinates(firm.full_street_address)
coordinates ? stat(:success) : stat(:failed)
firm.geocode!(coordinates)
end
private
def stat(key)
Stats.increment("radsignup.geocode_firm.#{key}")
end
end
<|file_sep|>current/app/jobs/geocode_firm_job.rb
require 'geocoder'
class GeocodeFirmJob < ActiveJob::Base
def perform(firm)
coordinates = Geocoder.coordinates(firm.full_street_address)
coordinates ? stat(:success) : stat(:failed)
firm.geocode!(coordinates)
end
private
def stat(key)
Stats.increment("radsignup.geocode_firm.#{key}")
end
end
<|file_sep|>updated/app/jobs/geocode_firm_job.rb | require 'geocoder'
class GeocodeFirmJob < ActiveJob::Base
def perform(firm)
coordinates = Geocoder.coordinates(firm.full_street_address)
coordinates ? stat(:success) : stat(:failed)
firm.geocode!(coordinates)
end
private
def stat(key)
Stats.increment("radsignup.geocode.firm.#{key}")
end
end | <|file_sep|>original/app/jobs/geocode_firm_job.rb
require 'geocoder'
class GeocodeFirmJob < ActiveJob::Base
def perform(firm)
coordinates = Geocoder.coordinates(firm.full_street_address)
coordinates ? stat(:success) : stat(:failed)
firm.geocode!(coordinates)
end
private
def stat(key)
Stats.increment("radsignup.geocode_firm.#{key}")
end
end
<|file_sep|>current/app/jobs/geocode_firm_job.rb
require 'geocoder'
class GeocodeFirmJob < ActiveJob::Base
def perform(firm)
coordinates = Geocoder.coordinates(firm.full_street_address)
coordinates ? stat(:success) : stat(:failed)
firm.geocode!(coordinates)
end
private
def stat(key)
Stats.increment("radsignup.geocode_firm.#{key}")
end
end
<|file_sep|>updated/app/jobs/geocode_firm_job.rb
require 'geocoder'
class GeocodeFirmJob < ActiveJob::Base
def perform(firm)
coordinates = Geocoder.coordinates(firm.full_street_address)
coordinates ? stat(:success) : stat(:failed)
firm.geocode!(coordinates)
end
private
def stat(key)
Stats.increment("radsignup.geocode.firm.#{key}")
end
end | 3d03fc9be9ab1e2e661ca56ef87bf4f10b2fe993 | app/jobs/geocode_firm_job.rb | app/jobs/geocode_firm_job.rb | Ruby |
<|file_sep|>original/2d/skeleton/README.md
# Skeleton2D Demo
This demo shows how to create a rigged and animated character in 2D using
Godot's Skeleton2D node. There are several movement-related animations and
there is a simple character controller that controls the animations.
Language: GDScript
Renderer: GLES 2
## Licenses
GBot character Copyright © circa 2020 Andreas Esau, MIT License.
Initial rigging and animating Copyright © 2020 RustyStriker, MIT License.
## Screenshots

<|file_sep|>current/2d/skeleton/README.md
# Skeleton2D Demo
This demo shows how to create a rigged and animated character in 2D using
Godot's Skeleton2D node. There are several movement-related animations and
there is a simple character controller that controls the animations.
Language: GDScript
Renderer: GLES 2
## Licenses
GBot character Copyright © circa 2020 Andreas Esau, MIT License.
Initial rigging and animating Copyright © 2020 RustyStriker, MIT License.
## Screenshots

<|file_sep|>updated/2d/skeleton/README.md | # Skeleton2D Demo
This demo shows how to create a rigged and animated character in 2D using
Godot's Skeleton2D node. There are several movement-related animations and
there is a simple character controller that controls the animations.
Language: GDScript
Renderer: GLES 2
Check out this demo on the asset library: https://godotengine.org/asset-library/asset/1027
## Licenses
GBot character Copyright © circa 2020 Andreas Esau, MIT License.
Initial rigging and animating Copyright © 2020 RustyStriker, MIT License.
## Screenshots
 | <|file_sep|>original/2d/skeleton/README.md
# Skeleton2D Demo
This demo shows how to create a rigged and animated character in 2D using
Godot's Skeleton2D node. There are several movement-related animations and
there is a simple character controller that controls the animations.
Language: GDScript
Renderer: GLES 2
## Licenses
GBot character Copyright © circa 2020 Andreas Esau, MIT License.
Initial rigging and animating Copyright © 2020 RustyStriker, MIT License.
## Screenshots

<|file_sep|>current/2d/skeleton/README.md
# Skeleton2D Demo
This demo shows how to create a rigged and animated character in 2D using
Godot's Skeleton2D node. There are several movement-related animations and
there is a simple character controller that controls the animations.
Language: GDScript
Renderer: GLES 2
## Licenses
GBot character Copyright © circa 2020 Andreas Esau, MIT License.
Initial rigging and animating Copyright © 2020 RustyStriker, MIT License.
## Screenshots

<|file_sep|>updated/2d/skeleton/README.md
# Skeleton2D Demo
This demo shows how to create a rigged and animated character in 2D using
Godot's Skeleton2D node. There are several movement-related animations and
there is a simple character controller that controls the animations.
Language: GDScript
Renderer: GLES 2
Check out this demo on the asset library: https://godotengine.org/asset-library/asset/1027
## Licenses
GBot character Copyright © circa 2020 Andreas Esau, MIT License.
Initial rigging and animating Copyright © 2020 RustyStriker, MIT License.
## Screenshots
 | 9ea18370d3ba71d2bad7b944eb521dc309b3f513 | 2d/skeleton/README.md | 2d/skeleton/README.md | Markdown |
<|file_sep|>original/requirements/prod.txt
beautifulsoup4==4.8.2
eventlet==0.25.1
gunicorn==20.0.4
# CSS/JS Assets
flask_assets==2.0
yuicompressor==2.4.8
# Sockets
flask_socketio==4.2.1
python_socketio==4.4.0
python_engineio==3.11.2
# Celery
celery==4.4.1
redis==3.4.1
# Octave interaction
scipy==1.4.1
numpy==1.18.1
octave-kernel==0.31.0
<|file_sep|>current/requirements/prod.txt
beautifulsoup4==4.8.2
eventlet==0.25.1
gunicorn==20.0.4
# CSS/JS Assets
flask_assets==2.0
yuicompressor==2.4.8
# Sockets
flask_socketio==4.2.1
python_socketio==4.4.0
python_engineio==3.11.2
# Celery
celery==4.4.1
redis==3.4.1
# Octave interaction
scipy==1.4.1
numpy==1.18.1
octave-kernel==0.31.0
<|file_sep|>updated/requirements/prod.txt | beautifulsoup4==4.8.2
eventlet==0.25.1
gunicorn==20.0.4
# CSS/JS Assets
flask_assets==2.0
yuicompressor==2.4.8
# Sockets
flask_socketio==4.2.1
python_socketio==4.5.0
python_engineio==3.11.2
# Celery
celery==4.4.1
redis==3.4.1
# Octave interaction
scipy==1.4.1
numpy==1.18.1
octave-kernel==0.31.0 | <|file_sep|>original/requirements/prod.txt
beautifulsoup4==4.8.2
eventlet==0.25.1
gunicorn==20.0.4
# CSS/JS Assets
flask_assets==2.0
yuicompressor==2.4.8
# Sockets
flask_socketio==4.2.1
python_socketio==4.4.0
python_engineio==3.11.2
# Celery
celery==4.4.1
redis==3.4.1
# Octave interaction
scipy==1.4.1
numpy==1.18.1
octave-kernel==0.31.0
<|file_sep|>current/requirements/prod.txt
beautifulsoup4==4.8.2
eventlet==0.25.1
gunicorn==20.0.4
# CSS/JS Assets
flask_assets==2.0
yuicompressor==2.4.8
# Sockets
flask_socketio==4.2.1
python_socketio==4.4.0
python_engineio==3.11.2
# Celery
celery==4.4.1
redis==3.4.1
# Octave interaction
scipy==1.4.1
numpy==1.18.1
octave-kernel==0.31.0
<|file_sep|>updated/requirements/prod.txt
beautifulsoup4==4.8.2
eventlet==0.25.1
gunicorn==20.0.4
# CSS/JS Assets
flask_assets==2.0
yuicompressor==2.4.8
# Sockets
flask_socketio==4.2.1
python_socketio==4.5.0
python_engineio==3.11.2
# Celery
celery==4.4.1
redis==3.4.1
# Octave interaction
scipy==1.4.1
numpy==1.18.1
octave-kernel==0.31.0 | 3c0e8bc67d8226e121f18ff917cde81ed46e8a94 | requirements/prod.txt | requirements/prod.txt | Text |
<|file_sep|>original/package.json
"main": "index.js",
"scripts": {
"test": "find . \\( -name node_modules -prune \\) -o -name \"*.js\" -exec jshint \"{}\" \\;"
},
"author": "Edward Faulkner <ef@alum.mit.edu>",
"license": "MIT",
"repository": "https://github.com/ef4/ember-code-snippet",
"dependencies": {
"highlight.js": "^8.1.0",
"broccoli-static-compiler": "^0.1.4",
"broccoli-merge-trees": "^0.1.4",
"broccoli-browserify": "^0.1.0",
"broccoli-spelunk": "^0.1.2",
"es6-promise": "^1.0.0",
"broccoli-writer": "^0.1.1",
"glob": "^4.0.4"
},
"keywords": [
"ember-addon"
]
}
<|file_sep|>current/package.json
"main": "index.js",
"scripts": {
"test": "find . \\( -name node_modules -prune \\) -o -name \"*.js\" -exec jshint \"{}\" \\;"
},
"author": "Edward Faulkner <ef@alum.mit.edu>",
"license": "MIT",
"repository": "https://github.com/ef4/ember-code-snippet",
"dependencies": {
"highlight.js": "^8.1.0",
"broccoli-static-compiler": "^0.1.4",
"broccoli-merge-trees": "^0.1.4",
"broccoli-browserify": "^0.1.0",
"broccoli-spelunk": "^0.1.2",
"es6-promise": "^1.0.0",
"broccoli-writer": "^0.1.1",
"glob": "^4.0.4"
},
"keywords": [
"ember-addon"
]
}
<|file_sep|>updated/package.json | "main": "index.js",
"scripts": {
"test": "find . \\( -name node_modules -prune \\) -o -name \"*.js\" -exec jshint \"{}\" \\;"
},
"author": "Edward Faulkner <ef@alum.mit.edu>",
"license": "MIT",
"repository": "https://github.com/ef4/ember-code-snippet",
"dependencies": {
"highlight.js": "^8.1.0",
"broccoli-static-compiler": "^0.1.4",
"broccoli-merge-trees": "^0.1.4",
"broccoli-browserify": "^0.1.0",
"broccoli-flatiron": "^0.0.0",
"es6-promise": "^1.0.0",
"broccoli-writer": "^0.1.1",
"glob": "^4.0.4"
},
"keywords": [
"ember-addon"
]
} | <|file_sep|>original/package.json
"main": "index.js",
"scripts": {
"test": "find . \\( -name node_modules -prune \\) -o -name \"*.js\" -exec jshint \"{}\" \\;"
},
"author": "Edward Faulkner <ef@alum.mit.edu>",
"license": "MIT",
"repository": "https://github.com/ef4/ember-code-snippet",
"dependencies": {
"highlight.js": "^8.1.0",
"broccoli-static-compiler": "^0.1.4",
"broccoli-merge-trees": "^0.1.4",
"broccoli-browserify": "^0.1.0",
"broccoli-spelunk": "^0.1.2",
"es6-promise": "^1.0.0",
"broccoli-writer": "^0.1.1",
"glob": "^4.0.4"
},
"keywords": [
"ember-addon"
]
}
<|file_sep|>current/package.json
"main": "index.js",
"scripts": {
"test": "find . \\( -name node_modules -prune \\) -o -name \"*.js\" -exec jshint \"{}\" \\;"
},
"author": "Edward Faulkner <ef@alum.mit.edu>",
"license": "MIT",
"repository": "https://github.com/ef4/ember-code-snippet",
"dependencies": {
"highlight.js": "^8.1.0",
"broccoli-static-compiler": "^0.1.4",
"broccoli-merge-trees": "^0.1.4",
"broccoli-browserify": "^0.1.0",
"broccoli-spelunk": "^0.1.2",
"es6-promise": "^1.0.0",
"broccoli-writer": "^0.1.1",
"glob": "^4.0.4"
},
"keywords": [
"ember-addon"
]
}
<|file_sep|>updated/package.json
"main": "index.js",
"scripts": {
"test": "find . \\( -name node_modules -prune \\) -o -name \"*.js\" -exec jshint \"{}\" \\;"
},
"author": "Edward Faulkner <ef@alum.mit.edu>",
"license": "MIT",
"repository": "https://github.com/ef4/ember-code-snippet",
"dependencies": {
"highlight.js": "^8.1.0",
"broccoli-static-compiler": "^0.1.4",
"broccoli-merge-trees": "^0.1.4",
"broccoli-browserify": "^0.1.0",
"broccoli-flatiron": "^0.0.0",
"es6-promise": "^1.0.0",
"broccoli-writer": "^0.1.1",
"glob": "^4.0.4"
},
"keywords": [
"ember-addon"
]
} | 91f8e4f0f4a5d06e4a6fd9b622966df633c2e34b | package.json | package.json | JSON |
<|file_sep|>lib/alchemy/logger.rb.diff
original:
# Logs a warning to the Rails standard logger and adds some nicer formatting
updated:
# Logs a debug message to the Rails standard logger and adds some nicer formatting
<|file_sep|>original/lib/alchemy/logger.rb
module Alchemy
module Logger
# Logs a warning to the Rails standard logger and adds some nicer formatting
def self.warn(message, caller_string)
Rails.logger.warn %(\n++++ WARNING: #{message}\nCalled from: #{caller_string}\n)
return nil
end
def log_warning(message)
Alchemy::Logger.warn(message, caller.first)
end
end
end
<|file_sep|>current/lib/alchemy/logger.rb
module Alchemy
module Logger
# Logs a debug message to the Rails standard logger and adds some nicer formatting
def self.warn(message, caller_string)
Rails.logger.warn %(\n++++ WARNING: #{message}\nCalled from: #{caller_string}\n)
return nil
end
def log_warning(message)
Alchemy::Logger.warn(message, caller.first)
end
end
end
<|file_sep|>updated/lib/alchemy/logger.rb | module Alchemy
module Logger
# Logs a debug message to the Rails standard logger and adds some nicer formatting
def self.warn(message, caller_string)
Rails.logger.debug %(\n++++ WARNING: #{message}\nCalled from: #{caller_string}\n)
return nil
end
def log_warning(message)
Alchemy::Logger.warn(message, caller.first)
end
end
end | <|file_sep|>lib/alchemy/logger.rb.diff
original:
# Logs a warning to the Rails standard logger and adds some nicer formatting
updated:
# Logs a debug message to the Rails standard logger and adds some nicer formatting
<|file_sep|>original/lib/alchemy/logger.rb
module Alchemy
module Logger
# Logs a warning to the Rails standard logger and adds some nicer formatting
def self.warn(message, caller_string)
Rails.logger.warn %(\n++++ WARNING: #{message}\nCalled from: #{caller_string}\n)
return nil
end
def log_warning(message)
Alchemy::Logger.warn(message, caller.first)
end
end
end
<|file_sep|>current/lib/alchemy/logger.rb
module Alchemy
module Logger
# Logs a debug message to the Rails standard logger and adds some nicer formatting
def self.warn(message, caller_string)
Rails.logger.warn %(\n++++ WARNING: #{message}\nCalled from: #{caller_string}\n)
return nil
end
def log_warning(message)
Alchemy::Logger.warn(message, caller.first)
end
end
end
<|file_sep|>updated/lib/alchemy/logger.rb
module Alchemy
module Logger
# Logs a debug message to the Rails standard logger and adds some nicer formatting
def self.warn(message, caller_string)
Rails.logger.debug %(\n++++ WARNING: #{message}\nCalled from: #{caller_string}\n)
return nil
end
def log_warning(message)
Alchemy::Logger.warn(message, caller.first)
end
end
end | 07f8826eddc964d46e3476c279f7819ec9e1cfd1 | lib/alchemy/logger.rb | lib/alchemy/logger.rb | Ruby |
<|file_sep|>src/Psy/Command/TimeitCommand.php.diff
original:
updated:
use Psy\Input\CodeArgument;
<|file_sep|>src/Psy/Command/TimeitCommand.php.diff
original:
updated:
use Symfony\Component\Console\Input\InputArgument;
<|file_sep|>src/Psy/Command/TimeitCommand.php.diff
original:
use Symfony\Component\Console\Input\InputArgument;
updated:
<|file_sep|>src/Psy/Command/TimeitCommand.php.diff
original:
new InputArgument('target', InputArgument::REQUIRED, 'A target object or primitive to profile.', null),
updated:
new CodeArgument('code', InputArgument::REQUIRED, 'Code to execute.'),
<|file_sep|>src/Psy/Command/TimeitCommand.php.diff
original:
$target = $input->getArgument('target');
$start = microtime(true);
updated:
$code = $input->getArgument('code');
<|file_sep|>original/src/Psy/Command/TimeitCommand.php
}
/**
* {@inheritdoc}
*/
protected function execute(InputInterface $input, OutputInterface $output)
{
$target = $input->getArgument('target');
$start = microtime(true);
/** @var Shell $shell */
$shell = $this->getApplication();
$sh = new Shell(new Configuration());
$sh->setOutput($output);
$sh->setScopeVariables($shell->getScopeVariables());
$sh->execute($target);
$end = microtime(true);
$output->writeln(sprintf('<info>Command took %.6f seconds to complete.</info>', $end-$start));
}
<|file_sep|>current/src/Psy/Command/TimeitCommand.php
);
}
/**
* {@inheritdoc}
*/
protected function execute(InputInterface $input, OutputInterface $output)
{
$code = $input->getArgument('code');
/** @var Shell $shell */
$shell = $this->getApplication();
$sh = new Shell(new Configuration());
$sh->setOutput($output);
$sh->setScopeVariables($shell->getScopeVariables());
$end = microtime(true);
$output->writeln(sprintf('<info>Command took %.6f seconds to complete.</info>', $end-$start));
}
}
<|file_sep|>updated/src/Psy/Command/TimeitCommand.php |
/**
* {@inheritdoc}
*/
protected function execute(InputInterface $input, OutputInterface $output)
{
$code = $input->getArgument('code');
/** @var Shell $shell */
$shell = $this->getApplication();
$sh = new Shell(new Configuration());
$sh->setOutput($output);
$sh->setScopeVariables($shell->getScopeVariables());
$start = microtime(true);
$sh->execute($code);
$end = microtime(true);
$output->writeln(sprintf('<info>Command took %.6f seconds to complete.</info>', $end-$start));
}
} | <|file_sep|>src/Psy/Command/TimeitCommand.php.diff
original:
updated:
use Psy\Input\CodeArgument;
<|file_sep|>src/Psy/Command/TimeitCommand.php.diff
original:
updated:
use Symfony\Component\Console\Input\InputArgument;
<|file_sep|>src/Psy/Command/TimeitCommand.php.diff
original:
use Symfony\Component\Console\Input\InputArgument;
updated:
<|file_sep|>src/Psy/Command/TimeitCommand.php.diff
original:
new InputArgument('target', InputArgument::REQUIRED, 'A target object or primitive to profile.', null),
updated:
new CodeArgument('code', InputArgument::REQUIRED, 'Code to execute.'),
<|file_sep|>src/Psy/Command/TimeitCommand.php.diff
original:
$target = $input->getArgument('target');
$start = microtime(true);
updated:
$code = $input->getArgument('code');
<|file_sep|>original/src/Psy/Command/TimeitCommand.php
}
/**
* {@inheritdoc}
*/
protected function execute(InputInterface $input, OutputInterface $output)
{
$target = $input->getArgument('target');
$start = microtime(true);
/** @var Shell $shell */
$shell = $this->getApplication();
$sh = new Shell(new Configuration());
$sh->setOutput($output);
$sh->setScopeVariables($shell->getScopeVariables());
$sh->execute($target);
$end = microtime(true);
$output->writeln(sprintf('<info>Command took %.6f seconds to complete.</info>', $end-$start));
}
<|file_sep|>current/src/Psy/Command/TimeitCommand.php
);
}
/**
* {@inheritdoc}
*/
protected function execute(InputInterface $input, OutputInterface $output)
{
$code = $input->getArgument('code');
/** @var Shell $shell */
$shell = $this->getApplication();
$sh = new Shell(new Configuration());
$sh->setOutput($output);
$sh->setScopeVariables($shell->getScopeVariables());
$end = microtime(true);
$output->writeln(sprintf('<info>Command took %.6f seconds to complete.</info>', $end-$start));
}
}
<|file_sep|>updated/src/Psy/Command/TimeitCommand.php
/**
* {@inheritdoc}
*/
protected function execute(InputInterface $input, OutputInterface $output)
{
$code = $input->getArgument('code');
/** @var Shell $shell */
$shell = $this->getApplication();
$sh = new Shell(new Configuration());
$sh->setOutput($output);
$sh->setScopeVariables($shell->getScopeVariables());
$start = microtime(true);
$sh->execute($code);
$end = microtime(true);
$output->writeln(sprintf('<info>Command took %.6f seconds to complete.</info>', $end-$start));
}
} | eee4305f71d53dfaeed6b0042926e74f8530bece | src/Psy/Command/TimeitCommand.php | src/Psy/Command/TimeitCommand.php | PHP |
<|file_sep|>original/.travis.yml
- python: 2.7
env: TOXENV=py27
- python: 3.4
env: TOXENV=py34
- python: 3.5
env: TOXENV=py35
- python: 3.6
env: TOXENV=py36
- python: 3.7
env: TOXENV=py37
- python: 3.8
env: TOXENV=py38
- python: 3.5
env: TOXENV=flake
- python: 3.5
env: TOXENV=cover
install:
- pip install tox
script:
- tox
after_script:
<|file_sep|>current/.travis.yml
- python: 2.7
env: TOXENV=py27
- python: 3.4
env: TOXENV=py34
- python: 3.5
env: TOXENV=py35
- python: 3.6
env: TOXENV=py36
- python: 3.7
env: TOXENV=py37
- python: 3.8
env: TOXENV=py38
- python: 3.5
env: TOXENV=flake
- python: 3.5
env: TOXENV=cover
install:
- pip install tox
script:
- tox
after_script:
<|file_sep|>updated/.travis.yml | - python: 2.7
env: TOXENV=py27
- python: 3.4
env: TOXENV=py34
- python: 3.5
env: TOXENV=py35
- python: 3.6
env: TOXENV=py36
- python: 3.7
env: TOXENV=py37
- python: 3.8-dev
env: TOXENV=py38
- python: 3.5
env: TOXENV=flake
- python: 3.5
env: TOXENV=cover
install:
- pip install tox
script:
- tox
after_script: | <|file_sep|>original/.travis.yml
- python: 2.7
env: TOXENV=py27
- python: 3.4
env: TOXENV=py34
- python: 3.5
env: TOXENV=py35
- python: 3.6
env: TOXENV=py36
- python: 3.7
env: TOXENV=py37
- python: 3.8
env: TOXENV=py38
- python: 3.5
env: TOXENV=flake
- python: 3.5
env: TOXENV=cover
install:
- pip install tox
script:
- tox
after_script:
<|file_sep|>current/.travis.yml
- python: 2.7
env: TOXENV=py27
- python: 3.4
env: TOXENV=py34
- python: 3.5
env: TOXENV=py35
- python: 3.6
env: TOXENV=py36
- python: 3.7
env: TOXENV=py37
- python: 3.8
env: TOXENV=py38
- python: 3.5
env: TOXENV=flake
- python: 3.5
env: TOXENV=cover
install:
- pip install tox
script:
- tox
after_script:
<|file_sep|>updated/.travis.yml
- python: 2.7
env: TOXENV=py27
- python: 3.4
env: TOXENV=py34
- python: 3.5
env: TOXENV=py35
- python: 3.6
env: TOXENV=py36
- python: 3.7
env: TOXENV=py37
- python: 3.8-dev
env: TOXENV=py38
- python: 3.5
env: TOXENV=flake
- python: 3.5
env: TOXENV=cover
install:
- pip install tox
script:
- tox
after_script: | 269134b873751b0fe68e5442f04e0ef36656568c | .travis.yml | .travis.yml | YAML |
<|file_sep|>original/recipes/trigger_recipe_roller.txt
No-op file. Edit this to kick recipes.
This is a beginning of a story in this silly file.
Once upon a time, a budding web browser dev team needed a CI system.
All they had was one poor machine under a desk, and its name was Batty,
the Build and Test Yeti.
As the CI needs of the browser grew, Batty, the Build and Test Yeti, got
a new friend:
The End.
<|file_sep|>current/recipes/trigger_recipe_roller.txt
No-op file. Edit this to kick recipes.
This is a beginning of a story in this silly file.
Once upon a time, a budding web browser dev team needed a CI system.
All they had was one poor machine under a desk, and its name was Batty,
the Build and Test Yeti.
As the CI needs of the browser grew, Batty, the Build and Test Yeti, got
a new friend:
The End.
<|file_sep|>updated/recipes/trigger_recipe_roller.txt | No-op file. Edit this to kick recipes.
This is a beginning of a story in this silly file.
Once upon a time, a budding web browser dev team needed a CI system.
All they had was one poor machine under a desk, and its name was Batty,
the Build and Test Yeti.
As the CI needs of the browser grew, Batty, the Build and Test Yeti, got
a new friend:
The End! | <|file_sep|>original/recipes/trigger_recipe_roller.txt
No-op file. Edit this to kick recipes.
This is a beginning of a story in this silly file.
Once upon a time, a budding web browser dev team needed a CI system.
All they had was one poor machine under a desk, and its name was Batty,
the Build and Test Yeti.
As the CI needs of the browser grew, Batty, the Build and Test Yeti, got
a new friend:
The End.
<|file_sep|>current/recipes/trigger_recipe_roller.txt
No-op file. Edit this to kick recipes.
This is a beginning of a story in this silly file.
Once upon a time, a budding web browser dev team needed a CI system.
All they had was one poor machine under a desk, and its name was Batty,
the Build and Test Yeti.
As the CI needs of the browser grew, Batty, the Build and Test Yeti, got
a new friend:
The End.
<|file_sep|>updated/recipes/trigger_recipe_roller.txt
No-op file. Edit this to kick recipes.
This is a beginning of a story in this silly file.
Once upon a time, a budding web browser dev team needed a CI system.
All they had was one poor machine under a desk, and its name was Batty,
the Build and Test Yeti.
As the CI needs of the browser grew, Batty, the Build and Test Yeti, got
a new friend:
The End! | b63c4663d478bcf8d4fd163a52dc42eb85861ea7 | recipes/trigger_recipe_roller.txt | recipes/trigger_recipe_roller.txt | Text |
<|file_sep|>teknologr/members/static/js/applicant.js.diff
original:
$('#unknown_degree input').val('');
updated:
$('#id_degree_programme').val('');
<|file_sep|>original/teknologr/members/static/js/applicant.js
$(document).ready(function() {
$('#id_degree_programme_options').change(function() {
console.log('hello');
if (this.value === 'extra') {
$('#unknown_degree').show();
$('#unknown_degree input').val('');
} else {
$('#unknown_degree').hide();
$('#unknown_degree input').val(this.value);
}
});
});
<|file_sep|>current/teknologr/members/static/js/applicant.js
$(document).ready(function() {
$('#id_degree_programme_options').change(function() {
console.log('hello');
if (this.value === 'extra') {
$('#unknown_degree').show();
$('#id_degree_programme').val('');
} else {
$('#unknown_degree').hide();
$('#unknown_degree input').val(this.value);
}
});
});
<|file_sep|>updated/teknologr/members/static/js/applicant.js | $(document).ready(function() {
$('#id_degree_programme_options').change(function() {
console.log('hello');
if (this.value === 'extra') {
$('#unknown_degree').show();
$('#id_degree_programme').val('');
} else {
$('#unknown_degree').hide();
$('#id_degree_programme').val(this.value);
}
});
}); | <|file_sep|>teknologr/members/static/js/applicant.js.diff
original:
$('#unknown_degree input').val('');
updated:
$('#id_degree_programme').val('');
<|file_sep|>original/teknologr/members/static/js/applicant.js
$(document).ready(function() {
$('#id_degree_programme_options').change(function() {
console.log('hello');
if (this.value === 'extra') {
$('#unknown_degree').show();
$('#unknown_degree input').val('');
} else {
$('#unknown_degree').hide();
$('#unknown_degree input').val(this.value);
}
});
});
<|file_sep|>current/teknologr/members/static/js/applicant.js
$(document).ready(function() {
$('#id_degree_programme_options').change(function() {
console.log('hello');
if (this.value === 'extra') {
$('#unknown_degree').show();
$('#id_degree_programme').val('');
} else {
$('#unknown_degree').hide();
$('#unknown_degree input').val(this.value);
}
});
});
<|file_sep|>updated/teknologr/members/static/js/applicant.js
$(document).ready(function() {
$('#id_degree_programme_options').change(function() {
console.log('hello');
if (this.value === 'extra') {
$('#unknown_degree').show();
$('#id_degree_programme').val('');
} else {
$('#unknown_degree').hide();
$('#id_degree_programme').val(this.value);
}
});
}); | d6ce1424c5d99ba80ad0f13532d20140ae195e60 | teknologr/members/static/js/applicant.js | teknologr/members/static/js/applicant.js | JavaScript |
<|file_sep|>original/LICENSE.txt
<|file_sep|>current/LICENSE.txt
<|file_sep|>updated/LICENSE.txt | Copyright (c) 2013, Sergey Vinokurov
All rights reserved.
Redistribution and use in source and binary forms, with or without
modification, are permitted provided that the following conditions are met:
* Redistributions of source code must retain the above copyright
notice, this list of conditions and the following disclaimer.
* Redistributions in binary form must reproduce the above copyright
notice, this list of conditions and the following disclaimer in the
documentation and/or other materials provided with the distribution.
* Neither the name of the <organization> nor the
names of its contributors may be used to endorse or promote products
derived from this software without specific prior written permission.
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND
ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
DISCLAIMED. IN NO EVENT SHALL <COPYRIGHT HOLDER> BE LIABLE FOR ANY
DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
(INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND | <|file_sep|>original/LICENSE.txt
<|file_sep|>current/LICENSE.txt
<|file_sep|>updated/LICENSE.txt
Copyright (c) 2013, Sergey Vinokurov
All rights reserved.
Redistribution and use in source and binary forms, with or without
modification, are permitted provided that the following conditions are met:
* Redistributions of source code must retain the above copyright
notice, this list of conditions and the following disclaimer.
* Redistributions in binary form must reproduce the above copyright
notice, this list of conditions and the following disclaimer in the
documentation and/or other materials provided with the distribution.
* Neither the name of the <organization> nor the
names of its contributors may be used to endorse or promote products
derived from this software without specific prior written permission.
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND
ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
DISCLAIMED. IN NO EVENT SHALL <COPYRIGHT HOLDER> BE LIABLE FOR ANY
DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
(INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND | 8654eba3046dcc0eab7ce43497e58df05f719480 | LICENSE.txt | LICENSE.txt | Text |
<|file_sep|>src/main/webapp/src/app/app.module.ts.diff
original:
import {TestPageComponent} from './test-page/test-page.component';
updated:
<|file_sep|>original/src/main/webapp/src/app/app.module.ts
import {BrowserModule} from '@angular/platform-browser';
import {NgModule} from '@angular/core';
import {AppRoutingModule} from './app-routing.module';
import {AppComponent} from './app.component';
import {MaterialComponents} from './material-components';
import {TestPageComponent} from './test-page/test-page.component';
import {BrowserAnimationsModule} from '@angular/platform-browser/animations';
@NgModule({
declarations: [AppComponent, TestPageComponent],
imports: [
MaterialComponents,
BrowserModule,
AppRoutingModule,
BrowserAnimationsModule,
],
providers: [],
bootstrap: [AppComponent],
})
<|file_sep|>current/src/main/webapp/src/app/app.module.ts
import {BrowserModule} from '@angular/platform-browser';
import {NgModule} from '@angular/core';
import {AppRoutingModule} from './app-routing.module';
import {AppComponent} from './app.component';
import {MaterialComponents} from './material-components';
import {BrowserAnimationsModule} from '@angular/platform-browser/animations';
@NgModule({
declarations: [AppComponent, TestPageComponent],
imports: [
MaterialComponents,
BrowserModule,
AppRoutingModule,
BrowserAnimationsModule,
],
providers: [],
bootstrap: [AppComponent],
})
export class AppModule {}
<|file_sep|>updated/src/main/webapp/src/app/app.module.ts |
import {BrowserModule} from '@angular/platform-browser';
import {NgModule} from '@angular/core';
import {AppRoutingModule} from './app-routing.module';
import {AppComponent} from './app.component';
import {MaterialComponents} from './material-components';
import {BrowserAnimationsModule} from '@angular/platform-browser/animations';
@NgModule({
declarations: [AppComponent],
imports: [
MaterialComponents,
BrowserModule,
AppRoutingModule,
BrowserAnimationsModule,
],
providers: [],
bootstrap: [AppComponent],
})
export class AppModule {} | <|file_sep|>src/main/webapp/src/app/app.module.ts.diff
original:
import {TestPageComponent} from './test-page/test-page.component';
updated:
<|file_sep|>original/src/main/webapp/src/app/app.module.ts
import {BrowserModule} from '@angular/platform-browser';
import {NgModule} from '@angular/core';
import {AppRoutingModule} from './app-routing.module';
import {AppComponent} from './app.component';
import {MaterialComponents} from './material-components';
import {TestPageComponent} from './test-page/test-page.component';
import {BrowserAnimationsModule} from '@angular/platform-browser/animations';
@NgModule({
declarations: [AppComponent, TestPageComponent],
imports: [
MaterialComponents,
BrowserModule,
AppRoutingModule,
BrowserAnimationsModule,
],
providers: [],
bootstrap: [AppComponent],
})
<|file_sep|>current/src/main/webapp/src/app/app.module.ts
import {BrowserModule} from '@angular/platform-browser';
import {NgModule} from '@angular/core';
import {AppRoutingModule} from './app-routing.module';
import {AppComponent} from './app.component';
import {MaterialComponents} from './material-components';
import {BrowserAnimationsModule} from '@angular/platform-browser/animations';
@NgModule({
declarations: [AppComponent, TestPageComponent],
imports: [
MaterialComponents,
BrowserModule,
AppRoutingModule,
BrowserAnimationsModule,
],
providers: [],
bootstrap: [AppComponent],
})
export class AppModule {}
<|file_sep|>updated/src/main/webapp/src/app/app.module.ts
import {BrowserModule} from '@angular/platform-browser';
import {NgModule} from '@angular/core';
import {AppRoutingModule} from './app-routing.module';
import {AppComponent} from './app.component';
import {MaterialComponents} from './material-components';
import {BrowserAnimationsModule} from '@angular/platform-browser/animations';
@NgModule({
declarations: [AppComponent],
imports: [
MaterialComponents,
BrowserModule,
AppRoutingModule,
BrowserAnimationsModule,
],
providers: [],
bootstrap: [AppComponent],
})
export class AppModule {} | 02e010d7e85345d26632897972bb735134bf225d | src/main/webapp/src/app/app.module.ts | src/main/webapp/src/app/app.module.ts | TypeScript |
<|file_sep|>original/dist_chrome/content-script.js
}
});
port.start();
}
// let ember-debug know that content script has executed
document.documentElement.dataset.emberExtension = 1;
// Allow older versions of Ember (< 1.4) to detect the extension.
if (document.body) {
document.body.dataset.emberExtension = 1;
}
// clear a possible previous Ember icon
chrome.extension.sendMessage({ type: 'resetEmberIcon' });
// inject JS into the page to check for an app on domready
var script = document.createElement('script');
script.type = "text/javascript";
script.src = chrome.extension.getURL("in-page-script.js");
<|file_sep|>current/dist_chrome/content-script.js
}
});
port.start();
}
// let ember-debug know that content script has executed
document.documentElement.dataset.emberExtension = 1;
// Allow older versions of Ember (< 1.4) to detect the extension.
if (document.body) {
document.body.dataset.emberExtension = 1;
}
// clear a possible previous Ember icon
chrome.extension.sendMessage({ type: 'resetEmberIcon' });
// inject JS into the page to check for an app on domready
var script = document.createElement('script');
script.type = "text/javascript";
script.src = chrome.extension.getURL("in-page-script.js");
<|file_sep|>updated/dist_chrome/content-script.js | port.postMessage(message);
}
});
port.start();
}
// let ember-debug know that content script has executed
document.documentElement.dataset.emberExtension = 1;
// clear a possible previous Ember icon
chrome.extension.sendMessage({ type: 'resetEmberIcon' });
// inject JS into the page to check for an app on domready
var script = document.createElement('script');
script.type = "text/javascript";
script.src = chrome.extension.getURL("in-page-script.js");
if (document.body) document.body.appendChild(script);
}()); | <|file_sep|>original/dist_chrome/content-script.js
}
});
port.start();
}
// let ember-debug know that content script has executed
document.documentElement.dataset.emberExtension = 1;
// Allow older versions of Ember (< 1.4) to detect the extension.
if (document.body) {
document.body.dataset.emberExtension = 1;
}
// clear a possible previous Ember icon
chrome.extension.sendMessage({ type: 'resetEmberIcon' });
// inject JS into the page to check for an app on domready
var script = document.createElement('script');
script.type = "text/javascript";
script.src = chrome.extension.getURL("in-page-script.js");
<|file_sep|>current/dist_chrome/content-script.js
}
});
port.start();
}
// let ember-debug know that content script has executed
document.documentElement.dataset.emberExtension = 1;
// Allow older versions of Ember (< 1.4) to detect the extension.
if (document.body) {
document.body.dataset.emberExtension = 1;
}
// clear a possible previous Ember icon
chrome.extension.sendMessage({ type: 'resetEmberIcon' });
// inject JS into the page to check for an app on domready
var script = document.createElement('script');
script.type = "text/javascript";
script.src = chrome.extension.getURL("in-page-script.js");
<|file_sep|>updated/dist_chrome/content-script.js
port.postMessage(message);
}
});
port.start();
}
// let ember-debug know that content script has executed
document.documentElement.dataset.emberExtension = 1;
// clear a possible previous Ember icon
chrome.extension.sendMessage({ type: 'resetEmberIcon' });
// inject JS into the page to check for an app on domready
var script = document.createElement('script');
script.type = "text/javascript";
script.src = chrome.extension.getURL("in-page-script.js");
if (document.body) document.body.appendChild(script);
}()); | ae530367b59848c2259bf50e85e087634dab9bb1 | dist_chrome/content-script.js | dist_chrome/content-script.js | JavaScript |
<|file_sep|>ElCapitan/brew-casks.sh.diff
original:
updated:
brew cask install cyberduck
<|file_sep|>original/ElCapitan/brew-casks.sh
echo "Installing brew casks"
brew cask install alfred
brew cask install android-studio
brew cask install appcleaner
brew cask install atom
brew cask install carbon-copy-cloner
brew cask install ccmenu
brew cask install dash
brew cask install dropbox
brew cask install firefox
brew cask install flux
brew cask install handbrake
brew cask install Caskroom/versions/intellij-idea-ce
brew cask install kindle
brew cask install macdown
brew cask install opera
brew cask install skitch
brew cask install steam
brew cask install textwrangler
brew cask install visual-studio-code
brew cask install vlc
<|file_sep|>current/ElCapitan/brew-casks.sh
brew cask install alfred
brew cask install android-studio
brew cask install appcleaner
brew cask install atom
brew cask install carbon-copy-cloner
brew cask install ccmenu
brew cask install cyberduck
brew cask install dash
brew cask install dropbox
brew cask install firefox
brew cask install flux
brew cask install handbrake
brew cask install Caskroom/versions/intellij-idea-ce
brew cask install kindle
brew cask install macdown
brew cask install opera
brew cask install skitch
brew cask install steam
brew cask install textwrangler
brew cask install visual-studio-code
brew cask install vlc
<|file_sep|>updated/ElCapitan/brew-casks.sh | brew cask install android-studio
brew cask install appcleaner
brew cask install atom
brew cask install carbon-copy-cloner
brew cask install ccmenu
brew cask install cyberduck
brew cask install dash
brew cask install dropbox
brew cask install firefox
brew cask install flux
brew cask install handbrake
brew cask install Caskroom/versions/intellij-idea-ce
brew cask install kindle
brew cask install macdown
brew cask install opera
brew cask install skitch
brew cask install steam
brew cask install textwrangler
brew cask install visual-studio-code
brew cask install vlc
brew cask install xscope | <|file_sep|>ElCapitan/brew-casks.sh.diff
original:
updated:
brew cask install cyberduck
<|file_sep|>original/ElCapitan/brew-casks.sh
echo "Installing brew casks"
brew cask install alfred
brew cask install android-studio
brew cask install appcleaner
brew cask install atom
brew cask install carbon-copy-cloner
brew cask install ccmenu
brew cask install dash
brew cask install dropbox
brew cask install firefox
brew cask install flux
brew cask install handbrake
brew cask install Caskroom/versions/intellij-idea-ce
brew cask install kindle
brew cask install macdown
brew cask install opera
brew cask install skitch
brew cask install steam
brew cask install textwrangler
brew cask install visual-studio-code
brew cask install vlc
<|file_sep|>current/ElCapitan/brew-casks.sh
brew cask install alfred
brew cask install android-studio
brew cask install appcleaner
brew cask install atom
brew cask install carbon-copy-cloner
brew cask install ccmenu
brew cask install cyberduck
brew cask install dash
brew cask install dropbox
brew cask install firefox
brew cask install flux
brew cask install handbrake
brew cask install Caskroom/versions/intellij-idea-ce
brew cask install kindle
brew cask install macdown
brew cask install opera
brew cask install skitch
brew cask install steam
brew cask install textwrangler
brew cask install visual-studio-code
brew cask install vlc
<|file_sep|>updated/ElCapitan/brew-casks.sh
brew cask install android-studio
brew cask install appcleaner
brew cask install atom
brew cask install carbon-copy-cloner
brew cask install ccmenu
brew cask install cyberduck
brew cask install dash
brew cask install dropbox
brew cask install firefox
brew cask install flux
brew cask install handbrake
brew cask install Caskroom/versions/intellij-idea-ce
brew cask install kindle
brew cask install macdown
brew cask install opera
brew cask install skitch
brew cask install steam
brew cask install textwrangler
brew cask install visual-studio-code
brew cask install vlc
brew cask install xscope | c561feb49f4d60bd4dc189f2617e6101cd717824 | ElCapitan/brew-casks.sh | ElCapitan/brew-casks.sh | Shell |
<|file_sep|>original/thrifty-bunny.gemspec
spec.email = ["bsiggelkow@me.com"]
spec.summary = "RabbitMQ adapter for Apache Thrift"
spec.description = "RabbitMQ adapter for Apache Thrift"
spec.homepage = "http://github.com/bsiggelkow/thrifty-bunny"
spec.license = 'MIT'
spec.files = `git ls-files -z`.split("\x0")
spec.executables = spec.files.grep(%r{^bin/}) { |f| File.basename(f) }
spec.test_files = spec.files.grep(%r{^(test|spec|features)/})
spec.require_paths = ["lib"]
spec.add_development_dependency "bundler", "~> 1.7"
spec.add_development_dependency "rake", "~> 10.0"
spec.add_development_dependency 'rspec'
spec.add_development_dependency 'pry'
spec.add_dependency 'thrift'
spec.add_dependency 'thin'
spec.add_dependency 'bunny'
spec.add_dependency 'uuidtools'
end
<|file_sep|>current/thrifty-bunny.gemspec
spec.email = ["bsiggelkow@me.com"]
spec.summary = "RabbitMQ adapter for Apache Thrift"
spec.description = "RabbitMQ adapter for Apache Thrift"
spec.homepage = "http://github.com/bsiggelkow/thrifty-bunny"
spec.license = 'MIT'
spec.files = `git ls-files -z`.split("\x0")
spec.executables = spec.files.grep(%r{^bin/}) { |f| File.basename(f) }
spec.test_files = spec.files.grep(%r{^(test|spec|features)/})
spec.require_paths = ["lib"]
spec.add_development_dependency "bundler", "~> 1.7"
spec.add_development_dependency "rake", "~> 10.0"
spec.add_development_dependency 'rspec'
spec.add_development_dependency 'pry'
spec.add_dependency 'thrift'
spec.add_dependency 'thin'
spec.add_dependency 'bunny'
spec.add_dependency 'uuidtools'
end
<|file_sep|>updated/thrifty-bunny.gemspec | spec.email = ["bsiggelkow@me.com"]
spec.summary = "RabbitMQ adapter for Apache Thrift"
spec.description = "RabbitMQ adapter for Apache Thrift"
spec.homepage = "http://github.com/bsiggelkow/thrifty-bunny"
spec.license = 'MIT'
spec.files = `git ls-files -z`.split("\x0")
spec.executables = spec.files.grep(%r{^bin/}) { |f| File.basename(f) }
spec.test_files = spec.files.grep(%r{^(test|spec|features)/})
spec.require_paths = ["lib"]
spec.add_development_dependency "bundler", "~> 1.7"
spec.add_development_dependency "rake", "~> 10.0"
spec.add_development_dependency 'rspec'
spec.add_development_dependency 'pry'
spec.add_dependency 'thrift'
spec.add_dependency 'thin'
spec.add_dependency 'bunny', '~> 1.6.3'
spec.add_dependency 'uuidtools'
end | <|file_sep|>original/thrifty-bunny.gemspec
spec.email = ["bsiggelkow@me.com"]
spec.summary = "RabbitMQ adapter for Apache Thrift"
spec.description = "RabbitMQ adapter for Apache Thrift"
spec.homepage = "http://github.com/bsiggelkow/thrifty-bunny"
spec.license = 'MIT'
spec.files = `git ls-files -z`.split("\x0")
spec.executables = spec.files.grep(%r{^bin/}) { |f| File.basename(f) }
spec.test_files = spec.files.grep(%r{^(test|spec|features)/})
spec.require_paths = ["lib"]
spec.add_development_dependency "bundler", "~> 1.7"
spec.add_development_dependency "rake", "~> 10.0"
spec.add_development_dependency 'rspec'
spec.add_development_dependency 'pry'
spec.add_dependency 'thrift'
spec.add_dependency 'thin'
spec.add_dependency 'bunny'
spec.add_dependency 'uuidtools'
end
<|file_sep|>current/thrifty-bunny.gemspec
spec.email = ["bsiggelkow@me.com"]
spec.summary = "RabbitMQ adapter for Apache Thrift"
spec.description = "RabbitMQ adapter for Apache Thrift"
spec.homepage = "http://github.com/bsiggelkow/thrifty-bunny"
spec.license = 'MIT'
spec.files = `git ls-files -z`.split("\x0")
spec.executables = spec.files.grep(%r{^bin/}) { |f| File.basename(f) }
spec.test_files = spec.files.grep(%r{^(test|spec|features)/})
spec.require_paths = ["lib"]
spec.add_development_dependency "bundler", "~> 1.7"
spec.add_development_dependency "rake", "~> 10.0"
spec.add_development_dependency 'rspec'
spec.add_development_dependency 'pry'
spec.add_dependency 'thrift'
spec.add_dependency 'thin'
spec.add_dependency 'bunny'
spec.add_dependency 'uuidtools'
end
<|file_sep|>updated/thrifty-bunny.gemspec
spec.email = ["bsiggelkow@me.com"]
spec.summary = "RabbitMQ adapter for Apache Thrift"
spec.description = "RabbitMQ adapter for Apache Thrift"
spec.homepage = "http://github.com/bsiggelkow/thrifty-bunny"
spec.license = 'MIT'
spec.files = `git ls-files -z`.split("\x0")
spec.executables = spec.files.grep(%r{^bin/}) { |f| File.basename(f) }
spec.test_files = spec.files.grep(%r{^(test|spec|features)/})
spec.require_paths = ["lib"]
spec.add_development_dependency "bundler", "~> 1.7"
spec.add_development_dependency "rake", "~> 10.0"
spec.add_development_dependency 'rspec'
spec.add_development_dependency 'pry'
spec.add_dependency 'thrift'
spec.add_dependency 'thin'
spec.add_dependency 'bunny', '~> 1.6.3'
spec.add_dependency 'uuidtools'
end | 334d3daf95eb8940163352ecb06ee5b3c372020b | thrifty-bunny.gemspec | thrifty-bunny.gemspec | Ruby |
<|file_sep|>original/.idea/artifacts/ProgrammingLife_jar.xml
<|file_sep|>current/.idea/artifacts/ProgrammingLife_jar.xml
<|file_sep|>updated/.idea/artifacts/ProgrammingLife_jar.xml | <component name="ArtifactManager">
<artifact type="jar" name="ProgrammingLife:jar">
<output-path>$PROJECT_DIR$/out/artifacts/ProgrammingLife_jar</output-path>
<root id="archive" name="ProgrammingLife.jar">
<element id="module-output" name="ProgrammingLife" />
<element id="extracted-dir" path="$MAVEN_REPOSITORY$/com/diffplug/durian/durian/3.4.0/durian-3.4.0.jar" path-in-jar="/" />
<element id="extracted-dir" path="$MAVEN_REPOSITORY$/org/apache/commons/commons-lang3/3.5/commons-lang3-3.5.jar" path-in-jar="/" />
</root>
</artifact>
</component> | <|file_sep|>original/.idea/artifacts/ProgrammingLife_jar.xml
<|file_sep|>current/.idea/artifacts/ProgrammingLife_jar.xml
<|file_sep|>updated/.idea/artifacts/ProgrammingLife_jar.xml
<component name="ArtifactManager">
<artifact type="jar" name="ProgrammingLife:jar">
<output-path>$PROJECT_DIR$/out/artifacts/ProgrammingLife_jar</output-path>
<root id="archive" name="ProgrammingLife.jar">
<element id="module-output" name="ProgrammingLife" />
<element id="extracted-dir" path="$MAVEN_REPOSITORY$/com/diffplug/durian/durian/3.4.0/durian-3.4.0.jar" path-in-jar="/" />
<element id="extracted-dir" path="$MAVEN_REPOSITORY$/org/apache/commons/commons-lang3/3.5/commons-lang3-3.5.jar" path-in-jar="/" />
</root>
</artifact>
</component> | 8d12bc080ef9f976533434f78a674578bdeef189 | .idea/artifacts/ProgrammingLife_jar.xml | .idea/artifacts/ProgrammingLife_jar.xml | XML |
<|file_sep|>app/Console/Kernel.php.diff
original:
->weeklyOn(5, '4:00');
updated:
->weeklyOn(6, '4:00'); // 6 = Saturday
<|file_sep|>original/app/Console/Kernel.php
/**
* Define the application's command schedule.
*
* @param \Illuminate\Console\Scheduling\Schedule $schedule
* @return void
*/
protected function schedule(Schedule $schedule) {
// Synchronise student list with the LDAP
$schedule->command('syncstudents')
->weeklyOn(5, '4:00');
// Send notification mails
$schedule->command('sendnotificationmails --subscribed=instantly')
->everyThirtyMinutes();
$schedule->command('sendnotificationmails --subscribed=daily')
->dailyAt('4:00');
$schedule->command('sendnotificationmails --subscribed=weekly')
->weeklyOn(6, '4:00');
}
}
<|file_sep|>current/app/Console/Kernel.php
/**
* Define the application's command schedule.
*
* @param \Illuminate\Console\Scheduling\Schedule $schedule
* @return void
*/
protected function schedule(Schedule $schedule) {
// Synchronise student list with the LDAP
$schedule->command('syncstudents')
->weeklyOn(6, '4:00'); // 6 = Saturday
// Send notification mails
$schedule->command('sendnotificationmails --subscribed=instantly')
->everyThirtyMinutes();
$schedule->command('sendnotificationmails --subscribed=daily')
->dailyAt('4:00');
$schedule->command('sendnotificationmails --subscribed=weekly')
->weeklyOn(6, '4:00');
}
}
<|file_sep|>updated/app/Console/Kernel.php |
/**
* Define the application's command schedule.
*
* @param \Illuminate\Console\Scheduling\Schedule $schedule
* @return void
*/
protected function schedule(Schedule $schedule) {
// Synchronise student list with the LDAP
$schedule->command('syncstudents')
->weeklyOn(6, '4:00'); // 6 = Saturday
// Send notification mails
$schedule->command('sendnotificationmails --subscribed=instantly')
->everyThirtyMinutes();
$schedule->command('sendnotificationmails --subscribed=daily')
->dailyAt('4:00');
$schedule->command('sendnotificationmails --subscribed=weekly')
->weeklyOn(0, '4:00'); // 0 = Sunday
}
} | <|file_sep|>app/Console/Kernel.php.diff
original:
->weeklyOn(5, '4:00');
updated:
->weeklyOn(6, '4:00'); // 6 = Saturday
<|file_sep|>original/app/Console/Kernel.php
/**
* Define the application's command schedule.
*
* @param \Illuminate\Console\Scheduling\Schedule $schedule
* @return void
*/
protected function schedule(Schedule $schedule) {
// Synchronise student list with the LDAP
$schedule->command('syncstudents')
->weeklyOn(5, '4:00');
// Send notification mails
$schedule->command('sendnotificationmails --subscribed=instantly')
->everyThirtyMinutes();
$schedule->command('sendnotificationmails --subscribed=daily')
->dailyAt('4:00');
$schedule->command('sendnotificationmails --subscribed=weekly')
->weeklyOn(6, '4:00');
}
}
<|file_sep|>current/app/Console/Kernel.php
/**
* Define the application's command schedule.
*
* @param \Illuminate\Console\Scheduling\Schedule $schedule
* @return void
*/
protected function schedule(Schedule $schedule) {
// Synchronise student list with the LDAP
$schedule->command('syncstudents')
->weeklyOn(6, '4:00'); // 6 = Saturday
// Send notification mails
$schedule->command('sendnotificationmails --subscribed=instantly')
->everyThirtyMinutes();
$schedule->command('sendnotificationmails --subscribed=daily')
->dailyAt('4:00');
$schedule->command('sendnotificationmails --subscribed=weekly')
->weeklyOn(6, '4:00');
}
}
<|file_sep|>updated/app/Console/Kernel.php
/**
* Define the application's command schedule.
*
* @param \Illuminate\Console\Scheduling\Schedule $schedule
* @return void
*/
protected function schedule(Schedule $schedule) {
// Synchronise student list with the LDAP
$schedule->command('syncstudents')
->weeklyOn(6, '4:00'); // 6 = Saturday
// Send notification mails
$schedule->command('sendnotificationmails --subscribed=instantly')
->everyThirtyMinutes();
$schedule->command('sendnotificationmails --subscribed=daily')
->dailyAt('4:00');
$schedule->command('sendnotificationmails --subscribed=weekly')
->weeklyOn(0, '4:00'); // 0 = Sunday
}
} | 21d3ccd3493a340fa0168f17df9ce2baab7f1347 | app/Console/Kernel.php | app/Console/Kernel.php | PHP |
<|file_sep|>original/test/lib/rubycritic/metric_adapters/flog_adapter_test.rb
describe Rubycritic::SmellAdapter::Flog do
before do
sample_path = "test/samples/flog/smelly.rb"
flog = Rubycritic::Analyser::Flog.new([sample_path])
@adapter = Rubycritic::SmellAdapter::Flog.new(flog)
end
it "detects smells" do
@adapter.smells.wont_be_empty
end
it "has smells with messages" do
smell = @adapter.smells.first
smell.message.must_be_kind_of String
end
it "has smells with scores" do
smell = @adapter.smells.first
smell.score.must_be_kind_of Numeric
end
end
<|file_sep|>current/test/lib/rubycritic/metric_adapters/flog_adapter_test.rb
describe Rubycritic::SmellAdapter::Flog do
before do
sample_path = "test/samples/flog/smelly.rb"
flog = Rubycritic::Analyser::Flog.new([sample_path])
@adapter = Rubycritic::SmellAdapter::Flog.new(flog)
end
it "detects smells" do
@adapter.smells.wont_be_empty
end
it "has smells with messages" do
smell = @adapter.smells.first
smell.message.must_be_kind_of String
end
it "has smells with scores" do
smell = @adapter.smells.first
smell.score.must_be_kind_of Numeric
end
end
<|file_sep|>updated/test/lib/rubycritic/metric_adapters/flog_adapter_test.rb | describe Rubycritic::SmellAdapter::Flog do
before do
sample_path = "test/samples/flog/smelly.rb"
flog = Rubycritic::Analyser::Flog.new([sample_path])
@adapter = Rubycritic::SmellAdapter::Flog.new(flog)
end
it "detects smells" do
@adapter.smells.wont_be_empty
end
it "has smells with messages" do
smell = @adapter.smells.first
smell.message.must_be_instance_of String
end
it "has smells with scores" do
smell = @adapter.smells.first
smell.score.must_be_kind_of Numeric
end
end | <|file_sep|>original/test/lib/rubycritic/metric_adapters/flog_adapter_test.rb
describe Rubycritic::SmellAdapter::Flog do
before do
sample_path = "test/samples/flog/smelly.rb"
flog = Rubycritic::Analyser::Flog.new([sample_path])
@adapter = Rubycritic::SmellAdapter::Flog.new(flog)
end
it "detects smells" do
@adapter.smells.wont_be_empty
end
it "has smells with messages" do
smell = @adapter.smells.first
smell.message.must_be_kind_of String
end
it "has smells with scores" do
smell = @adapter.smells.first
smell.score.must_be_kind_of Numeric
end
end
<|file_sep|>current/test/lib/rubycritic/metric_adapters/flog_adapter_test.rb
describe Rubycritic::SmellAdapter::Flog do
before do
sample_path = "test/samples/flog/smelly.rb"
flog = Rubycritic::Analyser::Flog.new([sample_path])
@adapter = Rubycritic::SmellAdapter::Flog.new(flog)
end
it "detects smells" do
@adapter.smells.wont_be_empty
end
it "has smells with messages" do
smell = @adapter.smells.first
smell.message.must_be_kind_of String
end
it "has smells with scores" do
smell = @adapter.smells.first
smell.score.must_be_kind_of Numeric
end
end
<|file_sep|>updated/test/lib/rubycritic/metric_adapters/flog_adapter_test.rb
describe Rubycritic::SmellAdapter::Flog do
before do
sample_path = "test/samples/flog/smelly.rb"
flog = Rubycritic::Analyser::Flog.new([sample_path])
@adapter = Rubycritic::SmellAdapter::Flog.new(flog)
end
it "detects smells" do
@adapter.smells.wont_be_empty
end
it "has smells with messages" do
smell = @adapter.smells.first
smell.message.must_be_instance_of String
end
it "has smells with scores" do
smell = @adapter.smells.first
smell.score.must_be_kind_of Numeric
end
end | e1da8c877977b5a6d5a42aa6d7d88694b2e92e34 | test/lib/rubycritic/metric_adapters/flog_adapter_test.rb | test/lib/rubycritic/metric_adapters/flog_adapter_test.rb | Ruby |
<|file_sep|>original/hieradata/common.yaml
---
message: "This node is using common data"
#Puppet Server Tuning
puppet_enterprise::master::puppetserver::jruby_max_requests_per_instance: 0
<|file_sep|>current/hieradata/common.yaml
---
message: "This node is using common data"
#Puppet Server Tuning
puppet_enterprise::master::puppetserver::jruby_max_requests_per_instance: 0
<|file_sep|>updated/hieradata/common.yaml | ---
message: "This node is using common data" | <|file_sep|>original/hieradata/common.yaml
---
message: "This node is using common data"
#Puppet Server Tuning
puppet_enterprise::master::puppetserver::jruby_max_requests_per_instance: 0
<|file_sep|>current/hieradata/common.yaml
---
message: "This node is using common data"
#Puppet Server Tuning
puppet_enterprise::master::puppetserver::jruby_max_requests_per_instance: 0
<|file_sep|>updated/hieradata/common.yaml
---
message: "This node is using common data" | 5497d199b3c851a317638905e77f44f2fafc64d4 | hieradata/common.yaml | hieradata/common.yaml | YAML |
<|file_sep|>original/packages/opencv/opencv-samples_git.bb
<|file_sep|>current/packages/opencv/opencv-samples_git.bb
<|file_sep|>updated/packages/opencv/opencv-samples_git.bb | DESCRIPTION = "Opencv : The Open Computer Vision Library"
HOMEPAGE = "http://sourceforge.net/projects/opencvlibrary"
SECTION = "libs"
PRIORITY = "optional"
LICENSE = "GPLv2"
PR = "r3"
DEPENDS = "opencv"
SRCREV = "fb7cc87ed6a7425f3edc58e2d1eb44599496ac09"
SRC_URI = "git://github.com/nzjrs/opencv.git;protocol=git;branch=libv4l"
PV = "1.1.0+git${SRCREV}"
S = "${WORKDIR}/git"
do_install() {
cd samples/c
install -d ${D}/${bindir}
for i in *.c; do | <|file_sep|>original/packages/opencv/opencv-samples_git.bb
<|file_sep|>current/packages/opencv/opencv-samples_git.bb
<|file_sep|>updated/packages/opencv/opencv-samples_git.bb
DESCRIPTION = "Opencv : The Open Computer Vision Library"
HOMEPAGE = "http://sourceforge.net/projects/opencvlibrary"
SECTION = "libs"
PRIORITY = "optional"
LICENSE = "GPLv2"
PR = "r3"
DEPENDS = "opencv"
SRCREV = "fb7cc87ed6a7425f3edc58e2d1eb44599496ac09"
SRC_URI = "git://github.com/nzjrs/opencv.git;protocol=git;branch=libv4l"
PV = "1.1.0+git${SRCREV}"
S = "${WORKDIR}/git"
do_install() {
cd samples/c
install -d ${D}/${bindir}
for i in *.c; do | 3c871af940959b5543908e909d7747ba0b4affaa | packages/opencv/opencv-samples_git.bb | packages/opencv/opencv-samples_git.bb | BitBake |
<|file_sep|>original/LICENSE.txt
Copyright (c) 2015 Bouke van der Bijl
MIT License
Permission is hereby granted, free of charge, to any person obtaining
a copy of this software and associated documentation files (the
"Software"), to deal in the Software without restriction, including
without limitation the rights to use, copy, modify, merge, publish,
distribute, sublicense, and/or sell copies of the Software, and to
permit persons to whom the Software is furnished to do so, subject to
the following conditions:
The above copyright notice and this permission notice shall be
included in all copies or substantial portions of the Software.
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE
LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION
OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION
<|file_sep|>current/LICENSE.txt
Copyright (c) 2015 Bouke van der Bijl
MIT License
Permission is hereby granted, free of charge, to any person obtaining
a copy of this software and associated documentation files (the
"Software"), to deal in the Software without restriction, including
without limitation the rights to use, copy, modify, merge, publish,
distribute, sublicense, and/or sell copies of the Software, and to
permit persons to whom the Software is furnished to do so, subject to
the following conditions:
The above copyright notice and this permission notice shall be
included in all copies or substantial portions of the Software.
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE
LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION
OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION
<|file_sep|>updated/LICENSE.txt | Copyright (c) 2015 Shopify
MIT License
Permission is hereby granted, free of charge, to any person obtaining
a copy of this software and associated documentation files (the
"Software"), to deal in the Software without restriction, including
without limitation the rights to use, copy, modify, merge, publish,
distribute, sublicense, and/or sell copies of the Software, and to
permit persons to whom the Software is furnished to do so, subject to
the following conditions:
The above copyright notice and this permission notice shall be
included in all copies or substantial portions of the Software.
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE
LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION
OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION | <|file_sep|>original/LICENSE.txt
Copyright (c) 2015 Bouke van der Bijl
MIT License
Permission is hereby granted, free of charge, to any person obtaining
a copy of this software and associated documentation files (the
"Software"), to deal in the Software without restriction, including
without limitation the rights to use, copy, modify, merge, publish,
distribute, sublicense, and/or sell copies of the Software, and to
permit persons to whom the Software is furnished to do so, subject to
the following conditions:
The above copyright notice and this permission notice shall be
included in all copies or substantial portions of the Software.
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE
LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION
OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION
<|file_sep|>current/LICENSE.txt
Copyright (c) 2015 Bouke van der Bijl
MIT License
Permission is hereby granted, free of charge, to any person obtaining
a copy of this software and associated documentation files (the
"Software"), to deal in the Software without restriction, including
without limitation the rights to use, copy, modify, merge, publish,
distribute, sublicense, and/or sell copies of the Software, and to
permit persons to whom the Software is furnished to do so, subject to
the following conditions:
The above copyright notice and this permission notice shall be
included in all copies or substantial portions of the Software.
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE
LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION
OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION
<|file_sep|>updated/LICENSE.txt
Copyright (c) 2015 Shopify
MIT License
Permission is hereby granted, free of charge, to any person obtaining
a copy of this software and associated documentation files (the
"Software"), to deal in the Software without restriction, including
without limitation the rights to use, copy, modify, merge, publish,
distribute, sublicense, and/or sell copies of the Software, and to
permit persons to whom the Software is furnished to do so, subject to
the following conditions:
The above copyright notice and this permission notice shall be
included in all copies or substantial portions of the Software.
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE
LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION
OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION | 336086d5b6c8a5289aa9a3cdd3751118d1f087dd | LICENSE.txt | LICENSE.txt | Text |
<|file_sep|>setup/optional_setup_steps.sh.diff
original:
updated:
# Dependencies for Sublime Text (from https://www.sublimetext.com/docs/3/linux_repositories.html#apt)
wget -qO - https://download.sublimetext.com/sublimehq-pub.gpg | sudo apt-key add -
sudo apt-get install -yq apt-transport-https
echo "deb https://download.sublimetext.com/ apt/dev/" | sudo tee /etc/apt/sources.list.d/sublime-text.list
sudo apt-get update
<|file_sep|>setup/optional_setup_steps.sh.diff
original:
updated:
sublime-text
<|file_sep|>original/setup/optional_setup_steps.sh
terminator \
vim \
wget
# Allow SSH into this machine
sudo apt-get install -yq openssh-server
# Set the default settings for gedit to be OK for programming
gsettings set org.gnome.gedit.preferences.editor auto-indent true
gsettings set org.gnome.gedit.preferences.editor bracket-matching true
gsettings set org.gnome.gedit.preferences.editor display-line-numbers true
gsettings set org.gnome.gedit.preferences.editor highlight-current-line true
gsettings set org.gnome.gedit.preferences.editor insert-spaces true
gsettings set org.gnome.gedit.preferences.editor tabs-size "uint32 2"
# Clean up the Unity launcher
gsettings set com.canonical.Unity.Launcher favorites "['application://firefox.desktop']"
# Set the cache to timeout after 1 hour (setting is in seconds)
git config --global credential.helper 'cache --timeout=3600'
<|file_sep|>current/setup/optional_setup_steps.sh
emacs \
gitk \
sublime-text
terminator \
vim \
wget
# Allow SSH into this machine
sudo apt-get install -yq openssh-server
# Set the default settings for gedit to be OK for programming
gsettings set org.gnome.gedit.preferences.editor auto-indent true
gsettings set org.gnome.gedit.preferences.editor bracket-matching true
gsettings set org.gnome.gedit.preferences.editor display-line-numbers true
gsettings set org.gnome.gedit.preferences.editor highlight-current-line true
gsettings set org.gnome.gedit.preferences.editor insert-spaces true
gsettings set org.gnome.gedit.preferences.editor tabs-size "uint32 2"
# Clean up the Unity launcher
gsettings set com.canonical.Unity.Launcher favorites "['application://firefox.desktop']"
<|file_sep|>updated/setup/optional_setup_steps.sh | emacs \
gitk \
sublime-text
terminator \
vim \
wget
# Allow SSH into this machine
sudo apt-get install -yq openssh-server
# Set the default settings for gedit to be OK for programming
gsettings set org.gnome.gedit.preferences.editor auto-indent true
gsettings set org.gnome.gedit.preferences.editor bracket-matching true
gsettings set org.gnome.gedit.preferences.editor display-line-numbers true
gsettings set org.gnome.gedit.preferences.editor highlight-current-line true
gsettings set org.gnome.gedit.preferences.editor insert-spaces true
gsettings set org.gnome.gedit.preferences.editor tabs-size "uint32 2"
# Clean up the Unity launcher
gsettings set com.canonical.Unity.Launcher favorites "['application://firefox.desktop']"
| <|file_sep|>setup/optional_setup_steps.sh.diff
original:
updated:
# Dependencies for Sublime Text (from https://www.sublimetext.com/docs/3/linux_repositories.html#apt)
wget -qO - https://download.sublimetext.com/sublimehq-pub.gpg | sudo apt-key add -
sudo apt-get install -yq apt-transport-https
echo "deb https://download.sublimetext.com/ apt/dev/" | sudo tee /etc/apt/sources.list.d/sublime-text.list
sudo apt-get update
<|file_sep|>setup/optional_setup_steps.sh.diff
original:
updated:
sublime-text
<|file_sep|>original/setup/optional_setup_steps.sh
terminator \
vim \
wget
# Allow SSH into this machine
sudo apt-get install -yq openssh-server
# Set the default settings for gedit to be OK for programming
gsettings set org.gnome.gedit.preferences.editor auto-indent true
gsettings set org.gnome.gedit.preferences.editor bracket-matching true
gsettings set org.gnome.gedit.preferences.editor display-line-numbers true
gsettings set org.gnome.gedit.preferences.editor highlight-current-line true
gsettings set org.gnome.gedit.preferences.editor insert-spaces true
gsettings set org.gnome.gedit.preferences.editor tabs-size "uint32 2"
# Clean up the Unity launcher
gsettings set com.canonical.Unity.Launcher favorites "['application://firefox.desktop']"
# Set the cache to timeout after 1 hour (setting is in seconds)
git config --global credential.helper 'cache --timeout=3600'
<|file_sep|>current/setup/optional_setup_steps.sh
emacs \
gitk \
sublime-text
terminator \
vim \
wget
# Allow SSH into this machine
sudo apt-get install -yq openssh-server
# Set the default settings for gedit to be OK for programming
gsettings set org.gnome.gedit.preferences.editor auto-indent true
gsettings set org.gnome.gedit.preferences.editor bracket-matching true
gsettings set org.gnome.gedit.preferences.editor display-line-numbers true
gsettings set org.gnome.gedit.preferences.editor highlight-current-line true
gsettings set org.gnome.gedit.preferences.editor insert-spaces true
gsettings set org.gnome.gedit.preferences.editor tabs-size "uint32 2"
# Clean up the Unity launcher
gsettings set com.canonical.Unity.Launcher favorites "['application://firefox.desktop']"
<|file_sep|>updated/setup/optional_setup_steps.sh
emacs \
gitk \
sublime-text
terminator \
vim \
wget
# Allow SSH into this machine
sudo apt-get install -yq openssh-server
# Set the default settings for gedit to be OK for programming
gsettings set org.gnome.gedit.preferences.editor auto-indent true
gsettings set org.gnome.gedit.preferences.editor bracket-matching true
gsettings set org.gnome.gedit.preferences.editor display-line-numbers true
gsettings set org.gnome.gedit.preferences.editor highlight-current-line true
gsettings set org.gnome.gedit.preferences.editor insert-spaces true
gsettings set org.gnome.gedit.preferences.editor tabs-size "uint32 2"
# Clean up the Unity launcher
gsettings set com.canonical.Unity.Launcher favorites "['application://firefox.desktop']"
| e63b0192ed841977faec8dd321edfae27e7c8fa2 | setup/optional_setup_steps.sh | setup/optional_setup_steps.sh | Shell |
<|file_sep|>original/src/com/kduda/battleships/Main.java
import javafx.application.Application;
import javafx.fxml.FXMLLoader;
import javafx.scene.Parent;
import javafx.scene.Scene;
import javafx.stage.Stage;
public class Main extends Application {
public static void main(String[] args) {
launch(args);
}
@Override
public void start(Stage primaryStage) throws Exception {
Parent root = FXMLLoader.load(getClass().getResource("views/battleshipsScene.fxml"));
// setUserAgentStylesheet(STYLESHEET_CASPIAN);
primaryStage.setTitle("Battleships");
primaryStage.setScene(new Scene(root,1024,768));
primaryStage.show();
}
}
<|file_sep|>current/src/com/kduda/battleships/Main.java
import javafx.application.Application;
import javafx.fxml.FXMLLoader;
import javafx.scene.Parent;
import javafx.scene.Scene;
import javafx.stage.Stage;
public class Main extends Application {
public static void main(String[] args) {
launch(args);
}
@Override
public void start(Stage primaryStage) throws Exception {
Parent root = FXMLLoader.load(getClass().getResource("views/battleshipsScene.fxml"));
// setUserAgentStylesheet(STYLESHEET_CASPIAN);
primaryStage.setTitle("Battleships");
primaryStage.setScene(new Scene(root,1024,768));
primaryStage.show();
}
}
<|file_sep|>updated/src/com/kduda/battleships/Main.java | import javafx.fxml.FXMLLoader;
import javafx.scene.Parent;
import javafx.scene.Scene;
import javafx.stage.Stage;
public class Main extends Application {
public static void main(String[] args) {
launch(args);
}
@Override
public void start(Stage primaryStage) throws Exception {
Parent root = FXMLLoader.load(getClass().getResource("views/battleshipsScene.fxml"));
// setUserAgentStylesheet(STYLESHEET_CASPIAN);
primaryStage.setTitle("Battleships");
primaryStage.setScene(new Scene(root);
primaryStage.setResizable(false);
primaryStage.show();
}
} | <|file_sep|>original/src/com/kduda/battleships/Main.java
import javafx.application.Application;
import javafx.fxml.FXMLLoader;
import javafx.scene.Parent;
import javafx.scene.Scene;
import javafx.stage.Stage;
public class Main extends Application {
public static void main(String[] args) {
launch(args);
}
@Override
public void start(Stage primaryStage) throws Exception {
Parent root = FXMLLoader.load(getClass().getResource("views/battleshipsScene.fxml"));
// setUserAgentStylesheet(STYLESHEET_CASPIAN);
primaryStage.setTitle("Battleships");
primaryStage.setScene(new Scene(root,1024,768));
primaryStage.show();
}
}
<|file_sep|>current/src/com/kduda/battleships/Main.java
import javafx.application.Application;
import javafx.fxml.FXMLLoader;
import javafx.scene.Parent;
import javafx.scene.Scene;
import javafx.stage.Stage;
public class Main extends Application {
public static void main(String[] args) {
launch(args);
}
@Override
public void start(Stage primaryStage) throws Exception {
Parent root = FXMLLoader.load(getClass().getResource("views/battleshipsScene.fxml"));
// setUserAgentStylesheet(STYLESHEET_CASPIAN);
primaryStage.setTitle("Battleships");
primaryStage.setScene(new Scene(root,1024,768));
primaryStage.show();
}
}
<|file_sep|>updated/src/com/kduda/battleships/Main.java
import javafx.fxml.FXMLLoader;
import javafx.scene.Parent;
import javafx.scene.Scene;
import javafx.stage.Stage;
public class Main extends Application {
public static void main(String[] args) {
launch(args);
}
@Override
public void start(Stage primaryStage) throws Exception {
Parent root = FXMLLoader.load(getClass().getResource("views/battleshipsScene.fxml"));
// setUserAgentStylesheet(STYLESHEET_CASPIAN);
primaryStage.setTitle("Battleships");
primaryStage.setScene(new Scene(root);
primaryStage.setResizable(false);
primaryStage.show();
}
} | 4a351c628fe8e9c16132c02592936232c54093a1 | src/com/kduda/battleships/Main.java | src/com/kduda/battleships/Main.java | Java |
<|file_sep|>original/Formula/php70-uuid.rb
<|file_sep|>current/Formula/php70-uuid.rb
<|file_sep|>updated/Formula/php70-uuid.rb | require File.expand_path("../../Abstract/abstract-php-extension", __FILE__)
class Php70Uuid < AbstractPhp70Extension
init
desc "UUID extension"
homepage "https://pecl.php.net/package/uuid"
stable do
url "https://pecl.php.net/get/uuid-1.0.4.tgz"
sha256 "63079b6a62a9d43691ecbcd4eb52e5e5fe17b5a3d0f8e46e3c17ff265c06a11f"
patch do
# let's fix the path to uuid.h (uuid/uuid.h on linux, ossp/uuid.h on OSX)
# uuid_mac & uuid_time might not be available on OSX, let's add test to avoid compiling issue on these functions
url "https://gist.githubusercontent.com/romainneutron/fe068c297413aee565d5/raw/28d6ba0b6e902e82e71bb9a1ed768c836a8161e4/php-uuid-1.0.4"
sha256 "5f0664d5c4f55d4f6c037dab9f198e697afa3f9266854ed3945d7697fdb692b2"
end
end
head do
url "https://git.php.net/repository/pecl/networking/uuid.git" | <|file_sep|>original/Formula/php70-uuid.rb
<|file_sep|>current/Formula/php70-uuid.rb
<|file_sep|>updated/Formula/php70-uuid.rb
require File.expand_path("../../Abstract/abstract-php-extension", __FILE__)
class Php70Uuid < AbstractPhp70Extension
init
desc "UUID extension"
homepage "https://pecl.php.net/package/uuid"
stable do
url "https://pecl.php.net/get/uuid-1.0.4.tgz"
sha256 "63079b6a62a9d43691ecbcd4eb52e5e5fe17b5a3d0f8e46e3c17ff265c06a11f"
patch do
# let's fix the path to uuid.h (uuid/uuid.h on linux, ossp/uuid.h on OSX)
# uuid_mac & uuid_time might not be available on OSX, let's add test to avoid compiling issue on these functions
url "https://gist.githubusercontent.com/romainneutron/fe068c297413aee565d5/raw/28d6ba0b6e902e82e71bb9a1ed768c836a8161e4/php-uuid-1.0.4"
sha256 "5f0664d5c4f55d4f6c037dab9f198e697afa3f9266854ed3945d7697fdb692b2"
end
end
head do
url "https://git.php.net/repository/pecl/networking/uuid.git" | 25c265f64b7c26bfd68c12c8179fa631b89c3617 | Formula/php70-uuid.rb | Formula/php70-uuid.rb | Ruby |
<|file_sep|>original/rules/management/commands/CorpusLogicUpdate.py
<|file_sep|>current/rules/management/commands/CorpusLogicUpdate.py
<|file_sep|>updated/rules/management/commands/CorpusLogicUpdate.py | import logging
from django.core.management.base import BaseCommand, CommandError
from plyara import YaraParser
from rules.models import YaraRule
# Configure Logging
logging.basicConfig(level=logging.INFO)
class Command(BaseCommand):
help = 'Recalculate the logic hashes of the entire rule corpus'
def handle(self, *args, **options):
corpus = YaraRule.objects.all()
rule_count = corpus.count()
message = 'Updating logic hashes for {} rules'.format(rule_count)
logging.info(message)
rule_index = 0
| <|file_sep|>original/rules/management/commands/CorpusLogicUpdate.py
<|file_sep|>current/rules/management/commands/CorpusLogicUpdate.py
<|file_sep|>updated/rules/management/commands/CorpusLogicUpdate.py
import logging
from django.core.management.base import BaseCommand, CommandError
from plyara import YaraParser
from rules.models import YaraRule
# Configure Logging
logging.basicConfig(level=logging.INFO)
class Command(BaseCommand):
help = 'Recalculate the logic hashes of the entire rule corpus'
def handle(self, *args, **options):
corpus = YaraRule.objects.all()
rule_count = corpus.count()
message = 'Updating logic hashes for {} rules'.format(rule_count)
logging.info(message)
rule_index = 0
| b3f8be5b6ab7e4e713004447a3cfbda743d80394 | rules/management/commands/CorpusLogicUpdate.py | rules/management/commands/CorpusLogicUpdate.py | Python |
<|file_sep|>original/scripts/roles/cikit-php/tasks/xdebug.yml
---
- name: Set name of directory for XDebug
set_fact:
cikit_php_xdebug: "xdebug-{{ php.xdebug }}"
- name: Download XDebug
get_url:
url: "https://xdebug.org/files/{{ cikit_php_xdebug }}.tgz"
dest: "/tmp/{{ cikit_php_xdebug }}.tgz"
- name: Unarchive XDebug
shell: "tar -xf /tmp/{{ cikit_php_xdebug }}.tgz -C /tmp/"
args:
# Unarchive module does not work properly in Ansible 2.1.1.0.
warn: no
- include_tasks: fixtures/compile.yml name="XDebug" dir="/tmp/{{ cikit_php_xdebug }}"
<|file_sep|>current/scripts/roles/cikit-php/tasks/xdebug.yml
---
- name: Set name of directory for XDebug
set_fact:
cikit_php_xdebug: "xdebug-{{ php.xdebug }}"
- name: Download XDebug
get_url:
url: "https://xdebug.org/files/{{ cikit_php_xdebug }}.tgz"
dest: "/tmp/{{ cikit_php_xdebug }}.tgz"
- name: Unarchive XDebug
shell: "tar -xf /tmp/{{ cikit_php_xdebug }}.tgz -C /tmp/"
args:
# Unarchive module does not work properly in Ansible 2.1.1.0.
warn: no
- include_tasks: fixtures/compile.yml name="XDebug" dir="/tmp/{{ cikit_php_xdebug }}"
<|file_sep|>updated/scripts/roles/cikit-php/tasks/xdebug.yml | get_url:
url: "https://xdebug.org/files/{{ cikit_php_xdebug }}.tgz"
dest: "/tmp/{{ cikit_php_xdebug }}.tgz"
- name: Unarchive XDebug
shell: "tar -xf /tmp/{{ cikit_php_xdebug }}.tgz -C /tmp/"
args:
# Unarchive module does not work properly in Ansible 2.1.1.0.
warn: no
- include_tasks: fixtures/compile.yml name="XDebug" dir="/tmp/{{ cikit_php_xdebug }}"
- name: Set environment variable
lineinfile:
dest: /etc/bash.bashrc
line: 'export {{ item.key }}="{{ item.value }}"'
regexp: "{{ item.value | regex_escape }}"
insertbefore: BOF
with_dict:
XDEBUG_CONFIG: "idekey={{ php.extensions.xdebug.config.idekey }}"
PHP_IDE_CONFIG: "serverName={{ ansible_fqdn }}" | <|file_sep|>original/scripts/roles/cikit-php/tasks/xdebug.yml
---
- name: Set name of directory for XDebug
set_fact:
cikit_php_xdebug: "xdebug-{{ php.xdebug }}"
- name: Download XDebug
get_url:
url: "https://xdebug.org/files/{{ cikit_php_xdebug }}.tgz"
dest: "/tmp/{{ cikit_php_xdebug }}.tgz"
- name: Unarchive XDebug
shell: "tar -xf /tmp/{{ cikit_php_xdebug }}.tgz -C /tmp/"
args:
# Unarchive module does not work properly in Ansible 2.1.1.0.
warn: no
- include_tasks: fixtures/compile.yml name="XDebug" dir="/tmp/{{ cikit_php_xdebug }}"
<|file_sep|>current/scripts/roles/cikit-php/tasks/xdebug.yml
---
- name: Set name of directory for XDebug
set_fact:
cikit_php_xdebug: "xdebug-{{ php.xdebug }}"
- name: Download XDebug
get_url:
url: "https://xdebug.org/files/{{ cikit_php_xdebug }}.tgz"
dest: "/tmp/{{ cikit_php_xdebug }}.tgz"
- name: Unarchive XDebug
shell: "tar -xf /tmp/{{ cikit_php_xdebug }}.tgz -C /tmp/"
args:
# Unarchive module does not work properly in Ansible 2.1.1.0.
warn: no
- include_tasks: fixtures/compile.yml name="XDebug" dir="/tmp/{{ cikit_php_xdebug }}"
<|file_sep|>updated/scripts/roles/cikit-php/tasks/xdebug.yml
get_url:
url: "https://xdebug.org/files/{{ cikit_php_xdebug }}.tgz"
dest: "/tmp/{{ cikit_php_xdebug }}.tgz"
- name: Unarchive XDebug
shell: "tar -xf /tmp/{{ cikit_php_xdebug }}.tgz -C /tmp/"
args:
# Unarchive module does not work properly in Ansible 2.1.1.0.
warn: no
- include_tasks: fixtures/compile.yml name="XDebug" dir="/tmp/{{ cikit_php_xdebug }}"
- name: Set environment variable
lineinfile:
dest: /etc/bash.bashrc
line: 'export {{ item.key }}="{{ item.value }}"'
regexp: "{{ item.value | regex_escape }}"
insertbefore: BOF
with_dict:
XDEBUG_CONFIG: "idekey={{ php.extensions.xdebug.config.idekey }}"
PHP_IDE_CONFIG: "serverName={{ ansible_fqdn }}" | ad7610ebdb647411307b2353fb01f7bc9e97b8ad | scripts/roles/cikit-php/tasks/xdebug.yml | scripts/roles/cikit-php/tasks/xdebug.yml | YAML |
<|file_sep|>recipes/_nginx.rb.diff
original:
service 'nginx' do
action [:enable, :start]
supports :restart => true, :start => true, :stop => true, :reload => true
end
updated:
<|file_sep|>original/recipes/_nginx.rb
include_recipe 'oc-graphite::_uwsgi'
service 'nginx' do
action [:enable, :start]
supports :restart => true, :start => true, :stop => true, :reload => true
end
template '/etc/nginx/sites-available/graphite' do
source 'nginx-graphite.erb'
owner 'root'
group 'root'
mode 0644
notifies :reload, 'service[nginx]', :delayed
end
link '/etc/nginx/sites-enabled/graphite' do
to '/etc/nginx/sites-available/graphite'
end
file '/etc/nginx/sites-enabled/default' do
<|file_sep|>current/recipes/_nginx.rb
include_recipe 'oc-graphite::_uwsgi'
template '/etc/nginx/sites-available/graphite' do
source 'nginx-graphite.erb'
owner 'root'
group 'root'
mode 0644
notifies :reload, 'service[nginx]', :delayed
end
link '/etc/nginx/sites-enabled/graphite' do
to '/etc/nginx/sites-available/graphite'
end
file '/etc/nginx/sites-enabled/default' do
action :delete
notifies :reload, 'service[nginx]', :delayed
only_if { node['oc-graphite']['nginx']['disable_default_vhost'] }
end
<|file_sep|>updated/recipes/_nginx.rb | group 'root'
mode 0644
notifies :reload, 'service[nginx]', :delayed
end
link '/etc/nginx/sites-enabled/graphite' do
to '/etc/nginx/sites-available/graphite'
end
file '/etc/nginx/sites-enabled/default' do
action :delete
notifies :reload, 'service[nginx]', :delayed
only_if { node['oc-graphite']['nginx']['disable_default_vhost'] }
end
service 'nginx' do
action [:enable, :start]
supports :restart => true, :reload => true
end | <|file_sep|>recipes/_nginx.rb.diff
original:
service 'nginx' do
action [:enable, :start]
supports :restart => true, :start => true, :stop => true, :reload => true
end
updated:
<|file_sep|>original/recipes/_nginx.rb
include_recipe 'oc-graphite::_uwsgi'
service 'nginx' do
action [:enable, :start]
supports :restart => true, :start => true, :stop => true, :reload => true
end
template '/etc/nginx/sites-available/graphite' do
source 'nginx-graphite.erb'
owner 'root'
group 'root'
mode 0644
notifies :reload, 'service[nginx]', :delayed
end
link '/etc/nginx/sites-enabled/graphite' do
to '/etc/nginx/sites-available/graphite'
end
file '/etc/nginx/sites-enabled/default' do
<|file_sep|>current/recipes/_nginx.rb
include_recipe 'oc-graphite::_uwsgi'
template '/etc/nginx/sites-available/graphite' do
source 'nginx-graphite.erb'
owner 'root'
group 'root'
mode 0644
notifies :reload, 'service[nginx]', :delayed
end
link '/etc/nginx/sites-enabled/graphite' do
to '/etc/nginx/sites-available/graphite'
end
file '/etc/nginx/sites-enabled/default' do
action :delete
notifies :reload, 'service[nginx]', :delayed
only_if { node['oc-graphite']['nginx']['disable_default_vhost'] }
end
<|file_sep|>updated/recipes/_nginx.rb
group 'root'
mode 0644
notifies :reload, 'service[nginx]', :delayed
end
link '/etc/nginx/sites-enabled/graphite' do
to '/etc/nginx/sites-available/graphite'
end
file '/etc/nginx/sites-enabled/default' do
action :delete
notifies :reload, 'service[nginx]', :delayed
only_if { node['oc-graphite']['nginx']['disable_default_vhost'] }
end
service 'nginx' do
action [:enable, :start]
supports :restart => true, :reload => true
end | 765a9399a7716e7b85e3c6f9a343bb602044f9b3 | recipes/_nginx.rb | recipes/_nginx.rb | Ruby |
<|file_sep|>original/install-manpages.sh
<|file_sep|>current/install-manpages.sh
<|file_sep|>updated/install-manpages.sh | #!/bin/sh -e
[ -z "$PREFIX" ] && PREFIX="/usr/local"
[ -z "$MANDIR" ] && MANDIR="$PREFIX/share/man"
install -d "$DESTDIR/$MANDIR/man1"
for manpage in man/*; do
install "$manpage" "$DESTDIR/$MANDIR/man1/`basename $manpage`"
done | <|file_sep|>original/install-manpages.sh
<|file_sep|>current/install-manpages.sh
<|file_sep|>updated/install-manpages.sh
#!/bin/sh -e
[ -z "$PREFIX" ] && PREFIX="/usr/local"
[ -z "$MANDIR" ] && MANDIR="$PREFIX/share/man"
install -d "$DESTDIR/$MANDIR/man1"
for manpage in man/*; do
install "$manpage" "$DESTDIR/$MANDIR/man1/`basename $manpage`"
done | 2dc7414413bd2a17e7bb65a0cb1ba1ae8bf91dc9 | install-manpages.sh | install-manpages.sh | Shell |
<|file_sep|>original/tests/test_server.py
from unittest.mock import ANY
from queue_functions import do_work
from server import handle_post
from uploaders.s3 import get_url
from uploaders.s3 import upload
def test_post():
q = Mock()
filename = 'afakefilename'
files = {'file': [{'body': b'a-fake-file-body', 'filename': filename}]}
hash_object = hashlib.md5(filename.encode())
audio_filename = hash_object.hexdigest() + "-" + filename
analysis_filename = audio_filename + '.analysis.json'
expected = {'analysis': get_url(analysis_filename), 'audio': get_url(audio_filename)}
actual = json.reads(handle_post(q, files, get_url, upload))
q.enqueue.assert_called_with(do_work, (ANY, audio_filename, analysis_filename, upload))
assert expected == actual
<|file_sep|>current/tests/test_server.py
from unittest.mock import ANY
from queue_functions import do_work
from server import handle_post
from uploaders.s3 import get_url
from uploaders.s3 import upload
def test_post():
q = Mock()
filename = 'afakefilename'
files = {'file': [{'body': b'a-fake-file-body', 'filename': filename}]}
hash_object = hashlib.md5(filename.encode())
audio_filename = hash_object.hexdigest() + "-" + filename
analysis_filename = audio_filename + '.analysis.json'
expected = {'analysis': get_url(analysis_filename), 'audio': get_url(audio_filename)}
actual = json.reads(handle_post(q, files, get_url, upload))
q.enqueue.assert_called_with(do_work, (ANY, audio_filename, analysis_filename, upload))
assert expected == actual
<|file_sep|>updated/tests/test_server.py | from unittest.mock import ANY
from queue_functions import do_work
from server import handle_post
from uploaders.s3 import get_url
from uploaders.s3 import upload
def test_post():
q = Mock()
filename = 'afakefilename'
files = {'file': [{'body': b'a-fake-file-body', 'filename': filename}]}
hash_object = hashlib.md5(filename.encode())
audio_filename = hash_object.hexdigest() + "-" + filename
analysis_filename = audio_filename + '.analysis.json'
expected = {'analysis': get_url(analysis_filename), 'audio': get_url(audio_filename)}
actual = json.loads(handle_post(q, files, get_url, upload))
q.enqueue.assert_called_with(do_work, (ANY, audio_filename, analysis_filename, upload))
assert expected == actual | <|file_sep|>original/tests/test_server.py
from unittest.mock import ANY
from queue_functions import do_work
from server import handle_post
from uploaders.s3 import get_url
from uploaders.s3 import upload
def test_post():
q = Mock()
filename = 'afakefilename'
files = {'file': [{'body': b'a-fake-file-body', 'filename': filename}]}
hash_object = hashlib.md5(filename.encode())
audio_filename = hash_object.hexdigest() + "-" + filename
analysis_filename = audio_filename + '.analysis.json'
expected = {'analysis': get_url(analysis_filename), 'audio': get_url(audio_filename)}
actual = json.reads(handle_post(q, files, get_url, upload))
q.enqueue.assert_called_with(do_work, (ANY, audio_filename, analysis_filename, upload))
assert expected == actual
<|file_sep|>current/tests/test_server.py
from unittest.mock import ANY
from queue_functions import do_work
from server import handle_post
from uploaders.s3 import get_url
from uploaders.s3 import upload
def test_post():
q = Mock()
filename = 'afakefilename'
files = {'file': [{'body': b'a-fake-file-body', 'filename': filename}]}
hash_object = hashlib.md5(filename.encode())
audio_filename = hash_object.hexdigest() + "-" + filename
analysis_filename = audio_filename + '.analysis.json'
expected = {'analysis': get_url(analysis_filename), 'audio': get_url(audio_filename)}
actual = json.reads(handle_post(q, files, get_url, upload))
q.enqueue.assert_called_with(do_work, (ANY, audio_filename, analysis_filename, upload))
assert expected == actual
<|file_sep|>updated/tests/test_server.py
from unittest.mock import ANY
from queue_functions import do_work
from server import handle_post
from uploaders.s3 import get_url
from uploaders.s3 import upload
def test_post():
q = Mock()
filename = 'afakefilename'
files = {'file': [{'body': b'a-fake-file-body', 'filename': filename}]}
hash_object = hashlib.md5(filename.encode())
audio_filename = hash_object.hexdigest() + "-" + filename
analysis_filename = audio_filename + '.analysis.json'
expected = {'analysis': get_url(analysis_filename), 'audio': get_url(audio_filename)}
actual = json.loads(handle_post(q, files, get_url, upload))
q.enqueue.assert_called_with(do_work, (ANY, audio_filename, analysis_filename, upload))
assert expected == actual | 3b7dcc4d2a19b5ac03eebae35600c25dd038fe33 | tests/test_server.py | tests/test_server.py | Python |
<|file_sep|>original/perftest.py
<|file_sep|>current/perftest.py
<|file_sep|>updated/perftest.py | """
Simple peformance tests.
"""
import sys
import time
import couchdb
def main():
print 'sys.version : %r' % (sys.version,)
print 'sys.platform : %r' % (sys.platform,)
tests = [create_doc, create_bulk_docs]
if len(sys.argv) > 1:
tests = [test for test in tests if test.__name__ in sys.argv[1:]]
server = couchdb.Server()
for test in tests: | <|file_sep|>original/perftest.py
<|file_sep|>current/perftest.py
<|file_sep|>updated/perftest.py
"""
Simple peformance tests.
"""
import sys
import time
import couchdb
def main():
print 'sys.version : %r' % (sys.version,)
print 'sys.platform : %r' % (sys.platform,)
tests = [create_doc, create_bulk_docs]
if len(sys.argv) > 1:
tests = [test for test in tests if test.__name__ in sys.argv[1:]]
server = couchdb.Server()
for test in tests: | 877406927bc4754daeab10b9bfb0f7879e8f6092 | perftest.py | perftest.py | Python |
<|file_sep|>original/tests/testMustache.php
<|file_sep|>current/tests/testMustache.php
<|file_sep|>updated/tests/testMustache.php | <?php
/*
* Mustache PHP Compiler - Test the Mustache class
*
* (c) July 2011 - Manuel Odendahl - wesen@ruinwesen.com
*/
require_once(dirname(__FILE__)."/../vendor/simpletest/autorun.php");
require_once(dirname(__FILE__)."/../Mustache.php");
function __filename($file) {
return dirname(__FILE__).'/files/'.$file;
}
class TestMustache extends UnitTestCase {
function testPartial() {
$m = new Mustache();
$res = $m->partial(__filename("testPartial.mustache"));
$this->assertEqual($res, "partial {{mustache}}\n"); | <|file_sep|>original/tests/testMustache.php
<|file_sep|>current/tests/testMustache.php
<|file_sep|>updated/tests/testMustache.php
<?php
/*
* Mustache PHP Compiler - Test the Mustache class
*
* (c) July 2011 - Manuel Odendahl - wesen@ruinwesen.com
*/
require_once(dirname(__FILE__)."/../vendor/simpletest/autorun.php");
require_once(dirname(__FILE__)."/../Mustache.php");
function __filename($file) {
return dirname(__FILE__).'/files/'.$file;
}
class TestMustache extends UnitTestCase {
function testPartial() {
$m = new Mustache();
$res = $m->partial(__filename("testPartial.mustache"));
$this->assertEqual($res, "partial {{mustache}}\n"); | 266132d339ca83c84b54841ff14804a6932289ae | tests/testMustache.php | tests/testMustache.php | PHP |
<|file_sep|>original/.travis.yml
language: go
go:
- 1.6.x
- 1.7.x
- 1.8.x
- 1.9.x
- 1.10.x
- 1.11.x
- 1.12.x
- tip
gobuild_args: -tags no_pkg_config
dist: xenial
addons:
apt:
packages:
- bison
- flex
- automake
- autoconf
- libtool
- make
<|file_sep|>current/.travis.yml
language: go
go:
- 1.6.x
- 1.7.x
- 1.8.x
- 1.9.x
- 1.10.x
- 1.11.x
- 1.12.x
- tip
gobuild_args: -tags no_pkg_config
dist: xenial
addons:
apt:
packages:
- bison
- flex
- automake
- autoconf
- libtool
- make
<|file_sep|>updated/.travis.yml | language: go
go:
- 1.7.x
- 1.8.x
- 1.9.x
- 1.10.x
- 1.11.x
- 1.12.x
- tip
gobuild_args: -tags no_pkg_config
dist: xenial
addons:
apt:
packages:
- bison
- flex
- automake
- autoconf
- libtool
- make
- gcc | <|file_sep|>original/.travis.yml
language: go
go:
- 1.6.x
- 1.7.x
- 1.8.x
- 1.9.x
- 1.10.x
- 1.11.x
- 1.12.x
- tip
gobuild_args: -tags no_pkg_config
dist: xenial
addons:
apt:
packages:
- bison
- flex
- automake
- autoconf
- libtool
- make
<|file_sep|>current/.travis.yml
language: go
go:
- 1.6.x
- 1.7.x
- 1.8.x
- 1.9.x
- 1.10.x
- 1.11.x
- 1.12.x
- tip
gobuild_args: -tags no_pkg_config
dist: xenial
addons:
apt:
packages:
- bison
- flex
- automake
- autoconf
- libtool
- make
<|file_sep|>updated/.travis.yml
language: go
go:
- 1.7.x
- 1.8.x
- 1.9.x
- 1.10.x
- 1.11.x
- 1.12.x
- tip
gobuild_args: -tags no_pkg_config
dist: xenial
addons:
apt:
packages:
- bison
- flex
- automake
- autoconf
- libtool
- make
- gcc | ce52fb9bf92529e068f200314fd6cbbbb251c7c2 | .travis.yml | .travis.yml | YAML |
<|file_sep|>original/templates/ubuntu-14.04-server-amd64/vbox.sh
# Without libdbus virtualbox would not start automatically after compile
apt-get -y install --no-install-recommends libdbus-1-3
# Remove existing VirtualBox guest additions
/etc/init.d/virtualbox-ose-guest-utils stop
rmmod vboxguest
aptitude -y purge virtualbox-ose-guest-x11 virtualbox-ose-guest-dkms virtualbox-ose-guest-utils
aptitude -y install dkms
# Install the VirtualBox guest additions
VBOX_VERSION=$(cat /home/vagrant/.vbox_version)
VBOX_ISO=VBoxGuestAdditions_$VBOX_VERSION.iso
mount -o loop $VBOX_ISO /mnt
yes|sh /mnt/VBoxLinuxAdditions.run
umount /mnt
# Cleanup
rm $VBOX_ISO
<|file_sep|>current/templates/ubuntu-14.04-server-amd64/vbox.sh
# Without libdbus virtualbox would not start automatically after compile
apt-get -y install --no-install-recommends libdbus-1-3
# Remove existing VirtualBox guest additions
/etc/init.d/virtualbox-ose-guest-utils stop
rmmod vboxguest
aptitude -y purge virtualbox-ose-guest-x11 virtualbox-ose-guest-dkms virtualbox-ose-guest-utils
aptitude -y install dkms
# Install the VirtualBox guest additions
VBOX_VERSION=$(cat /home/vagrant/.vbox_version)
VBOX_ISO=VBoxGuestAdditions_$VBOX_VERSION.iso
mount -o loop $VBOX_ISO /mnt
yes|sh /mnt/VBoxLinuxAdditions.run
umount /mnt
# Cleanup
rm $VBOX_ISO
<|file_sep|>updated/templates/ubuntu-14.04-server-amd64/vbox.sh | apt-get -y install --no-install-recommends libdbus-1-3
# Remove existing VirtualBox guest additions
/etc/init.d/virtualbox-ose-guest-utils stop
rmmod vboxguest
aptitude -y purge virtualbox-ose-guest-x11 virtualbox-ose-guest-dkms virtualbox-ose-guest-utils
aptitude -y install dkms
# Install the VirtualBox guest additions
VBOX_VERSION=$(cat /home/vagrant/.vbox_version)
VBOX_ISO=VBoxGuestAdditions_$VBOX_VERSION.iso
mount -o loop $VBOX_ISO /mnt
yes|sh /mnt/VBoxLinuxAdditions.run
umount /mnt
# Temporary fix for VirtualBox Additions version 4.3.10
# issue #12879, see https://www.virtualbox.org/ticket/12879
[ -e /usr/lib/VBoxGuestAdditions ] || ln -s /opt/VBoxGuestAdditions-$VBOX_VERSION/lib/VBoxGuestAdditions /usr/lib/VBoxGuestAdditions
# Cleanup
rm $VBOX_ISO | <|file_sep|>original/templates/ubuntu-14.04-server-amd64/vbox.sh
# Without libdbus virtualbox would not start automatically after compile
apt-get -y install --no-install-recommends libdbus-1-3
# Remove existing VirtualBox guest additions
/etc/init.d/virtualbox-ose-guest-utils stop
rmmod vboxguest
aptitude -y purge virtualbox-ose-guest-x11 virtualbox-ose-guest-dkms virtualbox-ose-guest-utils
aptitude -y install dkms
# Install the VirtualBox guest additions
VBOX_VERSION=$(cat /home/vagrant/.vbox_version)
VBOX_ISO=VBoxGuestAdditions_$VBOX_VERSION.iso
mount -o loop $VBOX_ISO /mnt
yes|sh /mnt/VBoxLinuxAdditions.run
umount /mnt
# Cleanup
rm $VBOX_ISO
<|file_sep|>current/templates/ubuntu-14.04-server-amd64/vbox.sh
# Without libdbus virtualbox would not start automatically after compile
apt-get -y install --no-install-recommends libdbus-1-3
# Remove existing VirtualBox guest additions
/etc/init.d/virtualbox-ose-guest-utils stop
rmmod vboxguest
aptitude -y purge virtualbox-ose-guest-x11 virtualbox-ose-guest-dkms virtualbox-ose-guest-utils
aptitude -y install dkms
# Install the VirtualBox guest additions
VBOX_VERSION=$(cat /home/vagrant/.vbox_version)
VBOX_ISO=VBoxGuestAdditions_$VBOX_VERSION.iso
mount -o loop $VBOX_ISO /mnt
yes|sh /mnt/VBoxLinuxAdditions.run
umount /mnt
# Cleanup
rm $VBOX_ISO
<|file_sep|>updated/templates/ubuntu-14.04-server-amd64/vbox.sh
apt-get -y install --no-install-recommends libdbus-1-3
# Remove existing VirtualBox guest additions
/etc/init.d/virtualbox-ose-guest-utils stop
rmmod vboxguest
aptitude -y purge virtualbox-ose-guest-x11 virtualbox-ose-guest-dkms virtualbox-ose-guest-utils
aptitude -y install dkms
# Install the VirtualBox guest additions
VBOX_VERSION=$(cat /home/vagrant/.vbox_version)
VBOX_ISO=VBoxGuestAdditions_$VBOX_VERSION.iso
mount -o loop $VBOX_ISO /mnt
yes|sh /mnt/VBoxLinuxAdditions.run
umount /mnt
# Temporary fix for VirtualBox Additions version 4.3.10
# issue #12879, see https://www.virtualbox.org/ticket/12879
[ -e /usr/lib/VBoxGuestAdditions ] || ln -s /opt/VBoxGuestAdditions-$VBOX_VERSION/lib/VBoxGuestAdditions /usr/lib/VBoxGuestAdditions
# Cleanup
rm $VBOX_ISO | 3c0deaa545084ad95c0ce614b6ab01bfda2dcb0d | templates/ubuntu-14.04-server-amd64/vbox.sh | templates/ubuntu-14.04-server-amd64/vbox.sh | Shell |
<|file_sep|>original/src/OpenConext/EngineBlock/CompatibilityBundle/Bridge/MailSenderInterface.php
<?php
namespace OpenConext\EngineBlock\CompatibilityBundle\Bridge;
interface MailSenderInterface
{
/**
* @param MailMessage $message
* @return void
* @throws \Zend_Mail_Exception
*/
public function send(MailMessage $message);
}
<|file_sep|>current/src/OpenConext/EngineBlock/CompatibilityBundle/Bridge/MailSenderInterface.php
<?php
namespace OpenConext\EngineBlock\CompatibilityBundle\Bridge;
interface MailSenderInterface
{
/**
* @param MailMessage $message
* @return void
* @throws \Zend_Mail_Exception
*/
public function send(MailMessage $message);
}
<|file_sep|>updated/src/OpenConext/EngineBlock/CompatibilityBundle/Bridge/MailSenderInterface.php | <?php
namespace OpenConext\EngineBlock\CompatibilityBundle\Bridge;
interface MailSenderInterface
{
/**
* @param MailMessage $message
* @return void
*/
public function send(MailMessage $message);
} | <|file_sep|>original/src/OpenConext/EngineBlock/CompatibilityBundle/Bridge/MailSenderInterface.php
<?php
namespace OpenConext\EngineBlock\CompatibilityBundle\Bridge;
interface MailSenderInterface
{
/**
* @param MailMessage $message
* @return void
* @throws \Zend_Mail_Exception
*/
public function send(MailMessage $message);
}
<|file_sep|>current/src/OpenConext/EngineBlock/CompatibilityBundle/Bridge/MailSenderInterface.php
<?php
namespace OpenConext\EngineBlock\CompatibilityBundle\Bridge;
interface MailSenderInterface
{
/**
* @param MailMessage $message
* @return void
* @throws \Zend_Mail_Exception
*/
public function send(MailMessage $message);
}
<|file_sep|>updated/src/OpenConext/EngineBlock/CompatibilityBundle/Bridge/MailSenderInterface.php
<?php
namespace OpenConext\EngineBlock\CompatibilityBundle\Bridge;
interface MailSenderInterface
{
/**
* @param MailMessage $message
* @return void
*/
public function send(MailMessage $message);
} | a2532cd8abb5adccf29b5136c4ba23ecf46d3b36 | src/OpenConext/EngineBlock/CompatibilityBundle/Bridge/MailSenderInterface.php | src/OpenConext/EngineBlock/CompatibilityBundle/Bridge/MailSenderInterface.php | PHP |
<|file_sep|>original/package.json
"type": "git",
"url": "git://github.com/glaubinix/changelog-builder.git"
},
"license": "MIT",
"author": {
"name": "Stephan Vock",
"email": "stephan.vock@gmail.com"
},
"keywords": [
"git",
"svn",
"changelog"
],
"dependencies": {
"jira": "git://github.com/Woellchen/node-jira#develop"
},
"devDependencies": {
"precommit-hook": "*",
"tape": "^2.14.0"
}
}
<|file_sep|>current/package.json
"type": "git",
"url": "git://github.com/glaubinix/changelog-builder.git"
},
"license": "MIT",
"author": {
"name": "Stephan Vock",
"email": "stephan.vock@gmail.com"
},
"keywords": [
"git",
"svn",
"changelog"
],
"dependencies": {
"jira": "git://github.com/Woellchen/node-jira#develop"
},
"devDependencies": {
"precommit-hook": "*",
"tape": "^2.14.0"
}
}
<|file_sep|>updated/package.json | "type": "git",
"url": "git://github.com/glaubinix/changelog-builder.git"
},
"license": "MIT",
"author": {
"name": "Stephan Vock",
"email": "stephan.vock@gmail.com"
},
"keywords": [
"git",
"svn",
"changelog"
],
"dependencies": {
"jira": "^0.10.0"
},
"devDependencies": {
"precommit-hook": "*",
"tape": "^2.14.0"
}
} | <|file_sep|>original/package.json
"type": "git",
"url": "git://github.com/glaubinix/changelog-builder.git"
},
"license": "MIT",
"author": {
"name": "Stephan Vock",
"email": "stephan.vock@gmail.com"
},
"keywords": [
"git",
"svn",
"changelog"
],
"dependencies": {
"jira": "git://github.com/Woellchen/node-jira#develop"
},
"devDependencies": {
"precommit-hook": "*",
"tape": "^2.14.0"
}
}
<|file_sep|>current/package.json
"type": "git",
"url": "git://github.com/glaubinix/changelog-builder.git"
},
"license": "MIT",
"author": {
"name": "Stephan Vock",
"email": "stephan.vock@gmail.com"
},
"keywords": [
"git",
"svn",
"changelog"
],
"dependencies": {
"jira": "git://github.com/Woellchen/node-jira#develop"
},
"devDependencies": {
"precommit-hook": "*",
"tape": "^2.14.0"
}
}
<|file_sep|>updated/package.json
"type": "git",
"url": "git://github.com/glaubinix/changelog-builder.git"
},
"license": "MIT",
"author": {
"name": "Stephan Vock",
"email": "stephan.vock@gmail.com"
},
"keywords": [
"git",
"svn",
"changelog"
],
"dependencies": {
"jira": "^0.10.0"
},
"devDependencies": {
"precommit-hook": "*",
"tape": "^2.14.0"
}
} | b6462822ad53c586fd737e016fc92b9487a1182b | package.json | package.json | JSON |
<|file_sep|>original/composer.json
"homepage": "http://mickael-vieira.com"
}
],
"require": {
"php": ">=7.0",
"psr/link": "^1.0"
},
"require-dev": {
"phpspec/phpspec": "^3.0",
"squizlabs/php_codesniffer": "^3.0",
"henrikbjorn/phpspec-code-coverage": "^3.0",
"satooshi/php-coveralls": "^1.0"
},
"config": {
"bin-dir": "bin"
},
"autoload": {
"psr-4": {
"CollectionJson\\": "src/CollectionJson/"
}
},
<|file_sep|>current/composer.json
"homepage": "http://mickael-vieira.com"
}
],
"require": {
"php": ">=7.0",
"psr/link": "^1.0"
},
"require-dev": {
"phpspec/phpspec": "^3.0",
"squizlabs/php_codesniffer": "^3.0",
"henrikbjorn/phpspec-code-coverage": "^3.0",
"satooshi/php-coveralls": "^1.0"
},
"config": {
"bin-dir": "bin"
},
"autoload": {
"psr-4": {
"CollectionJson\\": "src/CollectionJson/"
}
},
<|file_sep|>updated/composer.json | "homepage": "http://mickael-vieira.com"
}
],
"require": {
"php": ">=7.0",
"psr/link": "^1.0"
},
"require-dev": {
"phpspec/phpspec": "^3.0",
"squizlabs/php_codesniffer": "^3.0",
"satooshi/php-coveralls": "^1.0",
"leanphp/phpspec-code-coverage": "^3.1"
},
"config": {
"bin-dir": "bin"
},
"autoload": {
"psr-4": {
"CollectionJson\\": "src/CollectionJson/"
}
}, | <|file_sep|>original/composer.json
"homepage": "http://mickael-vieira.com"
}
],
"require": {
"php": ">=7.0",
"psr/link": "^1.0"
},
"require-dev": {
"phpspec/phpspec": "^3.0",
"squizlabs/php_codesniffer": "^3.0",
"henrikbjorn/phpspec-code-coverage": "^3.0",
"satooshi/php-coveralls": "^1.0"
},
"config": {
"bin-dir": "bin"
},
"autoload": {
"psr-4": {
"CollectionJson\\": "src/CollectionJson/"
}
},
<|file_sep|>current/composer.json
"homepage": "http://mickael-vieira.com"
}
],
"require": {
"php": ">=7.0",
"psr/link": "^1.0"
},
"require-dev": {
"phpspec/phpspec": "^3.0",
"squizlabs/php_codesniffer": "^3.0",
"henrikbjorn/phpspec-code-coverage": "^3.0",
"satooshi/php-coveralls": "^1.0"
},
"config": {
"bin-dir": "bin"
},
"autoload": {
"psr-4": {
"CollectionJson\\": "src/CollectionJson/"
}
},
<|file_sep|>updated/composer.json
"homepage": "http://mickael-vieira.com"
}
],
"require": {
"php": ">=7.0",
"psr/link": "^1.0"
},
"require-dev": {
"phpspec/phpspec": "^3.0",
"squizlabs/php_codesniffer": "^3.0",
"satooshi/php-coveralls": "^1.0",
"leanphp/phpspec-code-coverage": "^3.1"
},
"config": {
"bin-dir": "bin"
},
"autoload": {
"psr-4": {
"CollectionJson\\": "src/CollectionJson/"
}
}, | 05b3d4913f7f95bf2d0a970a8a99d2656b3e2ea7 | composer.json | composer.json | JSON |
<|file_sep|>lib/learn_web/client/user/me.rb.diff
original:
exit
updated:
exit 1
<|file_sep|>lib/learn_web/client/user/me.rb.diff
original:
exit
updated:
exit 1
<|file_sep|>original/lib/learn_web/client/user/me.rb
populate_attributes!
elsif silent_output == false
case response.status
when 401
puts "It seems your OAuth token is incorrect. Please re-run config with: learn reset"
exit
when 500
puts "Something went wrong. Please try again."
exit
else
puts "Something went wrong. Please try again."
exit
end
end
self
end
end
end
end
end
<|file_sep|>current/lib/learn_web/client/user/me.rb
populate_attributes!
elsif silent_output == false
case response.status
when 401
puts "It seems your OAuth token is incorrect. Please re-run config with: learn reset"
exit 1
when 500
puts "Something went wrong. Please try again."
exit 1
else
puts "Something went wrong. Please try again."
exit
end
end
self
end
end
end
end
end
<|file_sep|>updated/lib/learn_web/client/user/me.rb | populate_attributes!
elsif silent_output == false
case response.status
when 401
puts "It seems your OAuth token is incorrect. Please re-run config with: learn reset"
exit 1
when 500
puts "Something went wrong. Please try again."
exit 1
else
puts "Something went wrong. Please try again."
exit 1
end
end
self
end
end
end
end
end | <|file_sep|>lib/learn_web/client/user/me.rb.diff
original:
exit
updated:
exit 1
<|file_sep|>lib/learn_web/client/user/me.rb.diff
original:
exit
updated:
exit 1
<|file_sep|>original/lib/learn_web/client/user/me.rb
populate_attributes!
elsif silent_output == false
case response.status
when 401
puts "It seems your OAuth token is incorrect. Please re-run config with: learn reset"
exit
when 500
puts "Something went wrong. Please try again."
exit
else
puts "Something went wrong. Please try again."
exit
end
end
self
end
end
end
end
end
<|file_sep|>current/lib/learn_web/client/user/me.rb
populate_attributes!
elsif silent_output == false
case response.status
when 401
puts "It seems your OAuth token is incorrect. Please re-run config with: learn reset"
exit 1
when 500
puts "Something went wrong. Please try again."
exit 1
else
puts "Something went wrong. Please try again."
exit
end
end
self
end
end
end
end
end
<|file_sep|>updated/lib/learn_web/client/user/me.rb
populate_attributes!
elsif silent_output == false
case response.status
when 401
puts "It seems your OAuth token is incorrect. Please re-run config with: learn reset"
exit 1
when 500
puts "Something went wrong. Please try again."
exit 1
else
puts "Something went wrong. Please try again."
exit 1
end
end
self
end
end
end
end
end | d0ad83967a024eb655b048d2a5cd48db1d54469a | lib/learn_web/client/user/me.rb | lib/learn_web/client/user/me.rb | Ruby |
<|file_sep|>data/building-hours/3-2-print-center.yaml.diff
original:
closedForChapelTime: true
updated:
# closedForChapelTime: true
<|file_sep|>original/data/building-hours/3-2-print-center.yaml
name: Print Center
image: print
category: Mail and Packages
schedule:
- title: Hours
closedForChapelTime: true
hours:
# - {days: [Mo, Tu, We, Th, Fr], from: '8:00am', to: '5:00pm'}
breakSchedule:
fall: []
thanksgiving: []
winter: []
interim: []
spring: []
easter: []
summer:
- title: Hours
hours:
- {days: [Mo, Tu, We, Th], from: '8:00am', to: '3:00pm'}
<|file_sep|>current/data/building-hours/3-2-print-center.yaml
name: Print Center
image: print
category: Mail and Packages
schedule:
- title: Hours
# closedForChapelTime: true
hours:
# - {days: [Mo, Tu, We, Th, Fr], from: '8:00am', to: '5:00pm'}
breakSchedule:
fall: []
thanksgiving: []
winter: []
interim: []
spring: []
easter: []
summer:
- title: Hours
hours:
- {days: [Mo, Tu, We, Th], from: '8:00am', to: '3:00pm'}
<|file_sep|>updated/data/building-hours/3-2-print-center.yaml | name: Print Center
image: print
category: Mail and Packages
schedule:
- title: Hours
# closedForChapelTime: true
hours:
# - {days: [Mo, Tu, We, Th, Fr], from: '8:00am', to: '5:00pm'}
- {days: [Mo, Tu, We, Th], from: '8:00am', to: '3:00pm'}
- {days: [Fr], from: '8:00am', to: '12:00pm'}
breakSchedule:
fall: []
thanksgiving: []
winter: []
interim: []
spring: []
easter: []
summer:
- title: Hours | <|file_sep|>data/building-hours/3-2-print-center.yaml.diff
original:
closedForChapelTime: true
updated:
# closedForChapelTime: true
<|file_sep|>original/data/building-hours/3-2-print-center.yaml
name: Print Center
image: print
category: Mail and Packages
schedule:
- title: Hours
closedForChapelTime: true
hours:
# - {days: [Mo, Tu, We, Th, Fr], from: '8:00am', to: '5:00pm'}
breakSchedule:
fall: []
thanksgiving: []
winter: []
interim: []
spring: []
easter: []
summer:
- title: Hours
hours:
- {days: [Mo, Tu, We, Th], from: '8:00am', to: '3:00pm'}
<|file_sep|>current/data/building-hours/3-2-print-center.yaml
name: Print Center
image: print
category: Mail and Packages
schedule:
- title: Hours
# closedForChapelTime: true
hours:
# - {days: [Mo, Tu, We, Th, Fr], from: '8:00am', to: '5:00pm'}
breakSchedule:
fall: []
thanksgiving: []
winter: []
interim: []
spring: []
easter: []
summer:
- title: Hours
hours:
- {days: [Mo, Tu, We, Th], from: '8:00am', to: '3:00pm'}
<|file_sep|>updated/data/building-hours/3-2-print-center.yaml
name: Print Center
image: print
category: Mail and Packages
schedule:
- title: Hours
# closedForChapelTime: true
hours:
# - {days: [Mo, Tu, We, Th, Fr], from: '8:00am', to: '5:00pm'}
- {days: [Mo, Tu, We, Th], from: '8:00am', to: '3:00pm'}
- {days: [Fr], from: '8:00am', to: '12:00pm'}
breakSchedule:
fall: []
thanksgiving: []
winter: []
interim: []
spring: []
easter: []
summer:
- title: Hours | 57c09ab37b8adcee99e375931404311acf0de878 | data/building-hours/3-2-print-center.yaml | data/building-hours/3-2-print-center.yaml | YAML |
<|file_sep|>original/generators/rspec/rspec_generator.rb
def initialize(runtime_args, runtime_options = {})
Dir.mkdir('lib/tasks') unless File.directory?('lib/tasks')
super
end
def manifest
record do |m|
script_options = { :chmod => 0755, :shebang => options[:shebang] == DEFAULT_SHEBANG ? nil : options[:shebang] }
if Rails::VERSION::STRING < "2.1.0"
m.file 'rspec.rake', 'lib/tasks/rspec.rake'
end
m.file 'script/autospec', 'script/autospec', script_options
m.file 'script/spec', 'script/spec', script_options
m.file 'script/spec_server', 'script/spec_server', script_options
m.directory 'spec'
m.file 'rcov.opts', 'spec/rcov.opts'
m.file 'spec.opts', 'spec/spec.opts'
m.template 'spec_helper.rb', 'spec/spec_helper.rb'
<|file_sep|>current/generators/rspec/rspec_generator.rb
def initialize(runtime_args, runtime_options = {})
Dir.mkdir('lib/tasks') unless File.directory?('lib/tasks')
super
end
def manifest
record do |m|
script_options = { :chmod => 0755, :shebang => options[:shebang] == DEFAULT_SHEBANG ? nil : options[:shebang] }
if Rails::VERSION::STRING < "2.1.0"
m.file 'rspec.rake', 'lib/tasks/rspec.rake'
end
m.file 'script/autospec', 'script/autospec', script_options
m.file 'script/spec', 'script/spec', script_options
m.file 'script/spec_server', 'script/spec_server', script_options
m.directory 'spec'
m.file 'rcov.opts', 'spec/rcov.opts'
m.file 'spec.opts', 'spec/spec.opts'
m.template 'spec_helper.rb', 'spec/spec_helper.rb'
<|file_sep|>updated/generators/rspec/rspec_generator.rb | def initialize(runtime_args, runtime_options = {})
Dir.mkdir('lib/tasks') unless File.directory?('lib/tasks')
super
end
def manifest
record do |m|
script_options = { :chmod => 0755, :shebang => options[:shebang] == DEFAULT_SHEBANG ? nil : options[:shebang] }
m.file 'rspec.rake', 'lib/tasks/rspec.rake'
m.file 'script/autospec', 'script/autospec', script_options
m.file 'script/spec', 'script/spec', script_options
m.file 'script/spec_server', 'script/spec_server', script_options
m.directory 'spec'
m.file 'rcov.opts', 'spec/rcov.opts'
m.file 'spec.opts', 'spec/spec.opts'
m.template 'spec_helper.rb', 'spec/spec_helper.rb'
m.directory 'stories' | <|file_sep|>original/generators/rspec/rspec_generator.rb
def initialize(runtime_args, runtime_options = {})
Dir.mkdir('lib/tasks') unless File.directory?('lib/tasks')
super
end
def manifest
record do |m|
script_options = { :chmod => 0755, :shebang => options[:shebang] == DEFAULT_SHEBANG ? nil : options[:shebang] }
if Rails::VERSION::STRING < "2.1.0"
m.file 'rspec.rake', 'lib/tasks/rspec.rake'
end
m.file 'script/autospec', 'script/autospec', script_options
m.file 'script/spec', 'script/spec', script_options
m.file 'script/spec_server', 'script/spec_server', script_options
m.directory 'spec'
m.file 'rcov.opts', 'spec/rcov.opts'
m.file 'spec.opts', 'spec/spec.opts'
m.template 'spec_helper.rb', 'spec/spec_helper.rb'
<|file_sep|>current/generators/rspec/rspec_generator.rb
def initialize(runtime_args, runtime_options = {})
Dir.mkdir('lib/tasks') unless File.directory?('lib/tasks')
super
end
def manifest
record do |m|
script_options = { :chmod => 0755, :shebang => options[:shebang] == DEFAULT_SHEBANG ? nil : options[:shebang] }
if Rails::VERSION::STRING < "2.1.0"
m.file 'rspec.rake', 'lib/tasks/rspec.rake'
end
m.file 'script/autospec', 'script/autospec', script_options
m.file 'script/spec', 'script/spec', script_options
m.file 'script/spec_server', 'script/spec_server', script_options
m.directory 'spec'
m.file 'rcov.opts', 'spec/rcov.opts'
m.file 'spec.opts', 'spec/spec.opts'
m.template 'spec_helper.rb', 'spec/spec_helper.rb'
<|file_sep|>updated/generators/rspec/rspec_generator.rb
def initialize(runtime_args, runtime_options = {})
Dir.mkdir('lib/tasks') unless File.directory?('lib/tasks')
super
end
def manifest
record do |m|
script_options = { :chmod => 0755, :shebang => options[:shebang] == DEFAULT_SHEBANG ? nil : options[:shebang] }
m.file 'rspec.rake', 'lib/tasks/rspec.rake'
m.file 'script/autospec', 'script/autospec', script_options
m.file 'script/spec', 'script/spec', script_options
m.file 'script/spec_server', 'script/spec_server', script_options
m.directory 'spec'
m.file 'rcov.opts', 'spec/rcov.opts'
m.file 'spec.opts', 'spec/spec.opts'
m.template 'spec_helper.rb', 'spec/spec_helper.rb'
m.directory 'stories' | 469315829dce46c37c337b260d67ca581189763f | generators/rspec/rspec_generator.rb | generators/rspec/rspec_generator.rb | Ruby |
<|file_sep|>libs/utils.py.diff
original:
#get the cache key for storage
updated:
<|file_sep|>libs/utils.py.diff
original:
updated:
"""Get the cache key for storage"""
<|file_sep|>libs/utils.py.diff
original:
#decorator for caching functions
updated:
<|file_sep|>libs/utils.py.diff
original:
updated:
"""Decorator for caching functions"""
<|file_sep|>original/libs/utils.py
serialise.append(str(arg))
for key,arg in kwargs.items():
if key == "clear_cache":
continue
serialise.append(str(key))
serialise.append(str(arg))
key = hashlib.md5("".join(serialise)).hexdigest()
return key
#decorator for caching functions
def cache_for(time):
def decorator(fn):
def wrapper(*args, **kwargs):
key = cache_get_key(fn.__name__, *args, **kwargs)
result = cache.get(key)
if not result or "clear_cache" in kwargs and kwargs["clear_cache"]:
result = fn(*args, **kwargs)
cache.set(key, result, time)
return result
return wrapper
return decorator
<|file_sep|>current/libs/utils.py
for key,arg in kwargs.items():
if key == "clear_cache":
continue
serialise.append(str(key))
serialise.append(str(arg))
key = hashlib.md5("".join(serialise)).hexdigest()
return key
def cache_for(time):
"""Decorator for caching functions"""
def decorator(fn):
def wrapper(*args, **kwargs):
key = cache_get_key(fn.__name__, *args, **kwargs)
result = cache.get(key)
if not result or "clear_cache" in kwargs and kwargs["clear_cache"]:
result = fn(*args, **kwargs)
cache.set(key, result, time)
return result
return wrapper
return decorator
<|file_sep|>updated/libs/utils.py | if key == "clear_cache":
continue
serialise.append(str(key))
serialise.append(str(arg))
key = hashlib.md5("".join(serialise)).hexdigest()
return key
def cache_for(time):
"""Decorator for caching functions"""
def decorator(fn):
def wrapper(*args, **kwargs):
key = cache_get_key(fn.__name__, *args, **kwargs)
result = cache.get(key)
if not result or "clear_cache" in kwargs and kwargs["clear_cache"]:
cache.delete(key)
result = fn(*args, **kwargs)
cache.set(key, result, time)
return result
return wrapper
return decorator | <|file_sep|>libs/utils.py.diff
original:
#get the cache key for storage
updated:
<|file_sep|>libs/utils.py.diff
original:
updated:
"""Get the cache key for storage"""
<|file_sep|>libs/utils.py.diff
original:
#decorator for caching functions
updated:
<|file_sep|>libs/utils.py.diff
original:
updated:
"""Decorator for caching functions"""
<|file_sep|>original/libs/utils.py
serialise.append(str(arg))
for key,arg in kwargs.items():
if key == "clear_cache":
continue
serialise.append(str(key))
serialise.append(str(arg))
key = hashlib.md5("".join(serialise)).hexdigest()
return key
#decorator for caching functions
def cache_for(time):
def decorator(fn):
def wrapper(*args, **kwargs):
key = cache_get_key(fn.__name__, *args, **kwargs)
result = cache.get(key)
if not result or "clear_cache" in kwargs and kwargs["clear_cache"]:
result = fn(*args, **kwargs)
cache.set(key, result, time)
return result
return wrapper
return decorator
<|file_sep|>current/libs/utils.py
for key,arg in kwargs.items():
if key == "clear_cache":
continue
serialise.append(str(key))
serialise.append(str(arg))
key = hashlib.md5("".join(serialise)).hexdigest()
return key
def cache_for(time):
"""Decorator for caching functions"""
def decorator(fn):
def wrapper(*args, **kwargs):
key = cache_get_key(fn.__name__, *args, **kwargs)
result = cache.get(key)
if not result or "clear_cache" in kwargs and kwargs["clear_cache"]:
result = fn(*args, **kwargs)
cache.set(key, result, time)
return result
return wrapper
return decorator
<|file_sep|>updated/libs/utils.py
if key == "clear_cache":
continue
serialise.append(str(key))
serialise.append(str(arg))
key = hashlib.md5("".join(serialise)).hexdigest()
return key
def cache_for(time):
"""Decorator for caching functions"""
def decorator(fn):
def wrapper(*args, **kwargs):
key = cache_get_key(fn.__name__, *args, **kwargs)
result = cache.get(key)
if not result or "clear_cache" in kwargs and kwargs["clear_cache"]:
cache.delete(key)
result = fn(*args, **kwargs)
cache.set(key, result, time)
return result
return wrapper
return decorator | 709b9e57d8ea664715fd9bb89729f99324c3e0c2 | libs/utils.py | libs/utils.py | Python |
<|file_sep|>original/.travis.yml
addons:
firefox: latest
google-chrome: latest
install:
- npm install
- bower install
before_script:
- gulp lint version:check
- polymer lint --rules polymer-2 --input ./src/*.html ./theme/*.html
- xvfb-run -s '-screen 0 1024x768x24' wct
script:
- if [[ "$TRAVIS_EVENT_TYPE" != "pull_request" && "$TRAVIS_BRANCH" != quick/* ]]; then
wct --env saucelabs;
fi
- if [ "$TRAVIS_EVENT_TYPE" = "cron" ]; then
wct --env saucelabs-cron;
fi
<|file_sep|>current/.travis.yml
addons:
firefox: latest
google-chrome: latest
install:
- npm install
- bower install
before_script:
- gulp lint version:check
- polymer lint --rules polymer-2 --input ./src/*.html ./theme/*.html
- xvfb-run -s '-screen 0 1024x768x24' wct
script:
- if [[ "$TRAVIS_EVENT_TYPE" != "pull_request" && "$TRAVIS_BRANCH" != quick/* ]]; then
wct --env saucelabs;
fi
- if [ "$TRAVIS_EVENT_TYPE" = "cron" ]; then
wct --env saucelabs-cron;
fi
<|file_sep|>updated/.travis.yml |
addons:
firefox: latest
google-chrome: latest
install:
- npm install
- bower install
before_script:
- gulp lint version:check
- polymer lint --rules polymer-2 --input ./src/*.html ./theme/**/*.html
- xvfb-run -s '-screen 0 1024x768x24' wct
script:
- if [[ "$TRAVIS_EVENT_TYPE" != "pull_request" && "$TRAVIS_BRANCH" != quick/* ]]; then
wct --env saucelabs;
fi
- if [ "$TRAVIS_EVENT_TYPE" = "cron" ]; then
wct --env saucelabs-cron;
fi | <|file_sep|>original/.travis.yml
addons:
firefox: latest
google-chrome: latest
install:
- npm install
- bower install
before_script:
- gulp lint version:check
- polymer lint --rules polymer-2 --input ./src/*.html ./theme/*.html
- xvfb-run -s '-screen 0 1024x768x24' wct
script:
- if [[ "$TRAVIS_EVENT_TYPE" != "pull_request" && "$TRAVIS_BRANCH" != quick/* ]]; then
wct --env saucelabs;
fi
- if [ "$TRAVIS_EVENT_TYPE" = "cron" ]; then
wct --env saucelabs-cron;
fi
<|file_sep|>current/.travis.yml
addons:
firefox: latest
google-chrome: latest
install:
- npm install
- bower install
before_script:
- gulp lint version:check
- polymer lint --rules polymer-2 --input ./src/*.html ./theme/*.html
- xvfb-run -s '-screen 0 1024x768x24' wct
script:
- if [[ "$TRAVIS_EVENT_TYPE" != "pull_request" && "$TRAVIS_BRANCH" != quick/* ]]; then
wct --env saucelabs;
fi
- if [ "$TRAVIS_EVENT_TYPE" = "cron" ]; then
wct --env saucelabs-cron;
fi
<|file_sep|>updated/.travis.yml
addons:
firefox: latest
google-chrome: latest
install:
- npm install
- bower install
before_script:
- gulp lint version:check
- polymer lint --rules polymer-2 --input ./src/*.html ./theme/**/*.html
- xvfb-run -s '-screen 0 1024x768x24' wct
script:
- if [[ "$TRAVIS_EVENT_TYPE" != "pull_request" && "$TRAVIS_BRANCH" != quick/* ]]; then
wct --env saucelabs;
fi
- if [ "$TRAVIS_EVENT_TYPE" = "cron" ]; then
wct --env saucelabs-cron;
fi | cb748883f94154aac65822eca34966a8b4ca83e8 | .travis.yml | .travis.yml | YAML |
<|file_sep|>original/remove_symlinks.sh
<|file_sep|>current/remove_symlinks.sh
<|file_sep|>updated/remove_symlinks.sh | #!/bin/bash
# -polka-
# Get the current directory
DIR="$( cd "$(dirname "$0" )" && pwd )"
# Link dot files to $HOME
for dotfile in $(find $DIR -maxdepth 1 -type f -name ".*"); do
if [[ $(uname) == 'Linux' && "$(basename $dotfile)" == '.tmux.linux.conf' ]]; then
rm -v "${HOME}/.tmux.conf"
cp -v $dotfile "${HOME}/.tmux.conf"
elif [[ $(uname) == 'Darwin' && "$(basename $dotfile)" == '.tmux.osx.conf' ]]; then
rm -v "${HOME}/.tmux.conf"
cp -v $dotfile "${HOME}/.tmux.conf"
elif [[ "$(basename $dotfile)" != ".tmux.osx.conf" && "$(basename $dotfile)" != ".tmux.linux.conf" ]]; then
rm -v "${HOME}/$(basename $dotfile)"
cp -v $dotfile "${HOME}/$(basename $dotfile)"
fi
done
# Link rc files in .bashrc.d | <|file_sep|>original/remove_symlinks.sh
<|file_sep|>current/remove_symlinks.sh
<|file_sep|>updated/remove_symlinks.sh
#!/bin/bash
# -polka-
# Get the current directory
DIR="$( cd "$(dirname "$0" )" && pwd )"
# Link dot files to $HOME
for dotfile in $(find $DIR -maxdepth 1 -type f -name ".*"); do
if [[ $(uname) == 'Linux' && "$(basename $dotfile)" == '.tmux.linux.conf' ]]; then
rm -v "${HOME}/.tmux.conf"
cp -v $dotfile "${HOME}/.tmux.conf"
elif [[ $(uname) == 'Darwin' && "$(basename $dotfile)" == '.tmux.osx.conf' ]]; then
rm -v "${HOME}/.tmux.conf"
cp -v $dotfile "${HOME}/.tmux.conf"
elif [[ "$(basename $dotfile)" != ".tmux.osx.conf" && "$(basename $dotfile)" != ".tmux.linux.conf" ]]; then
rm -v "${HOME}/$(basename $dotfile)"
cp -v $dotfile "${HOME}/$(basename $dotfile)"
fi
done
# Link rc files in .bashrc.d | 8b15d01bb2c722bded40046c57fe052993d78ef4 | remove_symlinks.sh | remove_symlinks.sh | Shell |
<|file_sep|>original/DevMachineInstall/devApps.ps1
# Dev apps to install
# Visual studio stuff
cinstm VisualStudio2013Ultimate
cinstm VisualStudio2012Ultimate
# Visual studio add-ins
cinstm xUnit.vs2012
cinstm WebEssentials2012
cinstm resharper -Version 7.1.3000.2254
cinstm typescript
cinstm typescript.vs
# databases and tools
cinstm MsSqlServer2012Express
cinstm mongodb
cinstm MongoVUE
# browsers
cinstm Firefox
cinstm GoogleChrome
<|file_sep|>current/DevMachineInstall/devApps.ps1
# Dev apps to install
# Visual studio stuff
cinstm VisualStudio2013Ultimate
cinstm VisualStudio2012Ultimate
# Visual studio add-ins
cinstm xUnit.vs2012
cinstm WebEssentials2012
cinstm resharper -Version 7.1.3000.2254
cinstm typescript
cinstm typescript.vs
# databases and tools
cinstm MsSqlServer2012Express
cinstm mongodb
cinstm MongoVUE
# browsers
cinstm Firefox
cinstm GoogleChrome
<|file_sep|>updated/DevMachineInstall/devApps.ps1 | # Dev apps to install
# Visual studio stuff
cinstm VisualStudio2013Ultimate
cinstm VisualStudio2012Ultimate
# Visual studio add-ins
cinstm xUnit.vs2012
cinstm WebEssentials2012
cinstm resharper -Version 7.1.3000.2254 # Use only resharper if you want the latests. This is what I have a license for
cinstm typescript
cinstm typescript.vs
# databases and tools
cinstm MsSqlServer2012Express
cinstm mongodb
cinstm MongoVUE
# browsers
cinstm Firefox
cinstm GoogleChrome | <|file_sep|>original/DevMachineInstall/devApps.ps1
# Dev apps to install
# Visual studio stuff
cinstm VisualStudio2013Ultimate
cinstm VisualStudio2012Ultimate
# Visual studio add-ins
cinstm xUnit.vs2012
cinstm WebEssentials2012
cinstm resharper -Version 7.1.3000.2254
cinstm typescript
cinstm typescript.vs
# databases and tools
cinstm MsSqlServer2012Express
cinstm mongodb
cinstm MongoVUE
# browsers
cinstm Firefox
cinstm GoogleChrome
<|file_sep|>current/DevMachineInstall/devApps.ps1
# Dev apps to install
# Visual studio stuff
cinstm VisualStudio2013Ultimate
cinstm VisualStudio2012Ultimate
# Visual studio add-ins
cinstm xUnit.vs2012
cinstm WebEssentials2012
cinstm resharper -Version 7.1.3000.2254
cinstm typescript
cinstm typescript.vs
# databases and tools
cinstm MsSqlServer2012Express
cinstm mongodb
cinstm MongoVUE
# browsers
cinstm Firefox
cinstm GoogleChrome
<|file_sep|>updated/DevMachineInstall/devApps.ps1
# Dev apps to install
# Visual studio stuff
cinstm VisualStudio2013Ultimate
cinstm VisualStudio2012Ultimate
# Visual studio add-ins
cinstm xUnit.vs2012
cinstm WebEssentials2012
cinstm resharper -Version 7.1.3000.2254 # Use only resharper if you want the latests. This is what I have a license for
cinstm typescript
cinstm typescript.vs
# databases and tools
cinstm MsSqlServer2012Express
cinstm mongodb
cinstm MongoVUE
# browsers
cinstm Firefox
cinstm GoogleChrome | 2fd17380c9cac9b5fd69af3fddf1b01807cdbd0e | DevMachineInstall/devApps.ps1 | DevMachineInstall/devApps.ps1 | PowerShell |
<|file_sep|>original/tests/units/FinderTest.php
<|file_sep|>current/tests/units/FinderTest.php
<|file_sep|>updated/tests/units/FinderTest.php | <?php
use Ktp\Finder;
use Ktp\Contracts\HttpClient as HttpClientContract;
class FinderTest extends PHPUnit_Framework_TestCase {
/** @test */
function finder_has_http_client()
{
$finder = new Finder;
$this->assertInstanceOf(HttpClientContract::class, $finder->httpClient());
}
/** @test */
function finder_find_return_null_when_nik_is_not_found()
{
$finder = new Finder;
$this->assertNull($finder->findByNik(123));
} | <|file_sep|>original/tests/units/FinderTest.php
<|file_sep|>current/tests/units/FinderTest.php
<|file_sep|>updated/tests/units/FinderTest.php
<?php
use Ktp\Finder;
use Ktp\Contracts\HttpClient as HttpClientContract;
class FinderTest extends PHPUnit_Framework_TestCase {
/** @test */
function finder_has_http_client()
{
$finder = new Finder;
$this->assertInstanceOf(HttpClientContract::class, $finder->httpClient());
}
/** @test */
function finder_find_return_null_when_nik_is_not_found()
{
$finder = new Finder;
$this->assertNull($finder->findByNik(123));
} | 96cb061f49e469957c440a2dae5a01bf6bda39fd | tests/units/FinderTest.php | tests/units/FinderTest.php | PHP |
<|file_sep|>original/README.md
#### TaskRunner / Building
* Grunt
* Webpack
* Babel
#### Testing
* Mocha
* Sinon
* Chai
#### Backend + DB
* Node.js
* Express
* Mongoose
* MongoDB
#### Map / Geolocation
* Leaflet.js
#### Graphic Icon
Front Lawn House by Alfredo Hernandez from the Noun Project
<|file_sep|>current/README.md
#### TaskRunner / Building
* Grunt
* Webpack
* Babel
#### Testing
* Mocha
* Sinon
* Chai
#### Backend + DB
* Node.js
* Express
* Mongoose
* MongoDB
#### Map / Geolocation
* Leaflet.js
#### Graphic Icon
Front Lawn House by Alfredo Hernandez from the Noun Project
<|file_sep|>updated/README.md | #### TaskRunner / Building
* Grunt
* Webpack
* Babel
#### Testing
* Mocha
* Sinon
* Chai
#### Backend + DB
* Node.js
* Express
* Mongoose
* MongoDB
#### Map / Geolocation
* Leaflet.js
#### Graphic Icon
* House by Vishwas Shetty from the Noun Project | <|file_sep|>original/README.md
#### TaskRunner / Building
* Grunt
* Webpack
* Babel
#### Testing
* Mocha
* Sinon
* Chai
#### Backend + DB
* Node.js
* Express
* Mongoose
* MongoDB
#### Map / Geolocation
* Leaflet.js
#### Graphic Icon
Front Lawn House by Alfredo Hernandez from the Noun Project
<|file_sep|>current/README.md
#### TaskRunner / Building
* Grunt
* Webpack
* Babel
#### Testing
* Mocha
* Sinon
* Chai
#### Backend + DB
* Node.js
* Express
* Mongoose
* MongoDB
#### Map / Geolocation
* Leaflet.js
#### Graphic Icon
Front Lawn House by Alfredo Hernandez from the Noun Project
<|file_sep|>updated/README.md
#### TaskRunner / Building
* Grunt
* Webpack
* Babel
#### Testing
* Mocha
* Sinon
* Chai
#### Backend + DB
* Node.js
* Express
* Mongoose
* MongoDB
#### Map / Geolocation
* Leaflet.js
#### Graphic Icon
* House by Vishwas Shetty from the Noun Project | 89858de26d73818bf48ff9915087cdea34a3fd21 | README.md | README.md | Markdown |
<|file_sep|>original/PATREON.md
Patreon Support
---------------
The Espruino Web IDE (and Espruino in general) is supported by generous
donations from many supporters on <a href="https://www.patreon.com/espruino" target="_blank">Patreon</a>
Supporters that pay $10/month or more to help support Espruino are:
Matt Eckman, Michael Leibman, Orry S, Eric Betts, MaBe, Sacha Gloor, INT3GRATION, James Palmer, Andreas Dröscher,
Dimitri Deniusov, John Kovach, Spence Konde, Ron Cromberge, Noah A Neibaron, Joshua Taylor, Thomas Smith, Ryan Paull
<|file_sep|>current/PATREON.md
Patreon Support
---------------
The Espruino Web IDE (and Espruino in general) is supported by generous
donations from many supporters on <a href="https://www.patreon.com/espruino" target="_blank">Patreon</a>
Supporters that pay $10/month or more to help support Espruino are:
Matt Eckman, Michael Leibman, Orry S, Eric Betts, MaBe, Sacha Gloor, INT3GRATION, James Palmer, Andreas Dröscher,
Dimitri Deniusov, John Kovach, Spence Konde, Ron Cromberge, Noah A Neibaron, Joshua Taylor, Thomas Smith, Ryan Paull
<|file_sep|>updated/PATREON.md | Patreon Support
---------------
The Espruino Web IDE (and Espruino in general) is supported by generous
donations from many supporters on <a href="https://www.patreon.com/espruino" target="_blank">Patreon</a>
Supporters that pay $10/month or more to help support Espruino are:
Matt Eckman, Michael Leibman, Orry S, Eric Betts, MaBe, Sacha Gloor, INT3GRATION, James Palmer, Andreas Dröscher,
Dimitri Deniusov, John Kovach, Spence Konde, Ron Cromberge, Noah A Neibaron, Joshua Taylor, Thomas Smith, Ryan Paull,
Gito Nirmolo
| <|file_sep|>original/PATREON.md
Patreon Support
---------------
The Espruino Web IDE (and Espruino in general) is supported by generous
donations from many supporters on <a href="https://www.patreon.com/espruino" target="_blank">Patreon</a>
Supporters that pay $10/month or more to help support Espruino are:
Matt Eckman, Michael Leibman, Orry S, Eric Betts, MaBe, Sacha Gloor, INT3GRATION, James Palmer, Andreas Dröscher,
Dimitri Deniusov, John Kovach, Spence Konde, Ron Cromberge, Noah A Neibaron, Joshua Taylor, Thomas Smith, Ryan Paull
<|file_sep|>current/PATREON.md
Patreon Support
---------------
The Espruino Web IDE (and Espruino in general) is supported by generous
donations from many supporters on <a href="https://www.patreon.com/espruino" target="_blank">Patreon</a>
Supporters that pay $10/month or more to help support Espruino are:
Matt Eckman, Michael Leibman, Orry S, Eric Betts, MaBe, Sacha Gloor, INT3GRATION, James Palmer, Andreas Dröscher,
Dimitri Deniusov, John Kovach, Spence Konde, Ron Cromberge, Noah A Neibaron, Joshua Taylor, Thomas Smith, Ryan Paull
<|file_sep|>updated/PATREON.md
Patreon Support
---------------
The Espruino Web IDE (and Espruino in general) is supported by generous
donations from many supporters on <a href="https://www.patreon.com/espruino" target="_blank">Patreon</a>
Supporters that pay $10/month or more to help support Espruino are:
Matt Eckman, Michael Leibman, Orry S, Eric Betts, MaBe, Sacha Gloor, INT3GRATION, James Palmer, Andreas Dröscher,
Dimitri Deniusov, John Kovach, Spence Konde, Ron Cromberge, Noah A Neibaron, Joshua Taylor, Thomas Smith, Ryan Paull,
Gito Nirmolo
| 33af531a529df7382a32aeb79d1057808239a870 | PATREON.md | PATREON.md | Markdown |
<|file_sep|>app/views/questions/_votes.html.erb.diff
original:
<%= form_for question, url: questions_upvote_path, method: :post do |f| %>
updated:
<%= form_for question, url: questions_upvote_path, method: :post, html: {class: "upvote_question"} do |f| %>
<|file_sep|>original/app/views/questions/_votes.html.erb
<div class="vote">
<%= form_for question, url: questions_upvote_path, method: :post do |f| %>
<input type="image" src="<%= image_path("upArrow.png") %>" alt="submit"/>
<input type="hidden" name="id" value="<%= question.id %>"/>
<% end %>
<%= form_for question, url: questions_downvote_path, method: :post do |f| %>
<input type="image" src="<%= image_path("downArrow.png") %>" alt="submit"/>
<input type="hidden" name="id" value="<%= question.id %>"/>
<% end %>
<span class="votes_count">
<%= question.votes_count %>
</span>
</div>
<|file_sep|>current/app/views/questions/_votes.html.erb
<div class="vote">
<%= form_for question, url: questions_upvote_path, method: :post, html: {class: "upvote_question"} do |f| %>
<input type="image" src="<%= image_path("upArrow.png") %>" alt="submit"/>
<input type="hidden" name="id" value="<%= question.id %>"/>
<% end %>
<%= form_for question, url: questions_downvote_path, method: :post do |f| %>
<input type="image" src="<%= image_path("downArrow.png") %>" alt="submit"/>
<input type="hidden" name="id" value="<%= question.id %>"/>
<% end %>
<span class="votes_count">
<%= question.votes_count %>
</span>
</div>
<|file_sep|>updated/app/views/questions/_votes.html.erb | <div class="vote">
<%= form_for question, url: questions_upvote_path, method: :post, html: {class: "upvote_question"} do |f| %>
<input type="image" src="<%= image_path("upArrow.png") %>" alt="submit"/>
<input type="hidden" name="id" value="<%= question.id %>"/>
<% end %>
<%= form_for question, url: questions_downvote_path, method: :post, html: {class: "downvote_question"} do |f| %>
<input type="image" src="<%= image_path("downArrow.png") %>" alt="submit"/>
<input type="hidden" name="id" value="<%= question.id %>"/>
<% end %>
<span class="votes_count">
<%= question.votes_count %>
</span>
</div> | <|file_sep|>app/views/questions/_votes.html.erb.diff
original:
<%= form_for question, url: questions_upvote_path, method: :post do |f| %>
updated:
<%= form_for question, url: questions_upvote_path, method: :post, html: {class: "upvote_question"} do |f| %>
<|file_sep|>original/app/views/questions/_votes.html.erb
<div class="vote">
<%= form_for question, url: questions_upvote_path, method: :post do |f| %>
<input type="image" src="<%= image_path("upArrow.png") %>" alt="submit"/>
<input type="hidden" name="id" value="<%= question.id %>"/>
<% end %>
<%= form_for question, url: questions_downvote_path, method: :post do |f| %>
<input type="image" src="<%= image_path("downArrow.png") %>" alt="submit"/>
<input type="hidden" name="id" value="<%= question.id %>"/>
<% end %>
<span class="votes_count">
<%= question.votes_count %>
</span>
</div>
<|file_sep|>current/app/views/questions/_votes.html.erb
<div class="vote">
<%= form_for question, url: questions_upvote_path, method: :post, html: {class: "upvote_question"} do |f| %>
<input type="image" src="<%= image_path("upArrow.png") %>" alt="submit"/>
<input type="hidden" name="id" value="<%= question.id %>"/>
<% end %>
<%= form_for question, url: questions_downvote_path, method: :post do |f| %>
<input type="image" src="<%= image_path("downArrow.png") %>" alt="submit"/>
<input type="hidden" name="id" value="<%= question.id %>"/>
<% end %>
<span class="votes_count">
<%= question.votes_count %>
</span>
</div>
<|file_sep|>updated/app/views/questions/_votes.html.erb
<div class="vote">
<%= form_for question, url: questions_upvote_path, method: :post, html: {class: "upvote_question"} do |f| %>
<input type="image" src="<%= image_path("upArrow.png") %>" alt="submit"/>
<input type="hidden" name="id" value="<%= question.id %>"/>
<% end %>
<%= form_for question, url: questions_downvote_path, method: :post, html: {class: "downvote_question"} do |f| %>
<input type="image" src="<%= image_path("downArrow.png") %>" alt="submit"/>
<input type="hidden" name="id" value="<%= question.id %>"/>
<% end %>
<span class="votes_count">
<%= question.votes_count %>
</span>
</div> | 2edd043877ba222f7ad0761761a9379a660378cd | app/views/questions/_votes.html.erb | app/views/questions/_votes.html.erb | HTML+ERB |
<|file_sep|>README.md.diff
original:
updated:
[](https://travis-ci.org/praekeltfoundation/debian-wheel-mirror)
[](https://requires.io/github/praekeltfoundation/debian-wheel-mirror/requirements/?branch=develop)
<|file_sep|>original/README.md
# debian-wheel-mirror
Automated Python wheel building and deployment for https://pypi.p16n.org
This repository contains a list of Python packages that we build into wheels and upload to our PyPi mirror. These wheels are used in our Debian-based Docker images: [`praekeltfoundation/dockerfiles`](https://github.com/praekeltfoundation/dockerfiles).
These packages are dependencies of software that we use or maintain. They have native extensions and so would typically require build tools to be installed before they could be used. We build the packages into binary wheels so that build tools are not required. Generally, other dependencies that don't require build tools should be fetched from the standard PyPi repository (https://pypi.python.org).
The packages are built inside Debian Jessie-based Docker containers. As such, the wheels that are produced are only guaranteed to be compatible with Debian Jessie. For some of these wheels, it may still be necessary to install runtime dependencies using `apt`.
<|file_sep|>current/README.md
# debian-wheel-mirror
[](https://travis-ci.org/praekeltfoundation/debian-wheel-mirror)
[](https://requires.io/github/praekeltfoundation/debian-wheel-mirror/requirements/?branch=develop)
Automated Python wheel building and deployment for https://pypi.p16n.org
This repository contains a list of Python packages that we build into wheels and upload to our PyPi mirror. These wheels are used in our Debian-based Docker images: [`praekeltfoundation/dockerfiles`](https://github.com/praekeltfoundation/dockerfiles).
These packages are dependencies of software that we use or maintain. They have native extensions and so would typically require build tools to be installed before they could be used. We build the packages into binary wheels so that build tools are not required. Generally, other dependencies that don't require build tools should be fetched from the standard PyPi repository (https://pypi.python.org).
The packages are built inside Debian Jessie-based Docker containers. As such, the wheels that are produced are only guaranteed to be compatible with Debian Jessie. For some of these wheels, it may still be necessary to install runtime dependencies using `apt`.
<|file_sep|>updated/README.md | # debian-wheel-mirror
[](https://travis-ci.org/praekeltfoundation/debian-wheel-mirror)
[](https://requires.io/github/praekeltfoundation/debian-wheel-mirror/requirements/?branch=develop)
Automated Python wheel building and deployment for https://pypi.p16n.org
This repository contains a list of Python packages that we build into wheels and upload to our PyPi mirror. These wheels are used in our Debian-based Docker images: [`praekeltfoundation/dockerfiles`](https://github.com/praekeltfoundation/dockerfiles).
The packages are kept up-to-date by automated pull requests made by [Requires.io](https://requires.io/github/praekeltfoundation/debian-wheel-mirror/requirements/?branch=develop). Packages are built inside Docker containers (currently using the [`python:2`](https://hub.docker.com/_/python/) image) and uploaded to our PyPi mirror by [Travis CI](https://travis-ci.org/praekeltfoundation/debian-wheel-mirror).
These packages are dependencies of software that we use or maintain. They have native extensions and so would typically require build tools to be installed before they could be used. We build the packages into binary wheels so that build tools are not required. Generally, other dependencies that don't require build tools should be fetched from the standard PyPi repository (https://pypi.python.org).
The packages are built inside Debian Jessie-based Docker containers. As such, the wheels that are produced are only guaranteed to be compatible with Debian Jessie. For some of these wheels, it may still be necessary to install runtime dependencies using `apt`. | <|file_sep|>README.md.diff
original:
updated:
[](https://travis-ci.org/praekeltfoundation/debian-wheel-mirror)
[](https://requires.io/github/praekeltfoundation/debian-wheel-mirror/requirements/?branch=develop)
<|file_sep|>original/README.md
# debian-wheel-mirror
Automated Python wheel building and deployment for https://pypi.p16n.org
This repository contains a list of Python packages that we build into wheels and upload to our PyPi mirror. These wheels are used in our Debian-based Docker images: [`praekeltfoundation/dockerfiles`](https://github.com/praekeltfoundation/dockerfiles).
These packages are dependencies of software that we use or maintain. They have native extensions and so would typically require build tools to be installed before they could be used. We build the packages into binary wheels so that build tools are not required. Generally, other dependencies that don't require build tools should be fetched from the standard PyPi repository (https://pypi.python.org).
The packages are built inside Debian Jessie-based Docker containers. As such, the wheels that are produced are only guaranteed to be compatible with Debian Jessie. For some of these wheels, it may still be necessary to install runtime dependencies using `apt`.
<|file_sep|>current/README.md
# debian-wheel-mirror
[](https://travis-ci.org/praekeltfoundation/debian-wheel-mirror)
[](https://requires.io/github/praekeltfoundation/debian-wheel-mirror/requirements/?branch=develop)
Automated Python wheel building and deployment for https://pypi.p16n.org
This repository contains a list of Python packages that we build into wheels and upload to our PyPi mirror. These wheels are used in our Debian-based Docker images: [`praekeltfoundation/dockerfiles`](https://github.com/praekeltfoundation/dockerfiles).
These packages are dependencies of software that we use or maintain. They have native extensions and so would typically require build tools to be installed before they could be used. We build the packages into binary wheels so that build tools are not required. Generally, other dependencies that don't require build tools should be fetched from the standard PyPi repository (https://pypi.python.org).
The packages are built inside Debian Jessie-based Docker containers. As such, the wheels that are produced are only guaranteed to be compatible with Debian Jessie. For some of these wheels, it may still be necessary to install runtime dependencies using `apt`.
<|file_sep|>updated/README.md
# debian-wheel-mirror
[](https://travis-ci.org/praekeltfoundation/debian-wheel-mirror)
[](https://requires.io/github/praekeltfoundation/debian-wheel-mirror/requirements/?branch=develop)
Automated Python wheel building and deployment for https://pypi.p16n.org
This repository contains a list of Python packages that we build into wheels and upload to our PyPi mirror. These wheels are used in our Debian-based Docker images: [`praekeltfoundation/dockerfiles`](https://github.com/praekeltfoundation/dockerfiles).
The packages are kept up-to-date by automated pull requests made by [Requires.io](https://requires.io/github/praekeltfoundation/debian-wheel-mirror/requirements/?branch=develop). Packages are built inside Docker containers (currently using the [`python:2`](https://hub.docker.com/_/python/) image) and uploaded to our PyPi mirror by [Travis CI](https://travis-ci.org/praekeltfoundation/debian-wheel-mirror).
These packages are dependencies of software that we use or maintain. They have native extensions and so would typically require build tools to be installed before they could be used. We build the packages into binary wheels so that build tools are not required. Generally, other dependencies that don't require build tools should be fetched from the standard PyPi repository (https://pypi.python.org).
The packages are built inside Debian Jessie-based Docker containers. As such, the wheels that are produced are only guaranteed to be compatible with Debian Jessie. For some of these wheels, it may still be necessary to install runtime dependencies using `apt`. | 3fa729e7fb11f05173ea14b0fed3294aff715c71 | README.md | README.md | Markdown |
<|file_sep|>benchmark-perf.gemspec.diff
original:
# coding: utf-8
updated:
<|file_sep|>original/benchmark-perf.gemspec
lib = File.expand_path('../lib', __FILE__)
$LOAD_PATH.unshift(lib) unless $LOAD_PATH.include?(lib)
require 'benchmark/perf/version'
Gem::Specification.new do |spec|
spec.name = "benchmark-perf"
spec.version = Benchmark::Perf::VERSION
spec.authors = ["Piotr Murach"]
spec.email = [""]
spec.summary = %q{Execution time and iteration performance benchmarking}
spec.description = %q{Execution time and iteration performance benchmarking}
spec.homepage = ""
spec.license = "MIT"
spec.files = `git ls-files -z`.split("\x0")
spec.executables = spec.files.grep(%r{^bin/}) { |f| File.basename(f) }
spec.test_files = spec.files.grep(%r{^spec/})
spec.require_paths = ["lib"]
spec.add_development_dependency 'bundler', '>= 1.5.0', '< 2.0'
spec.add_development_dependency 'rake'
<|file_sep|>current/benchmark-perf.gemspec
$LOAD_PATH.unshift(lib) unless $LOAD_PATH.include?(lib)
require 'benchmark/perf/version'
Gem::Specification.new do |spec|
spec.name = "benchmark-perf"
spec.version = Benchmark::Perf::VERSION
spec.authors = ["Piotr Murach"]
spec.email = [""]
spec.summary = %q{Execution time and iteration performance benchmarking}
spec.description = %q{Execution time and iteration performance benchmarking}
spec.homepage = ""
spec.license = "MIT"
spec.files = `git ls-files -z`.split("\x0")
spec.executables = spec.files.grep(%r{^bin/}) { |f| File.basename(f) }
spec.test_files = spec.files.grep(%r{^spec/})
spec.require_paths = ["lib"]
spec.add_development_dependency 'bundler', '>= 1.5.0', '< 2.0'
spec.add_development_dependency 'rake'
end
<|file_sep|>updated/benchmark-perf.gemspec | require 'benchmark/perf/version'
Gem::Specification.new do |spec|
spec.name = "benchmark-perf"
spec.version = Benchmark::Perf::VERSION
spec.authors = ["Piotr Murach"]
spec.email = [""]
spec.summary = %q{Execution time and iteration performance benchmarking}
spec.description = %q{Execution time and iteration performance benchmarking}
spec.homepage = ""
spec.license = "MIT"
spec.files = `git ls-files -z`.split("\x0")
spec.executables = spec.files.grep(%r{^bin/}) { |f| File.basename(f) }
spec.test_files = spec.files.grep(%r{^spec/})
spec.require_paths = ["lib"]
spec.add_development_dependency 'bundler', '~> 1.16'
spec.add_development_dependency 'rspec', '~> 3.0'
spec.add_development_dependency 'rake', '~> 10.0'
end | <|file_sep|>benchmark-perf.gemspec.diff
original:
# coding: utf-8
updated:
<|file_sep|>original/benchmark-perf.gemspec
lib = File.expand_path('../lib', __FILE__)
$LOAD_PATH.unshift(lib) unless $LOAD_PATH.include?(lib)
require 'benchmark/perf/version'
Gem::Specification.new do |spec|
spec.name = "benchmark-perf"
spec.version = Benchmark::Perf::VERSION
spec.authors = ["Piotr Murach"]
spec.email = [""]
spec.summary = %q{Execution time and iteration performance benchmarking}
spec.description = %q{Execution time and iteration performance benchmarking}
spec.homepage = ""
spec.license = "MIT"
spec.files = `git ls-files -z`.split("\x0")
spec.executables = spec.files.grep(%r{^bin/}) { |f| File.basename(f) }
spec.test_files = spec.files.grep(%r{^spec/})
spec.require_paths = ["lib"]
spec.add_development_dependency 'bundler', '>= 1.5.0', '< 2.0'
spec.add_development_dependency 'rake'
<|file_sep|>current/benchmark-perf.gemspec
$LOAD_PATH.unshift(lib) unless $LOAD_PATH.include?(lib)
require 'benchmark/perf/version'
Gem::Specification.new do |spec|
spec.name = "benchmark-perf"
spec.version = Benchmark::Perf::VERSION
spec.authors = ["Piotr Murach"]
spec.email = [""]
spec.summary = %q{Execution time and iteration performance benchmarking}
spec.description = %q{Execution time and iteration performance benchmarking}
spec.homepage = ""
spec.license = "MIT"
spec.files = `git ls-files -z`.split("\x0")
spec.executables = spec.files.grep(%r{^bin/}) { |f| File.basename(f) }
spec.test_files = spec.files.grep(%r{^spec/})
spec.require_paths = ["lib"]
spec.add_development_dependency 'bundler', '>= 1.5.0', '< 2.0'
spec.add_development_dependency 'rake'
end
<|file_sep|>updated/benchmark-perf.gemspec
require 'benchmark/perf/version'
Gem::Specification.new do |spec|
spec.name = "benchmark-perf"
spec.version = Benchmark::Perf::VERSION
spec.authors = ["Piotr Murach"]
spec.email = [""]
spec.summary = %q{Execution time and iteration performance benchmarking}
spec.description = %q{Execution time and iteration performance benchmarking}
spec.homepage = ""
spec.license = "MIT"
spec.files = `git ls-files -z`.split("\x0")
spec.executables = spec.files.grep(%r{^bin/}) { |f| File.basename(f) }
spec.test_files = spec.files.grep(%r{^spec/})
spec.require_paths = ["lib"]
spec.add_development_dependency 'bundler', '~> 1.16'
spec.add_development_dependency 'rspec', '~> 3.0'
spec.add_development_dependency 'rake', '~> 10.0'
end | d2f12484c620c7056d028fa552ee56a03233bf71 | benchmark-perf.gemspec | benchmark-perf.gemspec | Ruby |
<|file_sep|>.travis.yml.diff
original:
updated:
- "npm install -g markdownlint-cli"
<|file_sep|>original/.travis.yml
cache: "pip"
before_script:
# Run linters
- "flake8 ."
- "prospector --strictness veryhigh"
- "pylint src/vws"
- "mypy src tests"
# Make sure that imports are sorted as expected.
- "isort --recursive --check-only"
# Create an environment file used by `pytest-envfiles`.
# We don't add to this because we have environment variables set in the CI
# config.
- "touch vuforia_secrets.env"
script:
- "pytest --cov=src --cov=tests"
after_success:
- "coveralls"
<|file_sep|>current/.travis.yml
- "npm install -g markdownlint-cli"
cache: "pip"
before_script:
# Run linters
- "flake8 ."
- "prospector --strictness veryhigh"
- "pylint src/vws"
- "mypy src tests"
# Make sure that imports are sorted as expected.
- "isort --recursive --check-only"
# Create an environment file used by `pytest-envfiles`.
# We don't add to this because we have environment variables set in the CI
# config.
- "touch vuforia_secrets.env"
script:
- "pytest --cov=src --cov=tests"
after_success:
<|file_sep|>updated/.travis.yml | - "npm install -g markdownlint-cli"
cache: "pip"
before_script:
# Run linters
- "flake8 ."
- "prospector --strictness veryhigh"
- "pylint src/vws"
- "mypy src tests"
- "markdownlint README.md"
# Make sure that imports are sorted as expected.
- "isort --recursive --check-only"
# Create an environment file used by `pytest-envfiles`.
# We don't add to this because we have environment variables set in the CI
# config.
- "touch vuforia_secrets.env"
script:
- "pytest --cov=src --cov=tests"
| <|file_sep|>.travis.yml.diff
original:
updated:
- "npm install -g markdownlint-cli"
<|file_sep|>original/.travis.yml
cache: "pip"
before_script:
# Run linters
- "flake8 ."
- "prospector --strictness veryhigh"
- "pylint src/vws"
- "mypy src tests"
# Make sure that imports are sorted as expected.
- "isort --recursive --check-only"
# Create an environment file used by `pytest-envfiles`.
# We don't add to this because we have environment variables set in the CI
# config.
- "touch vuforia_secrets.env"
script:
- "pytest --cov=src --cov=tests"
after_success:
- "coveralls"
<|file_sep|>current/.travis.yml
- "npm install -g markdownlint-cli"
cache: "pip"
before_script:
# Run linters
- "flake8 ."
- "prospector --strictness veryhigh"
- "pylint src/vws"
- "mypy src tests"
# Make sure that imports are sorted as expected.
- "isort --recursive --check-only"
# Create an environment file used by `pytest-envfiles`.
# We don't add to this because we have environment variables set in the CI
# config.
- "touch vuforia_secrets.env"
script:
- "pytest --cov=src --cov=tests"
after_success:
<|file_sep|>updated/.travis.yml
- "npm install -g markdownlint-cli"
cache: "pip"
before_script:
# Run linters
- "flake8 ."
- "prospector --strictness veryhigh"
- "pylint src/vws"
- "mypy src tests"
- "markdownlint README.md"
# Make sure that imports are sorted as expected.
- "isort --recursive --check-only"
# Create an environment file used by `pytest-envfiles`.
# We don't add to this because we have environment variables set in the CI
# config.
- "touch vuforia_secrets.env"
script:
- "pytest --cov=src --cov=tests"
| 93e3ecf683d789df795c9bd1d7186268916ceeb0 | .travis.yml | .travis.yml | YAML |
<|file_sep|>packages/mw/mwc-probability-transition.yaml.diff
original:
hash: 42d82dd5cb90933637d7e247e88f0935fc75f8ea3e968e6abd1cec469ba9be8e
updated:
hash: abeae477b2f800f78cf4f4f9fc86b103472cdde5b2ef4761abfecbf2076b0e08
<|file_sep|>packages/mw/mwc-probability-transition.yaml.diff
original:
updated:
- '0.3.0.1'
<|file_sep|>original/packages/mw/mwc-probability-transition.yaml
mtl: -any
transformers: -any
mwc-probability: -any
primitive: -any
all-versions:
- '0.1.0.0'
- '0.2.0.0'
- '0.3.0.0'
author: Marco Zocca
latest: '0.3.0.0'
description-type: markdown
description: ! '# mwc-probability-transition
[](https://travis-ci.org/ocramz/mwc-probability-transition)
Types and primitives for stochastic simulation (e.g. integration of SDE, random
walks, Markov Chain Monte Carlo algorithms etc.)
'
<|file_sep|>current/packages/mw/mwc-probability-transition.yaml
mtl: -any
transformers: -any
mwc-probability: -any
primitive: -any
all-versions:
- '0.1.0.0'
- '0.2.0.0'
- '0.3.0.0'
- '0.3.0.1'
author: Marco Zocca
latest: '0.3.0.0'
description-type: markdown
description: ! '# mwc-probability-transition
[](https://travis-ci.org/ocramz/mwc-probability-transition)
Types and primitives for stochastic simulation (e.g. integration of SDE, random
walks, Markov Chain Monte Carlo algorithms etc.)
<|file_sep|>updated/packages/mw/mwc-probability-transition.yaml | mtl: -any
transformers: -any
mwc-probability: -any
primitive: -any
all-versions:
- '0.1.0.0'
- '0.2.0.0'
- '0.3.0.0'
- '0.3.0.1'
author: Marco Zocca
latest: '0.3.0.1'
description-type: markdown
description: ! '# mwc-probability-transition
[](https://travis-ci.org/ocramz/mwc-probability-transition)
Types and primitives for stochastic simulation (e.g. integration of SDE, random
walks, Markov Chain Monte Carlo algorithms etc.)
| <|file_sep|>packages/mw/mwc-probability-transition.yaml.diff
original:
hash: 42d82dd5cb90933637d7e247e88f0935fc75f8ea3e968e6abd1cec469ba9be8e
updated:
hash: abeae477b2f800f78cf4f4f9fc86b103472cdde5b2ef4761abfecbf2076b0e08
<|file_sep|>packages/mw/mwc-probability-transition.yaml.diff
original:
updated:
- '0.3.0.1'
<|file_sep|>original/packages/mw/mwc-probability-transition.yaml
mtl: -any
transformers: -any
mwc-probability: -any
primitive: -any
all-versions:
- '0.1.0.0'
- '0.2.0.0'
- '0.3.0.0'
author: Marco Zocca
latest: '0.3.0.0'
description-type: markdown
description: ! '# mwc-probability-transition
[](https://travis-ci.org/ocramz/mwc-probability-transition)
Types and primitives for stochastic simulation (e.g. integration of SDE, random
walks, Markov Chain Monte Carlo algorithms etc.)
'
<|file_sep|>current/packages/mw/mwc-probability-transition.yaml
mtl: -any
transformers: -any
mwc-probability: -any
primitive: -any
all-versions:
- '0.1.0.0'
- '0.2.0.0'
- '0.3.0.0'
- '0.3.0.1'
author: Marco Zocca
latest: '0.3.0.0'
description-type: markdown
description: ! '# mwc-probability-transition
[](https://travis-ci.org/ocramz/mwc-probability-transition)
Types and primitives for stochastic simulation (e.g. integration of SDE, random
walks, Markov Chain Monte Carlo algorithms etc.)
<|file_sep|>updated/packages/mw/mwc-probability-transition.yaml
mtl: -any
transformers: -any
mwc-probability: -any
primitive: -any
all-versions:
- '0.1.0.0'
- '0.2.0.0'
- '0.3.0.0'
- '0.3.0.1'
author: Marco Zocca
latest: '0.3.0.1'
description-type: markdown
description: ! '# mwc-probability-transition
[](https://travis-ci.org/ocramz/mwc-probability-transition)
Types and primitives for stochastic simulation (e.g. integration of SDE, random
walks, Markov Chain Monte Carlo algorithms etc.)
| bae4265c31082a932a8c4b3dfaa70bbab4635bdd | packages/mw/mwc-probability-transition.yaml | packages/mw/mwc-probability-transition.yaml | YAML |
<|file_sep|>original/server/requirements.txt
gunicorn==19.4.5
honcho==0.6.6
Eve==0.6.3
newrelic>=2.66,<2.67
-e git+git://github.com/superdesk/superdesk-core@959c6bb#egg=Superdesk-Core
<|file_sep|>current/server/requirements.txt
gunicorn==19.4.5
honcho==0.6.6
Eve==0.6.3
newrelic>=2.66,<2.67
-e git+git://github.com/superdesk/superdesk-core@959c6bb#egg=Superdesk-Core
<|file_sep|>updated/server/requirements.txt | gunicorn==19.4.5
honcho==0.6.6
Eve==0.6.3
newrelic>=2.66,<2.67
-e git+git://github.com/superdesk/superdesk-core@7fca944#egg=Superdesk-Core | <|file_sep|>original/server/requirements.txt
gunicorn==19.4.5
honcho==0.6.6
Eve==0.6.3
newrelic>=2.66,<2.67
-e git+git://github.com/superdesk/superdesk-core@959c6bb#egg=Superdesk-Core
<|file_sep|>current/server/requirements.txt
gunicorn==19.4.5
honcho==0.6.6
Eve==0.6.3
newrelic>=2.66,<2.67
-e git+git://github.com/superdesk/superdesk-core@959c6bb#egg=Superdesk-Core
<|file_sep|>updated/server/requirements.txt
gunicorn==19.4.5
honcho==0.6.6
Eve==0.6.3
newrelic>=2.66,<2.67
-e git+git://github.com/superdesk/superdesk-core@7fca944#egg=Superdesk-Core | 6bafeb289f56ddc41360384a42c0fa67c63ba5dd | server/requirements.txt | server/requirements.txt | Text |
<|file_sep|>original/iobuf/ibuf_readall.c
#include <iobuf/iobuf.h>
#include <str/str.h>
/** Read the remainder of the \c ibuf into the \c str. */
int ibuf_readall(ibuf* in, str* out)
{
for (;;) {
if (!str_catb(out,
in->io.buffer+in->io.bufstart,
in->io.buflen-in->io.bufstart))
return 0;
in->io.bufstart = in->io.buflen;
if (!ibuf_refill(in))
return ibuf_eof(in);
}
}
<|file_sep|>current/iobuf/ibuf_readall.c
#include <iobuf/iobuf.h>
#include <str/str.h>
/** Read the remainder of the \c ibuf into the \c str. */
int ibuf_readall(ibuf* in, str* out)
{
for (;;) {
if (!str_catb(out,
in->io.buffer+in->io.bufstart,
in->io.buflen-in->io.bufstart))
return 0;
in->io.bufstart = in->io.buflen;
if (!ibuf_refill(in))
return ibuf_eof(in);
}
}
<|file_sep|>updated/iobuf/ibuf_readall.c | #include <iobuf/iobuf.h>
#include <str/str.h>
/** Read the remainder of the \c ibuf into the \c str. */
int ibuf_readall(ibuf* in, str* out)
{
if (ibuf_eof(in)) return 1;
if (ibuf_error(in)) return 0;
for (;;) {
if (!str_catb(out,
in->io.buffer+in->io.bufstart,
in->io.buflen-in->io.bufstart))
return 0;
in->io.bufstart = in->io.buflen;
if (!ibuf_refill(in))
return ibuf_eof(in);
}
} | <|file_sep|>original/iobuf/ibuf_readall.c
#include <iobuf/iobuf.h>
#include <str/str.h>
/** Read the remainder of the \c ibuf into the \c str. */
int ibuf_readall(ibuf* in, str* out)
{
for (;;) {
if (!str_catb(out,
in->io.buffer+in->io.bufstart,
in->io.buflen-in->io.bufstart))
return 0;
in->io.bufstart = in->io.buflen;
if (!ibuf_refill(in))
return ibuf_eof(in);
}
}
<|file_sep|>current/iobuf/ibuf_readall.c
#include <iobuf/iobuf.h>
#include <str/str.h>
/** Read the remainder of the \c ibuf into the \c str. */
int ibuf_readall(ibuf* in, str* out)
{
for (;;) {
if (!str_catb(out,
in->io.buffer+in->io.bufstart,
in->io.buflen-in->io.bufstart))
return 0;
in->io.bufstart = in->io.buflen;
if (!ibuf_refill(in))
return ibuf_eof(in);
}
}
<|file_sep|>updated/iobuf/ibuf_readall.c
#include <iobuf/iobuf.h>
#include <str/str.h>
/** Read the remainder of the \c ibuf into the \c str. */
int ibuf_readall(ibuf* in, str* out)
{
if (ibuf_eof(in)) return 1;
if (ibuf_error(in)) return 0;
for (;;) {
if (!str_catb(out,
in->io.buffer+in->io.bufstart,
in->io.buflen-in->io.bufstart))
return 0;
in->io.bufstart = in->io.buflen;
if (!ibuf_refill(in))
return ibuf_eof(in);
}
} | 17530c3d7eceb106a74066446404117158e8aa93 | iobuf/ibuf_readall.c | iobuf/ibuf_readall.c | C |
<|file_sep|>original/.travis.yml
sudo: true
language: cpp
os:
- osx
compiler:
- gcc
cache:
directories:
- "$HOME/Library/Caches/Homebrew"
script:
- curl -L https://rawgit.com/InsightSoftwareConsortium/ITKPythonPackage/master/scripts/macpython-download-cache-and-build-module-wheels.sh -O
- chmod u+x macpython-download-cache-and-build-module-wheels.sh
- ./macpython-download-cache-and-build-module-wheels.sh 2.7 3.5
- tar -zcvf dist.tar.gz dist/
- curl -F file="@dist.tar.gz" https://file.io
<|file_sep|>current/.travis.yml
sudo: true
language: cpp
os:
- osx
compiler:
- gcc
cache:
directories:
- "$HOME/Library/Caches/Homebrew"
script:
- curl -L https://rawgit.com/InsightSoftwareConsortium/ITKPythonPackage/master/scripts/macpython-download-cache-and-build-module-wheels.sh -O
- chmod u+x macpython-download-cache-and-build-module-wheels.sh
- ./macpython-download-cache-and-build-module-wheels.sh 2.7 3.5
- tar -zcvf dist.tar.gz dist/
- curl -F file="@dist.tar.gz" https://file.io
<|file_sep|>updated/.travis.yml | sudo: true
language: cpp
os:
- osx
compiler:
- gcc
cache:
directories:
- "$HOME/Library/Caches/Homebrew"
script:
- curl -L https://rawgit.com/InsightSoftwareConsortium/ITKPythonPackage/master/scripts/macpython-download-cache-and-build-module-wheels.sh -O
- chmod u+x macpython-download-cache-and-build-module-wheels.sh
- ./macpython-download-cache-and-build-module-wheels.sh 2.7 3.5
- tar -zcvf dist.tar.gz dist/
- curl --upload-file dist.tar.gz https://transfer.sh/dist.tar.gz | <|file_sep|>original/.travis.yml
sudo: true
language: cpp
os:
- osx
compiler:
- gcc
cache:
directories:
- "$HOME/Library/Caches/Homebrew"
script:
- curl -L https://rawgit.com/InsightSoftwareConsortium/ITKPythonPackage/master/scripts/macpython-download-cache-and-build-module-wheels.sh -O
- chmod u+x macpython-download-cache-and-build-module-wheels.sh
- ./macpython-download-cache-and-build-module-wheels.sh 2.7 3.5
- tar -zcvf dist.tar.gz dist/
- curl -F file="@dist.tar.gz" https://file.io
<|file_sep|>current/.travis.yml
sudo: true
language: cpp
os:
- osx
compiler:
- gcc
cache:
directories:
- "$HOME/Library/Caches/Homebrew"
script:
- curl -L https://rawgit.com/InsightSoftwareConsortium/ITKPythonPackage/master/scripts/macpython-download-cache-and-build-module-wheels.sh -O
- chmod u+x macpython-download-cache-and-build-module-wheels.sh
- ./macpython-download-cache-and-build-module-wheels.sh 2.7 3.5
- tar -zcvf dist.tar.gz dist/
- curl -F file="@dist.tar.gz" https://file.io
<|file_sep|>updated/.travis.yml
sudo: true
language: cpp
os:
- osx
compiler:
- gcc
cache:
directories:
- "$HOME/Library/Caches/Homebrew"
script:
- curl -L https://rawgit.com/InsightSoftwareConsortium/ITKPythonPackage/master/scripts/macpython-download-cache-and-build-module-wheels.sh -O
- chmod u+x macpython-download-cache-and-build-module-wheels.sh
- ./macpython-download-cache-and-build-module-wheels.sh 2.7 3.5
- tar -zcvf dist.tar.gz dist/
- curl --upload-file dist.tar.gz https://transfer.sh/dist.tar.gz | 4757fb4442dfb1b1b4444a281fc3b246b0952cb7 | .travis.yml | .travis.yml | YAML |
<|file_sep|>original/spec/models/user_spec.rb
<|file_sep|>current/spec/models/user_spec.rb
<|file_sep|>updated/spec/models/user_spec.rb | require 'rails_helper'
require 'factory_girl_rails'
describe User do
let(:user){ FactoryGirl.create(:user) }
it "has a valid FACTORY" do
expect(FactoryGirl.create(:user)).to be_valid
end
it "should have an email address" do
#when left blank, this tests for truthiness
expect(user.email).to be
end
it "should have a valid email address" do
expect(user.email).to match(/^\w+.\w+@\w+.\w+$/)
end
it 'should be able to create snippets' do
expect(user.snippets).to be | <|file_sep|>original/spec/models/user_spec.rb
<|file_sep|>current/spec/models/user_spec.rb
<|file_sep|>updated/spec/models/user_spec.rb
require 'rails_helper'
require 'factory_girl_rails'
describe User do
let(:user){ FactoryGirl.create(:user) }
it "has a valid FACTORY" do
expect(FactoryGirl.create(:user)).to be_valid
end
it "should have an email address" do
#when left blank, this tests for truthiness
expect(user.email).to be
end
it "should have a valid email address" do
expect(user.email).to match(/^\w+.\w+@\w+.\w+$/)
end
it 'should be able to create snippets' do
expect(user.snippets).to be | a1a1830a4cee6d50d7a58b2ddf82105e9c5d255b | spec/models/user_spec.rb | spec/models/user_spec.rb | Ruby |
<|file_sep|>README.md.diff
original:
updated:
## Getting going
Here's how to give this a try. First, you'll need Postgres installed and
running. Next, run these commands:
```bash
$ git clone https://github.com/oestrich/hypermedia_rails.git
$ cd hypermedia_rails
$ bundle
$ RAILS_ENV=test bundle exec rake db:create db:migrate
$ RAILS_ENV=development bundle exec rake db:create db:migrate
$ bundle exec rake spec
```
You should see all the tests pass.
<|file_sep|>original/README.md
# Hypermedia API with Rails
This is a Hypermedia API example in Rails.
It uses several cool gems that are good for hypermedia and/or APIs:
* active_model_serializers
* rspec_api_documentation
* raddocs
## Hypermedia client
A basic example of a hypermedia client is included via Frenetic. Make sure to set up a user with the correct auth token.
<|file_sep|>current/README.md
* raddocs
## Getting going
Here's how to give this a try. First, you'll need Postgres installed and
running. Next, run these commands:
```bash
$ git clone https://github.com/oestrich/hypermedia_rails.git
$ cd hypermedia_rails
$ bundle
$ RAILS_ENV=test bundle exec rake db:create db:migrate
$ RAILS_ENV=development bundle exec rake db:create db:migrate
$ bundle exec rake spec
```
You should see all the tests pass.
## Hypermedia client
A basic example of a hypermedia client is included via Frenetic. Make sure to set up a user with the correct auth token.
<|file_sep|>updated/README.md |
```bash
$ git clone https://github.com/oestrich/hypermedia_rails.git
$ cd hypermedia_rails
$ bundle
$ RAILS_ENV=test bundle exec rake db:create db:migrate
$ RAILS_ENV=development bundle exec rake db:create db:migrate
$ bundle exec rake spec
```
You should see all the tests pass.
## Hypermedia client
A basic example of a hypermedia client is included via Frenetic. Make sure to set up a user with the correct auth token.
To try it, just
```bash
$ bundle exec ruby hypermedia.rb
``` | <|file_sep|>README.md.diff
original:
updated:
## Getting going
Here's how to give this a try. First, you'll need Postgres installed and
running. Next, run these commands:
```bash
$ git clone https://github.com/oestrich/hypermedia_rails.git
$ cd hypermedia_rails
$ bundle
$ RAILS_ENV=test bundle exec rake db:create db:migrate
$ RAILS_ENV=development bundle exec rake db:create db:migrate
$ bundle exec rake spec
```
You should see all the tests pass.
<|file_sep|>original/README.md
# Hypermedia API with Rails
This is a Hypermedia API example in Rails.
It uses several cool gems that are good for hypermedia and/or APIs:
* active_model_serializers
* rspec_api_documentation
* raddocs
## Hypermedia client
A basic example of a hypermedia client is included via Frenetic. Make sure to set up a user with the correct auth token.
<|file_sep|>current/README.md
* raddocs
## Getting going
Here's how to give this a try. First, you'll need Postgres installed and
running. Next, run these commands:
```bash
$ git clone https://github.com/oestrich/hypermedia_rails.git
$ cd hypermedia_rails
$ bundle
$ RAILS_ENV=test bundle exec rake db:create db:migrate
$ RAILS_ENV=development bundle exec rake db:create db:migrate
$ bundle exec rake spec
```
You should see all the tests pass.
## Hypermedia client
A basic example of a hypermedia client is included via Frenetic. Make sure to set up a user with the correct auth token.
<|file_sep|>updated/README.md
```bash
$ git clone https://github.com/oestrich/hypermedia_rails.git
$ cd hypermedia_rails
$ bundle
$ RAILS_ENV=test bundle exec rake db:create db:migrate
$ RAILS_ENV=development bundle exec rake db:create db:migrate
$ bundle exec rake spec
```
You should see all the tests pass.
## Hypermedia client
A basic example of a hypermedia client is included via Frenetic. Make sure to set up a user with the correct auth token.
To try it, just
```bash
$ bundle exec ruby hypermedia.rb
``` | 82ae82332ff37ffdbc17a4d1ae420a01ba4d88cc | README.md | README.md | Markdown |
<|file_sep|>original/composer.json
{
"name": "nexylan/paybox-direct-plus",
"description": "PayBox Direct Plus API wrapper",
"type": "library",
"license": "MIT",
"authors": [
{
"name": "Sullivan SENECHAL",
"email": "soullivaneuh@gmail.com"
}
],
"require": {
"php": "^5.5 || ^7.0",
"guzzlehttp/guzzle": "^6.0",
"symfony/options-resolver": "^2.7 || ^3.0"
},
"require-dev": {
"phpunit/phpunit": "^5.3",
"sllh/php-cs-fixer-styleci-bridge": "^2.0"
},
"autoload": {
<|file_sep|>current/composer.json
{
"name": "nexylan/paybox-direct-plus",
"description": "PayBox Direct Plus API wrapper",
"type": "library",
"license": "MIT",
"authors": [
{
"name": "Sullivan SENECHAL",
"email": "soullivaneuh@gmail.com"
}
],
"require": {
"php": "^5.5 || ^7.0",
"guzzlehttp/guzzle": "^6.0",
"symfony/options-resolver": "^2.7 || ^3.0"
},
"require-dev": {
"phpunit/phpunit": "^5.3",
"sllh/php-cs-fixer-styleci-bridge": "^2.0"
},
"autoload": {
<|file_sep|>updated/composer.json | {
"name": "nexylan/paybox-direct",
"description": "PayBox Direct and Direct Plus API wrapper",
"keywords": ["Paybox", "Direct", "API", "Symfony"],
"type": "library",
"license": "MIT",
"authors": [
{
"name": "Sullivan SENECHAL",
"email": "soullivaneuh@gmail.com"
}
],
"require": {
"php": "^5.5 || ^7.0",
"guzzlehttp/guzzle": "^6.0",
"symfony/options-resolver": "^2.7 || ^3.0"
},
"require-dev": {
"phpunit/phpunit": "^5.3",
"sllh/php-cs-fixer-styleci-bridge": "^2.0"
}, | <|file_sep|>original/composer.json
{
"name": "nexylan/paybox-direct-plus",
"description": "PayBox Direct Plus API wrapper",
"type": "library",
"license": "MIT",
"authors": [
{
"name": "Sullivan SENECHAL",
"email": "soullivaneuh@gmail.com"
}
],
"require": {
"php": "^5.5 || ^7.0",
"guzzlehttp/guzzle": "^6.0",
"symfony/options-resolver": "^2.7 || ^3.0"
},
"require-dev": {
"phpunit/phpunit": "^5.3",
"sllh/php-cs-fixer-styleci-bridge": "^2.0"
},
"autoload": {
<|file_sep|>current/composer.json
{
"name": "nexylan/paybox-direct-plus",
"description": "PayBox Direct Plus API wrapper",
"type": "library",
"license": "MIT",
"authors": [
{
"name": "Sullivan SENECHAL",
"email": "soullivaneuh@gmail.com"
}
],
"require": {
"php": "^5.5 || ^7.0",
"guzzlehttp/guzzle": "^6.0",
"symfony/options-resolver": "^2.7 || ^3.0"
},
"require-dev": {
"phpunit/phpunit": "^5.3",
"sllh/php-cs-fixer-styleci-bridge": "^2.0"
},
"autoload": {
<|file_sep|>updated/composer.json
{
"name": "nexylan/paybox-direct",
"description": "PayBox Direct and Direct Plus API wrapper",
"keywords": ["Paybox", "Direct", "API", "Symfony"],
"type": "library",
"license": "MIT",
"authors": [
{
"name": "Sullivan SENECHAL",
"email": "soullivaneuh@gmail.com"
}
],
"require": {
"php": "^5.5 || ^7.0",
"guzzlehttp/guzzle": "^6.0",
"symfony/options-resolver": "^2.7 || ^3.0"
},
"require-dev": {
"phpunit/phpunit": "^5.3",
"sllh/php-cs-fixer-styleci-bridge": "^2.0"
}, | 130cbf5603488eeea3fe6394353bad93ed1f7dd7 | composer.json | composer.json | JSON |
<|file_sep|>original/README.md
# HAProxy Documentation Converter
Made to convert the HAProxy documentation into HTML.
More than HTML, the main goal is to provide easy navigation.
## Documentations
A bot periodically fetches last commits for HAProxy 1.4 and 1.5 to produce up-to-date documentations.
Converted documentations are then stored online :
- [HAProxy 1.4 Configuration Manual](http://cbonte.github.com/haproxy-dconv/configuration-1.4.html)
- [HAProxy 1.5 Configuration Manual](http://cbonte.github.com/haproxy-dconv/configuration-1.5.html)
## Contribute
The project now lives by itself, as it is sufficiently useable. But I'm sure we can do even better.
Feel free to report feature requests or to provide patches !
<|file_sep|>current/README.md
# HAProxy Documentation Converter
Made to convert the HAProxy documentation into HTML.
More than HTML, the main goal is to provide easy navigation.
## Documentations
A bot periodically fetches last commits for HAProxy 1.4 and 1.5 to produce up-to-date documentations.
Converted documentations are then stored online :
- [HAProxy 1.4 Configuration Manual](http://cbonte.github.com/haproxy-dconv/configuration-1.4.html)
- [HAProxy 1.5 Configuration Manual](http://cbonte.github.com/haproxy-dconv/configuration-1.5.html)
## Contribute
The project now lives by itself, as it is sufficiently useable. But I'm sure we can do even better.
Feel free to report feature requests or to provide patches !
<|file_sep|>updated/README.md | # HAProxy Documentation Converter
Made to convert the HAProxy documentation into HTML.
More than HTML, the main goal is to provide easy navigation.
## Documentations
A bot periodically fetches last commits for HAProxy 1.4 and 1.5 to produce up-to-date documentations.
Converted documentations are then stored online :
- HAProxy 1.4 Configuration Manual [stable](http://cbonte.github.com/haproxy-dconv/configuration-1.4.html) [snapshot](http://cbonte.github.com/haproxy-dconv/snapshot/configuration-1.4.html)
- HAProxy 1.5 Configuration Manual [stable](http://cbonte.github.com/haproxy-dconv/configuration-1.5.html) [snapshot](http://cbonte.github.com/haproxy-dconv/snapshot/configuration-1.5.html)
## Contribute
The project now lives by itself, as it is sufficiently useable. But I'm sure we can do even better.
Feel free to report feature requests or to provide patches !
| <|file_sep|>original/README.md
# HAProxy Documentation Converter
Made to convert the HAProxy documentation into HTML.
More than HTML, the main goal is to provide easy navigation.
## Documentations
A bot periodically fetches last commits for HAProxy 1.4 and 1.5 to produce up-to-date documentations.
Converted documentations are then stored online :
- [HAProxy 1.4 Configuration Manual](http://cbonte.github.com/haproxy-dconv/configuration-1.4.html)
- [HAProxy 1.5 Configuration Manual](http://cbonte.github.com/haproxy-dconv/configuration-1.5.html)
## Contribute
The project now lives by itself, as it is sufficiently useable. But I'm sure we can do even better.
Feel free to report feature requests or to provide patches !
<|file_sep|>current/README.md
# HAProxy Documentation Converter
Made to convert the HAProxy documentation into HTML.
More than HTML, the main goal is to provide easy navigation.
## Documentations
A bot periodically fetches last commits for HAProxy 1.4 and 1.5 to produce up-to-date documentations.
Converted documentations are then stored online :
- [HAProxy 1.4 Configuration Manual](http://cbonte.github.com/haproxy-dconv/configuration-1.4.html)
- [HAProxy 1.5 Configuration Manual](http://cbonte.github.com/haproxy-dconv/configuration-1.5.html)
## Contribute
The project now lives by itself, as it is sufficiently useable. But I'm sure we can do even better.
Feel free to report feature requests or to provide patches !
<|file_sep|>updated/README.md
# HAProxy Documentation Converter
Made to convert the HAProxy documentation into HTML.
More than HTML, the main goal is to provide easy navigation.
## Documentations
A bot periodically fetches last commits for HAProxy 1.4 and 1.5 to produce up-to-date documentations.
Converted documentations are then stored online :
- HAProxy 1.4 Configuration Manual [stable](http://cbonte.github.com/haproxy-dconv/configuration-1.4.html) [snapshot](http://cbonte.github.com/haproxy-dconv/snapshot/configuration-1.4.html)
- HAProxy 1.5 Configuration Manual [stable](http://cbonte.github.com/haproxy-dconv/configuration-1.5.html) [snapshot](http://cbonte.github.com/haproxy-dconv/snapshot/configuration-1.5.html)
## Contribute
The project now lives by itself, as it is sufficiently useable. But I'm sure we can do even better.
Feel free to report feature requests or to provide patches !
| 584104a2b67cb2e775987a8321349b8e64341e66 | README.md | README.md | Markdown |
<|file_sep|>original/zeus/utils/revisions.py
) -> Revision:
"""
Attempt to transform a a commit-like reference into a valid revision.
"""
# try to find it from the database first
if len(ref) == 40:
revision = Revision.query.filter(
Revision.repository_id == repository.id, Revision.sha == ref
).first()
if revision:
return revision
if not with_vcs:
raise UnknownRevision
result = next(vcs_client.log(repository.id, parent=ref, limit=1))
revision = Revision.query.filter(
Revision.repository_id == repository.id, Revision.sha == result["sha"]
).first()
if not revision:
raise UnknownRevision
<|file_sep|>current/zeus/utils/revisions.py
) -> Revision:
"""
Attempt to transform a a commit-like reference into a valid revision.
"""
# try to find it from the database first
if len(ref) == 40:
revision = Revision.query.filter(
Revision.repository_id == repository.id, Revision.sha == ref
).first()
if revision:
return revision
if not with_vcs:
raise UnknownRevision
result = next(vcs_client.log(repository.id, parent=ref, limit=1))
revision = Revision.query.filter(
Revision.repository_id == repository.id, Revision.sha == result["sha"]
).first()
if not revision:
raise UnknownRevision
<|file_sep|>updated/zeus/utils/revisions.py | # try to find it from the database first
if len(ref) == 40:
revision = Revision.query.filter(
Revision.repository_id == repository.id, Revision.sha == ref
).first()
if revision:
return revision
if not with_vcs:
raise UnknownRevision
try:
result = vcs_client.log(repository.id, parent=ref, limit=1)[0]
except IndexError:
raise UnknownRevision
revision = Revision.query.filter(
Revision.repository_id == repository.id, Revision.sha == result["sha"]
).first()
if not revision:
raise UnknownRevision | <|file_sep|>original/zeus/utils/revisions.py
) -> Revision:
"""
Attempt to transform a a commit-like reference into a valid revision.
"""
# try to find it from the database first
if len(ref) == 40:
revision = Revision.query.filter(
Revision.repository_id == repository.id, Revision.sha == ref
).first()
if revision:
return revision
if not with_vcs:
raise UnknownRevision
result = next(vcs_client.log(repository.id, parent=ref, limit=1))
revision = Revision.query.filter(
Revision.repository_id == repository.id, Revision.sha == result["sha"]
).first()
if not revision:
raise UnknownRevision
<|file_sep|>current/zeus/utils/revisions.py
) -> Revision:
"""
Attempt to transform a a commit-like reference into a valid revision.
"""
# try to find it from the database first
if len(ref) == 40:
revision = Revision.query.filter(
Revision.repository_id == repository.id, Revision.sha == ref
).first()
if revision:
return revision
if not with_vcs:
raise UnknownRevision
result = next(vcs_client.log(repository.id, parent=ref, limit=1))
revision = Revision.query.filter(
Revision.repository_id == repository.id, Revision.sha == result["sha"]
).first()
if not revision:
raise UnknownRevision
<|file_sep|>updated/zeus/utils/revisions.py
# try to find it from the database first
if len(ref) == 40:
revision = Revision.query.filter(
Revision.repository_id == repository.id, Revision.sha == ref
).first()
if revision:
return revision
if not with_vcs:
raise UnknownRevision
try:
result = vcs_client.log(repository.id, parent=ref, limit=1)[0]
except IndexError:
raise UnknownRevision
revision = Revision.query.filter(
Revision.repository_id == repository.id, Revision.sha == result["sha"]
).first()
if not revision:
raise UnknownRevision | 209fef39f72a625e154f4455eaa6754d6a85e98b | zeus/utils/revisions.py | zeus/utils/revisions.py | Python |
<|file_sep|>original/package.json
{
"name": "time-logger",
"version": "1.0.0",
"description": "",
"main": "index.js",
"scripts": {
"test": "echo \"Error: no test specified\" && exit 1"
},
"repository": {
"type": "git",
"url": "git+https://github.com/tabalinas/time-logger.git"
},
"author": "Artem Tabalin <tabalinas@gmail.com> (http://tabalin.net/)",
"license": "MIT",
"bugs": {
"url": "https://github.com/tabalinas/time-logger/issues"
},
"homepage": "https://github.com/tabalinas/time-logger#readme"
}
<|file_sep|>current/package.json
{
"name": "time-logger",
"version": "1.0.0",
"description": "",
"main": "index.js",
"scripts": {
"test": "echo \"Error: no test specified\" && exit 1"
},
"repository": {
"type": "git",
"url": "git+https://github.com/tabalinas/time-logger.git"
},
"author": "Artem Tabalin <tabalinas@gmail.com> (http://tabalin.net/)",
"license": "MIT",
"bugs": {
"url": "https://github.com/tabalinas/time-logger/issues"
},
"homepage": "https://github.com/tabalinas/time-logger#readme"
}
<|file_sep|>updated/package.json | "description": "",
"main": "index.js",
"scripts": {
"test": "echo \"Error: no test specified\" && exit 1"
},
"repository": {
"type": "git",
"url": "git+https://github.com/tabalinas/time-logger.git"
},
"author": "Artem Tabalin <tabalinas@gmail.com> (http://tabalin.net/)",
"license": "MIT",
"bugs": {
"url": "https://github.com/tabalinas/time-logger/issues"
},
"homepage": "https://github.com/tabalinas/time-logger#readme",
"dependencies": {
"babel-standalone": "^6.21.1-0",
"react": "^15.4.1",
"react-dom": "^15.4.1"
}
} | <|file_sep|>original/package.json
{
"name": "time-logger",
"version": "1.0.0",
"description": "",
"main": "index.js",
"scripts": {
"test": "echo \"Error: no test specified\" && exit 1"
},
"repository": {
"type": "git",
"url": "git+https://github.com/tabalinas/time-logger.git"
},
"author": "Artem Tabalin <tabalinas@gmail.com> (http://tabalin.net/)",
"license": "MIT",
"bugs": {
"url": "https://github.com/tabalinas/time-logger/issues"
},
"homepage": "https://github.com/tabalinas/time-logger#readme"
}
<|file_sep|>current/package.json
{
"name": "time-logger",
"version": "1.0.0",
"description": "",
"main": "index.js",
"scripts": {
"test": "echo \"Error: no test specified\" && exit 1"
},
"repository": {
"type": "git",
"url": "git+https://github.com/tabalinas/time-logger.git"
},
"author": "Artem Tabalin <tabalinas@gmail.com> (http://tabalin.net/)",
"license": "MIT",
"bugs": {
"url": "https://github.com/tabalinas/time-logger/issues"
},
"homepage": "https://github.com/tabalinas/time-logger#readme"
}
<|file_sep|>updated/package.json
"description": "",
"main": "index.js",
"scripts": {
"test": "echo \"Error: no test specified\" && exit 1"
},
"repository": {
"type": "git",
"url": "git+https://github.com/tabalinas/time-logger.git"
},
"author": "Artem Tabalin <tabalinas@gmail.com> (http://tabalin.net/)",
"license": "MIT",
"bugs": {
"url": "https://github.com/tabalinas/time-logger/issues"
},
"homepage": "https://github.com/tabalinas/time-logger#readme",
"dependencies": {
"babel-standalone": "^6.21.1-0",
"react": "^15.4.1",
"react-dom": "^15.4.1"
}
} | 3a99f1c29284f0dd05db477080a117fc9699ce84 | package.json | package.json | JSON |
<|file_sep|>original/Sample-App/scripts/tsconfig.json
<|file_sep|>current/Sample-App/scripts/tsconfig.json
<|file_sep|>updated/Sample-App/scripts/tsconfig.json | {
"compilerOptions": {
"target": "es5",
"out": "www/scripts/appBundle.js",
"sourceMap": true,
"removeComments": true,
"sourceRoot": "/"
}
} | <|file_sep|>original/Sample-App/scripts/tsconfig.json
<|file_sep|>current/Sample-App/scripts/tsconfig.json
<|file_sep|>updated/Sample-App/scripts/tsconfig.json
{
"compilerOptions": {
"target": "es5",
"out": "www/scripts/appBundle.js",
"sourceMap": true,
"removeComments": true,
"sourceRoot": "/"
}
} | e612feecb82179877ce5abb959e42498eb138806 | Sample-App/scripts/tsconfig.json | Sample-App/scripts/tsconfig.json | JSON |
<|file_sep|>docs/4.0/extend/icons.md.diff
original:
updated:
<|file_sep|>docs/4.0/extend/icons.md.diff
original:
- [Entypo](http://www.entypo.com/)
updated:
<|file_sep|>original/docs/4.0/extend/icons.md
---
Bootstrap doesn't include an icon library by default, but we have a handful of recommendations for you to choose from. While most icon sets include multiple file formats, we prefer SVG implementations for their improved accessibility and vector support.
## Preferred
We've tested and used these icon sets ourselves.
- [Iconic](https://useiconic.com/open/)
- [Octicons](https://octicons.github.com/)
- [Entypo](http://www.entypo.com/)
## More options
While we haven't tried these out, they do look promising and provide multiple formats—including SVG.
- [Bytesize](https://github.com/danklammer/bytesize-icons)
- [Google Material icons](https://material.io/icons/)
- [Ionicons](http://ionicons.com/)
- [Feather](https://feathericons.com/)
- [Dripicons](http://demo.amitjakhu.com/dripicons/)
- [Ikons](http://ikons.piotrkwiatkowski.co.uk/)
- [Glyph](http://glyph.smarticons.co/)
<|file_sep|>current/docs/4.0/extend/icons.md
---
Bootstrap doesn't include an icon library by default, but we have a handful of recommendations for you to choose from. While most icon sets include multiple file formats, we prefer SVG implementations for their improved accessibility and vector support.
## Preferred
We've tested and used these icon sets ourselves.
- [Iconic](https://useiconic.com/open/)
- [Octicons](https://octicons.github.com/)
## More options
While we haven't tried these out, they do look promising and provide multiple formats—including SVG.
- [Bytesize](https://github.com/danklammer/bytesize-icons)
- [Google Material icons](https://material.io/icons/)
- [Ionicons](http://ionicons.com/)
- [Feather](https://feathericons.com/)
- [Dripicons](http://demo.amitjakhu.com/dripicons/)
- [Ikons](http://ikons.piotrkwiatkowski.co.uk/)
- [Glyph](http://glyph.smarticons.co/)
<|file_sep|>updated/docs/4.0/extend/icons.md |
Bootstrap doesn't include an icon library by default, but we have a handful of recommendations for you to choose from. While most icon sets include multiple file formats, we prefer SVG implementations for their improved accessibility and vector support.
## Preferred
We've tested and used these icon sets ourselves.
- [Iconic](https://useiconic.com/open/)
- [Octicons](https://octicons.github.com/)
## More options
While we haven't tried these out, they do look promising and provide multiple formats—including SVG.
- [Bytesize](https://github.com/danklammer/bytesize-icons)
- [Google Material icons](https://material.io/icons/)
- [Ionicons](http://ionicons.com/)
- [Feather](https://feathericons.com/)
- [Dripicons](http://demo.amitjakhu.com/dripicons/)
- [Ikons](http://ikons.piotrkwiatkowski.co.uk/)
- [Glyph](http://glyph.smarticons.co/) | <|file_sep|>docs/4.0/extend/icons.md.diff
original:
updated:
<|file_sep|>docs/4.0/extend/icons.md.diff
original:
- [Entypo](http://www.entypo.com/)
updated:
<|file_sep|>original/docs/4.0/extend/icons.md
---
Bootstrap doesn't include an icon library by default, but we have a handful of recommendations for you to choose from. While most icon sets include multiple file formats, we prefer SVG implementations for their improved accessibility and vector support.
## Preferred
We've tested and used these icon sets ourselves.
- [Iconic](https://useiconic.com/open/)
- [Octicons](https://octicons.github.com/)
- [Entypo](http://www.entypo.com/)
## More options
While we haven't tried these out, they do look promising and provide multiple formats—including SVG.
- [Bytesize](https://github.com/danklammer/bytesize-icons)
- [Google Material icons](https://material.io/icons/)
- [Ionicons](http://ionicons.com/)
- [Feather](https://feathericons.com/)
- [Dripicons](http://demo.amitjakhu.com/dripicons/)
- [Ikons](http://ikons.piotrkwiatkowski.co.uk/)
- [Glyph](http://glyph.smarticons.co/)
<|file_sep|>current/docs/4.0/extend/icons.md
---
Bootstrap doesn't include an icon library by default, but we have a handful of recommendations for you to choose from. While most icon sets include multiple file formats, we prefer SVG implementations for their improved accessibility and vector support.
## Preferred
We've tested and used these icon sets ourselves.
- [Iconic](https://useiconic.com/open/)
- [Octicons](https://octicons.github.com/)
## More options
While we haven't tried these out, they do look promising and provide multiple formats—including SVG.
- [Bytesize](https://github.com/danklammer/bytesize-icons)
- [Google Material icons](https://material.io/icons/)
- [Ionicons](http://ionicons.com/)
- [Feather](https://feathericons.com/)
- [Dripicons](http://demo.amitjakhu.com/dripicons/)
- [Ikons](http://ikons.piotrkwiatkowski.co.uk/)
- [Glyph](http://glyph.smarticons.co/)
<|file_sep|>updated/docs/4.0/extend/icons.md
Bootstrap doesn't include an icon library by default, but we have a handful of recommendations for you to choose from. While most icon sets include multiple file formats, we prefer SVG implementations for their improved accessibility and vector support.
## Preferred
We've tested and used these icon sets ourselves.
- [Iconic](https://useiconic.com/open/)
- [Octicons](https://octicons.github.com/)
## More options
While we haven't tried these out, they do look promising and provide multiple formats—including SVG.
- [Bytesize](https://github.com/danklammer/bytesize-icons)
- [Google Material icons](https://material.io/icons/)
- [Ionicons](http://ionicons.com/)
- [Feather](https://feathericons.com/)
- [Dripicons](http://demo.amitjakhu.com/dripicons/)
- [Ikons](http://ikons.piotrkwiatkowski.co.uk/)
- [Glyph](http://glyph.smarticons.co/) | dd041e7bd5760c74b23a9ca1e9a76dc3de1d7e24 | docs/4.0/extend/icons.md | docs/4.0/extend/icons.md | Markdown |
<|file_sep|>original/src/CMakeLists.txt
#
# Building libreset
#
#
# We use the var SOURCE_FILES to hold all source files to be linked together
# into libreset. Additionally, checks checky may be performed on that source
# files.
#
set(SOURCE_FILES
libreset/ll.c
)
#
# Libreset will be a shared object
#
add_library(reset SHARED ${SOURCE_FILES})
<|file_sep|>current/src/CMakeLists.txt
#
# Building libreset
#
#
# We use the var SOURCE_FILES to hold all source files to be linked together
# into libreset. Additionally, checks checky may be performed on that source
# files.
#
set(SOURCE_FILES
libreset/ll.c
)
#
# Libreset will be a shared object
#
add_library(reset SHARED ${SOURCE_FILES})
<|file_sep|>updated/src/CMakeLists.txt |
#
# We use the var SOURCE_FILES to hold all source files to be linked together
# into libreset. Additionally, checks checky may be performed on that source
# files.
#
set(SOURCE_FILES
libreset/ll.c
)
#
# General compile options
#
add_compile_options(-Wall -Wextra -Wunused -Wformat -Wno-unused-parameter -pedantic)
#
# Libreset will be a shared object
#
add_library(reset SHARED ${SOURCE_FILES})
| <|file_sep|>original/src/CMakeLists.txt
#
# Building libreset
#
#
# We use the var SOURCE_FILES to hold all source files to be linked together
# into libreset. Additionally, checks checky may be performed on that source
# files.
#
set(SOURCE_FILES
libreset/ll.c
)
#
# Libreset will be a shared object
#
add_library(reset SHARED ${SOURCE_FILES})
<|file_sep|>current/src/CMakeLists.txt
#
# Building libreset
#
#
# We use the var SOURCE_FILES to hold all source files to be linked together
# into libreset. Additionally, checks checky may be performed on that source
# files.
#
set(SOURCE_FILES
libreset/ll.c
)
#
# Libreset will be a shared object
#
add_library(reset SHARED ${SOURCE_FILES})
<|file_sep|>updated/src/CMakeLists.txt
#
# We use the var SOURCE_FILES to hold all source files to be linked together
# into libreset. Additionally, checks checky may be performed on that source
# files.
#
set(SOURCE_FILES
libreset/ll.c
)
#
# General compile options
#
add_compile_options(-Wall -Wextra -Wunused -Wformat -Wno-unused-parameter -pedantic)
#
# Libreset will be a shared object
#
add_library(reset SHARED ${SOURCE_FILES})
| 2ddd0aa55f44a25dfe08bc3dc6b01e9d8514520c | src/CMakeLists.txt | src/CMakeLists.txt | Text |
<|file_sep|>original/coverage.bash
#!/bin/bash
base_package=github.com/mitsuse/bullet
base_path=${GOPATH}/src/${base_package}
package_list=(
${base_package}
)
if [ ! -d ${base_path}/coverprofile ]
then
mkdir ${base_path}/coverprofile
else
rm ${base_path}/coverprofile/*.coverprofile
fi
for package in ${package_list[@]}
do
cover_name=$(echo ${package} | sed -e "s/\//__/g").coverprofile
cover_path=${base_path}/coverprofile/${cover_name}
go test -covermode=count -coverprofile ${cover_path} ${package}
<|file_sep|>current/coverage.bash
#!/bin/bash
base_package=github.com/mitsuse/bullet
base_path=${GOPATH}/src/${base_package}
package_list=(
${base_package}
)
if [ ! -d ${base_path}/coverprofile ]
then
mkdir ${base_path}/coverprofile
else
rm ${base_path}/coverprofile/*.coverprofile
fi
for package in ${package_list[@]}
do
cover_name=$(echo ${package} | sed -e "s/\//__/g").coverprofile
cover_path=${base_path}/coverprofile/${cover_name}
go test -covermode=count -coverprofile ${cover_path} ${package}
<|file_sep|>updated/coverage.bash | #!/bin/bash
base_package=github.com/mitsuse/bullet
base_path=${GOPATH}/src/${base_package}
package_list=(
${base_package}
${base_package}/pushbullet
${base_package}/pushbullet/pushes
)
if [ ! -d ${base_path}/coverprofile ]
then
mkdir ${base_path}/coverprofile
else
rm ${base_path}/coverprofile/*.coverprofile
fi
for package in ${package_list[@]}
do
cover_name=$(echo ${package} | sed -e "s/\//__/g").coverprofile | <|file_sep|>original/coverage.bash
#!/bin/bash
base_package=github.com/mitsuse/bullet
base_path=${GOPATH}/src/${base_package}
package_list=(
${base_package}
)
if [ ! -d ${base_path}/coverprofile ]
then
mkdir ${base_path}/coverprofile
else
rm ${base_path}/coverprofile/*.coverprofile
fi
for package in ${package_list[@]}
do
cover_name=$(echo ${package} | sed -e "s/\//__/g").coverprofile
cover_path=${base_path}/coverprofile/${cover_name}
go test -covermode=count -coverprofile ${cover_path} ${package}
<|file_sep|>current/coverage.bash
#!/bin/bash
base_package=github.com/mitsuse/bullet
base_path=${GOPATH}/src/${base_package}
package_list=(
${base_package}
)
if [ ! -d ${base_path}/coverprofile ]
then
mkdir ${base_path}/coverprofile
else
rm ${base_path}/coverprofile/*.coverprofile
fi
for package in ${package_list[@]}
do
cover_name=$(echo ${package} | sed -e "s/\//__/g").coverprofile
cover_path=${base_path}/coverprofile/${cover_name}
go test -covermode=count -coverprofile ${cover_path} ${package}
<|file_sep|>updated/coverage.bash
#!/bin/bash
base_package=github.com/mitsuse/bullet
base_path=${GOPATH}/src/${base_package}
package_list=(
${base_package}
${base_package}/pushbullet
${base_package}/pushbullet/pushes
)
if [ ! -d ${base_path}/coverprofile ]
then
mkdir ${base_path}/coverprofile
else
rm ${base_path}/coverprofile/*.coverprofile
fi
for package in ${package_list[@]}
do
cover_name=$(echo ${package} | sed -e "s/\//__/g").coverprofile | 8ae118a96e9c5b009e92bb694dc914108636edf5 | coverage.bash | coverage.bash | Shell |
<|file_sep|>resources/views/admin/modalities_index.blade.php.diff
original:
updated:
<p>Click a modality to edit</p>
<|file_sep|>resources/views/admin/modalities_index.blade.php.diff
original:
<td>{{ $modality->modality }}</td>
<form action="/admin/modalities/{{ $modality->id }}/edit" method="POST">
<button type="submit">Edit</button>
</form>
updated:
<td><a href="/admin/modalities/{{ $modality->id }}/edit">{{ $modality->modality }}</a></td>
<|file_sep|>original/resources/views/admin/modalities_index.blade.php
<tbody>
@foreach ($modalities->chunk(2) as $chunk )
<tr>
@foreach ($chunk as $modality)
<td>{{ $modality->id }}</td>
<td>{{ $modality->modality }}</td>
<form action="/admin/modalities/{{ $modality->id }}/edit" method="POST">
<button type="submit">Edit</button>
</form>
@endforeach
</tr>
@endforeach
</tbody>
</table>
<h2>Add a Modality</h2>
<!-- Add a new modality -->
<form action="/admin/modalities" method="POST">
{{ csrf_field() }}
New Modality: <input type="TEXT" name="modality" size="25" />
<button type="SUBMIT">Add modality</button> / <a href="/">Main</a>
<|file_sep|>current/resources/views/admin/modalities_index.blade.php
<tbody>
@foreach ($modalities->chunk(2) as $chunk )
<tr>
@foreach ($chunk as $modality)
<td>{{ $modality->id }}</td>
<td><a href="/admin/modalities/{{ $modality->id }}/edit">{{ $modality->modality }}</a></td>
@endforeach
</tr>
@endforeach
</tbody>
</table>
<h2>Add a Modality</h2>
<!-- Add a new modality -->
<form action="/admin/modalities" method="POST">
{{ csrf_field() }}
New Modality: <input type="TEXT" name="modality" size="25" />
<button type="SUBMIT">Add modality</button> / <a href="/">Main</a>
</form>
@endsection
<|file_sep|>updated/resources/views/admin/modalities_index.blade.php | <tr>
@foreach ($chunk as $modality)
<td>{{ $modality->id }}</td>
<td><a href="/admin/modalities/{{ $modality->id }}/edit">{{ $modality->modality }}</a></td>
@endforeach
</tr>
@endforeach
</tbody>
</table>
<h2>Add a Modality</h2>
<!-- Add a new modality -->
<form class="form-inline" action="/admin/modalities" method="POST">
<div class="form-group">
{{ csrf_field() }}
<label for="modality">New Modality:</label> <input type="TEXT" class="form-control" id="modality" name="modality" size="25" />
<button class="btn btn-default" type="SUBMIT">Add modality</button> / <a href="/">Main</a>
</div>
</form>
@endsection | <|file_sep|>resources/views/admin/modalities_index.blade.php.diff
original:
updated:
<p>Click a modality to edit</p>
<|file_sep|>resources/views/admin/modalities_index.blade.php.diff
original:
<td>{{ $modality->modality }}</td>
<form action="/admin/modalities/{{ $modality->id }}/edit" method="POST">
<button type="submit">Edit</button>
</form>
updated:
<td><a href="/admin/modalities/{{ $modality->id }}/edit">{{ $modality->modality }}</a></td>
<|file_sep|>original/resources/views/admin/modalities_index.blade.php
<tbody>
@foreach ($modalities->chunk(2) as $chunk )
<tr>
@foreach ($chunk as $modality)
<td>{{ $modality->id }}</td>
<td>{{ $modality->modality }}</td>
<form action="/admin/modalities/{{ $modality->id }}/edit" method="POST">
<button type="submit">Edit</button>
</form>
@endforeach
</tr>
@endforeach
</tbody>
</table>
<h2>Add a Modality</h2>
<!-- Add a new modality -->
<form action="/admin/modalities" method="POST">
{{ csrf_field() }}
New Modality: <input type="TEXT" name="modality" size="25" />
<button type="SUBMIT">Add modality</button> / <a href="/">Main</a>
<|file_sep|>current/resources/views/admin/modalities_index.blade.php
<tbody>
@foreach ($modalities->chunk(2) as $chunk )
<tr>
@foreach ($chunk as $modality)
<td>{{ $modality->id }}</td>
<td><a href="/admin/modalities/{{ $modality->id }}/edit">{{ $modality->modality }}</a></td>
@endforeach
</tr>
@endforeach
</tbody>
</table>
<h2>Add a Modality</h2>
<!-- Add a new modality -->
<form action="/admin/modalities" method="POST">
{{ csrf_field() }}
New Modality: <input type="TEXT" name="modality" size="25" />
<button type="SUBMIT">Add modality</button> / <a href="/">Main</a>
</form>
@endsection
<|file_sep|>updated/resources/views/admin/modalities_index.blade.php
<tr>
@foreach ($chunk as $modality)
<td>{{ $modality->id }}</td>
<td><a href="/admin/modalities/{{ $modality->id }}/edit">{{ $modality->modality }}</a></td>
@endforeach
</tr>
@endforeach
</tbody>
</table>
<h2>Add a Modality</h2>
<!-- Add a new modality -->
<form class="form-inline" action="/admin/modalities" method="POST">
<div class="form-group">
{{ csrf_field() }}
<label for="modality">New Modality:</label> <input type="TEXT" class="form-control" id="modality" name="modality" size="25" />
<button class="btn btn-default" type="SUBMIT">Add modality</button> / <a href="/">Main</a>
</div>
</form>
@endsection | 664be3667730e7e723fb4d50844da530640e66a3 | resources/views/admin/modalities_index.blade.php | resources/views/admin/modalities_index.blade.php | PHP |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.