text stringlengths 93 16.4k | id stringlengths 20 40 | metadata dict | input_ids listlengths 45 2.05k | attention_mask listlengths 45 2.05k | complexity int64 1 9 |
|---|---|---|---|---|---|
func TestLinkedInProviderOverrides(t *testing.T) {
p := NewLinkedInProvider(
&ProviderData{
LoginURL: &url.URL{
Scheme: "https",
Host: "example.com",
Path: "/oauth/auth"},
RedeemURL: &url.URL{
Scheme: "https",
Host: "example.com",
Path: "/oauth/token"},
ProfileURL: &url.URL{
Scheme: "https",
Host: "example.com",
Path: "/oauth/profile"},
ValidateURL: &url.URL{
Scheme: "https",
Host: "example.com",
Path: "/oauth/tokeninfo"},
Scope: "profile"})
assert.NotEqual(t, nil, p)
assert.Equal(t, "LinkedIn", p.Data().ProviderName)
assert.Equal(t, "https://example.com/oauth/auth",
p.Data().LoginURL.String())
assert.Equal(t, "https://example.com/oauth/token",
p.Data().RedeemURL.String())
assert.Equal(t, "https://example.com/oauth/profile",
p.Data().ProfileURL.String())
assert.Equal(t, "https://example.com/oauth/tokeninfo",
p.Data().ValidateURL.String())
assert.Equal(t, "profile", p.Data().Scope)
} | explode_data.jsonl/39461 | {
"file_path": "/home/dung/Study/Code/Cross_test_gen/training_dataset/dedup_data/clean_data_go/data/explode_data.jsonl",
"token_count": 439
} | [
2830,
3393,
75203,
5179,
80010,
1155,
353,
8840,
836,
8,
341,
3223,
1669,
1532,
75203,
5179,
1006,
197,
197,
5,
5179,
1043,
515,
298,
197,
6231,
3144,
25,
609,
1085,
20893,
515,
571,
7568,
8058,
25,
330,
2428,
756,
571,
197,
9296,
2... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | 1 |
func TestDLL(t *testing.T) {
tests := []struct {
name string
source string
expected int
}{
{
name: "none",
expected: 0,
source: `
package main
func main() {
println("Hello!")
}
`,
},
{
name: "for",
expected: 1,
source: `
package main
func main() {
for i := 0; i < 5; i++ {
defer println("defer")
}
}
`,
},
{
name: "range",
expected: 1,
source: `
package main
func main() {
list := []int{1, 2, 3, 4, 5, 6, 7}
for _, x := range list {
defer println(x)
}
}
`,
},
{
name: "nested",
expected: 1,
source: `
package main
func main() {
list := []int{1, 2, 3, 4, 5, 6, 7}
for _, i := range list {
for j := 0; j < i; j++ {
defer println(j)
}
}
}
`,
},
{
name: "block",
expected: 1,
source: `
package main
func main() {
for i := 0; i < 5; i++ {
{
defer println("defer")
}
}
}
`,
},
{
name: "if",
expected: 1,
source: `
package main
func main() {
for i := 0; i < 5; i++ {
if true {
defer println("defer")
}
}
}
`,
},
{
name: "funclit",
expected: 0,
source: `
package main
func main() {
for i := 0; i < 5; i++ {
func() {
defer println("defer")
}()
}
}
`,
},
}
for _, tt := range tests {
t.Run(tt.name, func(t *testing.T) {
reports, err := gather(tt.source, false)
if err != nil {
t.Fatalf("unexpected error: %v", err)
}
if len(reports) != tt.expected {
t.Fatalf("expected %d reports, got %d", tt.expected, len(reports))
}
})
}
} | explode_data.jsonl/503 | {
"file_path": "/home/dung/Study/Code/Cross_test_gen/training_dataset/dedup_data/clean_data_go/data/explode_data.jsonl",
"token_count": 996
} | [
2830,
3393,
64187,
1155,
353,
8840,
836,
8,
341,
78216,
1669,
3056,
1235,
341,
197,
11609,
257,
914,
198,
197,
47418,
256,
914,
198,
197,
42400,
526,
198,
197,
59403,
197,
197,
515,
298,
11609,
25,
257,
330,
6697,
756,
298,
42400,
2... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | 3 |
func TestGetConfig(t *testing.T) {
mockDB, mock, err := sqlmock.New()
if err != nil {
t.Fatalf("an error '%s' was not expected when opening a stub database connection", err)
}
defer mockDB.Close()
db := sqlx.NewDb(mockDB, "sqlmock")
defer db.Close()
config := map[string]interface{}{
"name0": "val0",
"name1": "val1",
}
rows := sqlmock.NewRows([]string{"name", "value"})
for name, val := range config {
rows = rows.AddRow(name, val)
}
mock.ExpectQuery("SELECT").WillReturnRows(rows)
sqlConfig, err := getConfig(db)
if err != nil {
t.Errorf("getProfiles expected: nil error, actual: %v", err)
}
if !reflect.DeepEqual(config, sqlConfig) {
t.Errorf("getConfig expected: %+v actual: %+v", config, sqlConfig)
}
if err := mock.ExpectationsWereMet(); err != nil {
t.Errorf("there were unfulfilled expections: %s", err)
}
} | explode_data.jsonl/31539 | {
"file_path": "/home/dung/Study/Code/Cross_test_gen/training_dataset/dedup_data/clean_data_go/data/explode_data.jsonl",
"token_count": 335
} | [
2830,
3393,
1949,
2648,
1155,
353,
8840,
836,
8,
341,
77333,
3506,
11,
7860,
11,
1848,
1669,
5704,
16712,
7121,
741,
743,
1848,
961,
2092,
341,
197,
3244,
30762,
445,
276,
1465,
7677,
82,
6,
572,
537,
3601,
979,
8568,
264,
13633,
46... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | 6 |
func TestUserStoreGetByEmail(t *testing.T) {
Setup()
teamid := model.NewId()
u1 := &model.User{}
u1.Email = model.NewId()
Must(store.User().Save(u1))
Must(store.Team().SaveMember(&model.TeamMember{TeamId: teamid, UserId: u1.Id}))
if err := (<-store.User().GetByEmail(u1.Email)).Err; err != nil {
t.Fatal(err)
}
if err := (<-store.User().GetByEmail("")).Err; err == nil {
t.Fatal("Should have failed because of missing email")
}
} | explode_data.jsonl/5098 | {
"file_path": "/home/dung/Study/Code/Cross_test_gen/training_dataset/dedup_data/clean_data_go/data/explode_data.jsonl",
"token_count": 181
} | [
2830,
3393,
1474,
6093,
1949,
87197,
1155,
353,
8840,
836,
8,
341,
197,
21821,
2822,
197,
9196,
307,
1669,
1614,
7121,
764,
2822,
10676,
16,
1669,
609,
2528,
7344,
16094,
10676,
16,
24066,
284,
1614,
7121,
764,
741,
9209,
590,
31200,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | 3 |
func TestExtractErrorCounts(t *testing.T) {
require.Equal(
t,
[]statusErrorCount{
statusErrorCount{"tank", "UNAVAIL", 0, 0, 0},
statusErrorCount{"c1t0d0", "ONLINE", 0, 0, 0},
statusErrorCount{"c1t1d0", "UNAVAIL", 4, 1, 0},
},
extractErrorCounts(sampleStatusErrorOutput),
)
} | explode_data.jsonl/18993 | {
"file_path": "/home/dung/Study/Code/Cross_test_gen/training_dataset/dedup_data/clean_data_go/data/explode_data.jsonl",
"token_count": 132
} | [
2830,
3393,
28959,
1454,
63731,
1155,
353,
8840,
836,
8,
341,
17957,
12808,
1006,
197,
3244,
345,
197,
197,
1294,
2829,
1454,
2507,
515,
298,
23847,
1454,
2507,
4913,
85171,
497,
330,
1861,
8093,
5965,
497,
220,
15,
11,
220,
15,
11,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | 1 |
func TestValidateServerHooks_ValidInput(t *testing.T) {
var tests = []struct {
name string
hooksInput map[string][]models.Hook
}{
{"One hook, one script",
map[string][]models.Hook{"ApplicationStop": {
{
Location: "script-location",
Timeout: "10",
Runas: "user-name",
},
}},
},
{"One hook, multiple scripts",
map[string][]models.Hook{"ApplicationStop": {
{
Location: "script-location",
Timeout: "10",
Runas: "user-name",
},
{
Location: "script-location",
Timeout: "10",
Runas: "user-name",
},
}},
},
{"Multiple hooks, multiple scripts",
map[string][]models.Hook{"ApplicationStop": {
{
Location: "script-location",
Timeout: "10",
Runas: "user-name",
},
{
Location: "script-location",
Timeout: "10",
Runas: "user-name",
},
},
"BeforeInstall": {
{
Location: "script-location",
Timeout: "10",
Runas: "user-name",
},
}},
},
}
for _, test := range tests {
output := validateServerHooks(test.hooksInput)
if output != true {
t.Errorf("The validateServerHooks function failed for: %v", test)
}
}
} | explode_data.jsonl/71210 | {
"file_path": "/home/dung/Study/Code/Cross_test_gen/training_dataset/dedup_data/clean_data_go/data/explode_data.jsonl",
"token_count": 583
} | [
2830,
3393,
17926,
5475,
67769,
97279,
2505,
1155,
353,
8840,
836,
8,
341,
2405,
7032,
284,
3056,
1235,
341,
197,
11609,
981,
914,
198,
197,
9598,
14685,
2505,
2415,
14032,
45725,
6507,
3839,
1941,
198,
197,
59403,
197,
197,
4913,
3966,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | 3 |
func TestParseImageStreamImageName(t *testing.T) {
tests := map[string]struct {
input string
expectedRepo string
expectedId string
expectError bool
}{
"empty string": {
input: "",
expectError: true,
},
"one part": {
input: "a",
expectError: true,
},
"more than 2 parts": {
input: "a@b@c",
expectError: true,
},
"empty name part": {
input: "@id",
expectError: true,
},
"empty id part": {
input: "name@",
expectError: true,
},
"valid input": {
input: "repo@id",
expectedRepo: "repo",
expectedId: "id",
expectError: false,
},
}
for name, test := range tests {
repo, id, err := ParseImageStreamImageName(test.input)
didError := err != nil
if e, a := test.expectError, didError; e != a {
t.Errorf("%s: expected error=%t, got=%t: %s", name, e, a, err)
continue
}
if test.expectError {
continue
}
if e, a := test.expectedRepo, repo; e != a {
t.Errorf("%s: repo: expected %q, got %q", name, e, a)
continue
}
if e, a := test.expectedId, id; e != a {
t.Errorf("%s: id: expected %q, got %q", name, e, a)
continue
}
}
} | explode_data.jsonl/40823 | {
"file_path": "/home/dung/Study/Code/Cross_test_gen/training_dataset/dedup_data/clean_data_go/data/explode_data.jsonl",
"token_count": 567
} | [
2830,
3393,
14463,
1906,
3027,
1906,
675,
1155,
353,
8840,
836,
8,
341,
78216,
1669,
2415,
14032,
60,
1235,
341,
197,
22427,
286,
914,
198,
197,
42400,
25243,
914,
198,
197,
42400,
764,
256,
914,
198,
197,
24952,
1454,
220,
1807,
198,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | 6 |
func TestDefragOutputExhaustion(t *testing.T) {
if testing.Short() || !build.VLONG {
t.SkipNow()
}
t.Parallel()
wt, err := createWalletTester(t.Name(), modules.ProdDependencies)
if err != nil {
t.Fatal(err)
}
defer func() {
if err := wt.closeWt(); err != nil {
t.Fatal(err)
}
}()
wt.wallet.mu.Lock()
var dest types.UnlockHash
for k := range wt.wallet.keys {
dest = k
break
}
wt.wallet.mu.Unlock()
_, err = wt.miner.AddBlock()
if err != nil {
t.Fatal(err)
}
// concurrently make a bunch of transactions with lots of outputs to keep the
// defragger running
closechan := make(chan struct{})
donechan := make(chan struct{})
go func() {
defer close(donechan)
for {
select {
case <-closechan:
return
case <-time.After(time.Millisecond * 100):
_, err := wt.miner.AddBlock()
if err != nil {
t.Error(err)
return
}
txnValue := types.SiacoinPrecision.Mul64(3000)
fee := types.SiacoinPrecision.Mul64(10)
numOutputs := defragThreshold + 1
tbuilder, err := wt.wallet.StartTransaction()
if err != nil {
t.Error(err)
return
}
tbuilder.FundSiacoins(txnValue.Mul64(uint64(numOutputs)).Add(fee))
for i := 0; i < numOutputs; i++ {
tbuilder.AddSiacoinOutput(types.SiacoinOutput{
Value: txnValue,
UnlockHash: dest,
})
}
tbuilder.AddMinerFee(fee)
txns, err := tbuilder.Sign(true)
if err != nil {
t.Error("Error signing fragmenting transaction:", err)
}
err = wt.tpool.AcceptTransactionSet(txns)
if err != nil {
t.Error("Error accepting fragmenting transaction:", err)
}
_, err = wt.miner.AddBlock()
if err != nil {
t.Error(err)
return
}
}
}
}()
time.Sleep(time.Second * 1)
// ensure we can still send transactions while receiving aggressively
// fragmented outputs
for i := 0; i < 30; i++ {
sendAmount := types.SiacoinPrecision.Mul64(2000)
_, err = wt.wallet.SendSiacoins(sendAmount, types.UnlockHash{})
if err != nil {
t.Errorf("%v: %v", i, err)
}
time.Sleep(time.Millisecond * 50)
}
close(closechan)
<-donechan
} | explode_data.jsonl/64329 | {
"file_path": "/home/dung/Study/Code/Cross_test_gen/training_dataset/dedup_data/clean_data_go/data/explode_data.jsonl",
"token_count": 957
} | [
2830,
3393,
2620,
4101,
5097,
840,
15074,
290,
1155,
353,
8840,
836,
8,
341,
743,
7497,
55958,
368,
1369,
753,
5834,
5058,
51306,
341,
197,
3244,
57776,
7039,
741,
197,
532,
3244,
41288,
7957,
741,
6692,
83,
11,
1848,
1669,
1855,
3825... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | 2 |
func Test_connectionDetails(t *testing.T) {
type args struct {
kube client.Client
connDetails []v1beta1.ConnectionDetail
relName string
relNamespace string
}
type want struct {
out managed.ConnectionDetails
err error
}
cases := map[string]struct {
args
want
}{
"Fail_NotPartOfRelease": {
args: args{
kube: &test.MockClient{
MockGet: func(ctx context.Context, key client.ObjectKey, obj runtime.Object) error {
if o, ok := obj.(*unstructured.Unstructured); o.GetKind() == "Secret" && ok && key.Name == testSecretName && key.Namespace == testNamespace {
*obj.(*unstructured.Unstructured) = unstructured.Unstructured{
Object: map[string]interface{}{
"data": map[string]interface{}{
"db-password": "MTIzNDU=",
},
},
}
}
return nil
},
},
connDetails: []v1beta1.ConnectionDetail{
{
ObjectReference: corev1.ObjectReference{
Kind: "Secret",
Namespace: testNamespace,
Name: testSecretName,
APIVersion: "v1",
FieldPath: "data.db-password",
},
ToConnectionSecretKey: "password",
},
},
relName: testReleaseName,
relNamespace: testNamespace,
},
want: want{
out: managed.ConnectionDetails{},
err: errors.Errorf(errObjectNotPartOfRelease, corev1.ObjectReference{
Kind: "Secret",
Namespace: testNamespace,
Name: testSecretName,
APIVersion: "v1",
FieldPath: "data.db-password",
}),
},
},
"Success_PartOfRelease": {
args: args{
kube: &test.MockClient{
MockGet: func(ctx context.Context, key client.ObjectKey, obj runtime.Object) error {
if o, ok := obj.(*unstructured.Unstructured); o.GetKind() == "Secret" && ok && key.Name == testSecretName && key.Namespace == testNamespace {
*obj.(*unstructured.Unstructured) = unstructured.Unstructured{
Object: map[string]interface{}{
"metadata": map[string]interface{}{
"annotations": map[string]interface{}{
helmReleaseNameAnnotation: testReleaseName,
helmReleaseNamespaceAnnotation: testNamespace,
},
},
"data": map[string]interface{}{
"db-password": "MTIzNDU=",
},
},
}
}
return nil
},
},
connDetails: []v1beta1.ConnectionDetail{
{
ObjectReference: corev1.ObjectReference{
Kind: "Secret",
Namespace: testNamespace,
Name: testSecretName,
APIVersion: "v1",
FieldPath: "data.db-password",
},
ToConnectionSecretKey: "password",
},
},
relName: testReleaseName,
relNamespace: testNamespace,
},
want: want{
out: managed.ConnectionDetails{
"password": []byte("12345"),
},
},
},
}
for name, tc := range cases {
t.Run(name, func(t *testing.T) {
got, gotErr := connectionDetails(context.Background(), tc.args.kube, tc.args.connDetails, tc.args.relName, tc.args.relNamespace)
if diff := cmp.Diff(tc.want.err, gotErr, test.EquateErrors()); diff != "" {
t.Fatalf("connectionDetails(...): -want error, +got error: %s", diff)
}
if diff := cmp.Diff(tc.want.out, got); diff != "" {
t.Errorf("connectionDetails(...): -want result, +got result: %s", diff)
}
})
}
} | explode_data.jsonl/21511 | {
"file_path": "/home/dung/Study/Code/Cross_test_gen/training_dataset/dedup_data/clean_data_go/data/explode_data.jsonl",
"token_count": 1572
} | [
2830,
3393,
15866,
7799,
1155,
353,
8840,
836,
8,
341,
13158,
2827,
2036,
341,
197,
16463,
3760,
260,
2943,
11716,
198,
197,
32917,
7799,
220,
3056,
85,
16,
19127,
16,
17463,
10649,
198,
197,
197,
3748,
675,
414,
914,
198,
197,
197,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | 5 |
func TestGetPodDetails(t *testing.T) {
// POD_NAME & POD_NAMESPACE not exist
os.Setenv("POD_NAME", "")
os.Setenv("POD_NAMESPACE", "")
_, err1 := GetPodDetails(testclient.NewSimpleClientset())
if err1 == nil {
t.Errorf("expected an error but returned nil")
}
// POD_NAME not exist
os.Setenv("POD_NAME", "")
os.Setenv("POD_NAMESPACE", apiv1.NamespaceDefault)
_, err2 := GetPodDetails(testclient.NewSimpleClientset())
if err2 == nil {
t.Errorf("expected an error but returned nil")
}
// POD_NAMESPACE not exist
os.Setenv("POD_NAME", "testpod")
os.Setenv("POD_NAMESPACE", "")
_, err3 := GetPodDetails(testclient.NewSimpleClientset())
if err3 == nil {
t.Errorf("expected an error but returned nil")
}
// POD not exist
os.Setenv("POD_NAME", "testpod")
os.Setenv("POD_NAMESPACE", apiv1.NamespaceDefault)
_, err4 := GetPodDetails(testclient.NewSimpleClientset())
if err4 == nil {
t.Errorf("expected an error but returned nil")
}
// success to get PodInfo
fkClient := testclient.NewSimpleClientset(
&apiv1.PodList{Items: []apiv1.Pod{{
ObjectMeta: metav1.ObjectMeta{
Name: "testpod",
Namespace: apiv1.NamespaceDefault,
Labels: map[string]string{
"first": "first_label",
"second": "second_label",
},
},
}}},
&apiv1.NodeList{Items: []apiv1.Node{{
ObjectMeta: metav1.ObjectMeta{
Name: "demo",
},
Status: apiv1.NodeStatus{
Addresses: []apiv1.NodeAddress{
{
Type: apiv1.NodeInternalIP,
Address: "10.0.0.1",
},
},
},
}}})
epi, err5 := GetPodDetails(fkClient)
if err5 != nil {
t.Errorf("expected a PodInfo but returned error")
return
}
if epi == nil {
t.Errorf("expected a PodInfo but returned nil")
}
} | explode_data.jsonl/5383 | {
"file_path": "/home/dung/Study/Code/Cross_test_gen/training_dataset/dedup_data/clean_data_go/data/explode_data.jsonl",
"token_count": 753
} | [
2830,
3393,
1949,
23527,
7799,
1155,
353,
8840,
836,
8,
341,
197,
322,
90501,
4708,
609,
90501,
34552,
537,
3000,
198,
25078,
4202,
3160,
445,
2045,
35,
4708,
497,
14676,
25078,
4202,
3160,
445,
2045,
35,
34552,
497,
14676,
197,
6878,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | 7 |
func TestCreateTables(t *testing.T) {
err := DB.CreateTables(UserProfile{}, Post{})
assert.Nil(t, err)
assert.True(t, DB.CheckIfTableExists("user_profiles"))
assert.True(t, DB.CheckIfTableExists("renamed_posts"))
} | explode_data.jsonl/59898 | {
"file_path": "/home/dung/Study/Code/Cross_test_gen/training_dataset/dedup_data/clean_data_go/data/explode_data.jsonl",
"token_count": 85
} | [
2830,
3393,
4021,
21670,
1155,
353,
8840,
836,
8,
341,
9859,
1669,
5952,
7251,
21670,
13087,
8526,
22655,
3877,
37790,
6948,
59678,
1155,
11,
1848,
340,
6948,
32443,
1155,
11,
5952,
10600,
2679,
2556,
15575,
445,
872,
64021,
5455,
6948,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1
] | 1 |
func TestSet_Join(t *testing.T) {
gtest.C(t, func(t *gtest.T) {
s1 := gset.New(true)
s1.Add("a", "a1", "b", "c")
str1 := s1.Join(",")
t.Assert(strings.Contains(str1, "a1"), true)
})
gtest.C(t, func(t *gtest.T) {
s1 := gset.New(true)
s1.Add("a", `"b"`, `\c`)
str1 := s1.Join(",")
t.Assert(strings.Contains(str1, `"b"`), true)
t.Assert(strings.Contains(str1, `\c`), true)
t.Assert(strings.Contains(str1, `a`), true)
})
} | explode_data.jsonl/34390 | {
"file_path": "/home/dung/Study/Code/Cross_test_gen/training_dataset/dedup_data/clean_data_go/data/explode_data.jsonl",
"token_count": 242
} | [
2830,
3393,
1649,
10598,
1961,
1155,
353,
8840,
836,
8,
341,
3174,
1944,
727,
1155,
11,
2915,
1155,
353,
82038,
836,
8,
341,
197,
1903,
16,
1669,
342,
746,
7121,
3715,
340,
197,
1903,
16,
1904,
445,
64,
497,
330,
64,
16,
497,
330,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | 1 |
func TestValidateImageGCPolicy(t *testing.T) {
testCases := []struct {
name string
imageGCPolicy ImageGCPolicy
expectErr string
}{
{
name: "Test for LowThresholdPercent < HighThresholdPercent",
imageGCPolicy: ImageGCPolicy{
HighThresholdPercent: 2,
LowThresholdPercent: 1,
},
},
{
name: "Test for HighThresholdPercent < 0,",
imageGCPolicy: ImageGCPolicy{
HighThresholdPercent: -1,
},
expectErr: "invalid HighThresholdPercent -1, must be in range [0-100]",
},
{
name: "Test for HighThresholdPercent > 100",
imageGCPolicy: ImageGCPolicy{
HighThresholdPercent: 101,
},
expectErr: "invalid HighThresholdPercent 101, must be in range [0-100]",
},
{
name: "Test for LowThresholdPercent < 0",
imageGCPolicy: ImageGCPolicy{
LowThresholdPercent: -1,
},
expectErr: "invalid LowThresholdPercent -1, must be in range [0-100]",
},
{
name: "Test for LowThresholdPercent > 100",
imageGCPolicy: ImageGCPolicy{
LowThresholdPercent: 101,
},
expectErr: "invalid LowThresholdPercent 101, must be in range [0-100]",
},
{
name: "Test for LowThresholdPercent > HighThresholdPercent",
imageGCPolicy: ImageGCPolicy{
HighThresholdPercent: 1,
LowThresholdPercent: 2,
},
expectErr: "LowThresholdPercent 2 can not be higher than HighThresholdPercent 1",
},
}
for _, tc := range testCases {
if _, err := NewImageGCManager(nil, nil, nil, nil, tc.imageGCPolicy, ""); err != nil {
if err.Error() != tc.expectErr {
t.Errorf("[%s:]Expected err:%v, but got:%v", tc.name, tc.expectErr, err.Error())
}
}
}
} | explode_data.jsonl/48117 | {
"file_path": "/home/dung/Study/Code/Cross_test_gen/training_dataset/dedup_data/clean_data_go/data/explode_data.jsonl",
"token_count": 692
} | [
2830,
3393,
17926,
1906,
38,
7123,
8018,
1155,
353,
8840,
836,
8,
341,
18185,
37302,
1669,
3056,
1235,
341,
197,
11609,
688,
914,
198,
197,
31426,
38,
7123,
8018,
4654,
38,
7123,
8018,
198,
197,
24952,
7747,
257,
914,
198,
197,
59403,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | 4 |
func TestMissingPatternDiagnostic(t *testing.T) {
testenv.NeedsGo1Point(t, 16)
const files = `
-- go.mod --
module example.com
-- x.go --
package x
import (
_ "embed"
)
// Issue 47436
func F() {}
//go:embed NONEXISTENT
var foo string
`
Run(t, files, func(t *testing.T, env *Env) {
env.OpenFile("x.go")
env.Await(env.DiagnosticAtRegexpWithMessage("x.go", `NONEXISTENT`, "no matching files found"))
env.RegexpReplace("x.go", `NONEXISTENT`, "x.go")
env.Await(EmptyDiagnostics("x.go"))
})
} | explode_data.jsonl/1719 | {
"file_path": "/home/dung/Study/Code/Cross_test_gen/training_dataset/dedup_data/clean_data_go/data/explode_data.jsonl",
"token_count": 216
} | [
2830,
3393,
25080,
15760,
79388,
1155,
353,
8840,
836,
8,
341,
18185,
3160,
2067,
68,
6767,
10850,
16,
2609,
1155,
11,
220,
16,
21,
340,
4777,
3542,
284,
22074,
313,
728,
10929,
39514,
4352,
3110,
905,
198,
313,
856,
18002,
39514,
172... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | 1 |
func TestLinkWithCreationModeError(t *testing.T) {
fakeIGs := instances.NewFakeInstanceGroups(sets.NewString(), defaultNamer)
fakeNodePool := instances.NewNodePool(fakeIGs, defaultNamer)
fakeGCE := gce.FakeGCECloud(gce.DefaultTestClusterValues())
linker := newTestIGLinker(fakeGCE, fakeNodePool)
sp := utils.ServicePort{NodePort: 8080, Protocol: annotations.ProtocolHTTP}
modes := []BalancingMode{Rate, Utilization}
// block the update of Backends with the given balancingMode
// and verify that a backend with the other balancingMode is
// updated properly.
for i, bm := range modes {
(fakeGCE.Compute().(*cloud.MockGCE)).MockBackendServices.UpdateHook = func(ctx context.Context, key *meta.Key, be *compute.BackendService, m *cloud.MockBackendServices) error {
for _, b := range be.Backends {
if b.BalancingMode == string(bm) {
return &googleapi.Error{Code: http.StatusBadRequest}
}
}
return mock.UpdateBackendServiceHook(ctx, key, be, m)
}
// Mimic the instance group being created
if _, err := linker.instancePool.EnsureInstanceGroupsAndPorts(defaultNamer.InstanceGroup(), []int64{sp.NodePort}); err != nil {
t.Fatalf("Did not expect error when ensuring IG for ServicePort %+v: %v", sp, err)
}
// Mimic the syncer creating the backend.
linker.backendPool.Create(sp, "fake-health-check-link")
if err := linker.Link(sp, []GroupKey{{Zone: defaultZone}}); err != nil {
t.Fatalf("%v", err)
}
be, err := fakeGCE.GetGlobalBackendService(sp.BackendName(defaultNamer))
if err != nil {
t.Fatalf("%v", err)
}
if len(be.Backends) == 0 {
t.Fatalf("Expected Backends to be created")
}
for _, b := range be.Backends {
if b.BalancingMode != string(modes[(i+1)%len(modes)]) {
t.Fatalf("Wrong balancing mode, expected %v got %v", modes[(i+1)%len(modes)], b.BalancingMode)
}
}
linker.backendPool.Delete(sp.BackendName(defaultNamer))
}
} | explode_data.jsonl/81934 | {
"file_path": "/home/dung/Study/Code/Cross_test_gen/training_dataset/dedup_data/clean_data_go/data/explode_data.jsonl",
"token_count": 717
} | [
2830,
3393,
3939,
2354,
32701,
3636,
1454,
1155,
353,
8840,
836,
8,
341,
1166,
726,
1914,
82,
1669,
13121,
7121,
52317,
2523,
22173,
7,
4917,
7121,
703,
1507,
1638,
45,
15232,
340,
1166,
726,
1955,
10551,
1669,
13121,
7121,
1955,
10551,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | 3 |
func TestCopyBundleWithNonCollocatedReferencedImagesToRepo(t *testing.T) {
env := BuildEnv(t)
imgpkg := Imgpkg{t, Logger{}, env.ImgpkgPath}
assetsPath := filepath.Join("assets", "simple-app")
randFile, err := addRandomFile(assetsPath)
if err != nil {
t.Fatalf("failed to create unuique file: %v", err)
}
defer os.Remove(randFile)
image := env.Image + "-image-outside-repo"
out := imgpkg.Run([]string{"push", "--tty", "-i", image, "-f", assetsPath})
imageDigest := fmt.Sprintf("@%s", extractDigest(out, t))
// image intentionally does not exist in bundle repo
imageDigestRef := image + imageDigest
imgsYml := fmt.Sprintf(`---
apiVersion: imgpkg.carvel.dev/v1alpha1
kind: ImagesLock
spec:
images:
- image: %s
`, imageDigestRef)
imgpkgDir, err := createBundleDir(assetsPath, bundleYAML, imgsYml)
if err != nil {
t.Fatalf("failed to create bundle dir: %v", err)
}
defer os.RemoveAll(imgpkgDir)
out = imgpkg.Run([]string{"push", "--tty", "-b", env.Image, "-f", assetsPath})
bundleDigest := fmt.Sprintf("@%s", extractDigest(out, t))
bundleDigestRef := env.Image + bundleDigest
imgpkg.Run([]string{"copy", "--bundle", bundleDigestRef, "--to-repo", env.RelocationRepo})
refs := []string{env.RelocationRepo + imageDigest, env.RelocationRepo + bundleDigest}
if err := validateImagePresence(refs); err != nil {
t.Fatalf("could not validate image presence: %v", err)
}
} | explode_data.jsonl/23215 | {
"file_path": "/home/dung/Study/Code/Cross_test_gen/training_dataset/dedup_data/clean_data_go/data/explode_data.jsonl",
"token_count": 524
} | [
2830,
3393,
12106,
8409,
2354,
8121,
6127,
39463,
47447,
5767,
14228,
1249,
25243,
1155,
353,
8840,
836,
8,
341,
57538,
1669,
7854,
14359,
1155,
340,
39162,
30069,
1669,
2362,
21888,
7351,
90,
83,
11,
9514,
22655,
6105,
13,
13033,
30069,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | 4 |
func TestCookieStore_Get_decrypt_failure(t *testing.T) {
cipher := &fakeCipher{
fakeDecrypt: func(ciphertext []byte) ([]byte, error) {
return nil, errors.New("test decrypt failure")
},
}
opt := CookieOption{}
cookieman := New(cipher, opt)
w := httptest.NewRecorder()
store := cookieman.NewCookieStore("n", nil)
if err := store.Set(w, []byte("v")); err != nil {
t.Error(err)
}
req := GetRequestWithCookie(w)
if _, err := store.Get(req); err == nil {
t.Error("got nil, but want error")
}
} | explode_data.jsonl/48010 | {
"file_path": "/home/dung/Study/Code/Cross_test_gen/training_dataset/dedup_data/clean_data_go/data/explode_data.jsonl",
"token_count": 206
} | [
2830,
3393,
20616,
6093,
13614,
80764,
43618,
1155,
353,
8840,
836,
8,
341,
1444,
10558,
1669,
609,
30570,
79460,
515,
197,
1166,
726,
89660,
25,
2915,
1337,
45043,
3056,
3782,
8,
34923,
3782,
11,
1465,
8,
341,
298,
853,
2092,
11,
597... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | 1 |
func TestValidateOnceAWeekWindowFormat(t *testing.T) {
cases := []struct {
Value string
ErrCount int
}{
{
// once a day window format
Value: "04:00-05:00",
ErrCount: 1,
},
{
// invalid day of week
Value: "san:04:00-san:05:00",
ErrCount: 1,
},
{
// invalid hour
Value: "sun:24:00-san:25:00",
ErrCount: 1,
},
{
// invalid min
Value: "sun:04:00-sun:04:60",
ErrCount: 1,
},
{
// valid format
Value: "sun:04:00-sun:05:00",
ErrCount: 0,
},
{
// "Sun" can also be used
Value: "Sun:04:00-Sun:05:00",
ErrCount: 0,
},
{
// valid format
Value: "",
ErrCount: 0,
},
}
for _, tc := range cases {
_, errors := validateOnceAWeekWindowFormat(tc.Value, "maintenance_window")
if len(errors) != tc.ErrCount {
t.Fatalf("Expected %d validation errors, But got %d errors for \"%s\"", tc.ErrCount, len(errors), tc.Value)
}
}
} | explode_data.jsonl/78592 | {
"file_path": "/home/dung/Study/Code/Cross_test_gen/training_dataset/dedup_data/clean_data_go/data/explode_data.jsonl",
"token_count": 468
} | [
2830,
3393,
17926,
12522,
32,
17053,
4267,
4061,
1155,
353,
8840,
836,
8,
341,
1444,
2264,
1669,
3056,
1235,
341,
197,
47399,
262,
914,
198,
197,
197,
7747,
2507,
526,
198,
197,
59403,
197,
197,
515,
298,
197,
322,
3055,
264,
1899,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | 3 |
func TestToleratingMissingFiles(t *testing.T) {
loadingRules := ClientConfigLoadingRules{
Precedence: []string{"bogus1", "bogus2", "bogus3"},
}
_, err := loadingRules.Load()
if err != nil {
t.Fatalf("Unexpected error: %v", err)
}
} | explode_data.jsonl/67847 | {
"file_path": "/home/dung/Study/Code/Cross_test_gen/training_dataset/dedup_data/clean_data_go/data/explode_data.jsonl",
"token_count": 95
} | [
2830,
3393,
51,
22072,
1095,
25080,
10809,
1155,
353,
8840,
836,
8,
341,
197,
10628,
26008,
1669,
8423,
2648,
8578,
26008,
515,
197,
197,
4703,
1998,
763,
25,
3056,
917,
4913,
65,
538,
355,
16,
497,
330,
65,
538,
355,
17,
497,
330,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | 2 |
func TestGetSenderWithSameIDsReturnsSameSender(t *testing.T) {
resetAggregator()
InitAggregator(nil, "")
sender1, err := GetSender(checkID1)
assert.Nil(t, err)
assert.Len(t, aggregatorInstance.checkSamplers, 1)
assert.Len(t, senderPool.senders, 1)
sender2, err := GetSender(checkID1)
assert.Nil(t, err)
assert.Equal(t, sender1, sender2)
assert.Len(t, aggregatorInstance.checkSamplers, 1)
assert.Len(t, senderPool.senders, 1)
} | explode_data.jsonl/78295 | {
"file_path": "/home/dung/Study/Code/Cross_test_gen/training_dataset/dedup_data/clean_data_go/data/explode_data.jsonl",
"token_count": 179
} | [
2830,
3393,
1949,
20381,
2354,
19198,
30466,
16446,
19198,
20381,
1155,
353,
8840,
836,
8,
341,
70343,
9042,
58131,
741,
98762,
9042,
58131,
27907,
11,
85617,
1903,
1659,
16,
11,
1848,
1669,
2126,
20381,
24077,
915,
16,
340,
6948,
59678,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | 1 |
func TestTaxonomyDataParse(t *testing.T) {
type testCase struct {
lang string
cnt int
}
cases := []testCase{
testCase{"cs-CZ", 5427},
testCase{"da-DK", 5427},
testCase{"de-CH", 5427},
testCase{"de-DE", 5427},
testCase{"en-US", 5427},
testCase{"es-ES", 5427},
testCase{"fr-FR", 5427},
testCase{"it-IT", 5427},
testCase{"ja-JP", 5442},
testCase{"pl-PL", 5427},
testCase{"pt-BR", 5427},
testCase{"sv-SE", 5427},
testCase{"zh-CN", 4586},
}
for _, cas := range cases {
t.Run(fmt.Sprintf("Load taxonomy data file in %s", cas.lang), func(t *testing.T) {
td := taxonomyData{
Language: cas.lang,
LoadFunc: data.Asset,
}
if td.Filename() != fmt.Sprintf("taxonomy-with-ids.%s.txt", cas.lang) {
t.Fatalf("%s is not valid taxonomy data filename", td.Filename())
}
if err := td.Parse(); err != nil {
t.Fatal(err)
}
if len(td.data) != cas.cnt {
t.Fatalf("%s contains %d records, not eq. to %d", td.Filename(), len(td.data), cas.cnt)
}
})
}
} | explode_data.jsonl/35846 | {
"file_path": "/home/dung/Study/Code/Cross_test_gen/training_dataset/dedup_data/clean_data_go/data/explode_data.jsonl",
"token_count": 466
} | [
2830,
3393,
31349,
16974,
1043,
14463,
1155,
353,
8840,
836,
8,
341,
13158,
54452,
2036,
341,
197,
197,
5205,
914,
198,
197,
60553,
220,
526,
198,
197,
532,
1444,
2264,
1669,
3056,
66194,
515,
197,
18185,
4207,
4913,
4837,
7658,
57,
4... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | 4 |
func TestCanSetIdempotencyToken(t *testing.T) {
cases := []struct {
CanSet bool
Case interface{}
}{
{
true,
struct {
Field *string `idempotencyToken:"true"`
}{},
},
{
true,
struct {
Field string `idempotencyToken:"true"`
}{},
},
{
false,
struct {
Field *string `idempotencyToken:"true"`
}{Field: new(string)},
},
{
false,
struct {
Field string `idempotencyToken:"true"`
}{Field: "value"},
},
{
false,
struct {
Field *int `idempotencyToken:"true"`
}{},
},
{
false,
struct {
Field *string
}{},
},
}
for i, c := range cases {
v := reflect.Indirect(reflect.ValueOf(c.Case))
ty := v.Type()
canSet := protocol.CanSetIdempotencyToken(v.Field(0), ty.Field(0))
if e, a := c.CanSet, canSet; e != a {
t.Errorf("%d, expect %v, got %v", i, e, a)
}
}
} | explode_data.jsonl/78825 | {
"file_path": "/home/dung/Study/Code/Cross_test_gen/training_dataset/dedup_data/clean_data_go/data/explode_data.jsonl",
"token_count": 423
} | [
2830,
3393,
6713,
1649,
764,
3262,
354,
2251,
3323,
1155,
353,
8840,
836,
8,
341,
1444,
2264,
1669,
3056,
1235,
341,
197,
6258,
276,
1649,
1807,
198,
197,
197,
4207,
256,
3749,
16094,
197,
59403,
197,
197,
515,
298,
42808,
345,
298,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | 3 |
func TestGetLabelsForVolume(t *testing.T) {
ctrl := gomock.NewController(t)
defer ctrl.Finish()
testCloud0 := GetTestCloud(ctrl)
diskName := "disk1"
diskURI := fmt.Sprintf("/subscriptions/%s/resourceGroups/%s/providers/Microsoft.Compute/disks/%s",
testCloud0.SubscriptionID, testCloud0.ResourceGroup, diskName)
diskSizeGB := int32(30)
fakeGetDiskFailed := "fakeGetDiskFailed"
fakeGetDiskFailedDiskURI := fmt.Sprintf("/subscriptions/%s/resourceGroups/%s/providers/Microsoft.Compute/disks/%s",
testCloud0.SubscriptionID, testCloud0.ResourceGroup, fakeGetDiskFailed)
testCases := []struct {
desc string
diskName string
pv *v1.PersistentVolume
existedDisk compute.Disk
expected map[string]string
expectedErr bool
expectedErrMsg error
}{
{
desc: "labels and no error shall be returned if everything is good",
diskName: diskName,
pv: &v1.PersistentVolume{
Spec: v1.PersistentVolumeSpec{
PersistentVolumeSource: v1.PersistentVolumeSource{
AzureDisk: &v1.AzureDiskVolumeSource{
DiskName: diskName,
DataDiskURI: diskURI,
},
},
},
},
existedDisk: compute.Disk{Name: to.StringPtr(diskName), DiskProperties: &compute.DiskProperties{DiskSizeGB: &diskSizeGB}, Zones: &[]string{"1"}},
expected: map[string]string{
v1.LabelTopologyRegion: testCloud0.Location,
v1.LabelTopologyZone: testCloud0.makeZone(testCloud0.Location, 1),
},
expectedErr: false,
},
{
desc: "an error shall be returned if everything is good with invalid zone",
diskName: diskName,
pv: &v1.PersistentVolume{
Spec: v1.PersistentVolumeSpec{
PersistentVolumeSource: v1.PersistentVolumeSource{
AzureDisk: &v1.AzureDiskVolumeSource{
DiskName: diskName,
DataDiskURI: diskURI,
},
},
},
},
existedDisk: compute.Disk{Name: to.StringPtr(diskName), DiskProperties: &compute.DiskProperties{DiskSizeGB: &diskSizeGB}, Zones: &[]string{"invalid"}},
expectedErr: true,
expectedErrMsg: fmt.Errorf("failed to parse zone [invalid] for AzureDisk %v: %v", diskName, "strconv.Atoi: parsing \"invalid\": invalid syntax"),
},
{
desc: "nil shall be returned if everything is good with null Zones",
diskName: diskName,
pv: &v1.PersistentVolume{
Spec: v1.PersistentVolumeSpec{
PersistentVolumeSource: v1.PersistentVolumeSource{
AzureDisk: &v1.AzureDiskVolumeSource{
DiskName: diskName,
DataDiskURI: diskURI,
},
},
},
},
existedDisk: compute.Disk{Name: to.StringPtr(diskName), DiskProperties: &compute.DiskProperties{DiskSizeGB: &diskSizeGB}},
expected: map[string]string{
v1.LabelTopologyRegion: testCloud0.Location,
},
expectedErr: false,
expectedErrMsg: nil,
},
{
desc: "an error shall be returned if everything is good with get disk failed",
diskName: fakeGetDiskFailed,
pv: &v1.PersistentVolume{
Spec: v1.PersistentVolumeSpec{
PersistentVolumeSource: v1.PersistentVolumeSource{
AzureDisk: &v1.AzureDiskVolumeSource{
DiskName: fakeGetDiskFailed,
DataDiskURI: fakeGetDiskFailedDiskURI,
},
},
},
},
existedDisk: compute.Disk{Name: to.StringPtr(fakeGetDiskFailed), DiskProperties: &compute.DiskProperties{DiskSizeGB: &diskSizeGB}, Zones: &[]string{"1"}},
expectedErr: true,
expectedErrMsg: fmt.Errorf("Retriable: false, RetryAfter: 0s, HTTPStatusCode: 0, RawError: %w", fmt.Errorf("Get Disk failed")),
},
{
desc: "an error shall be returned if everything is good with invalid DiskURI",
diskName: diskName,
pv: &v1.PersistentVolume{
Spec: v1.PersistentVolumeSpec{
PersistentVolumeSource: v1.PersistentVolumeSource{
AzureDisk: &v1.AzureDiskVolumeSource{
DiskName: diskName,
DataDiskURI: "invalidDiskURI",
},
},
},
},
existedDisk: compute.Disk{Name: to.StringPtr(diskName), DiskProperties: &compute.DiskProperties{DiskSizeGB: &diskSizeGB}, Zones: &[]string{"1"}},
expectedErr: true,
expectedErrMsg: fmt.Errorf("invalid disk URI: invalidDiskURI"),
},
{
desc: "nil shall be returned if everything is good but pv.Spec.AzureDisk.DiskName is cloudvolume.ProvisionedVolumeName",
diskName: diskName,
pv: &v1.PersistentVolume{
Spec: v1.PersistentVolumeSpec{
PersistentVolumeSource: v1.PersistentVolumeSource{
AzureDisk: &v1.AzureDiskVolumeSource{
DiskName: cloudvolume.ProvisionedVolumeName,
DataDiskURI: diskURI,
},
},
},
},
existedDisk: compute.Disk{Name: to.StringPtr(diskName), DiskProperties: &compute.DiskProperties{DiskSizeGB: &diskSizeGB}},
expected: nil,
expectedErr: false,
},
{
desc: "nil shall be returned if everything is good but pv.Spec.AzureDisk is nil",
diskName: diskName,
pv: &v1.PersistentVolume{
Spec: v1.PersistentVolumeSpec{
PersistentVolumeSource: v1.PersistentVolumeSource{},
},
},
existedDisk: compute.Disk{Name: to.StringPtr(diskName), DiskProperties: &compute.DiskProperties{DiskSizeGB: &diskSizeGB}},
expected: nil,
expectedErr: false,
},
}
for i, test := range testCases {
testCloud := GetTestCloud(ctrl)
mockDisksClient := testCloud.DisksClient.(*mockdiskclient.MockInterface)
if test.diskName == fakeGetDiskFailed {
mockDisksClient.EXPECT().Get(gomock.Any(), testCloud.ResourceGroup, test.diskName).Return(test.existedDisk, &retry.Error{RawError: fmt.Errorf("Get Disk failed")}).AnyTimes()
} else {
mockDisksClient.EXPECT().Get(gomock.Any(), testCloud.ResourceGroup, test.diskName).Return(test.existedDisk, nil).AnyTimes()
}
mockDisksClient.EXPECT().CreateOrUpdate(gomock.Any(), testCloud.ResourceGroup, test.diskName, gomock.Any()).Return(nil).AnyTimes()
result, err := testCloud.GetLabelsForVolume(context.TODO(), test.pv)
assert.Equal(t, test.expected, result, "TestCase[%d]: %s, expected: %v, return: %v", i, test.desc, test.expected, result)
assert.Equal(t, test.expectedErr, err != nil, "TestCase[%d]: %s, return error: %v", i, test.desc, err)
assert.Equal(t, test.expectedErrMsg, err, "TestCase[%d]: %s, expected: %v, return: %v", i, test.desc, test.expectedErrMsg, err)
}
} | explode_data.jsonl/36149 | {
"file_path": "/home/dung/Study/Code/Cross_test_gen/training_dataset/dedup_data/clean_data_go/data/explode_data.jsonl",
"token_count": 2593
} | [
2830,
3393,
1949,
23674,
2461,
18902,
1155,
353,
8840,
836,
8,
341,
84381,
1669,
342,
316,
1176,
7121,
2051,
1155,
340,
16867,
23743,
991,
18176,
2822,
18185,
16055,
15,
1669,
2126,
2271,
16055,
62100,
340,
2698,
3187,
675,
1669,
330,
3... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | 3 |
func TestOuterLinkV2WithMetadataPointerEmbedderDecode(t *testing.T) {
var o outerLinkV2WithMetadataPointerEmbedder
err := MsgpackDecode(&o, []byte{0x1, 0x2})
requireErrorHasSuffix(t, errCodecDecodeSelf, err)
} | explode_data.jsonl/72243 | {
"file_path": "/home/dung/Study/Code/Cross_test_gen/training_dataset/dedup_data/clean_data_go/data/explode_data.jsonl",
"token_count": 86
} | [
2830,
3393,
51322,
3939,
53,
17,
2354,
14610,
9084,
25486,
1107,
32564,
1155,
353,
8840,
836,
8,
341,
2405,
297,
15955,
3939,
53,
17,
2354,
14610,
9084,
25486,
1107,
198,
9859,
1669,
24205,
4748,
32564,
2099,
78,
11,
3056,
3782,
90,
1... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | 1 |
func TestParsersCStyleLog(t *testing.T) {
env := newInputTestingEnvironment(t)
testlogName := "test.log"
inp := env.mustCreateInput(map[string]interface{}{
"paths": []string{env.abspath(testlogName)},
"prospector.scanner.check_interval": "1ms",
"parsers": []map[string]interface{}{
map[string]interface{}{
"multiline": map[string]interface{}{
"type": "pattern",
"pattern": "\\\\$",
"negate": false,
"match": "before",
"timeout": "100ms", // set to lower value to speed up test
},
},
},
})
testlines := []byte(`The following are log messages
This is a C style log\\
file which is on multiple\\
lines
In addition it has normal lines
The total should be 4 lines covered
`)
env.mustWriteLinesToFile(testlogName, testlines)
ctx, cancelInput := context.WithCancel(context.Background())
env.startInput(ctx, inp)
env.waitUntilEventCount(4)
env.requireOffsetInRegistry(testlogName, len(testlines))
cancelInput()
env.waitUntilInputStops()
} | explode_data.jsonl/14917 | {
"file_path": "/home/dung/Study/Code/Cross_test_gen/training_dataset/dedup_data/clean_data_go/data/explode_data.jsonl",
"token_count": 409
} | [
2830,
3393,
47,
40488,
34,
2323,
2201,
1155,
353,
8840,
836,
8,
341,
57538,
1669,
501,
2505,
16451,
12723,
1155,
692,
18185,
839,
675,
1669,
330,
1944,
1665,
698,
17430,
79,
1669,
6105,
69419,
4021,
2505,
9147,
14032,
31344,
67066,
197,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | 1 |
func TestSetCurrentLogLevel(t *testing.T) {
var log = Init(INFO_LEVEL, false, false)
log.SetLogLevel(WARN_LEVEL)
var levelNum, _ = log.GetCurrentLogLevel()
if levelNum != WARN_LEVEL {
t.Errorf("Got: %d expected: %d", levelNum, WARN_LEVEL)
}
} | explode_data.jsonl/3657 | {
"file_path": "/home/dung/Study/Code/Cross_test_gen/training_dataset/dedup_data/clean_data_go/data/explode_data.jsonl",
"token_count": 99
} | [
2830,
3393,
1649,
5405,
72676,
1155,
353,
8840,
836,
8,
341,
2405,
1487,
284,
15690,
62318,
17415,
11,
895,
11,
895,
692,
6725,
4202,
72676,
14031,
9051,
17415,
340,
2405,
2188,
4651,
11,
716,
284,
1487,
44242,
72676,
741,
743,
2188,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | 2 |
func TestEmptyCallStack(t *testing.T) {
t.Parallel()
var buf bytes.Buffer
p := NewProfile("test18836")
p.Add("foo", 47674)
p.WriteTo(&buf, 1)
p.Remove("foo")
got := buf.String()
prefix := "test18836 profile: total 1\n"
if !strings.HasPrefix(got, prefix) {
t.Fatalf("got:\n\t%q\nwant prefix:\n\t%q\n", got, prefix)
}
lostevent := "lostProfileEvent"
if !strings.Contains(got, lostevent) {
t.Fatalf("got:\n\t%q\ndoes not contain:\n\t%q\n", got, lostevent)
}
} | explode_data.jsonl/13656 | {
"file_path": "/home/dung/Study/Code/Cross_test_gen/training_dataset/dedup_data/clean_data_go/data/explode_data.jsonl",
"token_count": 205
} | [
2830,
3393,
3522,
7220,
4336,
1155,
353,
8840,
836,
8,
341,
3244,
41288,
7957,
741,
2405,
6607,
5820,
22622,
198,
3223,
1669,
1532,
8526,
445,
1944,
16,
23,
23,
18,
21,
1138,
3223,
1904,
445,
7975,
497,
220,
19,
22,
21,
22,
19,
34... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | 3 |
func TestWorkspace_ReadPipelineManifest(t *testing.T) {
copilotDir := "/copilot"
testCases := map[string]struct {
fs func() afero.Fs
expectedError error
}{
"reads existing pipeline manifest": {
fs: func() afero.Fs {
fs := afero.NewMemMapFs()
fs.MkdirAll("/copilot", 0755)
manifest, _ := fs.Create("/copilot/pipeline.yml")
defer manifest.Close()
manifest.Write([]byte("hello"))
return fs
},
expectedError: nil,
},
"when no pipeline file exists": {
fs: func() afero.Fs {
fs := afero.NewMemMapFs()
fs.Mkdir(copilotDir, 0755)
return fs
},
expectedError: ErrNoPipelineInWorkspace,
},
}
for name, tc := range testCases {
t.Run(name, func(t *testing.T) {
// GIVEN
fs := tc.fs()
ws := &Workspace{
copilotDir: copilotDir,
fsUtils: &afero.Afero{Fs: fs},
}
// WHEN
_, err := ws.ReadPipelineManifest()
// THEN
if tc.expectedError != nil {
require.Equal(t, tc.expectedError.Error(), err.Error())
} else {
require.NoError(t, err)
}
})
}
} | explode_data.jsonl/30121 | {
"file_path": "/home/dung/Study/Code/Cross_test_gen/training_dataset/dedup_data/clean_data_go/data/explode_data.jsonl",
"token_count": 501
} | [
2830,
3393,
45981,
38381,
34656,
38495,
1155,
353,
8840,
836,
8,
341,
1444,
453,
23958,
6184,
1669,
3521,
37728,
23958,
698,
18185,
37302,
1669,
2415,
14032,
60,
1235,
341,
197,
53584,
310,
2915,
368,
264,
802,
78,
991,
82,
198,
197,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | 1 |
func TestAlignment(t *testing.T) {
type T1inner struct {
a int
}
type T1 struct {
T1inner
f int
}
type T2inner struct {
a, b int
}
type T2 struct {
T2inner
f int
}
x := T1{T1inner{2}, 17}
check2ndField(x, uintptr(unsafe.Pointer(&x.f))-uintptr(unsafe.Pointer(&x)), t)
x1 := T2{T2inner{2, 3}, 17}
check2ndField(x1, uintptr(unsafe.Pointer(&x1.f))-uintptr(unsafe.Pointer(&x1)), t)
} | explode_data.jsonl/29543 | {
"file_path": "/home/dung/Study/Code/Cross_test_gen/training_dataset/dedup_data/clean_data_go/data/explode_data.jsonl",
"token_count": 198
} | [
2830,
3393,
7033,
1155,
353,
8840,
836,
8,
341,
13158,
350,
16,
4382,
2036,
341,
197,
11323,
526,
198,
197,
532,
13158,
350,
16,
2036,
341,
197,
10261,
16,
4382,
198,
197,
1166,
526,
198,
197,
532,
13158,
350,
17,
4382,
2036,
341,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | 1 |
func TestVersion(t *testing.T) {
version1 := "20.02.15"
version2 := "20.03.25"
fmt.Println(CompareVersion(version1, version2))
version1 = "1.0.13"
version2 = "1.0.1a"
fmt.Println(CompareVersion(version1, version2))
version1 = "1.0.131"
version2 = "1.0.1a"
fmt.Println(CompareVersion(version1, version2))
version1 = "1.1.131"
version2 = "1.10.1a"
fmt.Println(CompareVersion(version1, version2))
version1 = "1.0.4"
version2 = "1.1.1"
fmt.Println(CompareVersion(version1, version2))
} | explode_data.jsonl/1125 | {
"file_path": "/home/dung/Study/Code/Cross_test_gen/training_dataset/dedup_data/clean_data_go/data/explode_data.jsonl",
"token_count": 221
} | [
2830,
3393,
5637,
1155,
353,
8840,
836,
8,
341,
74954,
16,
1669,
330,
17,
15,
13,
15,
17,
13,
16,
20,
698,
74954,
17,
1669,
330,
17,
15,
13,
15,
18,
13,
17,
20,
698,
11009,
12419,
7,
27374,
5637,
37770,
16,
11,
2319,
17,
4390,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | 1 |
func TestNatSort(t *testing.T) {
// Validate that the order of SRV records returned by a DNS
// lookup for a k8s StatefulSet are ordered as expected when
// a natsort is done.
input := []string{
"memcached-10.memcached.cortex.svc.cluster.local.",
"memcached-1.memcached.cortex.svc.cluster.local.",
"memcached-6.memcached.cortex.svc.cluster.local.",
"memcached-3.memcached.cortex.svc.cluster.local.",
"memcached-25.memcached.cortex.svc.cluster.local.",
}
expected := []string{
"memcached-1.memcached.cortex.svc.cluster.local.",
"memcached-3.memcached.cortex.svc.cluster.local.",
"memcached-6.memcached.cortex.svc.cluster.local.",
"memcached-10.memcached.cortex.svc.cluster.local.",
"memcached-25.memcached.cortex.svc.cluster.local.",
}
natsort.Sort(input)
require.Equal(t, expected, input)
} | explode_data.jsonl/2661 | {
"file_path": "/home/dung/Study/Code/Cross_test_gen/training_dataset/dedup_data/clean_data_go/data/explode_data.jsonl",
"token_count": 350
} | [
2830,
3393,
65214,
10231,
1155,
353,
8840,
836,
8,
341,
197,
322,
23282,
429,
279,
1973,
315,
20880,
53,
7424,
5927,
553,
264,
27598,
198,
197,
322,
18615,
369,
264,
595,
23,
82,
3234,
1262,
1649,
525,
11457,
438,
3601,
979,
198,
19... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | 1 |
func TestMetricsMiddleware(t *testing.T) {
cfg := newFakeKeycloakConfig()
cfg.EnableMetrics = true
cfg.LocalhostMetrics = true
requests := []fakeRequest{
{
URI: cfg.WithOAuthURI(metricsURL),
Headers: map[string]string{
"X-Forwarded-For": "10.0.0.1",
},
ExpectedCode: http.StatusForbidden,
},
// Some request must run before this one to generate request status numbers
{
URI: cfg.WithOAuthURI(metricsURL),
ExpectedCode: http.StatusOK,
ExpectedContentContains: "proxy_request_status_total",
},
}
newFakeProxy(cfg).RunTests(t, requests)
} | explode_data.jsonl/14746 | {
"file_path": "/home/dung/Study/Code/Cross_test_gen/training_dataset/dedup_data/clean_data_go/data/explode_data.jsonl",
"token_count": 259
} | [
2830,
3393,
27328,
24684,
1155,
353,
8840,
836,
8,
341,
50286,
1669,
501,
52317,
1592,
88751,
2648,
741,
50286,
32287,
27328,
284,
830,
198,
50286,
20856,
3790,
27328,
284,
830,
198,
23555,
82,
1669,
3056,
30570,
1900,
515,
197,
197,
51... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | 1 |
func TestProviderRateLimiterHandler_Handle_SourceMicroService(t *testing.T) {
t.Log("testing providerratelimiter handler with source microservice and qps enabled as true")
initEnv()
c := handler.Chain{}
c.AddHandler(&handler.ProviderRateLimiterHandler{})
config.GlobalDefinition = &model.GlobalCfg{}
config.GlobalDefinition.Cse.FlowControl.Provider.QPS.Enabled = true
i := &invocation.Invocation{
SourceMicroService: "service1",
SchemaID: "schema1",
OperationID: "SayHello",
Args: &helloworld.HelloRequest{Name: "peter"},
}
c.Next(i, func(r *invocation.Response) error {
assert.NoError(t, r.Err)
log.Println(r.Result)
return r.Err
})
} | explode_data.jsonl/62821 | {
"file_path": "/home/dung/Study/Code/Cross_test_gen/training_dataset/dedup_data/clean_data_go/data/explode_data.jsonl",
"token_count": 267
} | [
2830,
3393,
5179,
11564,
43,
17700,
3050,
42714,
48122,
34609,
1860,
1155,
353,
8840,
836,
8,
341,
3244,
5247,
445,
8840,
2059,
59609,
301,
17700,
7013,
448,
2530,
8003,
7936,
323,
2804,
1690,
8970,
438,
830,
5130,
28248,
14359,
741,
14... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | 1 |
func TestWatchNewFile(t *testing.T) {
t.Parallel()
dir, ks := tmpKeyStore(t, false)
defer os.RemoveAll(dir)
// Ensure the watcher is started before adding any files.
ks.Accounts()
time.Sleep(1000 * time.Millisecond)
// Move in the files.
wantAccounts := make([]accounts.Account, len(cachetestAccounts))
for i := range cachetestAccounts {
wantAccounts[i] = accounts.Account{
Address: cachetestAccounts[i].Address,
URL: accounts.URL{Scheme: KeyStoreScheme, Path: filepath.Join(dir, filepath.Base(cachetestAccounts[i].URL.Path))},
}
if err := cp.CopyFile(wantAccounts[i].URL.Path, cachetestAccounts[i].URL.Path); err != nil {
t.Fatal(err)
}
}
// ks should see the accounts.
var list []accounts.Account
for d := 200 * time.Millisecond; d < 5*time.Second; d *= 2 {
list = ks.Accounts()
if reflect.DeepEqual(list, wantAccounts) {
// ks should have also received change notifications
select {
case <-ks.changes:
default:
t.Fatalf("wasn't notified of new accounts")
}
return
}
time.Sleep(d)
}
t.Errorf("got %s, want %s", spew.Sdump(list), spew.Sdump(wantAccounts))
} | explode_data.jsonl/36027 | {
"file_path": "/home/dung/Study/Code/Cross_test_gen/training_dataset/dedup_data/clean_data_go/data/explode_data.jsonl",
"token_count": 448
} | [
2830,
3393,
14247,
3564,
1703,
1155,
353,
8840,
836,
8,
341,
3244,
41288,
7957,
2822,
48532,
11,
41282,
1669,
4174,
1592,
6093,
1155,
11,
895,
340,
16867,
2643,
84427,
14161,
692,
197,
322,
29279,
279,
55727,
374,
3855,
1573,
7842,
894,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | 6 |
func TestNewPairs(t *testing.T) {
for _, tc := range []struct {
Name string
Local Candidates
Remote Candidates
Result Pairs
}{
{
Name: "Blank",
},
{
Name: "No pairs",
Local: Candidates{
{
Addr: Addr{
IP: net.ParseIP("1.1.1.1"),
},
},
},
Remote: Candidates{
{
Addr: Addr{
IP: net.ParseIP("2001:11:12:13:14:15:16:17"),
},
},
},
},
{
Name: "Simple",
Local: Candidates{
{
Addr: Addr{
IP: net.ParseIP("1.1.1.1"),
},
},
},
Remote: Candidates{
{
Addr: Addr{
IP: net.ParseIP("1.1.1.2"),
},
},
},
Result: Pairs{
{
Local: Candidate{
Addr: Addr{
IP: net.ParseIP("1.1.1.1"),
},
},
Remote: Candidate{
Addr: Addr{
IP: net.ParseIP("1.1.1.2"),
},
},
},
},
},
} {
t.Run(tc.Name, func(t *testing.T) {
got := NewPairs(tc.Local, tc.Remote)
if len(got) != len(tc.Result) {
t.Fatalf("bad length: %d (got) != %d (expected)", len(got), len(tc.Result))
}
for i := range tc.Result {
expectedAddr := tc.Result[i].Remote.Addr
gotAddr := got[i].Remote.Addr
if !gotAddr.Equal(expectedAddr) {
t.Errorf("[%d]: remote addr mismatch: %s (got) != %s (expected)", i, gotAddr, expectedAddr)
}
expectedAddr = tc.Result[i].Local.Addr
gotAddr = got[i].Local.Addr
if !gotAddr.Equal(expectedAddr) {
t.Errorf("[%d]: local addr mismatch: %s (got) != %s (expected)", i, gotAddr, expectedAddr)
}
}
})
}
} | explode_data.jsonl/9182 | {
"file_path": "/home/dung/Study/Code/Cross_test_gen/training_dataset/dedup_data/clean_data_go/data/explode_data.jsonl",
"token_count": 862
} | [
2830,
3393,
3564,
54228,
1155,
353,
8840,
836,
8,
341,
2023,
8358,
17130,
1669,
2088,
3056,
1235,
341,
197,
21297,
256,
914,
198,
197,
82404,
220,
64438,
198,
197,
197,
24703,
64438,
198,
197,
56503,
393,
4720,
198,
197,
59403,
197,
1... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | 5 |
func TestReconcilePrivateDNS(t *testing.T) {
testcases := []struct {
name string
expectedError string
expect func(s *mock_privatedns.MockScopeMockRecorder, m *mock_privatedns.MockclientMockRecorder)
}{
{
name: "no private dns",
expectedError: "",
expect: func(s *mock_privatedns.MockScopeMockRecorder, m *mock_privatedns.MockclientMockRecorder) {
s.PrivateDNSSpec().Return(nil)
},
},
{
name: "create ipv4 private dns successfully",
expectedError: "",
expect: func(s *mock_privatedns.MockScopeMockRecorder, m *mock_privatedns.MockclientMockRecorder) {
s.V(gomock.AssignableToTypeOf(2)).AnyTimes().Return(klogr.New())
s.PrivateDNSSpec().Return(&azure.PrivateDNSSpec{
ZoneName: "my-dns-zone",
VNetName: "my-vnet",
VNetResourceGroup: "vnet-rg",
LinkName: "my-link",
Records: []infrav1.AddressRecord{
{
Hostname: "hostname-1",
IP: "10.0.0.8",
},
},
})
s.ResourceGroup().AnyTimes().Return("my-rg")
s.SubscriptionID().Return("123")
m.CreateOrUpdateZone(gomockinternal.AContext(), "my-rg", "my-dns-zone", privatedns.PrivateZone{Location: to.StringPtr(azure.Global)})
m.CreateOrUpdateLink(gomockinternal.AContext(), "my-rg", "my-dns-zone", "my-link", privatedns.VirtualNetworkLink{
VirtualNetworkLinkProperties: &privatedns.VirtualNetworkLinkProperties{
VirtualNetwork: &privatedns.SubResource{
ID: to.StringPtr("/subscriptions/123/resourceGroups/vnet-rg/providers/Microsoft.Network/virtualNetworks/my-vnet"),
},
RegistrationEnabled: to.BoolPtr(false),
},
Location: to.StringPtr(azure.Global),
})
m.CreateOrUpdateRecordSet(gomockinternal.AContext(), "my-rg", "my-dns-zone", privatedns.A, "hostname-1", privatedns.RecordSet{
RecordSetProperties: &privatedns.RecordSetProperties{
TTL: to.Int64Ptr(300),
ARecords: &[]privatedns.ARecord{
{
Ipv4Address: to.StringPtr("10.0.0.8"),
},
},
},
})
},
},
{
name: "create ipv6 private dns successfully",
expectedError: "",
expect: func(s *mock_privatedns.MockScopeMockRecorder, m *mock_privatedns.MockclientMockRecorder) {
s.V(gomock.AssignableToTypeOf(2)).AnyTimes().Return(klogr.New())
s.PrivateDNSSpec().Return(&azure.PrivateDNSSpec{
ZoneName: "my-dns-zone",
VNetName: "my-vnet",
VNetResourceGroup: "vnet-rg",
LinkName: "my-link",
Records: []infrav1.AddressRecord{
{
Hostname: "hostname-2",
IP: "2603:1030:805:2::b",
},
},
})
s.ResourceGroup().AnyTimes().Return("my-rg")
s.SubscriptionID().Return("123")
m.CreateOrUpdateZone(gomockinternal.AContext(), "my-rg", "my-dns-zone", privatedns.PrivateZone{Location: to.StringPtr(azure.Global)})
m.CreateOrUpdateLink(gomockinternal.AContext(), "my-rg", "my-dns-zone", "my-link", privatedns.VirtualNetworkLink{
VirtualNetworkLinkProperties: &privatedns.VirtualNetworkLinkProperties{
VirtualNetwork: &privatedns.SubResource{
ID: to.StringPtr("/subscriptions/123/resourceGroups/vnet-rg/providers/Microsoft.Network/virtualNetworks/my-vnet"),
},
RegistrationEnabled: to.BoolPtr(false),
},
Location: to.StringPtr(azure.Global),
})
m.CreateOrUpdateRecordSet(gomockinternal.AContext(), "my-rg", "my-dns-zone", privatedns.AAAA, "hostname-2", privatedns.RecordSet{
RecordSetProperties: &privatedns.RecordSetProperties{
TTL: to.Int64Ptr(300),
AaaaRecords: &[]privatedns.AaaaRecord{
{
Ipv6Address: to.StringPtr("2603:1030:805:2::b"),
},
},
},
})
},
},
{
name: "link creation fails",
expectedError: "failed to create virtual network link my-link: #: Internal Server Error: StatusCode=500",
expect: func(s *mock_privatedns.MockScopeMockRecorder, m *mock_privatedns.MockclientMockRecorder) {
s.V(gomock.AssignableToTypeOf(2)).AnyTimes().Return(klogr.New())
s.PrivateDNSSpec().Return(&azure.PrivateDNSSpec{
ZoneName: "my-dns-zone",
VNetName: "my-vnet",
VNetResourceGroup: "vnet-rg",
LinkName: "my-link",
Records: []infrav1.AddressRecord{
{
Hostname: "hostname-1",
IP: "10.0.0.8",
},
},
})
s.ResourceGroup().AnyTimes().Return("my-rg")
s.SubscriptionID().Return("123")
m.CreateOrUpdateZone(gomockinternal.AContext(), "my-rg", "my-dns-zone", privatedns.PrivateZone{Location: to.StringPtr(azure.Global)})
m.CreateOrUpdateLink(gomockinternal.AContext(), "my-rg", "my-dns-zone", "my-link", privatedns.VirtualNetworkLink{
VirtualNetworkLinkProperties: &privatedns.VirtualNetworkLinkProperties{
VirtualNetwork: &privatedns.SubResource{
ID: to.StringPtr("/subscriptions/123/resourceGroups/vnet-rg/providers/Microsoft.Network/virtualNetworks/my-vnet"),
},
RegistrationEnabled: to.BoolPtr(false),
},
Location: to.StringPtr(azure.Global),
}).Return(autorest.NewErrorWithResponse("", "", &http.Response{StatusCode: 500}, "Internal Server Error"))
},
},
}
for _, tc := range testcases {
tc := tc
t.Run(tc.name, func(t *testing.T) {
g := NewWithT(t)
t.Parallel()
mockCtrl := gomock.NewController(t)
defer mockCtrl.Finish()
scopeMock := mock_privatedns.NewMockScope(mockCtrl)
clientMock := mock_privatedns.NewMockclient(mockCtrl)
tc.expect(scopeMock.EXPECT(), clientMock.EXPECT())
s := &Service{
Scope: scopeMock,
client: clientMock,
}
err := s.Reconcile(context.TODO())
if tc.expectedError != "" {
g.Expect(err).To(HaveOccurred())
g.Expect(err).To(MatchError(tc.expectedError))
} else {
g.Expect(err).NotTo(HaveOccurred())
}
})
}
} | explode_data.jsonl/70329 | {
"file_path": "/home/dung/Study/Code/Cross_test_gen/training_dataset/dedup_data/clean_data_go/data/explode_data.jsonl",
"token_count": 2724
} | [
2830,
3393,
693,
40446,
457,
16787,
61088,
1155,
353,
8840,
836,
8,
341,
18185,
23910,
1669,
3056,
1235,
341,
197,
11609,
688,
914,
198,
197,
42400,
1454,
914,
198,
197,
24952,
286,
2915,
1141,
353,
16712,
24726,
657,
4412,
24664,
10803... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | 1 |
func TestAvailableDevices(t *testing.T) {
executor := &exectest.MockExecutor{}
// set up a mock function to return "rook owned" partitions on the device and it does not have a filesystem
executor.MockExecuteCommandWithOutput = func(debug bool, name string, command string, args ...string) (string, error) {
logger.Infof("OUTPUT for %s. %s %+v", name, command, args)
if command == "lsblk" {
if strings.Index(name, "sdb") != -1 {
// /dev/sdb has a partition
return `NAME="sdb" SIZE="65" TYPE="disk" PKNAME=""
NAME="sdb1" SIZE="30" TYPE="part" PKNAME="sdb"`, nil
}
return "", nil
} else if command == "blkid" {
if strings.Index(name, "sdb1") != -1 {
// partition sdb1 has a label MY-PART
return "MY-PART", nil
}
} else if command == "df" {
if strings.Index(name, "sdc") != -1 {
// /dev/sdc has a file system
return "/dev/sdc ext4", nil
}
return "", nil
}
return "", fmt.Errorf("unknown command %s %+v", command, args)
}
context := &clusterd.Context{Executor: executor}
context.Devices = []*clusterd.LocalDisk{
{Name: "sda"},
{Name: "sdb"},
{Name: "sdc"},
{Name: "sdd"},
{Name: "nvme01"},
{Name: "rda"},
{Name: "rdb"},
}
// select all devices, including nvme01 for metadata
mapping, err := getAvailableDevices(context, "all", "nvme01", true)
assert.Nil(t, err)
assert.Equal(t, 5, len(mapping.Entries))
assert.Equal(t, -1, mapping.Entries["sda"].Data)
assert.Equal(t, -1, mapping.Entries["sdd"].Data)
assert.Equal(t, -1, mapping.Entries["rda"].Data)
assert.Equal(t, -1, mapping.Entries["rdb"].Data)
assert.Equal(t, -1, mapping.Entries["nvme01"].Data)
assert.NotNil(t, mapping.Entries["nvme01"].Metadata)
assert.Equal(t, 0, len(mapping.Entries["nvme01"].Metadata))
// select no devices both using and not using a filter
mapping, err = getAvailableDevices(context, "", "", false)
assert.Nil(t, err)
assert.Equal(t, 0, len(mapping.Entries))
mapping, err = getAvailableDevices(context, "", "", true)
assert.Nil(t, err)
assert.Equal(t, 0, len(mapping.Entries))
// select the sd* devices
mapping, err = getAvailableDevices(context, "^sd.$", "", true)
assert.Nil(t, err)
assert.Equal(t, 2, len(mapping.Entries))
assert.Equal(t, -1, mapping.Entries["sda"].Data)
assert.Equal(t, -1, mapping.Entries["sdd"].Data)
// select an exact device
mapping, err = getAvailableDevices(context, "sdd", "", false)
assert.Nil(t, err)
assert.Equal(t, 1, len(mapping.Entries))
assert.Equal(t, -1, mapping.Entries["sdd"].Data)
// select all devices except those that have a prefix of "s"
mapping, err = getAvailableDevices(context, "^[^s]", "", true)
assert.Nil(t, err)
assert.Equal(t, 3, len(mapping.Entries))
assert.Equal(t, -1, mapping.Entries["rda"].Data)
assert.Equal(t, -1, mapping.Entries["rdb"].Data)
assert.Equal(t, -1, mapping.Entries["nvme01"].Data)
} | explode_data.jsonl/53727 | {
"file_path": "/home/dung/Study/Code/Cross_test_gen/training_dataset/dedup_data/clean_data_go/data/explode_data.jsonl",
"token_count": 1179
} | [
2830,
3393,
16485,
40835,
1155,
353,
8840,
836,
8,
341,
67328,
4831,
1669,
609,
327,
439,
477,
24664,
25255,
16094,
197,
322,
738,
705,
264,
7860,
729,
311,
470,
330,
299,
562,
12938,
1,
46688,
389,
279,
3671,
323,
432,
1558,
537,
6... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | 7 |
func TestDnsStatisticsRecord(t *testing.T) {
config := dnsutils.GetFakeConfig()
stats := NewStatsPerStream(config, "test")
dm := dnsutils.DnsMessage{}
dm.Init()
dm.DNS.Type = dnsutils.DnsQuery
dm.NetworkInfo.Family = "INET"
dm.NetworkInfo.Protocol = "UDP"
dm.DNS.Qname = "dnscollector.test."
stats.Record(dm)
nb := stats.GetTotalDomains()
if nb != 1 {
t.Errorf("invalid number of domains, expected 1, got %d", nb)
}
} | explode_data.jsonl/75653 | {
"file_path": "/home/dung/Study/Code/Cross_test_gen/training_dataset/dedup_data/clean_data_go/data/explode_data.jsonl",
"token_count": 177
} | [
2830,
3393,
35,
4412,
38599,
6471,
1155,
353,
8840,
836,
8,
341,
25873,
1669,
44077,
6031,
2234,
52317,
2648,
741,
79659,
1669,
1532,
16635,
3889,
3027,
8754,
11,
330,
1944,
5130,
2698,
76,
1669,
44077,
6031,
909,
4412,
2052,
16094,
269... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | 2 |
func TestRuleError(t *testing.T) {
tests := []struct {
in ticketdb.DBError
want string
}{
{ticketdb.DBError{Description: "duplicate block"},
"duplicate block",
},
{ticketdb.DBError{Description: "human-readable error"},
"human-readable error",
},
}
t.Logf("Running %d tests", len(tests))
for i, test := range tests {
result := test.in.Error()
if result != test.want {
t.Errorf("Error #%d\n got: %s want: %s", i, result,
test.want)
continue
}
}
} | explode_data.jsonl/80670 | {
"file_path": "/home/dung/Study/Code/Cross_test_gen/training_dataset/dedup_data/clean_data_go/data/explode_data.jsonl",
"token_count": 203
} | [
2830,
3393,
11337,
1454,
1155,
353,
8840,
836,
8,
341,
78216,
1669,
3056,
1235,
341,
197,
17430,
256,
11727,
1999,
22537,
1454,
198,
197,
50780,
914,
198,
197,
59403,
197,
197,
90,
26534,
1999,
22537,
1454,
90,
5009,
25,
330,
63826,
2... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | 3 |
func TestReconcileWorkspaceMissing(t *testing.T) {
taskWithWorkspace := tb.Task("test-task-with-workspace",
tb.TaskSpec(
tb.TaskWorkspace("ws1", "a test task workspace", "", true),
), tb.TaskNamespace("foo"))
taskRun := tb.TaskRun("test-taskrun-missing-workspace", tb.TaskRunNamespace("foo"), tb.TaskRunSpec(
tb.TaskRunTaskRef(taskWithWorkspace.Name, tb.TaskRefAPIVersion("a1")),
))
d := test.Data{
Tasks: []*v1alpha1.Task{taskWithWorkspace},
TaskRuns: []*v1alpha1.TaskRun{taskRun},
ClusterTasks: nil,
PipelineResources: nil,
}
names.TestingSeed()
testAssets, cancel := getTaskRunController(t, d)
defer cancel()
clients := testAssets.Clients
if err := testAssets.Controller.Reconciler.Reconcile(context.Background(), getRunName(taskRun)); err != nil {
t.Errorf("expected no error reconciling valid TaskRun but got %v", err)
}
tr, err := clients.Pipeline.TektonV1alpha1().TaskRuns(taskRun.Namespace).Get(taskRun.Name, metav1.GetOptions{})
if err != nil {
t.Fatalf("Expected TaskRun %s to exist but instead got error when getting it: %v", taskRun.Name, err)
}
failedCorrectly := false
for _, c := range tr.Status.Conditions {
if c.Type == apis.ConditionSucceeded && c.Status == corev1.ConditionFalse && c.Reason == podconvert.ReasonFailedValidation {
failedCorrectly = true
}
}
if !failedCorrectly {
t.Errorf("Expected TaskRun to fail validation but it did not. Final conditions were:\n%#v", tr.Status.Conditions)
}
} | explode_data.jsonl/895 | {
"file_path": "/home/dung/Study/Code/Cross_test_gen/training_dataset/dedup_data/clean_data_go/data/explode_data.jsonl",
"token_count": 575
} | [
2830,
3393,
693,
40446,
457,
45981,
25080,
1155,
353,
8840,
836,
8,
341,
49115,
2354,
45981,
1669,
16363,
28258,
445,
1944,
52579,
26189,
28621,
8746,
756,
197,
62842,
28258,
8327,
1006,
298,
62842,
28258,
45981,
445,
8915,
16,
497,
330,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | 8 |
func TestEncoderWriterStruct1(t *testing.T) {
var buf bytes.Buffer
e := velocypack.NewEncoder(&buf)
for i := 0; i < 1000; i++ {
input := Struct1{
Field1: i,
}
must(e.Encode(input))
}
r := bytes.NewReader(buf.Bytes())
d := velocypack.NewDecoder(r)
for i := 0; i < 1000; i++ {
var v Struct1
must(d.Decode(&v))
expected := Struct1{
Field1: i,
}
ASSERT_EQ(v, expected, t)
}
} | explode_data.jsonl/52698 | {
"file_path": "/home/dung/Study/Code/Cross_test_gen/training_dataset/dedup_data/clean_data_go/data/explode_data.jsonl",
"token_count": 189
} | [
2830,
3393,
19921,
6492,
9422,
16,
1155,
353,
8840,
836,
8,
341,
2405,
6607,
5820,
22622,
198,
7727,
1669,
40509,
1082,
473,
7121,
19921,
2099,
5909,
340,
2023,
600,
1669,
220,
15,
26,
600,
366,
220,
16,
15,
15,
15,
26,
600,
1027,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | 3 |
func TestPythonInterop(t *testing.T) {
var r Record
if err := rlp.DecodeBytes(pyRecord, &r); err != nil {
t.Fatalf("can't decode: %v", err)
}
var (
wantAddr, _ = hex.DecodeString("caaa1485d83b18b32ed9ad666026151bf0cae8a0a88c857ae2d4c5be2daa6726")
wantSeq = uint64(1)
wantIP = IP4{127, 0, 0, 1}
wantDiscport = DiscPort(60606)
)
if r.Seq() != wantSeq {
t.Errorf("wrong seq: got %d, want %d", r.Seq(), wantSeq)
}
if addr := r.NodeAddr(); !bytes.Equal(addr, wantAddr) {
t.Errorf("wrong addr: got %x, want %x", addr, wantAddr)
}
want := map[Entry]interface{}{new(IP4): &wantIP, new(DiscPort): &wantDiscport}
for k, v := range want {
desc := fmt.Sprintf("loading key %q", k.ENRKey())
if assert.NoError(t, r.Load(k), desc) {
assert.Equal(t, k, v, desc)
}
}
} | explode_data.jsonl/39495 | {
"file_path": "/home/dung/Study/Code/Cross_test_gen/training_dataset/dedup_data/clean_data_go/data/explode_data.jsonl",
"token_count": 381
} | [
2830,
3393,
30280,
94000,
1155,
353,
8840,
836,
8,
341,
2405,
435,
13583,
198,
743,
1848,
1669,
435,
13545,
56372,
7078,
46827,
6471,
11,
609,
81,
1215,
1848,
961,
2092,
341,
197,
3244,
30762,
445,
4814,
944,
16895,
25,
1018,
85,
497,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | 6 |
func TestGreaterThanOrEqualToMatcherDatetime(t *testing.T) {
logger := logging.NewLogger(&logging.LoggerOptions{})
attrName := "value"
dto := &dtos.MatcherDTO{
MatcherType: "GREATER_THAN_OR_EQUAL_TO",
UnaryNumeric: &dtos.UnaryNumericMatcherDataDTO{
DataType: "DATETIME",
Value: int64(960293532000), // 06/06/2000
},
KeySelector: &dtos.KeySelectorDTO{
Attribute: &attrName,
},
}
matcher, err := BuildMatcher(dto, nil, logger)
if err != nil {
t.Error("There should be no errors when building the matcher")
t.Error(err)
}
matcherType := reflect.TypeOf(matcher).String()
if matcherType != "*matchers.GreaterThanOrEqualToMatcher" {
t.Errorf("Incorrect matcher constructed. Should be *matchers.GreaterThanOrEqualToMatcher and was %s", matcherType)
}
attributes := make(map[string]interface{})
attributes["value"] = int64(960293532)
if !matcher.Match("asd", attributes, nil) {
t.Error("Equal should match")
}
attributes["value"] = int64(1275782400)
if !matcher.Match("asd", attributes, nil) {
t.Error("Greater should match")
}
attributes["value"] = int64(293532000)
if matcher.Match("asd", attributes, nil) {
t.Error("Lower should NOT match")
}
} | explode_data.jsonl/60371 | {
"file_path": "/home/dung/Study/Code/Cross_test_gen/training_dataset/dedup_data/clean_data_go/data/explode_data.jsonl",
"token_count": 479
} | [
2830,
3393,
41366,
89387,
37554,
94191,
1155,
353,
8840,
836,
8,
341,
17060,
1669,
8392,
7121,
7395,
2099,
25263,
12750,
3798,
37790,
60943,
675,
1669,
330,
957,
698,
98864,
1669,
609,
8047,
436,
76452,
14923,
515,
197,
197,
37554,
929,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | 6 |
func TestNilServiceResolvers(t *testing.T) {
ctx := logger.WithLogger(broker.NewBackground())
resolvers := ServiceResolvers{nil}
_, err := resolvers.Resolve(ctx)
if err != nil {
t.Fatal(err)
}
} | explode_data.jsonl/2209 | {
"file_path": "/home/dung/Study/Code/Cross_test_gen/training_dataset/dedup_data/clean_data_go/data/explode_data.jsonl",
"token_count": 82
} | [
2830,
3393,
19064,
1860,
1061,
39435,
1155,
353,
8840,
836,
8,
341,
20985,
1669,
5925,
26124,
7395,
1883,
45985,
7121,
8706,
2398,
10202,
39435,
1669,
5362,
1061,
39435,
90,
8385,
532,
197,
6878,
1848,
1669,
592,
39435,
57875,
7502,
340,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1
] | 2 |
func TestIgnorePackages_whenTypical(t *testing.T) {
assert := testifyassert.New(t)
arbitraryPackages := []string{"abc", "xyz/abc"}
actual := IgnorePackages(arbitraryPackages)
assert.Equal(arbitraryPackages, actual)
} | explode_data.jsonl/73324 | {
"file_path": "/home/dung/Study/Code/Cross_test_gen/training_dataset/dedup_data/clean_data_go/data/explode_data.jsonl",
"token_count": 79
} | [
2830,
3393,
12497,
69513,
47636,
12834,
938,
1155,
353,
8840,
836,
8,
341,
6948,
1669,
48162,
2207,
7121,
1155,
692,
69340,
87851,
69513,
1669,
3056,
917,
4913,
13683,
497,
330,
28854,
14,
13683,
63159,
88814,
1669,
38971,
69513,
37544,
8... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1
] | 1 |
func TestFeeStats(t *testing.T) {
hmock := httptest.NewClient()
client := &Client{
HorizonURL: "https://localhost/",
HTTP: hmock,
}
// happy path
hmock.On(
"GET",
"https://localhost/fee_stats",
).ReturnString(200, feesResponse)
fees, err := client.FeeStats()
if assert.NoError(t, err) {
assert.Equal(t, fees.LastLedger, 22606298)
assert.Equal(t, fees.LastLedgerBaseFee, 100)
assert.Equal(t, fees.LedgerCapacityUsage, 0.97)
assert.Equal(t, fees.MinAcceptedFee, 130)
assert.Equal(t, fees.ModeAcceptedFee, 250)
assert.Equal(t, fees.P10AcceptedFee, 150)
assert.Equal(t, fees.P20AcceptedFee, 200)
assert.Equal(t, fees.P30AcceptedFee, 300)
assert.Equal(t, fees.P40AcceptedFee, 400)
assert.Equal(t, fees.P50AcceptedFee, 500)
assert.Equal(t, fees.P60AcceptedFee, 1000)
assert.Equal(t, fees.P70AcceptedFee, 2000)
assert.Equal(t, fees.P80AcceptedFee, 3000)
assert.Equal(t, fees.P90AcceptedFee, 4000)
assert.Equal(t, fees.P95AcceptedFee, 5000)
assert.Equal(t, fees.P99AcceptedFee, 8000)
}
// connection error
hmock.On(
"GET",
"https://localhost/metrics",
).ReturnError("http.Client error")
_, err = client.Metrics()
if assert.Error(t, err) {
assert.Contains(t, err.Error(), "http.Client error")
_, ok := err.(*Error)
assert.Equal(t, ok, false)
}
} | explode_data.jsonl/34857 | {
"file_path": "/home/dung/Study/Code/Cross_test_gen/training_dataset/dedup_data/clean_data_go/data/explode_data.jsonl",
"token_count": 581
} | [
2830,
3393,
41941,
16635,
1155,
353,
8840,
836,
8,
341,
9598,
16712,
1669,
54320,
70334,
7121,
2959,
741,
25291,
1669,
609,
2959,
515,
197,
13292,
269,
16973,
3144,
25,
330,
2428,
1110,
8301,
35075,
197,
197,
9230,
25,
981,
305,
16712,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | 3 |
func Test_Mock_AssertCalled_WithArguments(t *testing.T) {
var mockedService = new(TestExampleImplementation)
mockedService.On("Test_Mock_AssertCalled_WithArguments", 1, 2, 3).Return(5, 6, 7)
mockedService.Called(1, 2, 3)
tt := new(testing.T)
assert.True(t, mockedService.AssertCalled(tt, "Test_Mock_AssertCalled_WithArguments", 1, 2, 3))
assert.False(t, mockedService.AssertCalled(tt, "Test_Mock_AssertCalled_WithArguments", 2, 3, 4))
} | explode_data.jsonl/8609 | {
"file_path": "/home/dung/Study/Code/Cross_test_gen/training_dataset/dedup_data/clean_data_go/data/explode_data.jsonl",
"token_count": 193
} | [
2830,
3393,
1245,
1176,
62222,
529,
20960,
62,
2354,
19139,
1155,
353,
8840,
836,
8,
8022,
2405,
46149,
1860,
284,
501,
31159,
13314,
36850,
7229,
2109,
67385,
1860,
8071,
445,
2271,
1245,
1176,
62222,
529,
20960,
62,
2354,
19139,
497,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | 1 |
func TestGetRelation(t *testing.T) {
t.Parallel()
resolver := newMockResolver()
expectedParams := &LocalGetClassParams{
Kind: kind.ACTION_KIND,
ClassName: "SomeAction",
Properties: []SelectProperty{
{
Name: "HasAction",
IsPrimitive: false,
Refs: []SelectClass{
{
ClassName: "SomeAction",
RefProperties: []SelectProperty{
{
Name: "intField",
IsPrimitive: true,
},
{
Name: "HasAction",
IsPrimitive: false,
Refs: []SelectClass{
{
ClassName: "SomeAction",
RefProperties: []SelectProperty{
{
Name: "intField",
IsPrimitive: true,
},
},
},
},
},
},
},
},
},
},
}
resolver.On("LocalGetClass", expectedParams).
Return(test_helper.EmptyListThunk(), nil).Once()
query := "{ Get { Actions { SomeAction { HasAction { ... on SomeAction { intField, HasAction { ... on SomeAction { intField } } } } } } } }"
resolver.AssertResolve(t, query)
} | explode_data.jsonl/10590 | {
"file_path": "/home/dung/Study/Code/Cross_test_gen/training_dataset/dedup_data/clean_data_go/data/explode_data.jsonl",
"token_count": 566
} | [
2830,
3393,
1949,
33790,
1155,
353,
8840,
836,
8,
341,
3244,
41288,
7957,
2822,
10202,
7921,
1669,
501,
11571,
18190,
2822,
42400,
4870,
1669,
609,
7319,
1949,
1957,
4870,
515,
197,
197,
10629,
25,
414,
3093,
28934,
72959,
345,
197,
197... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | 1 |
func TestFindIndexOfEmail_Successful(t *testing.T) {
x := []string{"testingATtest.com",
"testing@Test.com",
"testingATtest.com",
"testingATtest.com"}
got := FindIndexOfEmail(x)
want := 1
if got != want {
t.Errorf("Got: %d, Wanted: %d", got, want)
}
} | explode_data.jsonl/45939 | {
"file_path": "/home/dung/Study/Code/Cross_test_gen/training_dataset/dedup_data/clean_data_go/data/explode_data.jsonl",
"token_count": 110
} | [
2830,
3393,
9885,
27376,
4781,
87161,
1262,
1155,
353,
8840,
836,
8,
341,
10225,
1669,
3056,
917,
4913,
8840,
828,
1944,
905,
756,
197,
197,
42097,
18229,
905,
756,
197,
197,
42097,
828,
1944,
905,
756,
197,
197,
42097,
828,
1944,
905... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | 2 |
func TestReturns200IfThereAreNoChecks(t *testing.T) {
recorder := httptest.NewRecorder()
req, err := http.NewRequest("GET", "https://fakeurl.com/debug/health", nil)
if err != nil {
t.Errorf("Failed to create request.")
}
StatusHandler(recorder, req)
if recorder.Code != 200 {
t.Errorf("Did not get a 200.")
}
} | explode_data.jsonl/56023 | {
"file_path": "/home/dung/Study/Code/Cross_test_gen/training_dataset/dedup_data/clean_data_go/data/explode_data.jsonl",
"token_count": 120
} | [
2830,
3393,
16446,
17,
15,
15,
2679,
3862,
11526,
2753,
49820,
1155,
353,
8840,
836,
8,
341,
67904,
1358,
1669,
54320,
70334,
7121,
47023,
2822,
24395,
11,
1848,
1669,
1758,
75274,
445,
3806,
497,
330,
2428,
1110,
30570,
1085,
905,
6747... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | 3 |
func TestJvInvalidWithMessage_string(t *testing.T) {
errMsg := "Error message 1"
jv := jq.JvInvalidWithMessage(jq.JvFromString(errMsg))
if jv.IsValid() == true {
t.Errorf("IsValid() returned true for JvInvalidWithMessage()")
}
msg := jv.Copy().GetInvalidMessage()
if msg.Kind() != jq.JvKindString {
t.Errorf("JvInvalidWithMessage().GetInvalidMessage().Kind() returned a kind other than JvKindString")
}
msg.Free()
str, ok := jv.GetInvalidMessageAsString()
if !ok {
t.Errorf("JvInvalidWithMessage().JvGetInvalidMessageAsString() is not ok")
}
if str != errMsg {
t.Errorf("JvInvalidWithMessage().JvGetInvalidMessageAsString() did not return original error message")
}
} | explode_data.jsonl/38761 | {
"file_path": "/home/dung/Study/Code/Cross_test_gen/training_dataset/dedup_data/clean_data_go/data/explode_data.jsonl",
"token_count": 248
} | [
2830,
3393,
41,
85,
7928,
2354,
2052,
3904,
1155,
353,
8840,
836,
8,
341,
9859,
6611,
1669,
330,
1454,
1943,
220,
16,
698,
12428,
85,
1669,
44648,
3503,
85,
7928,
2354,
2052,
96887,
3503,
85,
44491,
3964,
6611,
1171,
743,
502,
85,
2... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | 5 |
func TestDeleteByQueryDocument(t *testing.T) {
c := &internal.MockedConnection{
MockSend: func(query []byte, options types.QueryOptions) *types.KuzzleResponse {
parsedQuery := &types.KuzzleRequest{}
json.Unmarshal(query, parsedQuery)
assert.Equal(t, "document", parsedQuery.Controller)
assert.Equal(t, "deleteByQuery", parsedQuery.Action)
assert.Equal(t, "index", parsedQuery.Index)
assert.Equal(t, "collection", parsedQuery.Collection)
return &types.KuzzleResponse{Result: []byte(`
{
"hits": ["id1", "id2"]
}`),
}
},
}
k, _ := kuzzle.NewKuzzle(c, nil)
d := document.NewDocument(k)
_, err := d.DeleteByQuery("index", "collection", json.RawMessage(`{"foo": "bar"}`), nil)
assert.Nil(t, err)
} | explode_data.jsonl/75168 | {
"file_path": "/home/dung/Study/Code/Cross_test_gen/training_dataset/dedup_data/clean_data_go/data/explode_data.jsonl",
"token_count": 295
} | [
2830,
3393,
6435,
1359,
2859,
7524,
1155,
353,
8840,
836,
8,
1476,
1444,
1669,
609,
10481,
24664,
291,
4526,
515,
197,
9209,
1176,
11505,
25,
2915,
10741,
3056,
3782,
11,
2606,
4494,
15685,
3798,
8,
353,
9242,
11352,
14945,
2582,
341,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | 1 |
func TestCancelOrder(t *testing.T) {
_, err := b.CancelExistingOrder([]int64{1337})
if err == nil {
t.Error("Test failed - CancelgOrder() error", err)
}
} | explode_data.jsonl/48999 | {
"file_path": "/home/dung/Study/Code/Cross_test_gen/training_dataset/dedup_data/clean_data_go/data/explode_data.jsonl",
"token_count": 64
} | [
2830,
3393,
9269,
4431,
1155,
353,
8840,
836,
8,
341,
197,
6878,
1848,
1669,
293,
36491,
53067,
4431,
10556,
396,
21,
19,
90,
16,
18,
18,
22,
8824,
743,
1848,
621,
2092,
341,
197,
3244,
6141,
445,
2271,
4641,
481,
23542,
70,
4431,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1
] | 2 |
func TestShimDefaults(t *testing.T) {
assert := assert.New(t)
tmpdir, err := ioutil.TempDir(testDir, "")
assert.NoError(err)
defer os.RemoveAll(tmpdir)
testShimPath := filepath.Join(tmpdir, "shim")
testShimLinkPath := filepath.Join(tmpdir, "shim-link")
err = createEmptyFile(testShimPath)
assert.NoError(err)
err = syscall.Symlink(testShimPath, testShimLinkPath)
assert.NoError(err)
savedShimPath := defaultShimPath
defer func() {
defaultShimPath = savedShimPath
}()
defaultShimPath = testShimPath
s := shim{}
p, err := s.path()
assert.NoError(err)
assert.Equal(p, defaultShimPath, "default shim path wrong")
// test path resolution
defaultShimPath = testShimLinkPath
s = shim{}
p, err = s.path()
assert.NoError(err)
assert.Equal(p, testShimPath)
assert.False(s.debug())
s.Debug = true
assert.True(s.debug())
} | explode_data.jsonl/5131 | {
"file_path": "/home/dung/Study/Code/Cross_test_gen/training_dataset/dedup_data/clean_data_go/data/explode_data.jsonl",
"token_count": 341
} | [
2830,
3393,
2016,
318,
16273,
1155,
353,
8840,
836,
8,
341,
6948,
1669,
2060,
7121,
1155,
692,
20082,
3741,
11,
1848,
1669,
43144,
65009,
6184,
8623,
6184,
11,
14676,
6948,
35699,
3964,
340,
16867,
2643,
84427,
10368,
3741,
692,
18185,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | 1 |
End of preview. Expand
in Data Studio
README.md exists but content is empty.
- Downloads last month
- 45