2017-04-25 16:24:51 +09:00
|
|
|
// Copyright 2017 The Gitea Authors. All rights reserved.
|
2024-01-12 19:50:29 +09:00
|
|
|
// Copyright 2024 The Forgejo Authors c/o Codeberg e.V.. All rights reserved.
|
2022-11-28 03:20:29 +09:00
|
|
|
// SPDX-License-Identifier: MIT
|
2017-04-25 16:24:51 +09:00
|
|
|
|
Enable forbidigo linter (#24278)
Enable [forbidigo](https://github.com/ashanbrown/forbidigo) linter which
forbids print statements. Will check how to integrate this with the
smallest impact possible, so a few `nolint` comments will likely be
required. Plan is to just go through the issues and either:
- Remove the print if it is nonsensical
- Add a `//nolint` directive if it makes sense
I don't plan on investigating the individual issues any further.
<details>
<summary>Initial Lint Results</summary>
```
modules/log/event.go:348:6: use of `fmt.Println` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Println(err)
^
modules/log/event.go:382:6: use of `fmt.Println` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Println(err)
^
modules/queue/unique_queue_disk_channel_test.go:20:2: use of `fmt.Printf` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Printf("TempDir %s\n", tmpDir)
^
contrib/backport/backport.go:168:2: use of `fmt.Printf` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Printf("* Backporting %s to %s as %s\n", pr, localReleaseBranch, backportBranch)
^
contrib/backport/backport.go:216:4: use of `fmt.Printf` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Printf("* Navigate to %s to open PR\n", url)
^
contrib/backport/backport.go:223:2: use of `fmt.Printf` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Printf("* `xdg-open %s`\n", url)
^
contrib/backport/backport.go:233:2: use of `fmt.Printf` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Printf("* `git push -u %s %s`\n", remote, backportBranch)
^
contrib/backport/backport.go:243:2: use of `fmt.Printf` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Printf("* Amending commit to prepend `Backport #%s` to body\n", pr)
^
contrib/backport/backport.go:272:3: use of `fmt.Println` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Println("* Attempting git cherry-pick --continue")
^
contrib/backport/backport.go:281:2: use of `fmt.Printf` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Printf("* Attempting git cherry-pick %s\n", sha)
^
contrib/backport/backport.go:297:2: use of `fmt.Printf` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Printf("* Current branch is %s\n", currentBranch)
^
contrib/backport/backport.go:299:3: use of `fmt.Printf` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Printf("* Current branch is %s - not checking out\n", currentBranch)
^
contrib/backport/backport.go:304:3: use of `fmt.Printf` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Printf("* Branch %s already exists. Checking it out...\n", backportBranch)
^
contrib/backport/backport.go:308:2: use of `fmt.Printf` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Printf("* `git checkout -b %s %s`\n", backportBranch, releaseBranch)
^
contrib/backport/backport.go:313:2: use of `fmt.Printf` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Printf("* `git fetch %s main`\n", remote)
^
contrib/backport/backport.go:316:3: use of `fmt.Println` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Println(string(out))
^
contrib/backport/backport.go:319:2: use of `fmt.Println` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Println(string(out))
^
contrib/backport/backport.go:321:2: use of `fmt.Printf` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Printf("* `git fetch %s %s`\n", remote, releaseBranch)
^
contrib/backport/backport.go:324:3: use of `fmt.Println` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Println(string(out))
^
contrib/backport/backport.go:327:2: use of `fmt.Println` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Println(string(out))
^
models/unittest/fixtures.go:50:3: use of `fmt.Println` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Println("Unsupported RDBMS for integration tests")
^
models/unittest/fixtures.go:89:3: use of `fmt.Printf` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Printf("LoadFixtures failed after retries: %v\n", err)
^
models/unittest/fixtures.go:110:4: use of `fmt.Printf` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Printf("Failed to generate sequence update: %v\n", err)
^
models/unittest/fixtures.go:117:6: use of `fmt.Printf` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Printf("Failed to update sequence: %s Error: %v\n", value, err)
^
models/migrations/base/tests.go:118:3: use of `fmt.Println` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Println("Environment variable $GITEA_ROOT not set")
^
models/migrations/base/tests.go:127:3: use of `fmt.Printf` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Printf("Could not find gitea binary at %s\n", setting.AppPath)
^
models/migrations/base/tests.go:134:3: use of `fmt.Printf` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Printf("Environment variable $GITEA_CONF not set - defaulting to %s\n", giteaConf)
^
models/migrations/base/tests.go:145:3: use of `fmt.Printf` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Printf("Unable to create temporary data path %v\n", err)
^
models/migrations/base/tests.go:154:3: use of `fmt.Printf` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Printf("Unable to InitFull: %v\n", err)
^
models/migrations/v1_11/v112.go:34:5: use of `fmt.Printf` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Printf("Error: %v", err)
^
contrib/fixtures/fixture_generation.go:36:3: use of `fmt.Printf` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Printf("CreateTestEngine: %+v", err)
^
contrib/fixtures/fixture_generation.go:40:3: use of `fmt.Printf` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Printf("PrepareTestDatabase: %+v\n", err)
^
contrib/fixtures/fixture_generation.go:46:5: use of `fmt.Printf` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Printf("generate '%s': %+v\n", r, err)
^
contrib/fixtures/fixture_generation.go:53:5: use of `fmt.Printf` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Printf("generate '%s': %+v\n", g.name, err)
^
contrib/fixtures/fixture_generation.go:71:4: use of `fmt.Printf` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Printf("%s created.\n", path)
^
services/gitdiff/gitdiff_test.go:543:2: use of `println` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
println(result)
^
services/gitdiff/gitdiff_test.go:560:2: use of `println` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
println(result)
^
services/gitdiff/gitdiff_test.go:577:2: use of `println` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
println(result)
^
modules/web/routing/logger_manager.go:34:2: use of `print` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
print Printer
^
modules/doctor/paths.go:109:3: use of `fmt.Printf` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Printf("Warning: can't remove temporary file: '%s'\n", tmpFile.Name())
^
tests/test_utils.go:33:2: use of `fmt.Printf` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Printf(format+"\n", args...)
^
tests/test_utils.go:61:3: use of `fmt.Printf` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Printf("Environment variable $GITEA_CONF not set, use default: %s\n", giteaConf)
^
cmd/actions.go:54:9: use of `fmt.Printf` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
_, _ = fmt.Printf("%s\n", respText)
^
cmd/admin_user_change_password.go:74:2: use of `fmt.Printf` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Printf("%s's password has been successfully updated!\n", user.Name)
^
cmd/admin_user_create.go:109:3: use of `fmt.Printf` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Printf("generated random password is '%s'\n", password)
^
cmd/admin_user_create.go:164:3: use of `fmt.Printf` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Printf("Access token was successfully created... %s\n", t.Token)
^
cmd/admin_user_create.go:167:2: use of `fmt.Printf` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Printf("New user '%s' has been successfully created!\n", username)
^
cmd/admin_user_generate_access_token.go:74:3: use of `fmt.Printf` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Printf("%s\n", t.Token)
^
cmd/admin_user_generate_access_token.go:76:3: use of `fmt.Printf` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Printf("Access token was successfully created: %s\n", t.Token)
^
cmd/admin_user_must_change_password.go:56:2: use of `fmt.Printf` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Printf("Updated %d users setting MustChangePassword to %t\n", n, mustChangePassword)
^
cmd/convert.go:44:3: use of `fmt.Println` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Println("Converted successfully, please confirm your database's character set is now utf8mb4")
^
cmd/convert.go:50:3: use of `fmt.Println` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Println("Converted successfully, please confirm your database's all columns character is NVARCHAR now")
^
cmd/convert.go:52:3: use of `fmt.Println` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Println("This command can only be used with a MySQL or MSSQL database")
^
cmd/doctor.go:104:3: use of `fmt.Println` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Println(err)
^
cmd/doctor.go:105:3: use of `fmt.Println` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Println("Check if you are using the right config file. You can use a --config directive to specify one.")
^
cmd/doctor.go:243:3: use of `fmt.Println` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Println(err)
^
cmd/embedded.go:154:3: use of `fmt.Println` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Println(a.path)
^
cmd/embedded.go:198:3: use of `fmt.Println` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Println("Using app.ini at", setting.CustomConf)
^
cmd/embedded.go:217:2: use of `fmt.Printf` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Printf("Extracting to %s:\n", destdir)
^
cmd/embedded.go:253:3: use of `fmt.Printf` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Printf("%s already exists; skipped.\n", dest)
^
cmd/embedded.go:275:2: use of `fmt.Println` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Println(dest)
^
cmd/generate.go:63:2: use of `fmt.Printf` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Printf("%s", internalToken)
^
cmd/generate.go:66:3: use of `fmt.Printf` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Printf("\n")
^
cmd/generate.go:78:2: use of `fmt.Printf` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Printf("%s", JWTSecretBase64)
^
cmd/generate.go:81:3: use of `fmt.Printf` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Printf("\n")
^
cmd/generate.go:93:2: use of `fmt.Printf` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Printf("%s", secretKey)
^
cmd/generate.go:96:3: use of `fmt.Printf` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Printf("\n")
^
cmd/keys.go:74:2: use of `fmt.Println` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Println(strings.TrimSpace(authorizedString))
^
cmd/mailer.go:32:4: use of `fmt.Print` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Print("warning: Content is empty")
^
cmd/mailer.go:35:3: use of `fmt.Print` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Print("Proceed with sending email? [Y/n] ")
^
cmd/mailer.go:40:4: use of `fmt.Println` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Println("The mail was not sent")
^
cmd/mailer.go:49:9: use of `fmt.Printf` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
_, _ = fmt.Printf("Sent %s email(s) to all users\n", respText)
^
cmd/serv.go:147:3: use of `println` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
println("Gitea: SSH has been disabled")
^
cmd/serv.go:153:4: use of `fmt.Printf` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Printf("error showing subcommand help: %v\n", err)
^
cmd/serv.go:175:4: use of `println` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
println("Hi there! You've successfully authenticated with the deploy key named " + key.Name + ", but Gitea does not provide shell access.")
^
cmd/serv.go:177:4: use of `println` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
println("Hi there! You've successfully authenticated with the principal " + key.Content + ", but Gitea does not provide shell access.")
^
cmd/serv.go:179:4: use of `println` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
println("Hi there, " + user.Name + "! You've successfully authenticated with the key named " + key.Name + ", but Gitea does not provide shell access.")
^
cmd/serv.go:181:3: use of `println` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
println("If this is unexpected, please log in with password and setup Gitea under another user.")
^
cmd/serv.go:196:5: use of `fmt.Print` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Print(`{"type":"gitea","version":1}`)
^
tests/e2e/e2e_test.go:54:3: use of `fmt.Printf` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Printf("Error initializing test database: %v\n", err)
^
tests/e2e/e2e_test.go:63:3: use of `fmt.Printf` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Printf("util.RemoveAll: %v\n", err)
^
tests/e2e/e2e_test.go:67:3: use of `fmt.Printf` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Printf("Unable to remove repo indexer: %v\n", err)
^
tests/e2e/e2e_test.go:109:6: use of `fmt.Printf` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Printf("%v", stdout.String())
^
tests/e2e/e2e_test.go:110:6: use of `fmt.Printf` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Printf("%v", stderr.String())
^
tests/e2e/e2e_test.go:113:6: use of `fmt.Printf` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Printf("%v", stdout.String())
^
tests/integration/integration_test.go:124:3: use of `fmt.Printf` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Printf("Error initializing test database: %v\n", err)
^
tests/integration/integration_test.go:135:3: use of `fmt.Printf` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Printf("util.RemoveAll: %v\n", err)
^
tests/integration/integration_test.go:139:3: use of `fmt.Printf` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Printf("Unable to remove repo indexer: %v\n", err)
^
tests/integration/repo_test.go:357:4: use of `fmt.Printf` forbidden by pattern `^(fmt\.Print(|f|ln)|print|println)$` (forbidigo)
fmt.Printf("%s", resp.Body)
^
```
</details>
---------
Co-authored-by: Giteabot <teabot@gitea.io>
2023-04-24 18:50:58 +09:00
|
|
|
//nolint:forbidigo
|
2022-09-03 04:18:23 +09:00
|
|
|
package integration
|
2017-04-25 16:24:51 +09:00
|
|
|
|
|
|
|
import (
|
|
|
|
"bytes"
|
2019-12-15 18:51:28 +09:00
|
|
|
"context"
|
2017-04-25 16:24:51 +09:00
|
|
|
"fmt"
|
2021-04-07 01:44:02 +09:00
|
|
|
"hash"
|
|
|
|
"hash/fnv"
|
2017-04-25 16:24:51 +09:00
|
|
|
"io"
|
|
|
|
"net/http"
|
2017-05-02 09:49:55 +09:00
|
|
|
"net/http/cookiejar"
|
2017-12-04 07:46:01 +09:00
|
|
|
"net/http/httptest"
|
2017-05-02 09:49:55 +09:00
|
|
|
"net/url"
|
2017-04-25 16:24:51 +09:00
|
|
|
"os"
|
2017-11-03 02:51:03 +09:00
|
|
|
"path/filepath"
|
2023-11-23 01:26:21 +09:00
|
|
|
"strconv"
|
2017-05-02 09:49:55 +09:00
|
|
|
"strings"
|
2022-10-21 03:20:01 +09:00
|
|
|
"sync/atomic"
|
2017-04-25 16:24:51 +09:00
|
|
|
"testing"
|
2020-06-02 10:39:44 +09:00
|
|
|
"time"
|
2017-04-25 16:24:51 +09:00
|
|
|
|
2023-01-18 06:46:03 +09:00
|
|
|
"code.gitea.io/gitea/models/auth"
|
2023-06-05 18:43:31 +09:00
|
|
|
"code.gitea.io/gitea/models/db"
|
2024-01-12 19:50:29 +09:00
|
|
|
repo_model "code.gitea.io/gitea/models/repo"
|
|
|
|
unit_model "code.gitea.io/gitea/models/unit"
|
2021-11-12 23:36:47 +09:00
|
|
|
"code.gitea.io/gitea/models/unittest"
|
2023-06-05 18:43:31 +09:00
|
|
|
user_model "code.gitea.io/gitea/models/user"
|
2024-01-12 19:50:29 +09:00
|
|
|
"code.gitea.io/gitea/modules/git"
|
2019-12-15 18:51:28 +09:00
|
|
|
"code.gitea.io/gitea/modules/graceful"
|
2021-07-25 01:03:58 +09:00
|
|
|
"code.gitea.io/gitea/modules/json"
|
2020-11-01 05:51:48 +09:00
|
|
|
"code.gitea.io/gitea/modules/log"
|
2017-04-25 16:24:51 +09:00
|
|
|
"code.gitea.io/gitea/modules/setting"
|
Rewrite queue (#24505)
# ⚠️ Breaking
Many deprecated queue config options are removed (actually, they should
have been removed in 1.18/1.19).
If you see the fatal message when starting Gitea: "Please update your
app.ini to remove deprecated config options", please follow the error
messages to remove these options from your app.ini.
Example:
```
2023/05/06 19:39:22 [E] Removed queue option: `[indexer].ISSUE_INDEXER_QUEUE_TYPE`. Use new options in `[queue.issue_indexer]`
2023/05/06 19:39:22 [E] Removed queue option: `[indexer].UPDATE_BUFFER_LEN`. Use new options in `[queue.issue_indexer]`
2023/05/06 19:39:22 [F] Please update your app.ini to remove deprecated config options
```
Many options in `[queue]` are are dropped, including:
`WRAP_IF_NECESSARY`, `MAX_ATTEMPTS`, `TIMEOUT`, `WORKERS`,
`BLOCK_TIMEOUT`, `BOOST_TIMEOUT`, `BOOST_WORKERS`, they can be removed
from app.ini.
# The problem
The old queue package has some legacy problems:
* complexity: I doubt few people could tell how it works.
* maintainability: Too many channels and mutex/cond are mixed together,
too many different structs/interfaces depends each other.
* stability: due to the complexity & maintainability, sometimes there
are strange bugs and difficult to debug, and some code doesn't have test
(indeed some code is difficult to test because a lot of things are mixed
together).
* general applicability: although it is called "queue", its behavior is
not a well-known queue.
* scalability: it doesn't seem easy to make it work with a cluster
without breaking its behaviors.
It came from some very old code to "avoid breaking", however, its
technical debt is too heavy now. It's a good time to introduce a better
"queue" package.
# The new queue package
It keeps using old config and concept as much as possible.
* It only contains two major kinds of concepts:
* The "base queue": channel, levelqueue, redis
* They have the same abstraction, the same interface, and they are
tested by the same testing code.
* The "WokerPoolQueue", it uses the "base queue" to provide "worker
pool" function, calls the "handler" to process the data in the base
queue.
* The new code doesn't do "PushBack"
* Think about a queue with many workers, the "PushBack" can't guarantee
the order for re-queued unhandled items, so in new code it just does
"normal push"
* The new code doesn't do "pause/resume"
* The "pause/resume" was designed to handle some handler's failure: eg:
document indexer (elasticsearch) is down
* If a queue is paused for long time, either the producers blocks or the
new items are dropped.
* The new code doesn't do such "pause/resume" trick, it's not a common
queue's behavior and it doesn't help much.
* If there are unhandled items, the "push" function just blocks for a
few seconds and then re-queue them and retry.
* The new code doesn't do "worker booster"
* Gitea's queue's handlers are light functions, the cost is only the
go-routine, so it doesn't make sense to "boost" them.
* The new code only use "max worker number" to limit the concurrent
workers.
* The new "Push" never blocks forever
* Instead of creating more and more blocking goroutines, return an error
is more friendly to the server and to the end user.
There are more details in code comments: eg: the "Flush" problem, the
strange "code.index" hanging problem, the "immediate" queue problem.
Almost ready for review.
TODO:
* [x] add some necessary comments during review
* [x] add some more tests if necessary
* [x] update documents and config options
* [x] test max worker / active worker
* [x] re-run the CI tasks to see whether any test is flaky
* [x] improve the `handleOldLengthConfiguration` to provide more
friendly messages
* [x] fine tune default config values (eg: length?)
## Code coverage:
![image](https://user-images.githubusercontent.com/2114189/236620635-55576955-f95d-4810-b12f-879026a3afdf.png)
2023-05-08 20:49:59 +09:00
|
|
|
"code.gitea.io/gitea/modules/testlogger"
|
2020-08-12 05:05:34 +09:00
|
|
|
"code.gitea.io/gitea/modules/util"
|
2021-01-27 00:36:53 +09:00
|
|
|
"code.gitea.io/gitea/modules/web"
|
2017-04-25 16:24:51 +09:00
|
|
|
"code.gitea.io/gitea/routers"
|
2024-02-27 16:12:22 +09:00
|
|
|
gitea_context "code.gitea.io/gitea/services/context"
|
2024-01-12 19:50:29 +09:00
|
|
|
repo_service "code.gitea.io/gitea/services/repository"
|
|
|
|
files_service "code.gitea.io/gitea/services/repository/files"
|
2023-06-05 18:43:31 +09:00
|
|
|
user_service "code.gitea.io/gitea/services/user"
|
2022-09-03 04:18:23 +09:00
|
|
|
"code.gitea.io/gitea/tests"
|
2017-04-25 16:24:51 +09:00
|
|
|
|
2017-12-11 11:15:27 +09:00
|
|
|
"github.com/PuerkitoBio/goquery"
|
2024-01-12 19:50:29 +09:00
|
|
|
gouuid "github.com/google/uuid"
|
2023-06-27 22:00:15 +09:00
|
|
|
"github.com/markbates/goth"
|
|
|
|
"github.com/markbates/goth/gothic"
|
2023-06-05 18:29:07 +09:00
|
|
|
goth_gitlab "github.com/markbates/goth/providers/gitlab"
|
2023-10-01 19:58:58 +09:00
|
|
|
"github.com/santhosh-tekuri/jsonschema/v5"
|
2017-04-28 22:23:28 +09:00
|
|
|
"github.com/stretchr/testify/assert"
|
2017-04-25 16:24:51 +09:00
|
|
|
)
|
|
|
|
|
2023-08-13 01:30:16 +09:00
|
|
|
var testWebRoutes *web.Route
|
2017-04-25 16:24:51 +09:00
|
|
|
|
2019-02-13 00:09:43 +09:00
|
|
|
type NilResponseRecorder struct {
|
|
|
|
httptest.ResponseRecorder
|
|
|
|
Length int
|
|
|
|
}
|
|
|
|
|
|
|
|
func (n *NilResponseRecorder) Write(b []byte) (int, error) {
|
2019-06-13 04:41:28 +09:00
|
|
|
n.Length += len(b)
|
2019-02-13 00:09:43 +09:00
|
|
|
return len(b), nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// NewRecorder returns an initialized ResponseRecorder.
|
|
|
|
func NewNilResponseRecorder() *NilResponseRecorder {
|
|
|
|
return &NilResponseRecorder{
|
|
|
|
ResponseRecorder: *httptest.NewRecorder(),
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2021-04-07 01:44:02 +09:00
|
|
|
type NilResponseHashSumRecorder struct {
|
|
|
|
httptest.ResponseRecorder
|
|
|
|
Hash hash.Hash
|
|
|
|
Length int
|
|
|
|
}
|
|
|
|
|
|
|
|
func (n *NilResponseHashSumRecorder) Write(b []byte) (int, error) {
|
|
|
|
_, _ = n.Hash.Write(b)
|
|
|
|
n.Length += len(b)
|
|
|
|
return len(b), nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// NewRecorder returns an initialized ResponseRecorder.
|
|
|
|
func NewNilResponseHashSumRecorder() *NilResponseHashSumRecorder {
|
|
|
|
return &NilResponseHashSumRecorder{
|
|
|
|
Hash: fnv.New32(),
|
|
|
|
ResponseRecorder: *httptest.NewRecorder(),
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2017-04-25 16:24:51 +09:00
|
|
|
func TestMain(m *testing.M) {
|
Rewrite logger system (#24726)
## ⚠️ Breaking
The `log.<mode>.<logger>` style config has been dropped. If you used it,
please check the new config manual & app.example.ini to make your
instance output logs as expected.
Although many legacy options still work, it's encouraged to upgrade to
the new options.
The SMTP logger is deleted because SMTP is not suitable to collect logs.
If you have manually configured Gitea log options, please confirm the
logger system works as expected after upgrading.
## Description
Close #12082 and maybe more log-related issues, resolve some related
FIXMEs in old code (which seems unfixable before)
Just like rewriting queue #24505 : make code maintainable, clear legacy
bugs, and add the ability to support more writers (eg: JSON, structured
log)
There is a new document (with examples): `logging-config.en-us.md`
This PR is safer than the queue rewriting, because it's just for
logging, it won't break other logic.
## The old problems
The logging system is quite old and difficult to maintain:
* Unclear concepts: Logger, NamedLogger, MultiChannelledLogger,
SubLogger, EventLogger, WriterLogger etc
* Some code is diffuclt to konw whether it is right:
`log.DelNamedLogger("console")` vs `log.DelNamedLogger(log.DEFAULT)` vs
`log.DelLogger("console")`
* The old system heavily depends on ini config system, it's difficult to
create new logger for different purpose, and it's very fragile.
* The "color" trick is difficult to use and read, many colors are
unnecessary, and in the future structured log could help
* It's difficult to add other log formats, eg: JSON format
* The log outputer doesn't have full control of its goroutine, it's
difficult to make outputer have advanced behaviors
* The logs could be lost in some cases: eg: no Fatal error when using
CLI.
* Config options are passed by JSON, which is quite fragile.
* INI package makes the KEY in `[log]` section visible in `[log.sub1]`
and `[log.sub1.subA]`, this behavior is quite fragile and would cause
more unclear problems, and there is no strong requirement to support
`log.<mode>.<logger>` syntax.
## The new design
See `logger.go` for documents.
## Screenshot
<details>
![image](https://github.com/go-gitea/gitea/assets/2114189/4462d713-ba39-41f5-bb08-de912e67e1ff)
![image](https://github.com/go-gitea/gitea/assets/2114189/b188035e-f691-428b-8b2d-ff7b2199b2f9)
![image](https://github.com/go-gitea/gitea/assets/2114189/132e9745-1c3b-4e00-9e0d-15eaea495dee)
</details>
## TODO
* [x] add some new tests
* [x] fix some tests
* [x] test some sub-commands (manually ....)
---------
Co-authored-by: Jason Song <i@wolfogre.com>
Co-authored-by: delvh <dev.lh@web.de>
Co-authored-by: Giteabot <teabot@gitea.io>
2023-05-22 07:35:11 +09:00
|
|
|
defer log.GetManager().Close()
|
2020-11-01 05:51:48 +09:00
|
|
|
|
2019-12-15 18:51:28 +09:00
|
|
|
managerCtx, cancel := context.WithCancel(context.Background())
|
|
|
|
graceful.InitManager(managerCtx)
|
|
|
|
defer cancel()
|
|
|
|
|
2022-09-03 04:18:23 +09:00
|
|
|
tests.InitTest(true)
|
2023-08-13 01:30:16 +09:00
|
|
|
testWebRoutes = routers.NormalRoutes()
|
2017-04-25 16:24:51 +09:00
|
|
|
|
2020-06-02 10:39:44 +09:00
|
|
|
// integration test settings...
|
2023-02-20 01:12:01 +09:00
|
|
|
if setting.CfgProvider != nil {
|
|
|
|
testingCfg := setting.CfgProvider.Section("integration-tests")
|
Rewrite queue (#24505)
# ⚠️ Breaking
Many deprecated queue config options are removed (actually, they should
have been removed in 1.18/1.19).
If you see the fatal message when starting Gitea: "Please update your
app.ini to remove deprecated config options", please follow the error
messages to remove these options from your app.ini.
Example:
```
2023/05/06 19:39:22 [E] Removed queue option: `[indexer].ISSUE_INDEXER_QUEUE_TYPE`. Use new options in `[queue.issue_indexer]`
2023/05/06 19:39:22 [E] Removed queue option: `[indexer].UPDATE_BUFFER_LEN`. Use new options in `[queue.issue_indexer]`
2023/05/06 19:39:22 [F] Please update your app.ini to remove deprecated config options
```
Many options in `[queue]` are are dropped, including:
`WRAP_IF_NECESSARY`, `MAX_ATTEMPTS`, `TIMEOUT`, `WORKERS`,
`BLOCK_TIMEOUT`, `BOOST_TIMEOUT`, `BOOST_WORKERS`, they can be removed
from app.ini.
# The problem
The old queue package has some legacy problems:
* complexity: I doubt few people could tell how it works.
* maintainability: Too many channels and mutex/cond are mixed together,
too many different structs/interfaces depends each other.
* stability: due to the complexity & maintainability, sometimes there
are strange bugs and difficult to debug, and some code doesn't have test
(indeed some code is difficult to test because a lot of things are mixed
together).
* general applicability: although it is called "queue", its behavior is
not a well-known queue.
* scalability: it doesn't seem easy to make it work with a cluster
without breaking its behaviors.
It came from some very old code to "avoid breaking", however, its
technical debt is too heavy now. It's a good time to introduce a better
"queue" package.
# The new queue package
It keeps using old config and concept as much as possible.
* It only contains two major kinds of concepts:
* The "base queue": channel, levelqueue, redis
* They have the same abstraction, the same interface, and they are
tested by the same testing code.
* The "WokerPoolQueue", it uses the "base queue" to provide "worker
pool" function, calls the "handler" to process the data in the base
queue.
* The new code doesn't do "PushBack"
* Think about a queue with many workers, the "PushBack" can't guarantee
the order for re-queued unhandled items, so in new code it just does
"normal push"
* The new code doesn't do "pause/resume"
* The "pause/resume" was designed to handle some handler's failure: eg:
document indexer (elasticsearch) is down
* If a queue is paused for long time, either the producers blocks or the
new items are dropped.
* The new code doesn't do such "pause/resume" trick, it's not a common
queue's behavior and it doesn't help much.
* If there are unhandled items, the "push" function just blocks for a
few seconds and then re-queue them and retry.
* The new code doesn't do "worker booster"
* Gitea's queue's handlers are light functions, the cost is only the
go-routine, so it doesn't make sense to "boost" them.
* The new code only use "max worker number" to limit the concurrent
workers.
* The new "Push" never blocks forever
* Instead of creating more and more blocking goroutines, return an error
is more friendly to the server and to the end user.
There are more details in code comments: eg: the "Flush" problem, the
strange "code.index" hanging problem, the "immediate" queue problem.
Almost ready for review.
TODO:
* [x] add some necessary comments during review
* [x] add some more tests if necessary
* [x] update documents and config options
* [x] test max worker / active worker
* [x] re-run the CI tasks to see whether any test is flaky
* [x] improve the `handleOldLengthConfiguration` to provide more
friendly messages
* [x] fine tune default config values (eg: length?)
## Code coverage:
![image](https://user-images.githubusercontent.com/2114189/236620635-55576955-f95d-4810-b12f-879026a3afdf.png)
2023-05-08 20:49:59 +09:00
|
|
|
testlogger.SlowTest = testingCfg.Key("SLOW_TEST").MustDuration(testlogger.SlowTest)
|
|
|
|
testlogger.SlowFlush = testingCfg.Key("SLOW_FLUSH").MustDuration(testlogger.SlowFlush)
|
2020-06-02 10:39:44 +09:00
|
|
|
}
|
|
|
|
|
|
|
|
if os.Getenv("GITEA_SLOW_TEST_TIME") != "" {
|
|
|
|
duration, err := time.ParseDuration(os.Getenv("GITEA_SLOW_TEST_TIME"))
|
|
|
|
if err == nil {
|
Rewrite queue (#24505)
# ⚠️ Breaking
Many deprecated queue config options are removed (actually, they should
have been removed in 1.18/1.19).
If you see the fatal message when starting Gitea: "Please update your
app.ini to remove deprecated config options", please follow the error
messages to remove these options from your app.ini.
Example:
```
2023/05/06 19:39:22 [E] Removed queue option: `[indexer].ISSUE_INDEXER_QUEUE_TYPE`. Use new options in `[queue.issue_indexer]`
2023/05/06 19:39:22 [E] Removed queue option: `[indexer].UPDATE_BUFFER_LEN`. Use new options in `[queue.issue_indexer]`
2023/05/06 19:39:22 [F] Please update your app.ini to remove deprecated config options
```
Many options in `[queue]` are are dropped, including:
`WRAP_IF_NECESSARY`, `MAX_ATTEMPTS`, `TIMEOUT`, `WORKERS`,
`BLOCK_TIMEOUT`, `BOOST_TIMEOUT`, `BOOST_WORKERS`, they can be removed
from app.ini.
# The problem
The old queue package has some legacy problems:
* complexity: I doubt few people could tell how it works.
* maintainability: Too many channels and mutex/cond are mixed together,
too many different structs/interfaces depends each other.
* stability: due to the complexity & maintainability, sometimes there
are strange bugs and difficult to debug, and some code doesn't have test
(indeed some code is difficult to test because a lot of things are mixed
together).
* general applicability: although it is called "queue", its behavior is
not a well-known queue.
* scalability: it doesn't seem easy to make it work with a cluster
without breaking its behaviors.
It came from some very old code to "avoid breaking", however, its
technical debt is too heavy now. It's a good time to introduce a better
"queue" package.
# The new queue package
It keeps using old config and concept as much as possible.
* It only contains two major kinds of concepts:
* The "base queue": channel, levelqueue, redis
* They have the same abstraction, the same interface, and they are
tested by the same testing code.
* The "WokerPoolQueue", it uses the "base queue" to provide "worker
pool" function, calls the "handler" to process the data in the base
queue.
* The new code doesn't do "PushBack"
* Think about a queue with many workers, the "PushBack" can't guarantee
the order for re-queued unhandled items, so in new code it just does
"normal push"
* The new code doesn't do "pause/resume"
* The "pause/resume" was designed to handle some handler's failure: eg:
document indexer (elasticsearch) is down
* If a queue is paused for long time, either the producers blocks or the
new items are dropped.
* The new code doesn't do such "pause/resume" trick, it's not a common
queue's behavior and it doesn't help much.
* If there are unhandled items, the "push" function just blocks for a
few seconds and then re-queue them and retry.
* The new code doesn't do "worker booster"
* Gitea's queue's handlers are light functions, the cost is only the
go-routine, so it doesn't make sense to "boost" them.
* The new code only use "max worker number" to limit the concurrent
workers.
* The new "Push" never blocks forever
* Instead of creating more and more blocking goroutines, return an error
is more friendly to the server and to the end user.
There are more details in code comments: eg: the "Flush" problem, the
strange "code.index" hanging problem, the "immediate" queue problem.
Almost ready for review.
TODO:
* [x] add some necessary comments during review
* [x] add some more tests if necessary
* [x] update documents and config options
* [x] test max worker / active worker
* [x] re-run the CI tasks to see whether any test is flaky
* [x] improve the `handleOldLengthConfiguration` to provide more
friendly messages
* [x] fine tune default config values (eg: length?)
## Code coverage:
![image](https://user-images.githubusercontent.com/2114189/236620635-55576955-f95d-4810-b12f-879026a3afdf.png)
2023-05-08 20:49:59 +09:00
|
|
|
testlogger.SlowTest = duration
|
2020-06-02 10:39:44 +09:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if os.Getenv("GITEA_SLOW_FLUSH_TIME") != "" {
|
|
|
|
duration, err := time.ParseDuration(os.Getenv("GITEA_SLOW_FLUSH_TIME"))
|
|
|
|
if err == nil {
|
Rewrite queue (#24505)
# ⚠️ Breaking
Many deprecated queue config options are removed (actually, they should
have been removed in 1.18/1.19).
If you see the fatal message when starting Gitea: "Please update your
app.ini to remove deprecated config options", please follow the error
messages to remove these options from your app.ini.
Example:
```
2023/05/06 19:39:22 [E] Removed queue option: `[indexer].ISSUE_INDEXER_QUEUE_TYPE`. Use new options in `[queue.issue_indexer]`
2023/05/06 19:39:22 [E] Removed queue option: `[indexer].UPDATE_BUFFER_LEN`. Use new options in `[queue.issue_indexer]`
2023/05/06 19:39:22 [F] Please update your app.ini to remove deprecated config options
```
Many options in `[queue]` are are dropped, including:
`WRAP_IF_NECESSARY`, `MAX_ATTEMPTS`, `TIMEOUT`, `WORKERS`,
`BLOCK_TIMEOUT`, `BOOST_TIMEOUT`, `BOOST_WORKERS`, they can be removed
from app.ini.
# The problem
The old queue package has some legacy problems:
* complexity: I doubt few people could tell how it works.
* maintainability: Too many channels and mutex/cond are mixed together,
too many different structs/interfaces depends each other.
* stability: due to the complexity & maintainability, sometimes there
are strange bugs and difficult to debug, and some code doesn't have test
(indeed some code is difficult to test because a lot of things are mixed
together).
* general applicability: although it is called "queue", its behavior is
not a well-known queue.
* scalability: it doesn't seem easy to make it work with a cluster
without breaking its behaviors.
It came from some very old code to "avoid breaking", however, its
technical debt is too heavy now. It's a good time to introduce a better
"queue" package.
# The new queue package
It keeps using old config and concept as much as possible.
* It only contains two major kinds of concepts:
* The "base queue": channel, levelqueue, redis
* They have the same abstraction, the same interface, and they are
tested by the same testing code.
* The "WokerPoolQueue", it uses the "base queue" to provide "worker
pool" function, calls the "handler" to process the data in the base
queue.
* The new code doesn't do "PushBack"
* Think about a queue with many workers, the "PushBack" can't guarantee
the order for re-queued unhandled items, so in new code it just does
"normal push"
* The new code doesn't do "pause/resume"
* The "pause/resume" was designed to handle some handler's failure: eg:
document indexer (elasticsearch) is down
* If a queue is paused for long time, either the producers blocks or the
new items are dropped.
* The new code doesn't do such "pause/resume" trick, it's not a common
queue's behavior and it doesn't help much.
* If there are unhandled items, the "push" function just blocks for a
few seconds and then re-queue them and retry.
* The new code doesn't do "worker booster"
* Gitea's queue's handlers are light functions, the cost is only the
go-routine, so it doesn't make sense to "boost" them.
* The new code only use "max worker number" to limit the concurrent
workers.
* The new "Push" never blocks forever
* Instead of creating more and more blocking goroutines, return an error
is more friendly to the server and to the end user.
There are more details in code comments: eg: the "Flush" problem, the
strange "code.index" hanging problem, the "immediate" queue problem.
Almost ready for review.
TODO:
* [x] add some necessary comments during review
* [x] add some more tests if necessary
* [x] update documents and config options
* [x] test max worker / active worker
* [x] re-run the CI tasks to see whether any test is flaky
* [x] improve the `handleOldLengthConfiguration` to provide more
friendly messages
* [x] fine tune default config values (eg: length?)
## Code coverage:
![image](https://user-images.githubusercontent.com/2114189/236620635-55576955-f95d-4810-b12f-879026a3afdf.png)
2023-05-08 20:49:59 +09:00
|
|
|
testlogger.SlowFlush = duration
|
2020-06-02 10:39:44 +09:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2022-04-27 05:28:45 +09:00
|
|
|
os.Unsetenv("GIT_AUTHOR_NAME")
|
|
|
|
os.Unsetenv("GIT_AUTHOR_EMAIL")
|
|
|
|
os.Unsetenv("GIT_AUTHOR_DATE")
|
|
|
|
os.Unsetenv("GIT_COMMITTER_NAME")
|
|
|
|
os.Unsetenv("GIT_COMMITTER_EMAIL")
|
|
|
|
os.Unsetenv("GIT_COMMITTER_DATE")
|
|
|
|
|
2021-11-12 23:36:47 +09:00
|
|
|
err := unittest.InitFixtures(
|
|
|
|
unittest.FixturesOptions{
|
2021-09-24 20:32:56 +09:00
|
|
|
Dir: filepath.Join(filepath.Dir(setting.AppPath), "models/fixtures/"),
|
|
|
|
},
|
2017-04-25 16:24:51 +09:00
|
|
|
)
|
|
|
|
if err != nil {
|
|
|
|
fmt.Printf("Error initializing test database: %v\n", err)
|
|
|
|
os.Exit(1)
|
|
|
|
}
|
2023-04-19 22:40:42 +09:00
|
|
|
|
|
|
|
// FIXME: the console logger is deleted by mistake, so if there is any `log.Fatal`, developers won't see any error message.
|
|
|
|
// Instead, "No tests were found", last nonsense log is "According to the configuration, subsequent logs will not be printed to the console"
|
2017-09-17 05:16:21 +09:00
|
|
|
exitCode := m.Run()
|
|
|
|
|
2024-03-14 22:45:41 +09:00
|
|
|
if err := testlogger.WriterCloser.Reset(); err != nil {
|
|
|
|
fmt.Printf("testlogger.WriterCloser.Reset: %v\n", err)
|
|
|
|
os.Exit(1)
|
|
|
|
}
|
2019-04-11 20:49:49 +09:00
|
|
|
|
2020-08-12 05:05:34 +09:00
|
|
|
if err = util.RemoveAll(setting.Indexer.IssuePath); err != nil {
|
|
|
|
fmt.Printf("util.RemoveAll: %v\n", err)
|
2017-09-17 05:16:21 +09:00
|
|
|
os.Exit(1)
|
|
|
|
}
|
2020-08-12 05:05:34 +09:00
|
|
|
if err = util.RemoveAll(setting.Indexer.RepoPath); err != nil {
|
2017-10-27 15:10:54 +09:00
|
|
|
fmt.Printf("Unable to remove repo indexer: %v\n", err)
|
|
|
|
os.Exit(1)
|
|
|
|
}
|
2017-09-17 05:16:21 +09:00
|
|
|
|
|
|
|
os.Exit(exitCode)
|
2017-04-25 16:24:51 +09:00
|
|
|
}
|
|
|
|
|
2017-05-02 09:49:55 +09:00
|
|
|
type TestSession struct {
|
|
|
|
jar http.CookieJar
|
|
|
|
}
|
|
|
|
|
|
|
|
func (s *TestSession) GetCookie(name string) *http.Cookie {
|
|
|
|
baseURL, err := url.Parse(setting.AppURL)
|
|
|
|
if err != nil {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
for _, c := range s.jar.Cookies(baseURL) {
|
|
|
|
if c.Name == name {
|
|
|
|
return c
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
2023-12-22 08:59:59 +09:00
|
|
|
func (s *TestSession) MakeRequest(t testing.TB, rw *RequestWrapper, expectedStatus int) *httptest.ResponseRecorder {
|
2019-07-29 13:15:18 +09:00
|
|
|
t.Helper()
|
2023-12-22 08:59:59 +09:00
|
|
|
req := rw.Request
|
2017-05-02 09:49:55 +09:00
|
|
|
baseURL, err := url.Parse(setting.AppURL)
|
|
|
|
assert.NoError(t, err)
|
|
|
|
for _, c := range s.jar.Cookies(baseURL) {
|
|
|
|
req.AddCookie(c)
|
|
|
|
}
|
2023-12-22 08:59:59 +09:00
|
|
|
resp := MakeRequest(t, rw, expectedStatus)
|
2017-05-02 09:49:55 +09:00
|
|
|
|
|
|
|
ch := http.Header{}
|
2019-06-13 04:41:28 +09:00
|
|
|
ch.Add("Cookie", strings.Join(resp.Header()["Set-Cookie"], ";"))
|
2017-05-02 09:49:55 +09:00
|
|
|
cr := http.Request{Header: ch}
|
|
|
|
s.jar.SetCookies(baseURL, cr.Cookies())
|
|
|
|
|
|
|
|
return resp
|
|
|
|
}
|
|
|
|
|
2023-12-22 08:59:59 +09:00
|
|
|
func (s *TestSession) MakeRequestNilResponseRecorder(t testing.TB, rw *RequestWrapper, expectedStatus int) *NilResponseRecorder {
|
2019-07-29 13:15:18 +09:00
|
|
|
t.Helper()
|
2023-12-22 08:59:59 +09:00
|
|
|
req := rw.Request
|
2019-02-13 00:09:43 +09:00
|
|
|
baseURL, err := url.Parse(setting.AppURL)
|
|
|
|
assert.NoError(t, err)
|
|
|
|
for _, c := range s.jar.Cookies(baseURL) {
|
|
|
|
req.AddCookie(c)
|
|
|
|
}
|
2023-12-22 08:59:59 +09:00
|
|
|
resp := MakeRequestNilResponseRecorder(t, rw, expectedStatus)
|
2019-02-13 00:09:43 +09:00
|
|
|
|
|
|
|
ch := http.Header{}
|
2019-06-13 04:41:28 +09:00
|
|
|
ch.Add("Cookie", strings.Join(resp.Header()["Set-Cookie"], ";"))
|
2019-02-13 00:09:43 +09:00
|
|
|
cr := http.Request{Header: ch}
|
|
|
|
s.jar.SetCookies(baseURL, cr.Cookies())
|
|
|
|
|
|
|
|
return resp
|
|
|
|
}
|
|
|
|
|
2023-12-22 08:59:59 +09:00
|
|
|
func (s *TestSession) MakeRequestNilResponseHashSumRecorder(t testing.TB, rw *RequestWrapper, expectedStatus int) *NilResponseHashSumRecorder {
|
2021-04-07 01:44:02 +09:00
|
|
|
t.Helper()
|
2023-12-22 08:59:59 +09:00
|
|
|
req := rw.Request
|
2021-04-07 01:44:02 +09:00
|
|
|
baseURL, err := url.Parse(setting.AppURL)
|
|
|
|
assert.NoError(t, err)
|
|
|
|
for _, c := range s.jar.Cookies(baseURL) {
|
|
|
|
req.AddCookie(c)
|
|
|
|
}
|
2023-12-22 08:59:59 +09:00
|
|
|
resp := MakeRequestNilResponseHashSumRecorder(t, rw, expectedStatus)
|
2021-04-07 01:44:02 +09:00
|
|
|
|
|
|
|
ch := http.Header{}
|
|
|
|
ch.Add("Cookie", strings.Join(resp.Header()["Set-Cookie"], ";"))
|
|
|
|
cr := http.Request{Header: ch}
|
|
|
|
s.jar.SetCookies(baseURL, cr.Cookies())
|
|
|
|
|
|
|
|
return resp
|
|
|
|
}
|
|
|
|
|
2017-06-17 13:49:45 +09:00
|
|
|
const userPassword = "password"
|
|
|
|
|
2017-08-23 16:30:33 +09:00
|
|
|
func emptyTestSession(t testing.TB) *TestSession {
|
2019-07-29 13:15:18 +09:00
|
|
|
t.Helper()
|
2017-08-23 16:30:33 +09:00
|
|
|
jar, err := cookiejar.New(nil)
|
|
|
|
assert.NoError(t, err)
|
|
|
|
|
|
|
|
return &TestSession{jar: jar}
|
|
|
|
}
|
|
|
|
|
2023-01-18 06:46:03 +09:00
|
|
|
func getUserToken(t testing.TB, userName string, scope ...auth.AccessTokenScope) string {
|
|
|
|
return getTokenForLoggedInUser(t, loginUser(t, userName), scope...)
|
2022-04-08 13:22:10 +09:00
|
|
|
}
|
|
|
|
|
2023-06-27 22:00:15 +09:00
|
|
|
func mockCompleteUserAuth(mock func(res http.ResponseWriter, req *http.Request) (goth.User, error)) func() {
|
|
|
|
old := gothic.CompleteUserAuth
|
|
|
|
gothic.CompleteUserAuth = mock
|
|
|
|
return func() {
|
|
|
|
gothic.CompleteUserAuth = old
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2023-06-05 18:29:07 +09:00
|
|
|
func addAuthSource(t *testing.T, payload map[string]string) *auth.Source {
|
|
|
|
session := loginUser(t, "user1")
|
|
|
|
payload["_csrf"] = GetCSRF(t, session, "/admin/auths/new")
|
|
|
|
req := NewRequestWithValues(t, "POST", "/admin/auths/new", payload)
|
|
|
|
session.MakeRequest(t, req, http.StatusSeeOther)
|
|
|
|
source, err := auth.GetSourceByName(context.Background(), payload["name"])
|
|
|
|
assert.NoError(t, err)
|
|
|
|
return source
|
|
|
|
}
|
|
|
|
|
|
|
|
func authSourcePayloadOAuth2(name string) map[string]string {
|
|
|
|
return map[string]string{
|
|
|
|
"type": fmt.Sprintf("%d", auth.OAuth2),
|
|
|
|
"name": name,
|
|
|
|
"is_active": "on",
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
func authSourcePayloadGitLab(name string) map[string]string {
|
|
|
|
payload := authSourcePayloadOAuth2(name)
|
|
|
|
payload["oauth2_provider"] = "gitlab"
|
|
|
|
return payload
|
|
|
|
}
|
|
|
|
|
|
|
|
func authSourcePayloadGitLabCustom(name string) map[string]string {
|
|
|
|
payload := authSourcePayloadGitLab(name)
|
|
|
|
payload["oauth2_use_custom_url"] = "on"
|
|
|
|
payload["oauth2_auth_url"] = goth_gitlab.AuthURL
|
|
|
|
payload["oauth2_token_url"] = goth_gitlab.TokenURL
|
|
|
|
payload["oauth2_profile_url"] = goth_gitlab.ProfileURL
|
|
|
|
return payload
|
|
|
|
}
|
|
|
|
|
2023-06-05 18:43:31 +09:00
|
|
|
func createUser(ctx context.Context, t testing.TB, user *user_model.User) func() {
|
|
|
|
user.MustChangePassword = false
|
|
|
|
user.LowerName = strings.ToLower(user.Name)
|
|
|
|
|
|
|
|
assert.NoError(t, db.Insert(ctx, user))
|
|
|
|
|
|
|
|
if len(user.Email) > 0 {
|
2024-02-05 23:06:51 +09:00
|
|
|
assert.NoError(t, user_service.ReplacePrimaryEmailAddress(ctx, user, user.Email))
|
2023-06-05 18:43:31 +09:00
|
|
|
}
|
|
|
|
|
|
|
|
return func() {
|
|
|
|
assert.NoError(t, user_service.DeleteUser(ctx, user, true))
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2017-06-18 00:01:03 +09:00
|
|
|
func loginUser(t testing.TB, userName string) *TestSession {
|
2019-07-29 13:15:18 +09:00
|
|
|
t.Helper()
|
2022-12-22 22:09:35 +09:00
|
|
|
|
|
|
|
return loginUserWithPassword(t, userName, userPassword)
|
2017-06-17 13:49:45 +09:00
|
|
|
}
|
|
|
|
|
2017-06-18 00:01:03 +09:00
|
|
|
func loginUserWithPassword(t testing.TB, userName, password string) *TestSession {
|
2019-07-29 13:15:18 +09:00
|
|
|
t.Helper()
|
2023-11-23 01:26:21 +09:00
|
|
|
|
|
|
|
return loginUserWithPasswordRemember(t, userName, password, false)
|
|
|
|
}
|
|
|
|
|
|
|
|
func loginUserWithPasswordRemember(t testing.TB, userName, password string, rememberMe bool) *TestSession {
|
|
|
|
t.Helper()
|
2017-06-10 09:41:36 +09:00
|
|
|
req := NewRequest(t, "GET", "/user/login")
|
2017-07-08 04:36:47 +09:00
|
|
|
resp := MakeRequest(t, req, http.StatusOK)
|
2017-05-02 09:49:55 +09:00
|
|
|
|
2017-06-18 01:29:59 +09:00
|
|
|
doc := NewHTMLParser(t, resp.Body)
|
2017-06-17 13:49:45 +09:00
|
|
|
req = NewRequestWithValues(t, "POST", "/user/login", map[string]string{
|
|
|
|
"_csrf": doc.GetCSRF(),
|
|
|
|
"user_name": userName,
|
|
|
|
"password": password,
|
2023-11-23 01:26:21 +09:00
|
|
|
"remember": strconv.FormatBool(rememberMe),
|
2017-06-17 13:49:45 +09:00
|
|
|
})
|
2022-03-23 13:54:07 +09:00
|
|
|
resp = MakeRequest(t, req, http.StatusSeeOther)
|
2017-05-02 09:49:55 +09:00
|
|
|
|
|
|
|
ch := http.Header{}
|
2019-06-13 04:41:28 +09:00
|
|
|
ch.Add("Cookie", strings.Join(resp.Header()["Set-Cookie"], ";"))
|
2017-05-02 09:49:55 +09:00
|
|
|
cr := http.Request{Header: ch}
|
|
|
|
|
2017-08-23 16:30:33 +09:00
|
|
|
session := emptyTestSession(t)
|
|
|
|
|
2017-05-02 09:49:55 +09:00
|
|
|
baseURL, err := url.Parse(setting.AppURL)
|
|
|
|
assert.NoError(t, err)
|
2017-08-23 16:30:33 +09:00
|
|
|
session.jar.SetCookies(baseURL, cr.Cookies())
|
2017-05-02 09:49:55 +09:00
|
|
|
|
2017-08-23 16:30:33 +09:00
|
|
|
return session
|
2017-05-02 09:49:55 +09:00
|
|
|
}
|
|
|
|
|
2022-01-21 02:46:10 +09:00
|
|
|
// token has to be unique this counter take care of
|
2020-04-14 04:02:48 +09:00
|
|
|
var tokenCounter int64
|
|
|
|
|
2023-01-18 06:46:03 +09:00
|
|
|
// getTokenForLoggedInUser returns a token for a logged in user.
|
|
|
|
// The scope is an optional list of snake_case strings like the frontend form fields,
|
|
|
|
// but without the "scope_" prefix.
|
|
|
|
func getTokenForLoggedInUser(t testing.TB, session *TestSession, scopes ...auth.AccessTokenScope) string {
|
2019-07-29 13:15:18 +09:00
|
|
|
t.Helper()
|
2022-12-21 10:46:16 +09:00
|
|
|
var token string
|
2018-09-11 01:15:52 +09:00
|
|
|
req := NewRequest(t, "GET", "/user/settings/applications")
|
|
|
|
resp := session.MakeRequest(t, req, http.StatusOK)
|
2022-12-21 10:46:16 +09:00
|
|
|
var csrf string
|
|
|
|
for _, cookie := range resp.Result().Cookies() {
|
|
|
|
if cookie.Name != "_csrf" {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
csrf = cookie.Value
|
|
|
|
break
|
|
|
|
}
|
|
|
|
if csrf == "" {
|
|
|
|
doc := NewHTMLParser(t, resp.Body)
|
|
|
|
csrf = doc.GetCSRF()
|
|
|
|
}
|
|
|
|
assert.NotEmpty(t, csrf)
|
2023-01-18 06:46:03 +09:00
|
|
|
urlValues := url.Values{}
|
|
|
|
urlValues.Add("_csrf", csrf)
|
|
|
|
urlValues.Add("name", fmt.Sprintf("api-testing-token-%d", atomic.AddInt64(&tokenCounter, 1)))
|
|
|
|
for _, scope := range scopes {
|
|
|
|
urlValues.Add("scope", string(scope))
|
|
|
|
}
|
|
|
|
req = NewRequestWithURLValues(t, "POST", "/user/settings/applications", urlValues)
|
2022-12-21 10:46:16 +09:00
|
|
|
resp = session.MakeRequest(t, req, http.StatusSeeOther)
|
|
|
|
|
|
|
|
// Log the flash values on failure
|
|
|
|
if !assert.Equal(t, resp.Result().Header["Location"], []string{"/user/settings/applications"}) {
|
|
|
|
for _, cookie := range resp.Result().Cookies() {
|
2023-04-14 04:45:33 +09:00
|
|
|
if cookie.Name != gitea_context.CookieNameFlash {
|
2022-12-21 10:46:16 +09:00
|
|
|
continue
|
|
|
|
}
|
|
|
|
flash, _ := url.ParseQuery(cookie.Value)
|
|
|
|
for key, value := range flash {
|
|
|
|
t.Logf("Flash %q: %q", key, value)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2018-09-11 01:15:52 +09:00
|
|
|
req = NewRequest(t, "GET", "/user/settings/applications")
|
|
|
|
resp = session.MakeRequest(t, req, http.StatusOK)
|
|
|
|
htmlDoc := NewHTMLParser(t, resp.Body)
|
2022-12-21 10:46:16 +09:00
|
|
|
token = htmlDoc.doc.Find(".ui.info p").Text()
|
2022-10-21 03:20:01 +09:00
|
|
|
assert.NotEmpty(t, token)
|
2018-09-11 01:15:52 +09:00
|
|
|
return token
|
|
|
|
}
|
|
|
|
|
2023-12-22 08:59:59 +09:00
|
|
|
type RequestWrapper struct {
|
|
|
|
*http.Request
|
|
|
|
}
|
|
|
|
|
|
|
|
func (req *RequestWrapper) AddBasicAuth(username string) *RequestWrapper {
|
|
|
|
req.Request.SetBasicAuth(username, userPassword)
|
|
|
|
return req
|
|
|
|
}
|
|
|
|
|
|
|
|
func (req *RequestWrapper) AddTokenAuth(token string) *RequestWrapper {
|
|
|
|
if token == "" {
|
|
|
|
return req
|
|
|
|
}
|
|
|
|
if !strings.HasPrefix(token, "Bearer ") {
|
|
|
|
token = "Bearer " + token
|
|
|
|
}
|
|
|
|
req.Request.Header.Set("Authorization", token)
|
|
|
|
return req
|
|
|
|
}
|
|
|
|
|
|
|
|
func (req *RequestWrapper) SetHeader(name, value string) *RequestWrapper {
|
|
|
|
req.Request.Header.Set(name, value)
|
|
|
|
return req
|
|
|
|
}
|
|
|
|
|
|
|
|
func NewRequest(t testing.TB, method, urlStr string) *RequestWrapper {
|
2019-07-29 13:15:18 +09:00
|
|
|
t.Helper()
|
2017-06-17 13:49:45 +09:00
|
|
|
return NewRequestWithBody(t, method, urlStr, nil)
|
|
|
|
}
|
|
|
|
|
2023-12-22 08:59:59 +09:00
|
|
|
func NewRequestf(t testing.TB, method, urlFormat string, args ...any) *RequestWrapper {
|
2019-07-29 13:15:18 +09:00
|
|
|
t.Helper()
|
2017-06-25 09:15:42 +09:00
|
|
|
return NewRequest(t, method, fmt.Sprintf(urlFormat, args...))
|
|
|
|
}
|
|
|
|
|
2023-12-22 08:59:59 +09:00
|
|
|
func NewRequestWithValues(t testing.TB, method, urlStr string, values map[string]string) *RequestWrapper {
|
2019-07-29 13:15:18 +09:00
|
|
|
t.Helper()
|
2017-06-17 13:49:45 +09:00
|
|
|
urlValues := url.Values{}
|
|
|
|
for key, value := range values {
|
|
|
|
urlValues[key] = []string{value}
|
|
|
|
}
|
2023-01-18 06:46:03 +09:00
|
|
|
return NewRequestWithURLValues(t, method, urlStr, urlValues)
|
|
|
|
}
|
|
|
|
|
2023-12-22 08:59:59 +09:00
|
|
|
func NewRequestWithURLValues(t testing.TB, method, urlStr string, urlValues url.Values) *RequestWrapper {
|
2023-01-18 06:46:03 +09:00
|
|
|
t.Helper()
|
2023-12-22 08:59:59 +09:00
|
|
|
return NewRequestWithBody(t, method, urlStr, bytes.NewBufferString(urlValues.Encode())).
|
|
|
|
SetHeader("Content-Type", "application/x-www-form-urlencoded")
|
2017-06-17 13:49:45 +09:00
|
|
|
}
|
|
|
|
|
2023-12-22 08:59:59 +09:00
|
|
|
func NewRequestWithJSON(t testing.TB, method, urlStr string, v any) *RequestWrapper {
|
2019-07-29 13:15:18 +09:00
|
|
|
t.Helper()
|
2021-03-02 06:08:10 +09:00
|
|
|
|
2017-06-17 13:49:45 +09:00
|
|
|
jsonBytes, err := json.Marshal(v)
|
|
|
|
assert.NoError(t, err)
|
2023-12-22 08:59:59 +09:00
|
|
|
return NewRequestWithBody(t, method, urlStr, bytes.NewBuffer(jsonBytes)).
|
|
|
|
SetHeader("Content-Type", "application/json")
|
2017-06-10 09:41:36 +09:00
|
|
|
}
|
|
|
|
|
2023-12-22 08:59:59 +09:00
|
|
|
func NewRequestWithBody(t testing.TB, method, urlStr string, body io.Reader) *RequestWrapper {
|
2019-07-29 13:15:18 +09:00
|
|
|
t.Helper()
|
2021-01-27 00:36:53 +09:00
|
|
|
if !strings.HasPrefix(urlStr, "http") && !strings.HasPrefix(urlStr, "/") {
|
|
|
|
urlStr = "/" + urlStr
|
|
|
|
}
|
2023-12-22 08:59:59 +09:00
|
|
|
req, err := http.NewRequest(method, urlStr, body)
|
2017-06-10 09:41:36 +09:00
|
|
|
assert.NoError(t, err)
|
2023-12-22 08:59:59 +09:00
|
|
|
req.RequestURI = urlStr
|
2017-06-10 09:41:36 +09:00
|
|
|
|
2023-12-22 08:59:59 +09:00
|
|
|
return &RequestWrapper{req}
|
2018-07-07 10:54:30 +09:00
|
|
|
}
|
|
|
|
|
2017-07-08 04:36:47 +09:00
|
|
|
const NoExpectedStatus = -1
|
|
|
|
|
2023-12-22 08:59:59 +09:00
|
|
|
func MakeRequest(t testing.TB, rw *RequestWrapper, expectedStatus int) *httptest.ResponseRecorder {
|
2019-07-29 13:15:18 +09:00
|
|
|
t.Helper()
|
2023-12-22 08:59:59 +09:00
|
|
|
req := rw.Request
|
2017-12-04 07:46:01 +09:00
|
|
|
recorder := httptest.NewRecorder()
|
2023-04-19 22:40:42 +09:00
|
|
|
if req.RemoteAddr == "" {
|
|
|
|
req.RemoteAddr = "test-mock:12345"
|
|
|
|
}
|
2023-08-13 01:30:16 +09:00
|
|
|
testWebRoutes.ServeHTTP(recorder, req)
|
2017-07-08 04:36:47 +09:00
|
|
|
if expectedStatus != NoExpectedStatus {
|
2023-04-19 22:40:42 +09:00
|
|
|
if !assert.EqualValues(t, expectedStatus, recorder.Code, "Request: %s %s", req.Method, req.URL.String()) {
|
2017-12-11 11:15:27 +09:00
|
|
|
logUnexpectedResponse(t, recorder)
|
|
|
|
}
|
2017-07-08 04:36:47 +09:00
|
|
|
}
|
2017-12-04 07:46:01 +09:00
|
|
|
return recorder
|
2017-04-25 16:24:51 +09:00
|
|
|
}
|
2017-06-18 18:06:17 +09:00
|
|
|
|
2023-12-22 08:59:59 +09:00
|
|
|
func MakeRequestNilResponseRecorder(t testing.TB, rw *RequestWrapper, expectedStatus int) *NilResponseRecorder {
|
2019-07-29 13:15:18 +09:00
|
|
|
t.Helper()
|
2023-12-22 08:59:59 +09:00
|
|
|
req := rw.Request
|
2019-02-13 00:09:43 +09:00
|
|
|
recorder := NewNilResponseRecorder()
|
2023-08-13 01:30:16 +09:00
|
|
|
testWebRoutes.ServeHTTP(recorder, req)
|
2019-02-13 00:09:43 +09:00
|
|
|
if expectedStatus != NoExpectedStatus {
|
2021-04-07 01:44:02 +09:00
|
|
|
if !assert.EqualValues(t, expectedStatus, recorder.Code,
|
|
|
|
"Request: %s %s", req.Method, req.URL.String()) {
|
|
|
|
logUnexpectedResponse(t, &recorder.ResponseRecorder)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return recorder
|
|
|
|
}
|
|
|
|
|
2023-12-22 08:59:59 +09:00
|
|
|
func MakeRequestNilResponseHashSumRecorder(t testing.TB, rw *RequestWrapper, expectedStatus int) *NilResponseHashSumRecorder {
|
2021-04-07 01:44:02 +09:00
|
|
|
t.Helper()
|
2023-12-22 08:59:59 +09:00
|
|
|
req := rw.Request
|
2021-04-07 01:44:02 +09:00
|
|
|
recorder := NewNilResponseHashSumRecorder()
|
2023-08-13 01:30:16 +09:00
|
|
|
testWebRoutes.ServeHTTP(recorder, req)
|
2021-04-07 01:44:02 +09:00
|
|
|
if expectedStatus != NoExpectedStatus {
|
2019-02-13 00:09:43 +09:00
|
|
|
if !assert.EqualValues(t, expectedStatus, recorder.Code,
|
|
|
|
"Request: %s %s", req.Method, req.URL.String()) {
|
|
|
|
logUnexpectedResponse(t, &recorder.ResponseRecorder)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return recorder
|
|
|
|
}
|
|
|
|
|
2017-12-11 11:15:27 +09:00
|
|
|
// logUnexpectedResponse logs the contents of an unexpected response.
|
|
|
|
func logUnexpectedResponse(t testing.TB, recorder *httptest.ResponseRecorder) {
|
2019-07-29 13:15:18 +09:00
|
|
|
t.Helper()
|
2017-12-11 11:15:27 +09:00
|
|
|
respBytes := recorder.Body.Bytes()
|
|
|
|
if len(respBytes) == 0 {
|
2024-02-02 18:18:42 +09:00
|
|
|
// log the content of the flash cookie
|
|
|
|
for _, cookie := range recorder.Result().Cookies() {
|
|
|
|
if cookie.Name != gitea_context.CookieNameFlash {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
flash, _ := url.ParseQuery(cookie.Value)
|
|
|
|
for key, value := range flash {
|
|
|
|
// the key is itself url-encoded
|
|
|
|
if flash, err := url.ParseQuery(key); err == nil {
|
|
|
|
for key, value := range flash {
|
|
|
|
t.Logf("FlashCookie %q: %q", key, value)
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
t.Logf("FlashCookie %q: %q", key, value)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2017-12-11 11:15:27 +09:00
|
|
|
return
|
|
|
|
} else if len(respBytes) < 500 {
|
|
|
|
// if body is short, just log the whole thing
|
2023-04-19 22:40:42 +09:00
|
|
|
t.Log("Response: ", string(respBytes))
|
2017-12-11 11:15:27 +09:00
|
|
|
return
|
|
|
|
}
|
2023-10-24 11:54:59 +09:00
|
|
|
t.Log("Response length: ", len(respBytes))
|
2017-12-11 11:15:27 +09:00
|
|
|
|
|
|
|
// log the "flash" error message, if one exists
|
|
|
|
// we must create a new buffer, so that we don't "use up" resp.Body
|
|
|
|
htmlDoc, err := goquery.NewDocumentFromReader(bytes.NewBuffer(respBytes))
|
|
|
|
if err != nil {
|
|
|
|
return // probably a non-HTML response
|
|
|
|
}
|
|
|
|
errMsg := htmlDoc.Find(".ui.negative.message").Text()
|
|
|
|
if len(errMsg) > 0 {
|
|
|
|
t.Log("A flash error message was found:", errMsg)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2023-07-05 03:36:08 +09:00
|
|
|
func DecodeJSON(t testing.TB, resp *httptest.ResponseRecorder, v any) {
|
2019-07-29 13:15:18 +09:00
|
|
|
t.Helper()
|
2021-03-02 06:08:10 +09:00
|
|
|
|
2017-12-04 07:46:01 +09:00
|
|
|
decoder := json.NewDecoder(resp.Body)
|
2017-06-18 18:06:17 +09:00
|
|
|
assert.NoError(t, decoder.Decode(v))
|
|
|
|
}
|
2017-07-08 04:36:47 +09:00
|
|
|
|
2022-12-17 15:22:34 +09:00
|
|
|
func VerifyJSONSchema(t testing.TB, resp *httptest.ResponseRecorder, schemaFile string) {
|
|
|
|
t.Helper()
|
|
|
|
|
|
|
|
schemaFilePath := filepath.Join(filepath.Dir(setting.AppPath), "tests", "integration", "schemas", schemaFile)
|
|
|
|
_, schemaFileErr := os.Stat(schemaFilePath)
|
|
|
|
assert.Nil(t, schemaFileErr)
|
|
|
|
|
2023-10-01 19:58:58 +09:00
|
|
|
schema, err := jsonschema.Compile(schemaFilePath)
|
|
|
|
assert.NoError(t, err)
|
|
|
|
|
|
|
|
var data interface{}
|
|
|
|
err = json.Unmarshal(resp.Body.Bytes(), &data)
|
|
|
|
assert.NoError(t, err)
|
2022-12-17 15:22:34 +09:00
|
|
|
|
2023-10-01 19:58:58 +09:00
|
|
|
schemaValidation := schema.Validate(data)
|
|
|
|
assert.Nil(t, schemaValidation)
|
2022-12-17 15:22:34 +09:00
|
|
|
}
|
|
|
|
|
2017-07-08 04:36:47 +09:00
|
|
|
func GetCSRF(t testing.TB, session *TestSession, urlStr string) string {
|
2019-07-29 13:15:18 +09:00
|
|
|
t.Helper()
|
2017-07-08 04:36:47 +09:00
|
|
|
req := NewRequest(t, "GET", urlStr)
|
|
|
|
resp := session.MakeRequest(t, req, http.StatusOK)
|
|
|
|
doc := NewHTMLParser(t, resp.Body)
|
|
|
|
return doc.GetCSRF()
|
|
|
|
}
|
2023-08-18 18:21:24 +09:00
|
|
|
|
|
|
|
func GetHTMLTitle(t testing.TB, session *TestSession, urlStr string) string {
|
|
|
|
t.Helper()
|
|
|
|
|
|
|
|
req := NewRequest(t, "GET", urlStr)
|
|
|
|
var resp *httptest.ResponseRecorder
|
|
|
|
if session == nil {
|
|
|
|
resp = MakeRequest(t, req, http.StatusOK)
|
|
|
|
} else {
|
|
|
|
resp = session.MakeRequest(t, req, http.StatusOK)
|
|
|
|
}
|
|
|
|
|
|
|
|
doc := NewHTMLParser(t, resp.Body)
|
|
|
|
return doc.Find("head title").Text()
|
|
|
|
}
|
2024-01-12 19:50:29 +09:00
|
|
|
|
|
|
|
func CreateDeclarativeRepo(t *testing.T, owner *user_model.User, name string, enabledUnits, disabledUnits []unit_model.Type, files []*files_service.ChangeRepoFile) (*repo_model.Repository, string, func()) {
|
|
|
|
t.Helper()
|
|
|
|
|
|
|
|
repoName := name
|
|
|
|
if repoName == "" {
|
|
|
|
repoName = gouuid.NewString()
|
|
|
|
}
|
|
|
|
|
|
|
|
// Create a new repository
|
|
|
|
repo, err := repo_service.CreateRepository(db.DefaultContext, owner, owner, repo_service.CreateRepoOptions{
|
|
|
|
Name: repoName,
|
|
|
|
Description: "Temporary Repo",
|
|
|
|
AutoInit: true,
|
|
|
|
Gitignores: "",
|
|
|
|
License: "WTFPL",
|
|
|
|
Readme: "Default",
|
|
|
|
DefaultBranch: "main",
|
|
|
|
})
|
|
|
|
assert.NoError(t, err)
|
|
|
|
assert.NotEmpty(t, repo)
|
|
|
|
|
|
|
|
if enabledUnits != nil || disabledUnits != nil {
|
|
|
|
units := make([]repo_model.RepoUnit, len(enabledUnits))
|
|
|
|
for i, unitType := range enabledUnits {
|
|
|
|
units[i] = repo_model.RepoUnit{
|
|
|
|
RepoID: repo.ID,
|
|
|
|
Type: unitType,
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2024-03-26 00:54:38 +09:00
|
|
|
err := repo_service.UpdateRepositoryUnits(db.DefaultContext, repo, units, disabledUnits)
|
2024-01-12 19:50:29 +09:00
|
|
|
assert.NoError(t, err)
|
|
|
|
}
|
|
|
|
|
|
|
|
var sha string
|
|
|
|
if len(files) > 0 {
|
|
|
|
resp, err := files_service.ChangeRepoFiles(git.DefaultContext, repo, owner, &files_service.ChangeRepoFilesOptions{
|
|
|
|
Files: files,
|
|
|
|
Message: "add files",
|
|
|
|
OldBranch: "main",
|
|
|
|
NewBranch: "main",
|
|
|
|
Author: &files_service.IdentityOptions{
|
|
|
|
Name: owner.Name,
|
|
|
|
Email: owner.Email,
|
|
|
|
},
|
|
|
|
Committer: &files_service.IdentityOptions{
|
|
|
|
Name: owner.Name,
|
|
|
|
Email: owner.Email,
|
|
|
|
},
|
|
|
|
Dates: &files_service.CommitDateOptions{
|
|
|
|
Author: time.Now(),
|
|
|
|
Committer: time.Now(),
|
|
|
|
},
|
|
|
|
})
|
|
|
|
assert.NoError(t, err)
|
|
|
|
assert.NotEmpty(t, resp)
|
|
|
|
|
|
|
|
sha = resp.Commit.SHA
|
|
|
|
}
|
|
|
|
|
|
|
|
return repo, sha, func() {
|
|
|
|
repo_service.DeleteRepository(db.DefaultContext, owner, repo, false)
|
|
|
|
}
|
|
|
|
}
|