Delete one more e2e test
This commit is contained in:
parent
34e37c5671
commit
ff7b52cab1
@ -966,54 +966,6 @@ describe("E2E Tests for API Routes", () => {
|
|||||||
}, 180000); // 120 seconds
|
}, 180000); // 120 seconds
|
||||||
|
|
||||||
|
|
||||||
it.concurrent("should return a successful response with max depth option for a valid crawl job", async () => {
|
|
||||||
const crawlResponse = await request(TEST_URL)
|
|
||||||
.post("/v0/crawl")
|
|
||||||
.set("Authorization", `Bearer ${process.env.TEST_API_KEY}`)
|
|
||||||
.set("Content-Type", "application/json")
|
|
||||||
.send({
|
|
||||||
url: "https://www.scrapethissite.com",
|
|
||||||
crawlerOptions: { maxDepth: 1 },
|
|
||||||
});
|
|
||||||
expect(crawlResponse.statusCode).toBe(200);
|
|
||||||
|
|
||||||
let isCompleted = false;
|
|
||||||
let completedResponse;
|
|
||||||
|
|
||||||
while (!isCompleted) {
|
|
||||||
const response = await request(TEST_URL)
|
|
||||||
.get(`/v0/crawl/status/${crawlResponse.body.jobId}`)
|
|
||||||
.set("Authorization", `Bearer ${process.env.TEST_API_KEY}`);
|
|
||||||
expect(response.statusCode).toBe(200);
|
|
||||||
expect(response.body).toHaveProperty("status");
|
|
||||||
|
|
||||||
if (response.body.status === "completed") {
|
|
||||||
isCompleted = true;
|
|
||||||
completedResponse = response;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
expect(completedResponse.statusCode).toBe(200);
|
|
||||||
expect(completedResponse.body).toHaveProperty("status");
|
|
||||||
expect(completedResponse.body.status).toBe("completed");
|
|
||||||
expect(completedResponse.body).toHaveProperty("data");
|
|
||||||
expect(completedResponse.body.data[0]).toHaveProperty("content");
|
|
||||||
expect(completedResponse.body.data[0]).toHaveProperty("markdown");
|
|
||||||
expect(completedResponse.body.data[0]).toHaveProperty("metadata");
|
|
||||||
expect(completedResponse.body.data[0].metadata.pageStatusCode).toBe(200);
|
|
||||||
expect(completedResponse.body.data[0].metadata.pageError).toBeUndefined();
|
|
||||||
|
|
||||||
const urls = completedResponse.body.data.map(
|
|
||||||
(item: any) => item.metadata?.sourceURL
|
|
||||||
);
|
|
||||||
expect(urls.length).toBeGreaterThan(1);
|
|
||||||
|
|
||||||
// Check if all URLs have a maximum depth of 1
|
|
||||||
urls.forEach((url) => {
|
|
||||||
const pathSplits = new URL(url).pathname.split('/');
|
|
||||||
const depth = pathSplits.length - (pathSplits[0].length === 0 && pathSplits[pathSplits.length - 1].length === 0 ? 1 : 0);
|
|
||||||
expect(depth).toBeLessThanOrEqual(2);
|
|
||||||
});
|
|
||||||
}, 180000);
|
|
||||||
|
|
||||||
it.concurrent("should return a successful response for a valid crawl job with includeHtml set to true option (2)", async () => {
|
it.concurrent("should return a successful response for a valid crawl job with includeHtml set to true option (2)", async () => {
|
||||||
const crawlResponse = await request(TEST_URL)
|
const crawlResponse = await request(TEST_URL)
|
||||||
|
Loading…
Reference in New Issue
Block a user