You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
This repository are going to contain early stages of the pracuj / otodom scrapers. Here we will create fundamentals for the future one big project and the done things are going to be merged into one.
3
3
4
-
The tasks you can find in the corresponding directories.
4
+
The tasks you can find in the corresponding directories.
Copy file name to clipboardExpand all lines: otodom/task_1/task_1.md
+2-2Lines changed: 2 additions & 2 deletions
Original file line number
Diff line number
Diff line change
@@ -25,7 +25,7 @@ If something is missing you can leave the value as an empty string.
25
25
The Bot should be able to iterate through all the listings pages. The listings should be again collected and the duplicates should be removed.
26
26
### Task 2
27
27
28
-
Create a **settings.json** file. It should contain things which are going to define what bot is going to scrap. An example may look like:
28
+
Create a **settings.json** file. It should contain things which are going to define what bot is going to scrap. An example may look like:
29
29
```json
30
30
{
31
31
"base_url": "str",
@@ -40,4 +40,4 @@ Create a **settings.json** file. It should contain things which are going to def
40
40
```
41
41
and so on. Anything what may be usefull **please try to include**. Dependingly on the data the URL should be somehow generated. Look into Url how the Url is changed accordingly to what search parameters you applied on the site.
42
42
43
-
**Solutions** you can create in the **pracuj/task1/<your_name>** file and then make create a pull request.
43
+
**Solutions** you can create in the **pracuj/task1/<your_name>** file and then make create a pull request.
Copy file name to clipboardExpand all lines: pracuj/task_1/task_1.md
+2-2Lines changed: 2 additions & 2 deletions
Original file line number
Diff line number
Diff line change
@@ -20,7 +20,7 @@ If something is missing you can leave the value as an empty string.
20
20
The Bot should be able to iterate through all the listings pages. The listings should be again collected and the duplicates should be removed.
21
21
### Task 2
22
22
23
-
Create a **settings.json** file. It should contain things which are going to define what bot is going to scrap. An example may look like:
23
+
Create a **settings.json** file. It should contain things which are going to define what bot is going to scrap. An example may look like:
24
24
```json
25
25
{
26
26
"base_url": "str",
@@ -33,4 +33,4 @@ Create a **settings.json** file. It should contain things which are going to def
33
33
```
34
34
and so on. Anything what may be usefull **please try to include**. Start with the most important things. Dependingly on the data the URL should be somehow generated. Look into Url how the Url is changed accordingly to what search parameters you applied on the site.
35
35
36
-
**Solutions** you can create in the **pracuj/task1/<your_name>** file and then make create a pull request.
36
+
**Solutions** you can create in the **pracuj/task1/<your_name>** file and then make create a pull request.
0 commit comments