added scraper for it-jobbank
This commit is contained in:
		
							parent
							
								
									979ed97738
								
							
						
					
					
						commit
						32f83e358b
					
				
							
								
								
									
										4
									
								
								.gitignore
									
									
									
									
										vendored
									
									
								
							
							
						
						
									
										4
									
								
								.gitignore
									
									
									
									
										vendored
									
									
								
							@ -1,2 +1,4 @@
 | 
				
			|||||||
/thehub_cache
 | 
					/thehub_cache
 | 
				
			||||||
/jobs.json
 | 
					/thehub.json
 | 
				
			||||||
 | 
					/itjobbank_cache
 | 
				
			||||||
 | 
					/it-jobbank.json
 | 
				
			||||||
							
								
								
									
										178
									
								
								main.go
									
									
									
									
									
								
							
							
						
						
									
										178
									
								
								main.go
									
									
									
									
									
								
							@ -12,15 +12,16 @@ import (
 | 
				
			|||||||
)
 | 
					)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
type job struct {
 | 
					type job struct {
 | 
				
			||||||
	Title       string `json:"title"`
 | 
						Title       string    `json:"title"`
 | 
				
			||||||
	Logo        string `json:"logo"`
 | 
						Logo        string    `json:"logo"`
 | 
				
			||||||
	Company     string `json:"company"`
 | 
						Company     string    `json:"company"`
 | 
				
			||||||
	Location    string `json:"location"`
 | 
						Location    string    `json:"location"`
 | 
				
			||||||
	Type        string `json:"type"`
 | 
						Type        string    `json:"type"`
 | 
				
			||||||
	Description string `json:"description"`
 | 
						Description string    `json:"description"`
 | 
				
			||||||
	Link        string `json:"link"`
 | 
						Link        string    `json:"link"`
 | 
				
			||||||
	Skills      skills `json:"skills"`
 | 
						Skills      skills    `json:"skills"`
 | 
				
			||||||
	FirstSeen  	time.Time `json:"firstSeen"`
 | 
						FirstSeen   time.Time `json:"firstSeen"`
 | 
				
			||||||
 | 
						Source      string    `json:"source"`
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
type skills struct {
 | 
					type skills struct {
 | 
				
			||||||
@ -33,6 +34,9 @@ type skills struct {
 | 
				
			|||||||
	Tailwind   bool `json:"tailwind"`
 | 
						Tailwind   bool `json:"tailwind"`
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					// Utility functions
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					// Checks if a string contains any of the given keywords
 | 
				
			||||||
func skillChecker(description string) skills {
 | 
					func skillChecker(description string) skills {
 | 
				
			||||||
	return skills{
 | 
						return skills{
 | 
				
			||||||
		React:      strings.Contains(description, "React"),
 | 
							React:      strings.Contains(description, "React"),
 | 
				
			||||||
@ -44,7 +48,9 @@ func skillChecker(description string) skills {
 | 
				
			|||||||
		Tailwind:   strings.Contains(description, "Tailwind"),
 | 
							Tailwind:   strings.Contains(description, "Tailwind"),
 | 
				
			||||||
	}
 | 
						}
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
func jobsToJson(file *os.File) {
 | 
					
 | 
				
			||||||
 | 
					// Converts job struct to json
 | 
				
			||||||
 | 
					func jobsToJson(file *os.File, jobs []job, fName string) {
 | 
				
			||||||
	// Encode jobs slice to JSON
 | 
						// Encode jobs slice to JSON
 | 
				
			||||||
	encoder := json.NewEncoder(file)
 | 
						encoder := json.NewEncoder(file)
 | 
				
			||||||
	encoder.SetIndent("", "  ") // Pretty-print with indentation
 | 
						encoder.SetIndent("", "  ") // Pretty-print with indentation
 | 
				
			||||||
@ -55,24 +61,47 @@ func jobsToJson(file *os.File) {
 | 
				
			|||||||
	fmt.Println("Job details successfully written to", fName)
 | 
						fmt.Println("Job details successfully written to", fName)
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					func checkIfPaid(description string) {
 | 
				
			||||||
 | 
						for _, keyword := range unpaidKeywords {
 | 
				
			||||||
 | 
							if strings.Contains(strings.ToLower(description), keyword) {
 | 
				
			||||||
 | 
								return
 | 
				
			||||||
 | 
							}
 | 
				
			||||||
 | 
						}
 | 
				
			||||||
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					func checkIfStudent(description string) string {
 | 
				
			||||||
 | 
						for _, keyword := range studentKeywords {
 | 
				
			||||||
 | 
							if strings.Contains(strings.ToLower(description), keyword) {
 | 
				
			||||||
 | 
								return "student"
 | 
				
			||||||
 | 
							}
 | 
				
			||||||
 | 
						}
 | 
				
			||||||
 | 
						return "full time"
 | 
				
			||||||
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
// Slice to store job details
 | 
					// Slice to store job details
 | 
				
			||||||
var (
 | 
					var (
 | 
				
			||||||
	jobs     []job
 | 
						excluded        = []string{"senior", "lead", "founder", "cto", "vp of", "erfaren", "arkitekt", "architect", "manager", "ulønnet", "unpaid", "praktik"}
 | 
				
			||||||
	jobCount int
 | 
						unpaidKeywords  = []string{"unpaid", "praktik", "ulønnet"}
 | 
				
			||||||
	maxJobs  int    = 30
 | 
						studentKeywords = []string{"studerende", "studenter", "student", "medhjælper"}
 | 
				
			||||||
	fName    string = "jobs.json"
 | 
					 | 
				
			||||||
	excluded        = []string{"senior", "lead", "founder", "cto", "vp of"}
 | 
					 | 
				
			||||||
)
 | 
					)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
func scrapeHub() {
 | 
					func scrapeHub() {
 | 
				
			||||||
 | 
						// declare and initialize variables
 | 
				
			||||||
 | 
						var (
 | 
				
			||||||
 | 
							jobs         []job
 | 
				
			||||||
 | 
							jobCount     int
 | 
				
			||||||
 | 
							fName        = "thehub.json"
 | 
				
			||||||
 | 
							maxJobs      = 30
 | 
				
			||||||
 | 
							baseUrl      = "https://thehub.io"
 | 
				
			||||||
 | 
							searchString = "https://thehub.io/jobs?roles=frontenddeveloper&roles=fullstackdeveloper&roles=backenddeveloper&roles=devops&paid=true&countryCode=DK&sorting=newJobs"
 | 
				
			||||||
 | 
							file, err    = os.Create(fName)
 | 
				
			||||||
 | 
						)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	file, err := os.Create(fName)
 | 
					 | 
				
			||||||
	if err != nil {
 | 
						if err != nil {
 | 
				
			||||||
		log.Fatalf("Cannot create file %q: %s", fName, err)
 | 
							log.Fatalf("Cannot create file %q: %s", fName, err)
 | 
				
			||||||
	}
 | 
						}
 | 
				
			||||||
	defer file.Close()
 | 
						defer file.Close()
 | 
				
			||||||
	baseUrl := "https://thehub.io"
 | 
					
 | 
				
			||||||
	searchString := "https://thehub.io/jobs?roles=frontenddeveloper&roles=fullstackdeveloper&roles=backenddeveloper&roles=devops&paid=true&countryCode=DK&sorting=newJobs"
 | 
					 | 
				
			||||||
	// Instantiate default collector
 | 
						// Instantiate default collector
 | 
				
			||||||
	c := colly.NewCollector(
 | 
						c := colly.NewCollector(
 | 
				
			||||||
		// visit only the hub
 | 
							// visit only the hub
 | 
				
			||||||
@ -86,7 +115,9 @@ func scrapeHub() {
 | 
				
			|||||||
	detailsCollector := c.Clone()
 | 
						detailsCollector := c.Clone()
 | 
				
			||||||
	// On every <div> element with class "card__content attribute call callback
 | 
						// On every <div> element with class "card__content attribute call callback
 | 
				
			||||||
	c.OnHTML("div[class=card__content]", func(e *colly.HTMLElement) {
 | 
						c.OnHTML("div[class=card__content]", func(e *colly.HTMLElement) {
 | 
				
			||||||
		//ensure only scrape the amount of jobs specified
 | 
							if jobCount >= maxJobs {
 | 
				
			||||||
 | 
								return
 | 
				
			||||||
 | 
							}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
		// Get the title and ensure it doesn't contain any excluded words
 | 
							// Get the title and ensure it doesn't contain any excluded words
 | 
				
			||||||
		title := e.ChildText("span.card-job-find-list__position")
 | 
							title := e.ChildText("span.card-job-find-list__position")
 | 
				
			||||||
@ -114,14 +145,14 @@ func scrapeHub() {
 | 
				
			|||||||
		cutRight := ");"
 | 
							cutRight := ");"
 | 
				
			||||||
		trimmedLogo := strings.Trim(logo, cutLeft+cutRight)
 | 
							trimmedLogo := strings.Trim(logo, cutLeft+cutRight)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
		// Get job description
 | 
							// Get the HTML of the description and check to see if it's paid
 | 
				
			||||||
		// Get the HTML of the description
 | 
					 | 
				
			||||||
		descriptionHTML, err := e.DOM.Find("content.text-block__content > span").Html()
 | 
							descriptionHTML, err := e.DOM.Find("content.text-block__content > span").Html()
 | 
				
			||||||
		if err != nil {
 | 
							if err != nil {
 | 
				
			||||||
			log.Printf("Error getting HTML of description: %s", err)
 | 
								log.Printf("Error getting HTML of description: %s", err)
 | 
				
			||||||
			return
 | 
								return
 | 
				
			||||||
		}
 | 
							}
 | 
				
			||||||
		// Get company name
 | 
					
 | 
				
			||||||
 | 
							// fill in the job struct
 | 
				
			||||||
		jobDetails := job{
 | 
							jobDetails := job{
 | 
				
			||||||
			Title:       e.ChildText("h2[class=view-job-details__title]"),
 | 
								Title:       e.ChildText("h2[class=view-job-details__title]"),
 | 
				
			||||||
			Logo:        trimmedLogo,
 | 
								Logo:        trimmedLogo,
 | 
				
			||||||
@ -132,13 +163,14 @@ func scrapeHub() {
 | 
				
			|||||||
			Link:        e.Request.URL.String(),
 | 
								Link:        e.Request.URL.String(),
 | 
				
			||||||
			Skills:      skillChecker(e.ChildText("content.text-block__content > span")),
 | 
								Skills:      skillChecker(e.ChildText("content.text-block__content > span")),
 | 
				
			||||||
			FirstSeen:   time.Now(),
 | 
								FirstSeen:   time.Now(),
 | 
				
			||||||
 | 
								Source:      baseUrl,
 | 
				
			||||||
		}
 | 
							}
 | 
				
			||||||
		jobs = append(jobs, jobDetails)
 | 
							jobs = append(jobs, jobDetails)
 | 
				
			||||||
		jobCount++
 | 
							jobCount++
 | 
				
			||||||
		fmt.Println("Scraped job", jobCount)
 | 
							fmt.Println("Scraped job", jobCount)
 | 
				
			||||||
		if jobCount == maxJobs {
 | 
							if jobCount >= maxJobs {
 | 
				
			||||||
			jobsToJson(file)
 | 
								jobsToJson(file, jobs, fName)
 | 
				
			||||||
			os.Exit(0)
 | 
								return
 | 
				
			||||||
		}
 | 
							}
 | 
				
			||||||
	})
 | 
						})
 | 
				
			||||||
	// Handle pagination
 | 
						// Handle pagination
 | 
				
			||||||
@ -146,7 +178,6 @@ func scrapeHub() {
 | 
				
			|||||||
		nextPage := e.Attr("href")
 | 
							nextPage := e.Attr("href")
 | 
				
			||||||
		if nextPage != "" {
 | 
							if nextPage != "" {
 | 
				
			||||||
			fullNextPage := baseUrl + nextPage
 | 
								fullNextPage := baseUrl + nextPage
 | 
				
			||||||
			fmt.Println("Visiting next page:", fullNextPage)
 | 
					 | 
				
			||||||
			e.Request.Visit(fullNextPage)
 | 
								e.Request.Visit(fullNextPage)
 | 
				
			||||||
		}
 | 
							}
 | 
				
			||||||
	})
 | 
						})
 | 
				
			||||||
@ -155,8 +186,103 @@ func scrapeHub() {
 | 
				
			|||||||
 | 
					
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					func scrapeItJobBank() {
 | 
				
			||||||
 | 
						// declare and initialize variables
 | 
				
			||||||
 | 
						var (
 | 
				
			||||||
 | 
							jobs         []job
 | 
				
			||||||
 | 
							jobCount     int
 | 
				
			||||||
 | 
							fName        = "it-jobbank.json"
 | 
				
			||||||
 | 
							maxJobs      = 30
 | 
				
			||||||
 | 
							baseUrl      = "https://www.it-jobbank.dk"
 | 
				
			||||||
 | 
							searchString = "https://www.it-jobbank.dk/jobsoegning"
 | 
				
			||||||
 | 
							file, err    = os.Create(fName)
 | 
				
			||||||
 | 
						)
 | 
				
			||||||
 | 
						if err != nil {
 | 
				
			||||||
 | 
							log.Fatalf("Cannot create file %q: %s", fName, err)
 | 
				
			||||||
 | 
						}
 | 
				
			||||||
 | 
						defer file.Close()
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
						// Instantiate default collector
 | 
				
			||||||
 | 
						c := colly.NewCollector(
 | 
				
			||||||
 | 
							// visit only the hub
 | 
				
			||||||
 | 
							colly.AllowedDomains("www.it-jobbank.dk", "it-jobbank.dk"),
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
							// Cache responses to prevent multiple requests
 | 
				
			||||||
 | 
							colly.CacheDir("./itjobbank_cache"),
 | 
				
			||||||
 | 
						)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
						// Instantiate a new collector to visit the job details page
 | 
				
			||||||
 | 
						detailsCollector := c.Clone()
 | 
				
			||||||
 | 
						// On every <div> element with class "card__content attribute call callback
 | 
				
			||||||
 | 
						c.OnHTML("div[class=result]", func(e *colly.HTMLElement) {
 | 
				
			||||||
 | 
							if jobCount >= maxJobs {
 | 
				
			||||||
 | 
								return
 | 
				
			||||||
 | 
							}
 | 
				
			||||||
 | 
							// Get the title and ensure it doesn't contain any excluded words
 | 
				
			||||||
 | 
							title := e.ChildText("h3.job-title > a")
 | 
				
			||||||
 | 
							for _, excludedWord := range excluded {
 | 
				
			||||||
 | 
								if strings.Contains(strings.ToLower(title), excludedWord) {
 | 
				
			||||||
 | 
									return
 | 
				
			||||||
 | 
								}
 | 
				
			||||||
 | 
							}
 | 
				
			||||||
 | 
							fullLink := e.ChildAttr("h3.job-title > a", "href")
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
							detailsCollector.Visit(fullLink)
 | 
				
			||||||
 | 
						})
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
						detailsCollector.OnRequest(func(r *colly.Request) {
 | 
				
			||||||
 | 
							fmt.Println("Visiting", r.URL.String())
 | 
				
			||||||
 | 
						})
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
						detailsCollector.OnHTML("section > div", func(e *colly.HTMLElement) {
 | 
				
			||||||
 | 
							// get the description as html
 | 
				
			||||||
 | 
							descriptionHTML, err := e.DOM.Find("div[id=job_ad]").Html()
 | 
				
			||||||
 | 
							if err != nil {
 | 
				
			||||||
 | 
								log.Printf("Error getting HTML of description: %s", err)
 | 
				
			||||||
 | 
								return
 | 
				
			||||||
 | 
							}
 | 
				
			||||||
 | 
							// Check if the job is paid
 | 
				
			||||||
 | 
							checkIfPaid(descriptionHTML)
 | 
				
			||||||
 | 
							// fill in the job struct
 | 
				
			||||||
 | 
							title := e.ChildText("h1.title")
 | 
				
			||||||
 | 
							if title == "" {
 | 
				
			||||||
 | 
								title = e.ChildText("h1[id=jobtitle]")
 | 
				
			||||||
 | 
							}
 | 
				
			||||||
 | 
							jobDetails := job{
 | 
				
			||||||
 | 
								Title:       title,
 | 
				
			||||||
 | 
								Logo:        baseUrl + e.ChildAttr("div.companmony-logo > img", "src"),
 | 
				
			||||||
 | 
								Company:     e.ChildText("p.published"),
 | 
				
			||||||
 | 
								Location:    e.ChildText("div.job-location > p.caption"),
 | 
				
			||||||
 | 
								Type:        checkIfStudent(descriptionHTML),
 | 
				
			||||||
 | 
								Description: descriptionHTML,
 | 
				
			||||||
 | 
								Link:        e.Request.URL.String(),
 | 
				
			||||||
 | 
								Skills:      skillChecker(descriptionHTML),
 | 
				
			||||||
 | 
								FirstSeen:   time.Now(),
 | 
				
			||||||
 | 
								Source:      baseUrl,
 | 
				
			||||||
 | 
							}
 | 
				
			||||||
 | 
							jobs = append(jobs, jobDetails)
 | 
				
			||||||
 | 
							jobCount++
 | 
				
			||||||
 | 
							fmt.Println("Scraped job", jobCount)
 | 
				
			||||||
 | 
							if jobCount >= maxJobs {
 | 
				
			||||||
 | 
								jobsToJson(file, jobs, fName)
 | 
				
			||||||
 | 
								return
 | 
				
			||||||
 | 
							}
 | 
				
			||||||
 | 
						})
 | 
				
			||||||
 | 
						// Handle pagination
 | 
				
			||||||
 | 
						c.OnHTML("a.page-link", func(e *colly.HTMLElement) {
 | 
				
			||||||
 | 
							if jobCount >= maxJobs {
 | 
				
			||||||
 | 
								return
 | 
				
			||||||
 | 
							}
 | 
				
			||||||
 | 
							nextPage := e.Attr("href")
 | 
				
			||||||
 | 
							if nextPage != "" {
 | 
				
			||||||
 | 
								e.Request.Visit(nextPage)
 | 
				
			||||||
 | 
							}
 | 
				
			||||||
 | 
						})
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
						c.Visit(searchString)
 | 
				
			||||||
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
func main() {
 | 
					func main() {
 | 
				
			||||||
	scrapeHub()
 | 
						scrapeHub()
 | 
				
			||||||
 | 
						scrapeItJobBank()
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
				
			|||||||
		Loading…
	
	
			
			x
			
			
		
	
		Reference in New Issue
	
	Block a user