JobOffers_ESP / LinksObtainer.py
Miguel Gómez Prieto
Datast upload
4c4be9f
from playwright.sync_api import sync_playwright
from urllib.parse import urljoin
import time, sys
BASE_URL = "https://europa.eu/eures/portal/jv-se/search"
DEFAULT_PARAMS = {
"page": "1",
"resultsPerPage": "50",
"orderBy": "BEST_MATCH",
"locationCodes": "es",
"lang": "en",
}
def build_search_url(params=None):
params = params or DEFAULT_PARAMS
query = "&".join(f"{k}={v}" for k, v in params.items())
return f"{BASE_URL}?{query}"
def get_offer_links(max_pages=3, wait_ms=1500, headless=True, params=None, start_page=1):
all_links = []
try:
with sync_playwright() as p:
browser = p.chromium.launch(headless=headless)
context = browser.new_context()
page = context.new_page()
for page_num in range(start_page, max_pages + 1):
try:
local_params = dict(params or DEFAULT_PARAMS)
local_params["page"] = str(page_num)
url = build_search_url(local_params)
page.goto(url, wait_until="networkidle", timeout=15000)
time.sleep(wait_ms / 1000.0)
anchors = page.locator("a[id^='jv-result-summary-title-']").all()
if not anchors:
print(f"\n[!] Página {page_num} sin resultados, deteniendo.")
break
for a in anchors:
href = a.get_attribute("href")
if href:
full_url = urljoin("https://europa.eu", href)
all_links.append(full_url)
# Progreso en vivo
progress = (page_num / max_pages) * 100
sys.stdout.write(
f"\rProgreso: {progress:.1f}% | URLs recogidas: {len(all_links)}"
)
sys.stdout.flush()
except Exception as e:
print(f"\n[!] Error en página {page_num}: {e}")
break
browser.close()
except Exception as e:
print(f"\n[!] Error general: {e}")
# Deduplicar
seen, deduped = set(), []
for link in all_links:
if link not in seen:
seen.add(link)
deduped.append(link)
print("\n[✓] Scraping completado.")
return deduped
if __name__ == "__main__":
from datetime import datetime
import os
inicio = datetime.now()
script_dir = os.path.dirname(os.path.abspath(__file__))
timestamp = datetime.now().strftime("%d_%H%M%S")
# Obtenemos los links
urls = get_offer_links(100000, 500, start_page=202)
print(len(urls))
print(timestamp)
print()
fin = datetime.now()
duracion = fin - inicio
print(f"Duración en segundos: {duracion.total_seconds()}")
file_path = os.path.join(script_dir, "links", f"links{timestamp}.txt")
with open(file_path, "w") as f:
f.writelines([url + "\n" for url in urls])
print()
print("URL guardadas en: "+ file_path)