Bottom Page

Thread Rating:
  • 0 Vote(s) - 0 Average
  • 1
  • 2
  • 3
  • 4
  • 5
 webscrapping links and then enter those links to scrape data
#1
Hi, I am having difficulty trying to scrap webscrapping links and then enter those links to scrape data.

The webscrapping links is done but to enter those links from the first time webscrapping and entering those links inside and then collect data is another difficulty.

I have attached my first tier codes but i cant figure out the second tier codes.

Appreciate any kind help.

Thanks.

from bs4 import BeautifulSoup
import requests
import pandas as pd
from urllib.request import urlopen,urlparse, Request,HTTPError
import urllib
import re
import numpy as np
import csv
from http.client import BadStatusLine
import ssl
import json
#from googlesearch import search

class Google:
    @classmethod
    def search1(self, search):
      url_list = []   #store all the extracted urls in a List
      title_list = [] #store all the extracted titles in a List
      description_list = []  #store all the extracted Description in a List
      all_links = []
 
      for start in range(0,10):
        #page = requests.get('https://www.google.com/search?rlz=1C1CHBF_enSG851SG851&ei=Nib2XI6FEcmLvQS1xb-wBQ&q=site%3Alinkedin.com+inurl%3Ain+%7C+inurl%3Apub+%7C+inurl%3Aprofile+-inurl%3Adir+-inurl%3Atitle+-inurl%3Agroups+-inurl%3Acompany+-inurl%3Ajobs+-inurl%3Ajobs2+VP&oq=site%3Alinkedin.com+inurl%3Ain+%7C+inurl%3Apub+%7C+inurl%3Aprofile+-inurl%3Adir+-inurl%3Atitle+-inurl%3Agroups+-inurl%3Acompany+-inurl%3Ajobs+-inurl%3Ajobs2'+search+str(start*10), verify = False)
        page = requests.get('http://www.google.com/search?q='+search+str(start*10), verify = False, timeout=5)  
     
        #page = requests.get('https://www.google.com/search?q='+search, verify = True)
        soup = BeautifulSoup(page.content, "lxml")
        #soup = BeautifulSoup(page.content)

        
        for link in soup.find_all("a",href=re.compile("(?<=/url\?q=)(htt.*://.*)")): #original working code
            a = (re.split(":(?=http)",link["href"].replace("/url?q=","")))    
            a = a[0].split("&")[0]       
            url_list.append(a)   
Quote
#2
You would loop your url_list and create a request for each one in there.
Quote
#3
Hi metulburr, thanks for your quick reply. I'll work on it, thanks alot.
Quote

Top Page

Possibly Related Threads...
Thread Author Replies Views Last Post
  Webscrapping of Images that requires Authentication junos4350 0 54 Nov-15-2019, 06:24 AM
Last Post: junos4350
  Want to scrape a table data and export it into CSV format tahir1990 9 290 Oct-22-2019, 08:03 AM
Last Post: buran
  Extracting links from website with selenium bs4 and python M1ck0 1 208 Jul-20-2019, 10:29 PM
Last Post: Larz60+
  webscrapping links from pandas dataframe Wolverin 1 226 Jun-19-2019, 11:22 PM
Last Post: Larz60+
  webscrapping lists to dataframe kirito85 3 334 Jun-10-2019, 06:55 AM
Last Post: kirito85
  Scrape ASPX data with python... hoff1022 0 951 Feb-26-2019, 06:16 PM
Last Post: hoff1022
  How to exclude certain links while webscraping basis on keywords Prince_Bhatia 0 457 Oct-31-2018, 07:00 AM
Last Post: Prince_Bhatia
  Need To Scrape Some Links digitalmatic7 2 559 Oct-09-2018, 02:33 AM
Last Post: digitalmatic7
  Project: Opening all links within a web page Truman 0 531 Aug-07-2018, 12:33 AM
Last Post: Truman
  Download all secret links from a map design website fyec 0 685 Jul-24-2018, 09:08 PM
Last Post: fyec

Forum Jump:


Users browsing this thread: 1 Guest(s)