[Tutor] Web scraping using selenium and navigating nested dictionaries / lists.

Peter Otten __peter__ at web.de
Sun Jan 27 06:13:10 EST 2019


mhysnm1964 at gmail.com wrote:

> All,
> 
>  
> 
> Goal of new project.
> 
> I want to scrape all my books from Audible.com that I have purchased.
> Eventually I want to export this as a CSV file or maybe Json. I have not
> got
> that far yet. The reasoning behind this is to  learn selenium  for my work
> and get the list of books I have purchased. Killing two birds with one
> stone
> here. The work focus is to see if selenium   can automate some of the
> testing I have to do and collect useful information from the web page for
> my reports. This part of the goal is in the future. As I need to build my
> python skills up.
> 
>  
> 
> Thus far, I have been successful in logging into Audible and showing the
> library of books. I am able to store the table of books and want to use
> BeautifulSoup to extract the relevant information. Information I will want
> from the table is:
> 
> *	Author
> *	Title
> *	Date purchased
> *	Length
> *	Is the book in a series (there is a link for this)
> *	Link to the page storing the publish details.
> *	Download link
> 
> Hopefully this has given you enough information on what I am trying to
> achieve at this stage. AS I learn more about what I am doing, I am adding
> possible extra's tasks. Such as verifying if I have the book already
> download via itunes.
> 
>  
> 
> Learning goals:
> 
> Using the BeautifulSoup  structure that I have extracted from the page
> source for the table. I want to navigate the tree structure. BeautifulSoup
> provides children, siblings and parents methods. This is where I get stuck
> with programming logic. BeautifulSoup does provide find_all method plus
> selectors which I do not want to use for this exercise. As I want to learn
> how to walk a tree starting at the root and visiting each node of the
> tree. 

I think you make your life harder than necessary if you avoid the tools 
provided by the library you are using.

> Then I can look at the attributes for the tag as I go. I believe I
> have to set up a recursive loop or function call. Not sure on how to do
> this. Pseudo code:
> 
>  
> 
> Build table structure
> 
> Start at the root node.
> 
> Check to see if there is any children.
> 
> Pass first child to function.
> 
> Print attributes for tag at this level
> 
> In function, check for any sibling nodes.
> 
> If exist, call function again
> 
> If no siblings, then start at first sibling and get its child.
> 
>  
> 
> This is where I get struck. Each sibling can have children and they can
> have siblings. So how do I ensure I visit each node in the tree?

The problem with your description is that siblings do not matter. Just

- process root
- iterate over its children and call the function recursively with every
  child as the new root.

To make the function more useful you can pass a function instead of hard-
coding what you want to do with the elements. Given

def process_elements(elem, do_stuff):
    do_stuff(elem)
    for child in elem.children:
        process_elements(child, do_stuff)

you can print all elements with

soup = BeautifulSoup(...)
process_elements(soup, print)

and

process_elements(soup, lambda elem: print(elem.name))
will print only the names.

You need a bit of error checking to make it work, though.

But wait -- Python's generators let you rewrite process_elements so that you 
can use it without a callback:

def gen_elements(elem):
    yield elem
    for child in elem.children:
        yield from gen_elements(child)

for elem in gen_elements(soup):
    print(elem.name)

Note that 'yield from iterable' is a shortcut for
'for x in iterable: yield x', so there are actually two loops in 
gen_elements().

> Any tips or tricks for this would be grateful. As I could use this in
> other situations.




More information about the Tutor mailing list