Cannot get table data - HTML











up vote
2
down vote

favorite












I am trying to get the 'Earnings Announcements table' from: https://www.zacks.com/stock/research/amzn/earnings-announcements



I am using different beautifulsoup options but none get the table.



table = soup.find('table', attrs={'class': 'earnings_announcements_earnings_table'})

table = soup.find_all('table')


When I inspect the table, the elements of the table are there.



I am pasting a portion of the code I am getting for the table (js, json?).



document.obj_data = {
"earnings_announcements_earnings_table" :
[ [ "10/26/2017", "9/2017", "$0.06", "--", "--", "--", "--" ] , [ "7/27/2017", "6/2017", "$1.40", "$0.40", "<div class="right neg negative neg_icon showinline down">-1.00</div>", "<div class="right neg negative neg_icon showinline down">-71.43%</div>", "After Close" ] , [ "4/27/2017", "3/2017", "$1.03", "$1.48", "<div class="right pos positive pos_icon showinline up">+0.45</div>", "<div class="right pos positive pos_icon showinline up">+43.69%</div>", "After Close" ] , [ "2/2/2017", "12/2016", "$1.40", "$1.54", "<div class="right pos positive pos_icon showinline up">+0.14</div>", "<div class="right pos positive pos_icon showinline up">+10.00%</div>", "After Close" ] , [ "10/27/2016", "9/2016", "$0.85", "$0.52", "<div class="right neg negative neg_icon showinline down">-0.33</div>", "<div class="right neg negative neg_icon showinline down">-38.82%</div>", "After Close" ] , [ "7/28/2016", "6/2016", "$1.14", "$1.78", "<div class="right pos positive pos_icon showinline up">+0.64</div>", "<div class="right pos positive pos_icon showinline up">+56.14%</div>", "After Close" ] , [ "4/28/2016", "3/2016", "$0.61", "$1.07", "<div class="right pos positive pos_icon showinline up">+0.46</div>", "<div class="right pos positive pos_icon showinline up">+75.41%</div>", "After Close" ] , [ "1/28/2016", "12/2015", "$1.61", "$1.00", "<div class="right neg negative neg_icon showinline down">-0.61</div>", "<div class="right neg negative neg_icon showinline down">-37.89%</div>", "After Close" ] , [ "10/22/2015", "9/2015", "-$0.1", "$0.17", "<div class="right pos positive pos_icon showinline up">+0.27</div>", "<div class="right pos positive pos_icon showinline up">+270.00%</div>", "After Close" ] , [ "7/23/2015", "6/2015", "-$0.15", "$0.19", "<div class="right pos positive pos_icon showinline up">+0.34</div>", "<div class="right pos positive pos_icon showinline up">+226.67%</div>", "After Close" ] , [ "4/23/2015", "3/2015", "-$0.13", "-$0.12", "<div class="right pos positive pos_icon showinline up">+0.01</div>", "<div class="right pos positive pos_icon showinline up">+7.69%</div>", "After Close" ] , [ "1/29/2015", "12/2014", "$0.24", "$0.45", "<div class="right pos positive pos_icon showinline up">+0.21</div>", "<div class="right pos positive pos_icon showinline up">+87.50%</div>", "After Close" ] , [ "10/23/2014", "9/2014", "-$0.73", "-$0.95", "<div class="right neg negative neg_icon showinline down">-0.22</div>", "<div class="right neg negative neg_icon showinline down">-30.14%</div>", "After Close" ] , [ "7/24/2014", "6/2014", "-$0.13", "-$0.27", "<div class="right neg negative neg_icon showinline down">-0.14</div>", "<div class="right neg negative neg_icon showinline down">-107.69%</div>", "After Close" ] , [ "4/24/2014", "3/2014", "$0.22", "$0.23", "<div class="right pos positive pos_icon showinline up">+0.01</div>", "<div class="right pos positive pos_icon showinline up">+4.55%</div>", "After Close" ] , [ "1/30/2014", "12/2013", "$0.68", "$0.51", "<div class="right neg negative neg_icon showinline down">-0.17</div>", "<div class="right neg negative neg_icon showinline down">-25.00%</div>", "After Close" ] , [ "10/24/2013", "9/2013", "-$0.09", "-$0.09", "<div class="right pos_na showinline">0.00</div>", "<div class="right pos_na showinline">0.00%</div>", "After Close" ] , [ "7/25/2013", "6/2013", "$0.04", "-$0.02", "<div class="right neg negative neg_icon showinline down">-0.06</div>", "<div class="right neg negative neg_icon showinline down">-150.00%</div>", "After Close" ] , [ "4/25/2013", "3/2013", "$0.10", "$0.18", "<div class="right pos positive pos_icon showinline up">+0.08</div>", "<div class="right pos positive pos_icon showinline up">+80.00%</div>", "After Close" ] , [ "1/29/2013", "12/2012", "$0.28", "$0.21", "<div class="right neg negative neg_icon showinline down">-0.07</div>", "<div class="right neg negative neg_icon showinline down">-25.00%</div>", "After Close" ] , [ "10/25/2012", "9/2012", "-$0.08", "-$0.23", "<div class="right neg negative neg_icon showinline down">-0.15</div>", "<div class="right neg negative neg_icon showinline down">-187.50%</div>", "After Close" ] , [ "7/26/2012", "6/2012", "--", "--", "--", "--", "After Close" ] , [ "4/26/2012", "3/2012", "--", "--", "--", "--", "After Close" ] , [ "1/31/2012", "12/2011", "--", "--", "--", "--", "After Close" ] , [ "10/25/2011", "9/2011", "--", "--", "--", "--", "After Close" ] , [ "7/26/2011", "6/2011", "--", "--", "--", "--", "After Close" ] , [ "4/26/2011", "3/2011", "--", "--", "--", "--", "--" ] , [ "1/27/2011", "12/2010", "--", "--", "--", "--", "After Close" ] , [ "10/21/2010", "9/2010", "--", "--", "--", "--", "After Close" ] , [ "7/22/2010", "6/2010", "--", "--", "--", "--", "After Close" ] , [ "4/22/2010", "3/2010", "--", "--", "--", "--", "After Close" ] , [ "1/28/2010", "12/2009", "--", "--", "--", "--", "After Close" ] , [ "10/22/2009", "9/2009", "--", "--", "--", "--", "After Close" ] , [ "7/23/2009", "6/2009", "--", "--", "--", "--", "After Close" ] ]


How could I get this table?
Thanks!










share|improve this question




















  • 1




    the data is loaded dynamically and not sown in html, so you have to parse the data you are getting.
    – MD. Khairul Basar
    Sep 3 '17 at 14:54










  • Thanks!! PhantomJS, selenium?
    – Diego
    Sep 3 '17 at 15:16










  • I checked page source and still looked the same, so I don't think it will help. But, still can give it a try.
    – MD. Khairul Basar
    Sep 3 '17 at 15:18






  • 1




    It worked with selenium!
    – Diego
    Sep 3 '17 at 16:39

















up vote
2
down vote

favorite












I am trying to get the 'Earnings Announcements table' from: https://www.zacks.com/stock/research/amzn/earnings-announcements



I am using different beautifulsoup options but none get the table.



table = soup.find('table', attrs={'class': 'earnings_announcements_earnings_table'})

table = soup.find_all('table')


When I inspect the table, the elements of the table are there.



I am pasting a portion of the code I am getting for the table (js, json?).



document.obj_data = {
"earnings_announcements_earnings_table" :
[ [ "10/26/2017", "9/2017", "$0.06", "--", "--", "--", "--" ] , [ "7/27/2017", "6/2017", "$1.40", "$0.40", "<div class="right neg negative neg_icon showinline down">-1.00</div>", "<div class="right neg negative neg_icon showinline down">-71.43%</div>", "After Close" ] , [ "4/27/2017", "3/2017", "$1.03", "$1.48", "<div class="right pos positive pos_icon showinline up">+0.45</div>", "<div class="right pos positive pos_icon showinline up">+43.69%</div>", "After Close" ] , [ "2/2/2017", "12/2016", "$1.40", "$1.54", "<div class="right pos positive pos_icon showinline up">+0.14</div>", "<div class="right pos positive pos_icon showinline up">+10.00%</div>", "After Close" ] , [ "10/27/2016", "9/2016", "$0.85", "$0.52", "<div class="right neg negative neg_icon showinline down">-0.33</div>", "<div class="right neg negative neg_icon showinline down">-38.82%</div>", "After Close" ] , [ "7/28/2016", "6/2016", "$1.14", "$1.78", "<div class="right pos positive pos_icon showinline up">+0.64</div>", "<div class="right pos positive pos_icon showinline up">+56.14%</div>", "After Close" ] , [ "4/28/2016", "3/2016", "$0.61", "$1.07", "<div class="right pos positive pos_icon showinline up">+0.46</div>", "<div class="right pos positive pos_icon showinline up">+75.41%</div>", "After Close" ] , [ "1/28/2016", "12/2015", "$1.61", "$1.00", "<div class="right neg negative neg_icon showinline down">-0.61</div>", "<div class="right neg negative neg_icon showinline down">-37.89%</div>", "After Close" ] , [ "10/22/2015", "9/2015", "-$0.1", "$0.17", "<div class="right pos positive pos_icon showinline up">+0.27</div>", "<div class="right pos positive pos_icon showinline up">+270.00%</div>", "After Close" ] , [ "7/23/2015", "6/2015", "-$0.15", "$0.19", "<div class="right pos positive pos_icon showinline up">+0.34</div>", "<div class="right pos positive pos_icon showinline up">+226.67%</div>", "After Close" ] , [ "4/23/2015", "3/2015", "-$0.13", "-$0.12", "<div class="right pos positive pos_icon showinline up">+0.01</div>", "<div class="right pos positive pos_icon showinline up">+7.69%</div>", "After Close" ] , [ "1/29/2015", "12/2014", "$0.24", "$0.45", "<div class="right pos positive pos_icon showinline up">+0.21</div>", "<div class="right pos positive pos_icon showinline up">+87.50%</div>", "After Close" ] , [ "10/23/2014", "9/2014", "-$0.73", "-$0.95", "<div class="right neg negative neg_icon showinline down">-0.22</div>", "<div class="right neg negative neg_icon showinline down">-30.14%</div>", "After Close" ] , [ "7/24/2014", "6/2014", "-$0.13", "-$0.27", "<div class="right neg negative neg_icon showinline down">-0.14</div>", "<div class="right neg negative neg_icon showinline down">-107.69%</div>", "After Close" ] , [ "4/24/2014", "3/2014", "$0.22", "$0.23", "<div class="right pos positive pos_icon showinline up">+0.01</div>", "<div class="right pos positive pos_icon showinline up">+4.55%</div>", "After Close" ] , [ "1/30/2014", "12/2013", "$0.68", "$0.51", "<div class="right neg negative neg_icon showinline down">-0.17</div>", "<div class="right neg negative neg_icon showinline down">-25.00%</div>", "After Close" ] , [ "10/24/2013", "9/2013", "-$0.09", "-$0.09", "<div class="right pos_na showinline">0.00</div>", "<div class="right pos_na showinline">0.00%</div>", "After Close" ] , [ "7/25/2013", "6/2013", "$0.04", "-$0.02", "<div class="right neg negative neg_icon showinline down">-0.06</div>", "<div class="right neg negative neg_icon showinline down">-150.00%</div>", "After Close" ] , [ "4/25/2013", "3/2013", "$0.10", "$0.18", "<div class="right pos positive pos_icon showinline up">+0.08</div>", "<div class="right pos positive pos_icon showinline up">+80.00%</div>", "After Close" ] , [ "1/29/2013", "12/2012", "$0.28", "$0.21", "<div class="right neg negative neg_icon showinline down">-0.07</div>", "<div class="right neg negative neg_icon showinline down">-25.00%</div>", "After Close" ] , [ "10/25/2012", "9/2012", "-$0.08", "-$0.23", "<div class="right neg negative neg_icon showinline down">-0.15</div>", "<div class="right neg negative neg_icon showinline down">-187.50%</div>", "After Close" ] , [ "7/26/2012", "6/2012", "--", "--", "--", "--", "After Close" ] , [ "4/26/2012", "3/2012", "--", "--", "--", "--", "After Close" ] , [ "1/31/2012", "12/2011", "--", "--", "--", "--", "After Close" ] , [ "10/25/2011", "9/2011", "--", "--", "--", "--", "After Close" ] , [ "7/26/2011", "6/2011", "--", "--", "--", "--", "After Close" ] , [ "4/26/2011", "3/2011", "--", "--", "--", "--", "--" ] , [ "1/27/2011", "12/2010", "--", "--", "--", "--", "After Close" ] , [ "10/21/2010", "9/2010", "--", "--", "--", "--", "After Close" ] , [ "7/22/2010", "6/2010", "--", "--", "--", "--", "After Close" ] , [ "4/22/2010", "3/2010", "--", "--", "--", "--", "After Close" ] , [ "1/28/2010", "12/2009", "--", "--", "--", "--", "After Close" ] , [ "10/22/2009", "9/2009", "--", "--", "--", "--", "After Close" ] , [ "7/23/2009", "6/2009", "--", "--", "--", "--", "After Close" ] ]


How could I get this table?
Thanks!










share|improve this question




















  • 1




    the data is loaded dynamically and not sown in html, so you have to parse the data you are getting.
    – MD. Khairul Basar
    Sep 3 '17 at 14:54










  • Thanks!! PhantomJS, selenium?
    – Diego
    Sep 3 '17 at 15:16










  • I checked page source and still looked the same, so I don't think it will help. But, still can give it a try.
    – MD. Khairul Basar
    Sep 3 '17 at 15:18






  • 1




    It worked with selenium!
    – Diego
    Sep 3 '17 at 16:39















up vote
2
down vote

favorite









up vote
2
down vote

favorite











I am trying to get the 'Earnings Announcements table' from: https://www.zacks.com/stock/research/amzn/earnings-announcements



I am using different beautifulsoup options but none get the table.



table = soup.find('table', attrs={'class': 'earnings_announcements_earnings_table'})

table = soup.find_all('table')


When I inspect the table, the elements of the table are there.



I am pasting a portion of the code I am getting for the table (js, json?).



document.obj_data = {
"earnings_announcements_earnings_table" :
[ [ "10/26/2017", "9/2017", "$0.06", "--", "--", "--", "--" ] , [ "7/27/2017", "6/2017", "$1.40", "$0.40", "<div class="right neg negative neg_icon showinline down">-1.00</div>", "<div class="right neg negative neg_icon showinline down">-71.43%</div>", "After Close" ] , [ "4/27/2017", "3/2017", "$1.03", "$1.48", "<div class="right pos positive pos_icon showinline up">+0.45</div>", "<div class="right pos positive pos_icon showinline up">+43.69%</div>", "After Close" ] , [ "2/2/2017", "12/2016", "$1.40", "$1.54", "<div class="right pos positive pos_icon showinline up">+0.14</div>", "<div class="right pos positive pos_icon showinline up">+10.00%</div>", "After Close" ] , [ "10/27/2016", "9/2016", "$0.85", "$0.52", "<div class="right neg negative neg_icon showinline down">-0.33</div>", "<div class="right neg negative neg_icon showinline down">-38.82%</div>", "After Close" ] , [ "7/28/2016", "6/2016", "$1.14", "$1.78", "<div class="right pos positive pos_icon showinline up">+0.64</div>", "<div class="right pos positive pos_icon showinline up">+56.14%</div>", "After Close" ] , [ "4/28/2016", "3/2016", "$0.61", "$1.07", "<div class="right pos positive pos_icon showinline up">+0.46</div>", "<div class="right pos positive pos_icon showinline up">+75.41%</div>", "After Close" ] , [ "1/28/2016", "12/2015", "$1.61", "$1.00", "<div class="right neg negative neg_icon showinline down">-0.61</div>", "<div class="right neg negative neg_icon showinline down">-37.89%</div>", "After Close" ] , [ "10/22/2015", "9/2015", "-$0.1", "$0.17", "<div class="right pos positive pos_icon showinline up">+0.27</div>", "<div class="right pos positive pos_icon showinline up">+270.00%</div>", "After Close" ] , [ "7/23/2015", "6/2015", "-$0.15", "$0.19", "<div class="right pos positive pos_icon showinline up">+0.34</div>", "<div class="right pos positive pos_icon showinline up">+226.67%</div>", "After Close" ] , [ "4/23/2015", "3/2015", "-$0.13", "-$0.12", "<div class="right pos positive pos_icon showinline up">+0.01</div>", "<div class="right pos positive pos_icon showinline up">+7.69%</div>", "After Close" ] , [ "1/29/2015", "12/2014", "$0.24", "$0.45", "<div class="right pos positive pos_icon showinline up">+0.21</div>", "<div class="right pos positive pos_icon showinline up">+87.50%</div>", "After Close" ] , [ "10/23/2014", "9/2014", "-$0.73", "-$0.95", "<div class="right neg negative neg_icon showinline down">-0.22</div>", "<div class="right neg negative neg_icon showinline down">-30.14%</div>", "After Close" ] , [ "7/24/2014", "6/2014", "-$0.13", "-$0.27", "<div class="right neg negative neg_icon showinline down">-0.14</div>", "<div class="right neg negative neg_icon showinline down">-107.69%</div>", "After Close" ] , [ "4/24/2014", "3/2014", "$0.22", "$0.23", "<div class="right pos positive pos_icon showinline up">+0.01</div>", "<div class="right pos positive pos_icon showinline up">+4.55%</div>", "After Close" ] , [ "1/30/2014", "12/2013", "$0.68", "$0.51", "<div class="right neg negative neg_icon showinline down">-0.17</div>", "<div class="right neg negative neg_icon showinline down">-25.00%</div>", "After Close" ] , [ "10/24/2013", "9/2013", "-$0.09", "-$0.09", "<div class="right pos_na showinline">0.00</div>", "<div class="right pos_na showinline">0.00%</div>", "After Close" ] , [ "7/25/2013", "6/2013", "$0.04", "-$0.02", "<div class="right neg negative neg_icon showinline down">-0.06</div>", "<div class="right neg negative neg_icon showinline down">-150.00%</div>", "After Close" ] , [ "4/25/2013", "3/2013", "$0.10", "$0.18", "<div class="right pos positive pos_icon showinline up">+0.08</div>", "<div class="right pos positive pos_icon showinline up">+80.00%</div>", "After Close" ] , [ "1/29/2013", "12/2012", "$0.28", "$0.21", "<div class="right neg negative neg_icon showinline down">-0.07</div>", "<div class="right neg negative neg_icon showinline down">-25.00%</div>", "After Close" ] , [ "10/25/2012", "9/2012", "-$0.08", "-$0.23", "<div class="right neg negative neg_icon showinline down">-0.15</div>", "<div class="right neg negative neg_icon showinline down">-187.50%</div>", "After Close" ] , [ "7/26/2012", "6/2012", "--", "--", "--", "--", "After Close" ] , [ "4/26/2012", "3/2012", "--", "--", "--", "--", "After Close" ] , [ "1/31/2012", "12/2011", "--", "--", "--", "--", "After Close" ] , [ "10/25/2011", "9/2011", "--", "--", "--", "--", "After Close" ] , [ "7/26/2011", "6/2011", "--", "--", "--", "--", "After Close" ] , [ "4/26/2011", "3/2011", "--", "--", "--", "--", "--" ] , [ "1/27/2011", "12/2010", "--", "--", "--", "--", "After Close" ] , [ "10/21/2010", "9/2010", "--", "--", "--", "--", "After Close" ] , [ "7/22/2010", "6/2010", "--", "--", "--", "--", "After Close" ] , [ "4/22/2010", "3/2010", "--", "--", "--", "--", "After Close" ] , [ "1/28/2010", "12/2009", "--", "--", "--", "--", "After Close" ] , [ "10/22/2009", "9/2009", "--", "--", "--", "--", "After Close" ] , [ "7/23/2009", "6/2009", "--", "--", "--", "--", "After Close" ] ]


How could I get this table?
Thanks!










share|improve this question















I am trying to get the 'Earnings Announcements table' from: https://www.zacks.com/stock/research/amzn/earnings-announcements



I am using different beautifulsoup options but none get the table.



table = soup.find('table', attrs={'class': 'earnings_announcements_earnings_table'})

table = soup.find_all('table')


When I inspect the table, the elements of the table are there.



I am pasting a portion of the code I am getting for the table (js, json?).



document.obj_data = {
"earnings_announcements_earnings_table" :
[ [ "10/26/2017", "9/2017", "$0.06", "--", "--", "--", "--" ] , [ "7/27/2017", "6/2017", "$1.40", "$0.40", "<div class="right neg negative neg_icon showinline down">-1.00</div>", "<div class="right neg negative neg_icon showinline down">-71.43%</div>", "After Close" ] , [ "4/27/2017", "3/2017", "$1.03", "$1.48", "<div class="right pos positive pos_icon showinline up">+0.45</div>", "<div class="right pos positive pos_icon showinline up">+43.69%</div>", "After Close" ] , [ "2/2/2017", "12/2016", "$1.40", "$1.54", "<div class="right pos positive pos_icon showinline up">+0.14</div>", "<div class="right pos positive pos_icon showinline up">+10.00%</div>", "After Close" ] , [ "10/27/2016", "9/2016", "$0.85", "$0.52", "<div class="right neg negative neg_icon showinline down">-0.33</div>", "<div class="right neg negative neg_icon showinline down">-38.82%</div>", "After Close" ] , [ "7/28/2016", "6/2016", "$1.14", "$1.78", "<div class="right pos positive pos_icon showinline up">+0.64</div>", "<div class="right pos positive pos_icon showinline up">+56.14%</div>", "After Close" ] , [ "4/28/2016", "3/2016", "$0.61", "$1.07", "<div class="right pos positive pos_icon showinline up">+0.46</div>", "<div class="right pos positive pos_icon showinline up">+75.41%</div>", "After Close" ] , [ "1/28/2016", "12/2015", "$1.61", "$1.00", "<div class="right neg negative neg_icon showinline down">-0.61</div>", "<div class="right neg negative neg_icon showinline down">-37.89%</div>", "After Close" ] , [ "10/22/2015", "9/2015", "-$0.1", "$0.17", "<div class="right pos positive pos_icon showinline up">+0.27</div>", "<div class="right pos positive pos_icon showinline up">+270.00%</div>", "After Close" ] , [ "7/23/2015", "6/2015", "-$0.15", "$0.19", "<div class="right pos positive pos_icon showinline up">+0.34</div>", "<div class="right pos positive pos_icon showinline up">+226.67%</div>", "After Close" ] , [ "4/23/2015", "3/2015", "-$0.13", "-$0.12", "<div class="right pos positive pos_icon showinline up">+0.01</div>", "<div class="right pos positive pos_icon showinline up">+7.69%</div>", "After Close" ] , [ "1/29/2015", "12/2014", "$0.24", "$0.45", "<div class="right pos positive pos_icon showinline up">+0.21</div>", "<div class="right pos positive pos_icon showinline up">+87.50%</div>", "After Close" ] , [ "10/23/2014", "9/2014", "-$0.73", "-$0.95", "<div class="right neg negative neg_icon showinline down">-0.22</div>", "<div class="right neg negative neg_icon showinline down">-30.14%</div>", "After Close" ] , [ "7/24/2014", "6/2014", "-$0.13", "-$0.27", "<div class="right neg negative neg_icon showinline down">-0.14</div>", "<div class="right neg negative neg_icon showinline down">-107.69%</div>", "After Close" ] , [ "4/24/2014", "3/2014", "$0.22", "$0.23", "<div class="right pos positive pos_icon showinline up">+0.01</div>", "<div class="right pos positive pos_icon showinline up">+4.55%</div>", "After Close" ] , [ "1/30/2014", "12/2013", "$0.68", "$0.51", "<div class="right neg negative neg_icon showinline down">-0.17</div>", "<div class="right neg negative neg_icon showinline down">-25.00%</div>", "After Close" ] , [ "10/24/2013", "9/2013", "-$0.09", "-$0.09", "<div class="right pos_na showinline">0.00</div>", "<div class="right pos_na showinline">0.00%</div>", "After Close" ] , [ "7/25/2013", "6/2013", "$0.04", "-$0.02", "<div class="right neg negative neg_icon showinline down">-0.06</div>", "<div class="right neg negative neg_icon showinline down">-150.00%</div>", "After Close" ] , [ "4/25/2013", "3/2013", "$0.10", "$0.18", "<div class="right pos positive pos_icon showinline up">+0.08</div>", "<div class="right pos positive pos_icon showinline up">+80.00%</div>", "After Close" ] , [ "1/29/2013", "12/2012", "$0.28", "$0.21", "<div class="right neg negative neg_icon showinline down">-0.07</div>", "<div class="right neg negative neg_icon showinline down">-25.00%</div>", "After Close" ] , [ "10/25/2012", "9/2012", "-$0.08", "-$0.23", "<div class="right neg negative neg_icon showinline down">-0.15</div>", "<div class="right neg negative neg_icon showinline down">-187.50%</div>", "After Close" ] , [ "7/26/2012", "6/2012", "--", "--", "--", "--", "After Close" ] , [ "4/26/2012", "3/2012", "--", "--", "--", "--", "After Close" ] , [ "1/31/2012", "12/2011", "--", "--", "--", "--", "After Close" ] , [ "10/25/2011", "9/2011", "--", "--", "--", "--", "After Close" ] , [ "7/26/2011", "6/2011", "--", "--", "--", "--", "After Close" ] , [ "4/26/2011", "3/2011", "--", "--", "--", "--", "--" ] , [ "1/27/2011", "12/2010", "--", "--", "--", "--", "After Close" ] , [ "10/21/2010", "9/2010", "--", "--", "--", "--", "After Close" ] , [ "7/22/2010", "6/2010", "--", "--", "--", "--", "After Close" ] , [ "4/22/2010", "3/2010", "--", "--", "--", "--", "After Close" ] , [ "1/28/2010", "12/2009", "--", "--", "--", "--", "After Close" ] , [ "10/22/2009", "9/2009", "--", "--", "--", "--", "After Close" ] , [ "7/23/2009", "6/2009", "--", "--", "--", "--", "After Close" ] ]


How could I get this table?
Thanks!







html python-2.7 beautifulsoup






share|improve this question















share|improve this question













share|improve this question




share|improve this question








edited Sep 3 '17 at 14:49

























asked Sep 3 '17 at 14:40









Diego

3321315




3321315








  • 1




    the data is loaded dynamically and not sown in html, so you have to parse the data you are getting.
    – MD. Khairul Basar
    Sep 3 '17 at 14:54










  • Thanks!! PhantomJS, selenium?
    – Diego
    Sep 3 '17 at 15:16










  • I checked page source and still looked the same, so I don't think it will help. But, still can give it a try.
    – MD. Khairul Basar
    Sep 3 '17 at 15:18






  • 1




    It worked with selenium!
    – Diego
    Sep 3 '17 at 16:39
















  • 1




    the data is loaded dynamically and not sown in html, so you have to parse the data you are getting.
    – MD. Khairul Basar
    Sep 3 '17 at 14:54










  • Thanks!! PhantomJS, selenium?
    – Diego
    Sep 3 '17 at 15:16










  • I checked page source and still looked the same, so I don't think it will help. But, still can give it a try.
    – MD. Khairul Basar
    Sep 3 '17 at 15:18






  • 1




    It worked with selenium!
    – Diego
    Sep 3 '17 at 16:39










1




1




the data is loaded dynamically and not sown in html, so you have to parse the data you are getting.
– MD. Khairul Basar
Sep 3 '17 at 14:54




the data is loaded dynamically and not sown in html, so you have to parse the data you are getting.
– MD. Khairul Basar
Sep 3 '17 at 14:54












Thanks!! PhantomJS, selenium?
– Diego
Sep 3 '17 at 15:16




Thanks!! PhantomJS, selenium?
– Diego
Sep 3 '17 at 15:16












I checked page source and still looked the same, so I don't think it will help. But, still can give it a try.
– MD. Khairul Basar
Sep 3 '17 at 15:18




I checked page source and still looked the same, so I don't think it will help. But, still can give it a try.
– MD. Khairul Basar
Sep 3 '17 at 15:18




1




1




It worked with selenium!
– Diego
Sep 3 '17 at 16:39






It worked with selenium!
– Diego
Sep 3 '17 at 16:39














1 Answer
1






active

oldest

votes

















up vote
0
down vote



accepted










So the solution is to parse the whole HTML document using Python's string and RegExp functions instead of BeautifulSoup because we are not trying to get the data from HTML tags but instead we want to get them inside a JS code.



So this code basically, get the JS array inside "earnings_announcements_earnings_table" and since the JS Array is the same as Python's list structure, I just parse it using ast. The result is a list were you can loop into and it shows all data from all the pages of the table.



import urllib2
import re
import ast

user_agent = {'User-Agent': 'Mozilla/5.0 (Windows NT 6.1; WOW64; rv:12.0) Gecko/20100101 Firefox/12.0'}
req = urllib2.Request('https://www.zacks.com/stock/research/amzn/earnings-announcements', None, user_agent)
source = urllib2.urlopen(req).read()

compiled = re.compile('"earnings_announcements_earnings_table"s+:', flags=re.IGNORECASE | re.DOTALL)
match = re.search(compiled, source)
if match:
source = source[match.end(): len(source)]

compiled = re.compile('"earnings_announcements_webcasts_table"', flags=re.IGNORECASE | re.DOTALL)
match = re.search(compiled, source)
if match:
source = source[0: match.start()]

result = ast.literal_eval(str(source).strip('rnt, '))
print result


Let me know if you need clarifications.






share|improve this answer





















  • Thanks so much! It works great! Elements 4 and 5 on each list appear w html code.
    – Diego
    Sep 3 '17 at 16:36












  • Great! You can call BeautifulSoup if you like to clean those html codes.
    – chad
    Sep 3 '17 at 16:48











Your Answer






StackExchange.ifUsing("editor", function () {
StackExchange.using("externalEditor", function () {
StackExchange.using("snippets", function () {
StackExchange.snippets.init();
});
});
}, "code-snippets");

StackExchange.ready(function() {
var channelOptions = {
tags: "".split(" "),
id: "1"
};
initTagRenderer("".split(" "), "".split(" "), channelOptions);

StackExchange.using("externalEditor", function() {
// Have to fire editor after snippets, if snippets enabled
if (StackExchange.settings.snippets.snippetsEnabled) {
StackExchange.using("snippets", function() {
createEditor();
});
}
else {
createEditor();
}
});

function createEditor() {
StackExchange.prepareEditor({
heartbeatType: 'answer',
convertImagesToLinks: true,
noModals: true,
showLowRepImageUploadWarning: true,
reputationToPostImages: 10,
bindNavPrevention: true,
postfix: "",
imageUploader: {
brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
allowUrls: true
},
onDemand: true,
discardSelector: ".discard-answer"
,immediatelyShowMarkdownHelp:true
});


}
});














 

draft saved


draft discarded


















StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstackoverflow.com%2fquestions%2f46024536%2fcannot-get-table-data-html%23new-answer', 'question_page');
}
);

Post as a guest















Required, but never shown

























1 Answer
1






active

oldest

votes








1 Answer
1






active

oldest

votes









active

oldest

votes






active

oldest

votes








up vote
0
down vote



accepted










So the solution is to parse the whole HTML document using Python's string and RegExp functions instead of BeautifulSoup because we are not trying to get the data from HTML tags but instead we want to get them inside a JS code.



So this code basically, get the JS array inside "earnings_announcements_earnings_table" and since the JS Array is the same as Python's list structure, I just parse it using ast. The result is a list were you can loop into and it shows all data from all the pages of the table.



import urllib2
import re
import ast

user_agent = {'User-Agent': 'Mozilla/5.0 (Windows NT 6.1; WOW64; rv:12.0) Gecko/20100101 Firefox/12.0'}
req = urllib2.Request('https://www.zacks.com/stock/research/amzn/earnings-announcements', None, user_agent)
source = urllib2.urlopen(req).read()

compiled = re.compile('"earnings_announcements_earnings_table"s+:', flags=re.IGNORECASE | re.DOTALL)
match = re.search(compiled, source)
if match:
source = source[match.end(): len(source)]

compiled = re.compile('"earnings_announcements_webcasts_table"', flags=re.IGNORECASE | re.DOTALL)
match = re.search(compiled, source)
if match:
source = source[0: match.start()]

result = ast.literal_eval(str(source).strip('rnt, '))
print result


Let me know if you need clarifications.






share|improve this answer





















  • Thanks so much! It works great! Elements 4 and 5 on each list appear w html code.
    – Diego
    Sep 3 '17 at 16:36












  • Great! You can call BeautifulSoup if you like to clean those html codes.
    – chad
    Sep 3 '17 at 16:48















up vote
0
down vote



accepted










So the solution is to parse the whole HTML document using Python's string and RegExp functions instead of BeautifulSoup because we are not trying to get the data from HTML tags but instead we want to get them inside a JS code.



So this code basically, get the JS array inside "earnings_announcements_earnings_table" and since the JS Array is the same as Python's list structure, I just parse it using ast. The result is a list were you can loop into and it shows all data from all the pages of the table.



import urllib2
import re
import ast

user_agent = {'User-Agent': 'Mozilla/5.0 (Windows NT 6.1; WOW64; rv:12.0) Gecko/20100101 Firefox/12.0'}
req = urllib2.Request('https://www.zacks.com/stock/research/amzn/earnings-announcements', None, user_agent)
source = urllib2.urlopen(req).read()

compiled = re.compile('"earnings_announcements_earnings_table"s+:', flags=re.IGNORECASE | re.DOTALL)
match = re.search(compiled, source)
if match:
source = source[match.end(): len(source)]

compiled = re.compile('"earnings_announcements_webcasts_table"', flags=re.IGNORECASE | re.DOTALL)
match = re.search(compiled, source)
if match:
source = source[0: match.start()]

result = ast.literal_eval(str(source).strip('rnt, '))
print result


Let me know if you need clarifications.






share|improve this answer





















  • Thanks so much! It works great! Elements 4 and 5 on each list appear w html code.
    – Diego
    Sep 3 '17 at 16:36












  • Great! You can call BeautifulSoup if you like to clean those html codes.
    – chad
    Sep 3 '17 at 16:48













up vote
0
down vote



accepted







up vote
0
down vote



accepted






So the solution is to parse the whole HTML document using Python's string and RegExp functions instead of BeautifulSoup because we are not trying to get the data from HTML tags but instead we want to get them inside a JS code.



So this code basically, get the JS array inside "earnings_announcements_earnings_table" and since the JS Array is the same as Python's list structure, I just parse it using ast. The result is a list were you can loop into and it shows all data from all the pages of the table.



import urllib2
import re
import ast

user_agent = {'User-Agent': 'Mozilla/5.0 (Windows NT 6.1; WOW64; rv:12.0) Gecko/20100101 Firefox/12.0'}
req = urllib2.Request('https://www.zacks.com/stock/research/amzn/earnings-announcements', None, user_agent)
source = urllib2.urlopen(req).read()

compiled = re.compile('"earnings_announcements_earnings_table"s+:', flags=re.IGNORECASE | re.DOTALL)
match = re.search(compiled, source)
if match:
source = source[match.end(): len(source)]

compiled = re.compile('"earnings_announcements_webcasts_table"', flags=re.IGNORECASE | re.DOTALL)
match = re.search(compiled, source)
if match:
source = source[0: match.start()]

result = ast.literal_eval(str(source).strip('rnt, '))
print result


Let me know if you need clarifications.






share|improve this answer












So the solution is to parse the whole HTML document using Python's string and RegExp functions instead of BeautifulSoup because we are not trying to get the data from HTML tags but instead we want to get them inside a JS code.



So this code basically, get the JS array inside "earnings_announcements_earnings_table" and since the JS Array is the same as Python's list structure, I just parse it using ast. The result is a list were you can loop into and it shows all data from all the pages of the table.



import urllib2
import re
import ast

user_agent = {'User-Agent': 'Mozilla/5.0 (Windows NT 6.1; WOW64; rv:12.0) Gecko/20100101 Firefox/12.0'}
req = urllib2.Request('https://www.zacks.com/stock/research/amzn/earnings-announcements', None, user_agent)
source = urllib2.urlopen(req).read()

compiled = re.compile('"earnings_announcements_earnings_table"s+:', flags=re.IGNORECASE | re.DOTALL)
match = re.search(compiled, source)
if match:
source = source[match.end(): len(source)]

compiled = re.compile('"earnings_announcements_webcasts_table"', flags=re.IGNORECASE | re.DOTALL)
match = re.search(compiled, source)
if match:
source = source[0: match.start()]

result = ast.literal_eval(str(source).strip('rnt, '))
print result


Let me know if you need clarifications.







share|improve this answer












share|improve this answer



share|improve this answer










answered Sep 3 '17 at 16:16









chad

763215




763215












  • Thanks so much! It works great! Elements 4 and 5 on each list appear w html code.
    – Diego
    Sep 3 '17 at 16:36












  • Great! You can call BeautifulSoup if you like to clean those html codes.
    – chad
    Sep 3 '17 at 16:48


















  • Thanks so much! It works great! Elements 4 and 5 on each list appear w html code.
    – Diego
    Sep 3 '17 at 16:36












  • Great! You can call BeautifulSoup if you like to clean those html codes.
    – chad
    Sep 3 '17 at 16:48
















Thanks so much! It works great! Elements 4 and 5 on each list appear w html code.
– Diego
Sep 3 '17 at 16:36






Thanks so much! It works great! Elements 4 and 5 on each list appear w html code.
– Diego
Sep 3 '17 at 16:36














Great! You can call BeautifulSoup if you like to clean those html codes.
– chad
Sep 3 '17 at 16:48




Great! You can call BeautifulSoup if you like to clean those html codes.
– chad
Sep 3 '17 at 16:48


















 

draft saved


draft discarded



















































 


draft saved


draft discarded














StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstackoverflow.com%2fquestions%2f46024536%2fcannot-get-table-data-html%23new-answer', 'question_page');
}
);

Post as a guest















Required, but never shown





















































Required, but never shown














Required, but never shown












Required, but never shown







Required, but never shown

































Required, but never shown














Required, but never shown












Required, but never shown







Required, but never shown







Popular posts from this blog

Bressuire

Vorschmack

Quarantine