HI.
Its a bit tricky process and you need to build a specialized html parsing logic depending on which news paper site you are targeting.
Suppose you are targeting
http://epaperbeta.timesofindia.com/index.aspx?eid=31808&dt=20150905[
^]
Then figure out how you get url of PDF inside this html page.
Just for clue I can tell you that download link available at the top which invokes the below java script code.
function downloadpdftest() {
var getslidevalue = parseInt(sudoSlider.getValue("currentSlide"), 10);
var nextsudoslider = sudoSlider.getSlide(getslidevalue);
var nextslideid = nextsudoslider.find('img').attr('src');
var fPath = nextslideid.toString();
fPath = fPath.replace(".JPG", ".pdf");
fPath = fPath.replace(".jpg", ".pdf");
var currPDFName = fPath.replace("Page", "PagePrint");
window.open(currPDFName, 'PDF', 'left=150,top=10,width=750,height=700,scrollbars=yes,status=yes');
}
Now try to generate currPDFName (pdf url) by fetching the content of this page in c# using WebRequest or WebClient Class.
http://stackoverflow.com/questions/16642196/get-html-code-from-a-website-c-sharp[
^]code see
Then parse html to generate PDF url.
Once you get pdf url then below code to download PDF file.
using(WebClient client = new WebClient())
{
client.DownloadFile("http://www.irs.gov/pub/irs-pdf/fw4.pdf", @"C:\Temp.pdf");
}
http://stackoverflow.com/questions/2913830/download-pdf-programatically[
^]