Application to scrape data from SEC Edgar public info site
$30-100 USD
Σε Εξέλιξη
Αναρτήθηκε περίπου 12 χρόνια πριν
$30-100 USD
Πληρωμή κατά την παράδοση
I need an application - which may be in Java, R, or any other environment that I don't need to buy new software for! - that can go and retrieve documents from <[login to view URL]> (type a symbol in the description/ticker field and a list of all documents for that ticker is returned).
Inputs: list of company tickers, date range, document type from {10K, 10Q, 8k}, destination folder. Inputs can be in a manually-edited settings file rather than in a GUI.
Output: downloaded files of that document type for that company in that date range into a local folder (with filenames to indicate company ticker, document type, and year)
I think this is a fairly straightforward assignment for web-savvy programmers.
## Deliverables
* * *This broadcast message was sent to all bidders on Monday Feb 6, 2012 7:58:45 AM:
To clarify one answer that all bidders will need to know, concerning which file to scrape and what to do with it. Choose the .htm file that has a column matching the "Type" or report chosen (so 8K, 10K, etc). It will usually be the first item. Saving as .htm is okay. But I am not sure how to make the associated files show up as part of the .htm when offline? So you may have to save all of them to a single folder. I'm fuzzy on how this would work but those associated files are important. Alternatively, if there is a way to save the fully-integrated file as a .pdf, that's okay as well. Some of these submissions in fact will have a .pdf version, and it's okay to save that as well.