|
|
Posted by Colin McKinnon on 08/10/06 21:50
joe t. wrote:
> The subject may sound a little cryptic, so i'll try my best to explain.
> Details are unavailable, as i am under a nondisclosure agreement, but
> i'm looking for general principles and tips, not necessarily fixes for
> existing code.
>
<snip long winded explanation>
So you want to copy someone else's data and you've only got an HTTP
interface intended for humans to the remote system.
There's plenty of companies doing this already - no need to be shy.
How simple it is depends on how well their site is written - assuming it is
well written you should be able to parse the pages with the XML parser. How
to get the pages? That's rather up to you - you could use a site ripper
like pavuk or write your own spider, e.g. using snoopy.
HTH
C.
Navigation:
[Reply to this message]
|