Reply to Re: my Script Dies during fgetcsv

Your name:

Reply:


Posted by MS on 09/28/40 11:40

I Tried the set_time_limit(0);
and I have the same issue.

The particular piece of code is 4 calls deep would this have any baring in
the set_time_limit? do I need set_time_limit in each of the 4 calling
scripts?

Alex

> It looks like your script times out. Try set_time_limit(0);
>
> Hope this helps,
>
> DXTrim
>
> >
> > $handle = fopen($feed_url , "r");
> > while (($data = fgetcsv($handle, 2000, ",")) !== FALSE) {
> > set_time_limit(600);
> > ...
> > ...
> > }
> >
> > which basically reads the $feed_url link and processes the data and
inputs
> > into the database.
> >
> > this process is repeated for different $feed_urls
> >
> > And works for a few urls but the script just stops (with no
> error/warnings)
> > when it reaches around 12000 database entries
> >
> > I added set_time_limit(600); thinking that the script was timing out,
but
> > it doesn;t seem to have made any difference.
> >
> > My max_input_time is set to -1 (i dont know what the -1 denotes)
> >
> > Is there such a thing as timeout for the website i am recieving the data
> > from??
> >
> > Any other ideas what could be causing this?
> >
> > Alex
> > --
>
> --------------------------------------------------------------------------
> --
> > http://www.myclubweb.co.uk - The Home of Club Websites
>
> --------------------------------------------------------------------------

[Back to original message]


Удаленная работа для программистов  •  Как заработать на Google AdSense  •  England, UK  •  статьи на английском  •  PHP MySQL CMS Apache Oscommerce  •  Online Business Knowledge Base  •  DVD MP3 AVI MP4 players codecs conversion help
Home  •  Search  •  Site Map  •  Set as Homepage  •  Add to Favourites

Copyright © 2005-2006 Powered by Custom PHP Programming

Сайт изготовлен в Студии Валентина Петручека
изготовление и поддержка веб-сайтов, разработка программного обеспечения, поисковая оптимизация