1. I have a 5GB dump of crawled data that needs to be parsed.
2. It's contact data: name, workplace, phone number, fax, etc etc.. 250k contacts in total.
3. The data is very clean, all from ONE source, but it is pretty complex.
4. Each contact has many attributes, like 1 workplace, n skills, n references to other contacts in the database (colleagues, co-workers in the same workplace, co-worker in the same department etc.)
I need this data parsed. It is currently dumped into MongoDB and we have already written a shell script parsing much of the contact information, but not all. If you want you can continue with this script, or you can start over and create your own.
This should not take longer than a day of work if you know what you are doing.
This is a fair estimate from a developler who has crawled it and started parsing, but now doesn't have the time to complete the job.
There is no rush, so you can work on this over the next 3 weeks.
41 Freelancer bieten im Durchschnitt $233 für diesen Job
hi, please attach sample crawled data and list out all the fields to parse out I can write a Perl script to extract the crawled data into MySQL database/CSV file
Hi, if possible that you export you data to a csv or text file it will be more easy to parse the data correctly. I will download the file and parse the data using perl scripts. Best regards, Ilirjan