Large scale crawler with C/C++???
I have been given the task of overseeing the development of our companys new project which is the design and implementation of a large scale crawling, processing and serving system in a distributed environment. Tasks include implementing search algorithms, data mining, improving relevancy or search results, managing terabytes of data and scaling algorithms to work on very large data sets, and serving search results using a large network of LINUX servers. The twist is that the company wants to use C and C++ in the developement rather than Java which I am proficient in.
My problem is that I have only worked with small scale models and most of my experience is in Java. I would appreciate it if anyone can provide me some information on the subject and point me in the right direction